[ 466.876284] env[61806]: DEBUG os_vif [-] Loaded VIF plugin class '' with name 'linux_bridge' {{(pid=61806) initialize /opt/stack/data/venv/lib/python3.10/site-packages/os_vif/__init__.py:44}} [ 466.876606] env[61806]: DEBUG os_vif [-] Loaded VIF plugin class '' with name 'noop' {{(pid=61806) initialize /opt/stack/data/venv/lib/python3.10/site-packages/os_vif/__init__.py:44}} [ 466.876651] env[61806]: DEBUG os_vif [-] Loaded VIF plugin class '' with name 'ovs' {{(pid=61806) initialize /opt/stack/data/venv/lib/python3.10/site-packages/os_vif/__init__.py:44}} [ 466.876982] env[61806]: INFO os_vif [-] Loaded VIF plugins: linux_bridge, noop, ovs [ 466.970689] env[61806]: DEBUG oslo_concurrency.processutils [-] Running cmd (subprocess): grep -F node.session.scan /sbin/iscsiadm {{(pid=61806) execute /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/processutils.py:390}} [ 466.980744] env[61806]: DEBUG oslo_concurrency.processutils [-] CMD "grep -F node.session.scan /sbin/iscsiadm" returned: 0 in 0.010s {{(pid=61806) execute /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/processutils.py:428}} [ 467.583047] env[61806]: INFO nova.virt.driver [None req-dc7f331a-33fb-411a-b781-8d8651f91bc6 None None] Loading compute driver 'vmwareapi.VMwareVCDriver' [ 467.654509] env[61806]: DEBUG oslo_concurrency.lockutils [-] Acquiring lock "oslo_vmware_api_lock" by "oslo_vmware.api.VMwareAPISession._create_session" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 467.654674] env[61806]: DEBUG oslo_concurrency.lockutils [-] Lock "oslo_vmware_api_lock" acquired by "oslo_vmware.api.VMwareAPISession._create_session" :: waited 0.000s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 467.654776] env[61806]: DEBUG oslo_vmware.service [-] Creating suds client with soap_url='https://vc1.osci.c.eu-de-1.cloud.sap:443/sdk' and wsdl_url='https://vc1.osci.c.eu-de-1.cloud.sap:443/sdk/vimService.wsdl' {{(pid=61806) __init__ /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:242}} [ 470.758032] env[61806]: DEBUG oslo_vmware.service [-] Invoking ServiceInstance.RetrieveServiceContent with opID=oslo.vmware-0bce1ffc-69a3-49c2-b65b-bbb3ee71938d {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 470.773863] env[61806]: DEBUG oslo_vmware.api [-] Logging into host: vc1.osci.c.eu-de-1.cloud.sap. {{(pid=61806) _create_session /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:242}} [ 470.774032] env[61806]: DEBUG oslo_vmware.service [-] Invoking SessionManager.Login with opID=oslo.vmware-a22589c8-adac-42cb-9fde-49f01abbd139 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 470.813837] env[61806]: INFO oslo_vmware.api [-] Successfully established new session; session ID is 82d3f. [ 470.813993] env[61806]: DEBUG oslo_concurrency.lockutils [-] Lock "oslo_vmware_api_lock" "released" by "oslo_vmware.api.VMwareAPISession._create_session" :: held 3.159s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 470.814547] env[61806]: INFO nova.virt.vmwareapi.driver [None req-dc7f331a-33fb-411a-b781-8d8651f91bc6 None None] VMware vCenter version: 7.0.3 [ 470.817997] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0e14f349-18c3-4a89-8e0c-444337daf353 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 470.834952] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5d7588a0-8fb0-446f-9cd3-e2a1c96f6dcc {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 470.841055] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-336bf0dc-cdb1-4346-997f-3c0cd1a045c1 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 470.847657] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-20d0d48d-f422-411c-b775-c6b218eb0671 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 470.860616] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-72e0d09d-c95c-42d5-ae2e-ab8fe38f0bf8 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 470.866493] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0f5d4b60-de64-4484-b822-ae4e341f9065 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 470.896317] env[61806]: DEBUG oslo_vmware.service [-] Invoking ExtensionManager.FindExtension with opID=oslo.vmware-6991de76-3324-453c-ba66-e29f80113ef3 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 470.901307] env[61806]: DEBUG nova.virt.vmwareapi.driver [None req-dc7f331a-33fb-411a-b781-8d8651f91bc6 None None] Extension org.openstack.compute already exists. {{(pid=61806) _register_openstack_extension /opt/stack/nova/nova/virt/vmwareapi/driver.py:225}} [ 470.903924] env[61806]: INFO nova.compute.provider_config [None req-dc7f331a-33fb-411a-b781-8d8651f91bc6 None None] No provider configs found in /etc/nova/provider_config/. If files are present, ensure the Nova process has access. [ 471.407671] env[61806]: DEBUG nova.context [None req-dc7f331a-33fb-411a-b781-8d8651f91bc6 None None] Found 2 cells: 00000000-0000-0000-0000-000000000000(cell0),674c5d15-eba8-4fa9-a7e3-618d5869bdac(cell1) {{(pid=61806) load_cells /opt/stack/nova/nova/context.py:464}} [ 471.409871] env[61806]: DEBUG oslo_concurrency.lockutils [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] Acquiring lock "00000000-0000-0000-0000-000000000000" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 471.410098] env[61806]: DEBUG oslo_concurrency.lockutils [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] Lock "00000000-0000-0000-0000-000000000000" acquired by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: waited 0.000s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 471.410824] env[61806]: DEBUG oslo_concurrency.lockutils [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] Lock "00000000-0000-0000-0000-000000000000" "released" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: held 0.001s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 471.411259] env[61806]: DEBUG oslo_concurrency.lockutils [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] Acquiring lock "674c5d15-eba8-4fa9-a7e3-618d5869bdac" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 471.411450] env[61806]: DEBUG oslo_concurrency.lockutils [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] Lock "674c5d15-eba8-4fa9-a7e3-618d5869bdac" acquired by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: waited 0.000s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 471.412455] env[61806]: DEBUG oslo_concurrency.lockutils [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] Lock "674c5d15-eba8-4fa9-a7e3-618d5869bdac" "released" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: held 0.001s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 471.432377] env[61806]: INFO dbcounter [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] Registered counter for database nova_cell0 [ 471.440599] env[61806]: INFO dbcounter [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] Registered counter for database nova_cell1 [ 471.444035] env[61806]: DEBUG oslo_db.sqlalchemy.engines [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] MySQL server mode set to STRICT_TRANS_TABLES,STRICT_ALL_TABLES,NO_ZERO_IN_DATE,NO_ZERO_DATE,ERROR_FOR_DIVISION_BY_ZERO,TRADITIONAL,NO_AUTO_CREATE_USER,NO_ENGINE_SUBSTITUTION {{(pid=61806) _check_effective_sql_mode /opt/stack/data/venv/lib/python3.10/site-packages/oslo_db/sqlalchemy/engines.py:342}} [ 471.444660] env[61806]: DEBUG oslo_db.sqlalchemy.engines [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] MySQL server mode set to STRICT_TRANS_TABLES,STRICT_ALL_TABLES,NO_ZERO_IN_DATE,NO_ZERO_DATE,ERROR_FOR_DIVISION_BY_ZERO,TRADITIONAL,NO_AUTO_CREATE_USER,NO_ENGINE_SUBSTITUTION {{(pid=61806) _check_effective_sql_mode /opt/stack/data/venv/lib/python3.10/site-packages/oslo_db/sqlalchemy/engines.py:342}} [ 471.449094] env[61806]: ERROR nova.db.main.api [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] No DB access allowed in nova-compute: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 471.449094] env[61806]: result = function(*args, **kwargs) [ 471.449094] env[61806]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 471.449094] env[61806]: return func(*args, **kwargs) [ 471.449094] env[61806]: File "/opt/stack/nova/nova/context.py", line 422, in gather_result [ 471.449094] env[61806]: result = fn(*args, **kwargs) [ 471.449094] env[61806]: File "/opt/stack/nova/nova/db/main/api.py", line 179, in wrapper [ 471.449094] env[61806]: return f(*args, **kwargs) [ 471.449094] env[61806]: File "/opt/stack/nova/nova/objects/service.py", line 553, in _db_service_get_minimum_version [ 471.449094] env[61806]: return db.service_get_minimum_version(context, binaries) [ 471.449094] env[61806]: File "/opt/stack/nova/nova/db/main/api.py", line 238, in wrapper [ 471.449094] env[61806]: _check_db_access() [ 471.449094] env[61806]: File "/opt/stack/nova/nova/db/main/api.py", line 188, in _check_db_access [ 471.449094] env[61806]: stacktrace = ''.join(traceback.format_stack()) [ 471.449094] env[61806]: [ 471.450209] env[61806]: ERROR nova.db.main.api [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] No DB access allowed in nova-compute: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 471.450209] env[61806]: result = function(*args, **kwargs) [ 471.450209] env[61806]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 471.450209] env[61806]: return func(*args, **kwargs) [ 471.450209] env[61806]: File "/opt/stack/nova/nova/context.py", line 422, in gather_result [ 471.450209] env[61806]: result = fn(*args, **kwargs) [ 471.450209] env[61806]: File "/opt/stack/nova/nova/db/main/api.py", line 179, in wrapper [ 471.450209] env[61806]: return f(*args, **kwargs) [ 471.450209] env[61806]: File "/opt/stack/nova/nova/objects/service.py", line 553, in _db_service_get_minimum_version [ 471.450209] env[61806]: return db.service_get_minimum_version(context, binaries) [ 471.450209] env[61806]: File "/opt/stack/nova/nova/db/main/api.py", line 238, in wrapper [ 471.450209] env[61806]: _check_db_access() [ 471.450209] env[61806]: File "/opt/stack/nova/nova/db/main/api.py", line 188, in _check_db_access [ 471.450209] env[61806]: stacktrace = ''.join(traceback.format_stack()) [ 471.450209] env[61806]: [ 471.450810] env[61806]: WARNING nova.objects.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] Failed to get minimum service version for cell 00000000-0000-0000-0000-000000000000 [ 471.450810] env[61806]: WARNING nova.objects.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] Failed to get minimum service version for cell 674c5d15-eba8-4fa9-a7e3-618d5869bdac [ 471.451191] env[61806]: DEBUG oslo_concurrency.lockutils [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] Acquiring lock "singleton_lock" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 471.451364] env[61806]: DEBUG oslo_concurrency.lockutils [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] Acquired lock "singleton_lock" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 471.451610] env[61806]: DEBUG oslo_concurrency.lockutils [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] Releasing lock "singleton_lock" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 471.451927] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] Full set of CONF: {{(pid=61806) _wait_for_exit_or_signal /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/service.py:363}} [ 471.452081] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] ******************************************************************************** {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2806}} [ 471.452212] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] Configuration options gathered from: {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2807}} [ 471.452349] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] command line args: ['--config-file', '/etc/nova/nova.conf', '--config-file', '/etc/nova/nova-cpu-common.conf', '--config-file', '/etc/nova/nova-cpu-1.conf'] {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2808}} [ 471.452542] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] config files: ['/etc/nova/nova.conf', '/etc/nova/nova-cpu-common.conf', '/etc/nova/nova-cpu-1.conf'] {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2809}} [ 471.452672] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] ================================================================================ {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2811}} [ 471.452882] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] allow_resize_to_same_host = True {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 471.453067] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] arq_binding_timeout = 300 {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 471.453202] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] backdoor_port = None {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 471.453331] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] backdoor_socket = None {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 471.453495] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] block_device_allocate_retries = 60 {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 471.453657] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] block_device_allocate_retries_interval = 3 {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 471.453824] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] cert = self.pem {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 471.453989] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] compute_driver = vmwareapi.VMwareVCDriver {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 471.454173] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] compute_monitors = [] {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 471.454341] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] config_dir = [] {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 471.454516] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] config_drive_format = iso9660 {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 471.454647] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] config_file = ['/etc/nova/nova.conf', '/etc/nova/nova-cpu-common.conf', '/etc/nova/nova-cpu-1.conf'] {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 471.454810] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] config_source = [] {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 471.454975] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] console_host = devstack {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 471.455154] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] control_exchange = nova {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 471.455314] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] cpu_allocation_ratio = None {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 471.455474] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] daemon = False {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 471.455639] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] debug = True {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 471.455793] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] default_access_ip_network_name = None {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 471.455957] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] default_availability_zone = nova {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 471.456123] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] default_ephemeral_format = None {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 471.456317] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] default_green_pool_size = 1000 {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 471.456576] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] default_log_levels = ['amqp=WARN', 'amqplib=WARN', 'boto=WARN', 'qpid=WARN', 'sqlalchemy=WARN', 'suds=INFO', 'oslo.messaging=INFO', 'oslo_messaging=INFO', 'iso8601=WARN', 'requests.packages.urllib3.connectionpool=WARN', 'urllib3.connectionpool=WARN', 'websocket=WARN', 'requests.packages.urllib3.util.retry=WARN', 'urllib3.util.retry=WARN', 'keystonemiddleware=WARN', 'routes.middleware=WARN', 'stevedore=WARN', 'taskflow=WARN', 'keystoneauth=WARN', 'oslo.cache=INFO', 'oslo_policy=INFO', 'dogpile.core.dogpile=INFO', 'glanceclient=WARN', 'oslo.privsep.daemon=INFO'] {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 471.456746] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] default_schedule_zone = None {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 471.456907] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] disk_allocation_ratio = None {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 471.457081] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] enable_new_services = True {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 471.457264] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] enabled_apis = ['osapi_compute'] {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 471.457431] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] enabled_ssl_apis = [] {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 471.457595] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] flat_injected = False {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 471.457754] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] force_config_drive = False {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 471.457912] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] force_raw_images = True {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 471.458094] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] graceful_shutdown_timeout = 5 {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 471.458255] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] heal_instance_info_cache_interval = 60 {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 471.458498] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] host = cpu-1 {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 471.458686] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] initial_cpu_allocation_ratio = 4.0 {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 471.458855] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] initial_disk_allocation_ratio = 1.0 {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 471.459034] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] initial_ram_allocation_ratio = 1.0 {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 471.459254] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] injected_network_template = /opt/stack/nova/nova/virt/interfaces.template {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 471.459420] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] instance_build_timeout = 0 {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 471.459581] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] instance_delete_interval = 300 {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 471.459748] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] instance_format = [instance: %(uuid)s] {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 471.459915] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] instance_name_template = instance-%08x {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 471.460090] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] instance_usage_audit = False {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 471.460266] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] instance_usage_audit_period = month {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 471.460435] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] instance_uuid_format = [instance: %(uuid)s] {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 471.460604] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] instances_path = /opt/stack/data/nova/instances {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 471.460770] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] internal_service_availability_zone = internal {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 471.460923] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] key = None {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 471.461093] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] live_migration_retry_count = 30 {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 471.461262] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] log_color = False {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 471.461426] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] log_config_append = None {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 471.461597] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] log_date_format = %Y-%m-%d %H:%M:%S {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 471.461758] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] log_dir = None {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 471.461918] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] log_file = None {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 471.462058] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] log_options = True {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 471.462221] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] log_rotate_interval = 1 {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 471.462391] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] log_rotate_interval_type = days {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 471.462560] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] log_rotation_type = none {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 471.462686] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] logging_context_format_string = %(color)s%(levelname)s %(name)s [%(global_request_id)s %(request_id)s %(project_name)s %(user_name)s%(color)s] %(instance)s%(color)s%(message)s {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 471.462814] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] logging_debug_format_suffix = {{(pid=%(process)d) %(funcName)s %(pathname)s:%(lineno)d}} {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 471.462980] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] logging_default_format_string = %(color)s%(levelname)s %(name)s [-%(color)s] %(instance)s%(color)s%(message)s {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 471.463159] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] logging_exception_prefix = ERROR %(name)s %(instance)s {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 471.463288] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] logging_user_identity_format = %(user)s %(project)s %(domain)s %(system_scope)s %(user_domain)s %(project_domain)s {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 471.463449] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] long_rpc_timeout = 1800 {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 471.463608] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] max_concurrent_builds = 10 {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 471.463766] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] max_concurrent_live_migrations = 1 {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 471.463921] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] max_concurrent_snapshots = 5 {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 471.464089] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] max_local_block_devices = 3 {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 471.464246] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] max_logfile_count = 30 {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 471.464403] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] max_logfile_size_mb = 200 {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 471.464565] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] maximum_instance_delete_attempts = 5 {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 471.464726] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] metadata_listen = 0.0.0.0 {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 471.464891] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] metadata_listen_port = 8775 {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 471.465066] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] metadata_workers = 2 {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 471.465230] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] migrate_max_retries = -1 {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 471.465398] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] mkisofs_cmd = genisoimage {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 471.465608] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] my_block_storage_ip = 10.180.1.21 {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 471.465743] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] my_ip = 10.180.1.21 {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 471.465904] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] network_allocate_retries = 0 {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 471.466093] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] non_inheritable_image_properties = ['cache_in_nova', 'bittorrent'] {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 471.466282] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] osapi_compute_listen = 0.0.0.0 {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 471.466459] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] osapi_compute_listen_port = 8774 {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 471.466658] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] osapi_compute_unique_server_name_scope = {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 471.466827] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] osapi_compute_workers = 2 {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 471.466989] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] password_length = 12 {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 471.467163] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] periodic_enable = True {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 471.467320] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] periodic_fuzzy_delay = 60 {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 471.467489] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] pointer_model = usbtablet {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 471.467655] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] preallocate_images = none {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 471.467814] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] publish_errors = False {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 471.467944] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] pybasedir = /opt/stack/nova {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 471.468114] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] ram_allocation_ratio = None {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 471.468287] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] rate_limit_burst = 0 {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 471.468506] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] rate_limit_except_level = CRITICAL {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 471.468677] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] rate_limit_interval = 0 {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 471.468852] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] reboot_timeout = 0 {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 471.469025] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] reclaim_instance_interval = 0 {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 471.469186] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] record = None {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 471.469372] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] reimage_timeout_per_gb = 60 {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 471.469553] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] report_interval = 120 {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 471.469715] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] rescue_timeout = 0 {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 471.469876] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] reserved_host_cpus = 0 {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 471.470047] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] reserved_host_disk_mb = 0 {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 471.470210] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] reserved_host_memory_mb = 512 {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 471.470389] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] reserved_huge_pages = None {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 471.470563] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] resize_confirm_window = 0 {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 471.470775] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] resize_fs_using_block_device = False {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 471.470956] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] resume_guests_state_on_host_boot = False {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 471.471144] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] rootwrap_config = /etc/nova/rootwrap.conf {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 471.471307] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] rpc_response_timeout = 60 {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 471.471467] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] run_external_periodic_tasks = True {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 471.471638] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] running_deleted_instance_action = reap {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 471.471798] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] running_deleted_instance_poll_interval = 1800 {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 471.471956] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] running_deleted_instance_timeout = 0 {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 471.472128] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] scheduler_instance_sync_interval = 120 {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 471.472301] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] service_down_time = 720 {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 471.472471] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] servicegroup_driver = db {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 471.472631] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] shell_completion = None {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 471.472789] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] shelved_offload_time = 0 {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 471.472971] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] shelved_poll_interval = 3600 {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 471.473240] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] shutdown_timeout = 0 {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 471.473426] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] source_is_ipv6 = False {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 471.473597] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] ssl_only = False {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 471.473845] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] state_path = /opt/stack/data/n-cpu-1 {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 471.474024] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] sync_power_state_interval = 600 {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 471.474195] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] sync_power_state_pool_size = 1000 {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 471.474368] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] syslog_log_facility = LOG_USER {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 471.474532] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] tempdir = None {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 471.474692] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] timeout_nbd = 10 {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 471.474863] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] transport_url = **** {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 471.475033] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] update_resources_interval = 0 {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 471.475199] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] use_cow_images = True {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 471.475360] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] use_eventlog = False {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 471.475523] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] use_journal = False {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 471.475681] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] use_json = False {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 471.475839] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] use_rootwrap_daemon = False {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 471.476049] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] use_stderr = False {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 471.476267] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] use_syslog = False {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 471.476436] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] vcpu_pin_set = None {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 471.476611] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] vif_plugging_is_fatal = True {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 471.476782] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] vif_plugging_timeout = 300 {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 471.476952] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] virt_mkfs = [] {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 471.477132] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] volume_usage_poll_interval = 0 {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 471.477298] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] watch_log_file = False {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 471.477470] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] web = /usr/share/spice-html5 {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 471.477661] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] oslo_concurrency.disable_process_locking = False {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.478233] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] oslo_concurrency.lock_path = /opt/stack/data/n-cpu-1 {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.478465] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] oslo_messaging_metrics.metrics_buffer_size = 1000 {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.478651] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] oslo_messaging_metrics.metrics_enabled = False {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.478835] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] oslo_messaging_metrics.metrics_process_name = {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.479018] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] oslo_messaging_metrics.metrics_socket_file = /var/tmp/metrics_collector.sock {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.479202] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] oslo_messaging_metrics.metrics_thread_stop_timeout = 10 {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.479393] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] api.auth_strategy = keystone {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.479568] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] api.compute_link_prefix = None {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.479746] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] api.config_drive_skip_versions = 1.0 2007-01-19 2007-03-01 2007-08-29 2007-10-10 2007-12-15 2008-02-01 2008-09-01 {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.479923] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] api.dhcp_domain = novalocal {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.480119] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] api.enable_instance_password = True {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.480387] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] api.glance_link_prefix = None {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.480596] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] api.instance_list_cells_batch_fixed_size = 100 {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.480782] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] api.instance_list_cells_batch_strategy = distributed {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.480952] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] api.instance_list_per_project_cells = False {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.481139] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] api.list_records_by_skipping_down_cells = True {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.481312] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] api.local_metadata_per_cell = False {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.481489] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] api.max_limit = 1000 {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.481661] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] api.metadata_cache_expiration = 15 {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.481837] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] api.neutron_default_tenant_id = default {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.482021] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] api.response_validation = warn {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.482198] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] api.use_neutron_default_nets = False {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.482367] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] api.vendordata_dynamic_connect_timeout = 5 {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.482532] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] api.vendordata_dynamic_failure_fatal = False {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.482702] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] api.vendordata_dynamic_read_timeout = 5 {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.482877] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] api.vendordata_dynamic_ssl_certfile = {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.483060] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] api.vendordata_dynamic_targets = [] {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.483230] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] api.vendordata_jsonfile_path = None {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.483419] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] api.vendordata_providers = ['StaticJSON'] {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.483615] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] cache.backend = dogpile.cache.memcached {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.483788] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] cache.backend_argument = **** {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.483961] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] cache.config_prefix = cache.oslo {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.484150] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] cache.dead_timeout = 60.0 {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.484318] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] cache.debug_cache_backend = False {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.484486] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] cache.enable_retry_client = False {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.484652] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] cache.enable_socket_keepalive = False {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.484826] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] cache.enabled = True {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.484991] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] cache.enforce_fips_mode = False {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.485172] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] cache.expiration_time = 600 {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.485341] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] cache.hashclient_retry_attempts = 2 {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.485514] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] cache.hashclient_retry_delay = 1.0 {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.485681] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] cache.memcache_dead_retry = 300 {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.485842] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] cache.memcache_password = **** {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.486022] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] cache.memcache_pool_connection_get_timeout = 10 {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.486198] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] cache.memcache_pool_flush_on_reconnect = False {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.486388] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] cache.memcache_pool_maxsize = 10 {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.486562] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] cache.memcache_pool_unused_timeout = 60 {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.486728] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] cache.memcache_sasl_enabled = False {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.486912] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] cache.memcache_servers = ['localhost:11211'] {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.487100] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] cache.memcache_socket_timeout = 1.0 {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.487268] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] cache.memcache_username = None {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.487440] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] cache.proxies = [] {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.487612] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] cache.redis_db = 0 {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.487777] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] cache.redis_password = **** {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.487952] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] cache.redis_sentinel_service_name = mymaster {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.488147] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] cache.redis_sentinels = ['localhost:26379'] {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.488341] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] cache.redis_server = localhost:6379 {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.488524] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] cache.redis_socket_timeout = 1.0 {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.488689] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] cache.redis_username = None {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.488857] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] cache.retry_attempts = 2 {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.489037] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] cache.retry_delay = 0.0 {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.489208] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] cache.socket_keepalive_count = 1 {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.489374] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] cache.socket_keepalive_idle = 1 {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.489537] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] cache.socket_keepalive_interval = 1 {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.489696] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] cache.tls_allowed_ciphers = None {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.489855] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] cache.tls_cafile = None {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.490021] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] cache.tls_certfile = None {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.490189] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] cache.tls_enabled = False {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.490349] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] cache.tls_keyfile = None {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.490522] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] cinder.auth_section = None {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.490717] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] cinder.auth_type = password {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.490894] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] cinder.cafile = None {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.491087] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] cinder.catalog_info = volumev3::publicURL {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.491255] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] cinder.certfile = None {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.491422] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] cinder.collect_timing = False {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.491589] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] cinder.cross_az_attach = True {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.491753] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] cinder.debug = False {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.491915] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] cinder.endpoint_template = None {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.492093] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] cinder.http_retries = 3 {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.492262] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] cinder.insecure = False {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.492445] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] cinder.keyfile = None {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.492645] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] cinder.os_region_name = RegionOne {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.492817] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] cinder.split_loggers = False {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.492980] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] cinder.timeout = None {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.493169] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] compute.consecutive_build_service_disable_threshold = 10 {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.493335] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] compute.cpu_dedicated_set = None {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.493501] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] compute.cpu_shared_set = None {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.493669] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] compute.image_type_exclude_list = [] {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.493839] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] compute.live_migration_wait_for_vif_plug = True {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.494015] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] compute.max_concurrent_disk_ops = 0 {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.494188] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] compute.max_disk_devices_to_attach = -1 {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.494353] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] compute.packing_host_numa_cells_allocation_strategy = False {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.494527] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] compute.provider_config_location = /etc/nova/provider_config/ {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.494693] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] compute.resource_provider_association_refresh = 300 {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.494855] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] compute.sharing_providers_max_uuids_per_request = 200 {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.495028] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] compute.shutdown_retry_interval = 10 {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.495219] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] compute.vmdk_allowed_types = ['streamOptimized', 'monolithicSparse'] {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.495402] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] conductor.workers = 2 {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.495585] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] console.allowed_origins = [] {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.495750] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] console.ssl_ciphers = None {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.495922] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] console.ssl_minimum_version = default {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.496107] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] consoleauth.enforce_session_timeout = False {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.496300] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] consoleauth.token_ttl = 600 {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.496481] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] cyborg.cafile = None {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.496647] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] cyborg.certfile = None {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.496813] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] cyborg.collect_timing = False {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.496976] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] cyborg.connect_retries = None {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.497153] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] cyborg.connect_retry_delay = None {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.497313] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] cyborg.endpoint_override = None {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.497478] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] cyborg.insecure = False {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.497638] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] cyborg.keyfile = None {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.497799] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] cyborg.max_version = None {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.497959] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] cyborg.min_version = None {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.498130] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] cyborg.region_name = None {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.498294] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] cyborg.retriable_status_codes = None {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.498483] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] cyborg.service_name = None {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.498663] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] cyborg.service_type = accelerator {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.498828] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] cyborg.split_loggers = False {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.498990] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] cyborg.status_code_retries = None {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.499167] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] cyborg.status_code_retry_delay = None {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.499330] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] cyborg.timeout = None {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.499587] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] cyborg.valid_interfaces = ['internal', 'public'] {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.499791] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] cyborg.version = None {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.499984] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] database.backend = sqlalchemy {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.500177] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] database.connection = **** {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.500350] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] database.connection_debug = 0 {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.500529] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] database.connection_parameters = {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.500708] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] database.connection_recycle_time = 3600 {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.500880] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] database.connection_trace = False {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.501058] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] database.db_inc_retry_interval = True {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.501232] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] database.db_max_retries = 20 {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.501398] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] database.db_max_retry_interval = 10 {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.501566] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] database.db_retry_interval = 1 {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.501731] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] database.max_overflow = 50 {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.501896] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] database.max_pool_size = 5 {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.502071] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] database.max_retries = 10 {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.502247] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] database.mysql_sql_mode = TRADITIONAL {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.502410] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] database.mysql_wsrep_sync_wait = None {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.502577] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] database.pool_timeout = None {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.502741] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] database.retry_interval = 10 {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.502902] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] database.slave_connection = **** {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.503076] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] database.sqlite_synchronous = True {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.503245] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] database.use_db_reconnect = False {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.503427] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] api_database.backend = sqlalchemy {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.503665] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] api_database.connection = **** {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.503858] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] api_database.connection_debug = 0 {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.504048] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] api_database.connection_parameters = {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.504222] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] api_database.connection_recycle_time = 3600 {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.504395] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] api_database.connection_trace = False {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.504562] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] api_database.db_inc_retry_interval = True {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.504732] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] api_database.db_max_retries = 20 {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.504897] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] api_database.db_max_retry_interval = 10 {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.505070] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] api_database.db_retry_interval = 1 {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.505240] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] api_database.max_overflow = 50 {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.505406] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] api_database.max_pool_size = 5 {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.505573] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] api_database.max_retries = 10 {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.505745] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] api_database.mysql_sql_mode = TRADITIONAL {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.505908] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] api_database.mysql_wsrep_sync_wait = None {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.506079] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] api_database.pool_timeout = None {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.506263] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] api_database.retry_interval = 10 {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.506439] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] api_database.slave_connection = **** {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.506609] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] api_database.sqlite_synchronous = True {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.506786] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] devices.enabled_mdev_types = [] {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.506970] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] ephemeral_storage_encryption.cipher = aes-xts-plain64 {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.507166] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] ephemeral_storage_encryption.default_format = luks {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.507335] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] ephemeral_storage_encryption.enabled = False {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.507505] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] ephemeral_storage_encryption.key_size = 512 {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.507680] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] glance.api_servers = None {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.507848] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] glance.cafile = None {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.508019] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] glance.certfile = None {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.508190] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] glance.collect_timing = False {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.508382] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] glance.connect_retries = None {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.508578] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] glance.connect_retry_delay = None {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.508749] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] glance.debug = False {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.508919] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] glance.default_trusted_certificate_ids = [] {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.509097] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] glance.enable_certificate_validation = False {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.509265] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] glance.enable_rbd_download = False {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.509440] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] glance.endpoint_override = None {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.509626] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] glance.insecure = False {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.509794] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] glance.keyfile = None {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.509956] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] glance.max_version = None {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.510131] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] glance.min_version = None {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.510299] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] glance.num_retries = 3 {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.510473] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] glance.rbd_ceph_conf = {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.510682] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] glance.rbd_connect_timeout = 5 {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.510840] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] glance.rbd_pool = {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.511017] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] glance.rbd_user = {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.511185] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] glance.region_name = None {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.511352] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] glance.retriable_status_codes = None {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.511517] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] glance.service_name = None {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.511692] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] glance.service_type = image {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.511858] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] glance.split_loggers = False {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.512029] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] glance.status_code_retries = None {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.512196] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] glance.status_code_retry_delay = None {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.512359] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] glance.timeout = None {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.512573] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] glance.valid_interfaces = ['internal', 'public'] {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.512748] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] glance.verify_glance_signatures = False {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.512913] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] glance.version = None {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.513090] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] guestfs.debug = False {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.513261] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] mks.enabled = False {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.513620] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] mks.mksproxy_base_url = http://127.0.0.1:6090/ {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.513825] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] image_cache.manager_interval = 2400 {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.513987] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] image_cache.precache_concurrency = 1 {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.514172] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] image_cache.remove_unused_base_images = True {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.514343] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] image_cache.remove_unused_original_minimum_age_seconds = 86400 {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.514515] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] image_cache.remove_unused_resized_minimum_age_seconds = 3600 {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.514692] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] image_cache.subdirectory_name = _base {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.514868] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] ironic.api_max_retries = 60 {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.515057] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] ironic.api_retry_interval = 2 {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.515211] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] ironic.auth_section = None {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.515376] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] ironic.auth_type = None {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.515558] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] ironic.cafile = None {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.515799] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] ironic.certfile = None {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.515988] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] ironic.collect_timing = False {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.516173] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] ironic.conductor_group = None {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.516341] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] ironic.connect_retries = None {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.516506] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] ironic.connect_retry_delay = None {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.516665] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] ironic.endpoint_override = None {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.516832] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] ironic.insecure = False {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.516992] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] ironic.keyfile = None {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.517182] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] ironic.max_version = None {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.517351] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] ironic.min_version = None {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.517524] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] ironic.peer_list = [] {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.517688] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] ironic.region_name = None {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.517850] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] ironic.retriable_status_codes = None {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.518028] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] ironic.serial_console_state_timeout = 10 {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.518197] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] ironic.service_name = None {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.518405] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] ironic.service_type = baremetal {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.518586] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] ironic.shard = None {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.518792] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] ironic.split_loggers = False {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.518968] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] ironic.status_code_retries = None {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.519144] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] ironic.status_code_retry_delay = None {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.519312] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] ironic.timeout = None {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.519516] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] ironic.valid_interfaces = ['internal', 'public'] {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.519694] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] ironic.version = None {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.519882] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] key_manager.backend = nova.keymgr.conf_key_mgr.ConfKeyManager {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.520072] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] key_manager.fixed_key = **** {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.520262] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] barbican.auth_endpoint = http://localhost/identity/v3 {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.520429] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] barbican.barbican_api_version = None {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.520594] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] barbican.barbican_endpoint = None {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.520770] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] barbican.barbican_endpoint_type = public {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.520932] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] barbican.barbican_region_name = None {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.521108] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] barbican.cafile = None {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.521272] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] barbican.certfile = None {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.521437] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] barbican.collect_timing = False {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.521603] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] barbican.insecure = False {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.521763] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] barbican.keyfile = None {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.521927] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] barbican.number_of_retries = 60 {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.522101] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] barbican.retry_delay = 1 {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.522270] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] barbican.send_service_user_token = False {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.522432] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] barbican.split_loggers = False {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.522595] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] barbican.timeout = None {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.522757] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] barbican.verify_ssl = True {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.522915] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] barbican.verify_ssl_path = None {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.523093] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] barbican_service_user.auth_section = None {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.523261] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] barbican_service_user.auth_type = None {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.523423] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] barbican_service_user.cafile = None {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.523584] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] barbican_service_user.certfile = None {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.523747] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] barbican_service_user.collect_timing = False {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.523907] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] barbican_service_user.insecure = False {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.524077] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] barbican_service_user.keyfile = None {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.524244] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] barbican_service_user.split_loggers = False {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.524402] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] barbican_service_user.timeout = None {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.524578] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] vault.approle_role_id = **** {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.524801] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] vault.approle_secret_id = **** {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.524990] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] vault.kv_mountpoint = secret {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.525173] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] vault.kv_path = None {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.525345] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] vault.kv_version = 2 {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.525512] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] vault.namespace = None {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.525674] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] vault.root_token_id = **** {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.525835] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] vault.ssl_ca_crt_file = None {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.526008] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] vault.timeout = 60.0 {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.526185] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] vault.use_ssl = False {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.526396] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] vault.vault_url = http://127.0.0.1:8200 {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.526581] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] keystone.auth_section = None {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.526748] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] keystone.auth_type = None {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.526911] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] keystone.cafile = None {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.527086] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] keystone.certfile = None {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.527258] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] keystone.collect_timing = False {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.527420] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] keystone.connect_retries = None {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.527583] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] keystone.connect_retry_delay = None {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.527745] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] keystone.endpoint_override = None {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.527906] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] keystone.insecure = False {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.528076] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] keystone.keyfile = None {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.528241] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] keystone.max_version = None {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.528436] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] keystone.min_version = None {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.528607] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] keystone.region_name = None {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.528772] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] keystone.retriable_status_codes = None {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.528935] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] keystone.service_name = None {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.529133] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] keystone.service_type = identity {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.529326] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] keystone.split_loggers = False {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.529467] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] keystone.status_code_retries = None {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.529632] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] keystone.status_code_retry_delay = None {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.529792] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] keystone.timeout = None {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.529975] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] keystone.valid_interfaces = ['internal', 'public'] {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.530153] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] keystone.version = None {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.530358] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] libvirt.connection_uri = {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.530526] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] libvirt.cpu_mode = None {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.530705] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] libvirt.cpu_model_extra_flags = [] {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.530874] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] libvirt.cpu_models = [] {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.531061] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] libvirt.cpu_power_governor_high = performance {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.531237] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] libvirt.cpu_power_governor_low = powersave {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.531406] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] libvirt.cpu_power_management = False {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.531583] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] libvirt.cpu_power_management_strategy = cpu_state {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.531758] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] libvirt.device_detach_attempts = 8 {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.531920] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] libvirt.device_detach_timeout = 20 {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.532100] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] libvirt.disk_cachemodes = [] {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.532265] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] libvirt.disk_prefix = None {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.532435] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] libvirt.enabled_perf_events = [] {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.532604] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] libvirt.file_backed_memory = 0 {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.532771] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] libvirt.gid_maps = [] {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.532933] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] libvirt.hw_disk_discard = None {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.533108] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] libvirt.hw_machine_type = None {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.533283] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] libvirt.images_rbd_ceph_conf = {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.533449] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] libvirt.images_rbd_glance_copy_poll_interval = 15 {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.533616] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] libvirt.images_rbd_glance_copy_timeout = 600 {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.533788] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] libvirt.images_rbd_glance_store_name = {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.533965] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] libvirt.images_rbd_pool = rbd {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.534152] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] libvirt.images_type = default {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.534314] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] libvirt.images_volume_group = None {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.534481] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] libvirt.inject_key = False {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.534644] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] libvirt.inject_partition = -2 {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.534813] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] libvirt.inject_password = False {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.534969] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] libvirt.iscsi_iface = None {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.535145] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] libvirt.iser_use_multipath = False {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.535313] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] libvirt.live_migration_bandwidth = 0 {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.535479] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] libvirt.live_migration_completion_timeout = 800 {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.535642] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] libvirt.live_migration_downtime = 500 {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.535806] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] libvirt.live_migration_downtime_delay = 75 {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.535968] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] libvirt.live_migration_downtime_steps = 10 {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.536144] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] libvirt.live_migration_inbound_addr = None {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.536337] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] libvirt.live_migration_permit_auto_converge = False {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.536511] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] libvirt.live_migration_permit_post_copy = False {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.536674] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] libvirt.live_migration_scheme = None {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.536849] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] libvirt.live_migration_timeout_action = abort {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.537029] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] libvirt.live_migration_tunnelled = False {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.537199] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] libvirt.live_migration_uri = None {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.537366] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] libvirt.live_migration_with_native_tls = False {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.537532] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] libvirt.max_queues = None {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.537697] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] libvirt.mem_stats_period_seconds = 10 {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.537933] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] libvirt.migration_inbound_addr = 10.180.1.21 {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.538105] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] libvirt.nfs_mount_options = None {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.538464] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] libvirt.nfs_mount_point_base = /opt/stack/data/n-cpu-1/mnt {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.538652] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] libvirt.num_aoe_discover_tries = 3 {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.538824] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] libvirt.num_iser_scan_tries = 5 {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.538991] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] libvirt.num_memory_encrypted_guests = None {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.539176] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] libvirt.num_nvme_discover_tries = 5 {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.539365] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] libvirt.num_pcie_ports = 0 {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.539556] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] libvirt.num_volume_scan_tries = 5 {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.539730] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] libvirt.pmem_namespaces = [] {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.539895] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] libvirt.quobyte_client_cfg = None {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.540201] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] libvirt.quobyte_mount_point_base = /opt/stack/data/n-cpu-1/mnt {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.540380] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] libvirt.rbd_connect_timeout = 5 {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.540556] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] libvirt.rbd_destroy_volume_retries = 12 {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.540723] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] libvirt.rbd_destroy_volume_retry_interval = 5 {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.540887] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] libvirt.rbd_secret_uuid = None {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.541060] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] libvirt.rbd_user = None {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.541232] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] libvirt.realtime_scheduler_priority = 1 {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.541408] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] libvirt.remote_filesystem_transport = ssh {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.541575] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] libvirt.rescue_image_id = None {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.541737] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] libvirt.rescue_kernel_id = None {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.541898] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] libvirt.rescue_ramdisk_id = None {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.542080] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] libvirt.rng_dev_path = /dev/urandom {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.542246] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] libvirt.rx_queue_size = None {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.542420] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] libvirt.smbfs_mount_options = {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.542698] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] libvirt.smbfs_mount_point_base = /opt/stack/data/n-cpu-1/mnt {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.542874] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] libvirt.snapshot_compression = False {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.543050] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] libvirt.snapshot_image_format = None {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.543275] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] libvirt.snapshots_directory = /opt/stack/data/nova/instances/snapshots {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.543446] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] libvirt.sparse_logical_volumes = False {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.543615] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] libvirt.swtpm_enabled = False {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.543785] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] libvirt.swtpm_group = tss {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.543953] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] libvirt.swtpm_user = tss {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.544139] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] libvirt.sysinfo_serial = unique {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.544299] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] libvirt.tb_cache_size = None {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.544460] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] libvirt.tx_queue_size = None {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.544626] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] libvirt.uid_maps = [] {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.544791] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] libvirt.use_virtio_for_bridges = True {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.544958] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] libvirt.virt_type = kvm {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.545139] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] libvirt.volume_clear = zero {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.545341] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] libvirt.volume_clear_size = 0 {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.545470] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] libvirt.volume_use_multipath = False {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.545636] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] libvirt.vzstorage_cache_path = None {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.545806] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] libvirt.vzstorage_log_path = /var/log/vstorage/%(cluster_name)s/nova.log.gz {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.545977] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] libvirt.vzstorage_mount_group = qemu {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.546159] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] libvirt.vzstorage_mount_opts = [] {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.546362] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] libvirt.vzstorage_mount_perms = 0770 {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.546649] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] libvirt.vzstorage_mount_point_base = /opt/stack/data/n-cpu-1/mnt {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.546831] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] libvirt.vzstorage_mount_user = stack {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.547022] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] libvirt.wait_soft_reboot_seconds = 120 {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.547196] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] neutron.auth_section = None {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.547375] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] neutron.auth_type = password {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.547542] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] neutron.cafile = None {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.547707] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] neutron.certfile = None {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.547874] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] neutron.collect_timing = False {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.548046] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] neutron.connect_retries = None {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.548214] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] neutron.connect_retry_delay = None {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.548421] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] neutron.default_floating_pool = public {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.548590] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] neutron.endpoint_override = None {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.548761] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] neutron.extension_sync_interval = 600 {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.548921] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] neutron.http_retries = 3 {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.549095] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] neutron.insecure = False {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.549258] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] neutron.keyfile = None {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.549443] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] neutron.max_version = None {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.549627] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] neutron.metadata_proxy_shared_secret = **** {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.549789] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] neutron.min_version = None {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.549961] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] neutron.ovs_bridge = br-int {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.550142] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] neutron.physnets = [] {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.550325] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] neutron.region_name = RegionOne {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.550520] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] neutron.retriable_status_codes = None {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.550698] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] neutron.service_metadata_proxy = True {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.550861] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] neutron.service_name = None {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.551044] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] neutron.service_type = network {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.551214] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] neutron.split_loggers = False {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.551375] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] neutron.status_code_retries = None {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.551538] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] neutron.status_code_retry_delay = None {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.551698] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] neutron.timeout = None {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.551882] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] neutron.valid_interfaces = ['internal', 'public'] {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.552057] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] neutron.version = None {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.552237] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] notifications.bdms_in_notifications = False {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.552417] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] notifications.default_level = INFO {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.552596] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] notifications.notification_format = unversioned {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.552762] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] notifications.notify_on_state_change = None {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.552940] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] notifications.versioned_notifications_topics = ['versioned_notifications'] {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.553134] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] pci.alias = [] {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.553309] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] pci.device_spec = [] {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.553476] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] pci.report_in_placement = False {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.553653] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] placement.auth_section = None {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.553827] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] placement.auth_type = password {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.553997] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] placement.auth_url = http://10.180.1.21/identity {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.554173] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] placement.cafile = None {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.554335] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] placement.certfile = None {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.554501] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] placement.collect_timing = False {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.554660] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] placement.connect_retries = None {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.554822] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] placement.connect_retry_delay = None {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.554985] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] placement.default_domain_id = None {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.555158] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] placement.default_domain_name = None {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.555319] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] placement.domain_id = None {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.555483] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] placement.domain_name = None {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.555641] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] placement.endpoint_override = None {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.555803] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] placement.insecure = False {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.555964] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] placement.keyfile = None {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.556137] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] placement.max_version = None {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.556320] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] placement.min_version = None {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.556501] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] placement.password = **** {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.556665] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] placement.project_domain_id = None {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.556834] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] placement.project_domain_name = Default {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.557013] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] placement.project_id = None {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.557194] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] placement.project_name = service {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.557368] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] placement.region_name = RegionOne {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.557537] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] placement.retriable_status_codes = None {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.557701] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] placement.service_name = None {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.557875] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] placement.service_type = placement {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.558050] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] placement.split_loggers = False {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.558216] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] placement.status_code_retries = None {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.558418] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] placement.status_code_retry_delay = None {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.558671] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] placement.system_scope = None {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.558756] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] placement.timeout = None {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.558919] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] placement.trust_id = None {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.559093] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] placement.user_domain_id = None {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.559268] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] placement.user_domain_name = Default {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.559487] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] placement.user_id = None {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.559685] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] placement.username = nova {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.559875] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] placement.valid_interfaces = ['internal', 'public'] {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.560053] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] placement.version = None {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.560240] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] quota.cores = 20 {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.560409] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] quota.count_usage_from_placement = False {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.560586] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] quota.driver = nova.quota.DbQuotaDriver {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.560763] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] quota.injected_file_content_bytes = 10240 {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.560935] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] quota.injected_file_path_length = 255 {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.561118] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] quota.injected_files = 5 {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.561291] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] quota.instances = 10 {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.561461] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] quota.key_pairs = 100 {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.561633] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] quota.metadata_items = 128 {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.561802] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] quota.ram = 51200 {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.561968] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] quota.recheck_quota = True {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.562159] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] quota.server_group_members = 10 {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.562330] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] quota.server_groups = 10 {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.562606] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] scheduler.discover_hosts_in_cells_interval = -1 {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.562826] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] scheduler.enable_isolated_aggregate_filtering = False {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.562975] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] scheduler.image_metadata_prefilter = False {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.563162] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] scheduler.limit_tenants_to_placement_aggregate = False {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.563333] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] scheduler.max_attempts = 3 {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.563503] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] scheduler.max_placement_results = 1000 {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.563670] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] scheduler.placement_aggregate_required_for_tenants = False {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.563836] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] scheduler.query_placement_for_image_type_support = False {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.564008] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] scheduler.query_placement_for_routed_network_aggregates = False {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.564189] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] scheduler.workers = 2 {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.564363] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] filter_scheduler.aggregate_image_properties_isolation_namespace = None {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.564540] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] filter_scheduler.aggregate_image_properties_isolation_separator = . {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.564725] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] filter_scheduler.available_filters = ['nova.scheduler.filters.all_filters'] {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.564899] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] filter_scheduler.build_failure_weight_multiplier = 1000000.0 {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.565080] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] filter_scheduler.cpu_weight_multiplier = 1.0 {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.565253] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] filter_scheduler.cross_cell_move_weight_multiplier = 1000000.0 {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.565427] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] filter_scheduler.disk_weight_multiplier = 1.0 {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.565662] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] filter_scheduler.enabled_filters = ['ComputeFilter', 'ComputeCapabilitiesFilter', 'ImagePropertiesFilter', 'ServerGroupAntiAffinityFilter', 'ServerGroupAffinityFilter', 'SameHostFilter', 'DifferentHostFilter'] {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.565843] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] filter_scheduler.host_subset_size = 1 {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.566025] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] filter_scheduler.hypervisor_version_weight_multiplier = 1.0 {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.566235] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] filter_scheduler.image_properties_default_architecture = None {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.566388] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] filter_scheduler.io_ops_weight_multiplier = -1.0 {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.566562] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] filter_scheduler.isolated_hosts = [] {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.566733] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] filter_scheduler.isolated_images = [] {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.566898] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] filter_scheduler.max_instances_per_host = 50 {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.567072] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] filter_scheduler.max_io_ops_per_host = 8 {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.567240] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] filter_scheduler.num_instances_weight_multiplier = 0.0 {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.567403] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] filter_scheduler.pci_in_placement = False {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.567568] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] filter_scheduler.pci_weight_multiplier = 1.0 {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.567730] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] filter_scheduler.ram_weight_multiplier = 1.0 {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.567892] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] filter_scheduler.restrict_isolated_hosts_to_isolated_images = True {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.568062] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] filter_scheduler.shuffle_best_same_weighed_hosts = False {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.568231] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] filter_scheduler.soft_affinity_weight_multiplier = 1.0 {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.568422] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] filter_scheduler.soft_anti_affinity_weight_multiplier = 1.0 {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.568599] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] filter_scheduler.track_instance_changes = True {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.568778] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] filter_scheduler.weight_classes = ['nova.scheduler.weights.all_weighers'] {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.568951] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] metrics.required = True {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.569134] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] metrics.weight_multiplier = 1.0 {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.569309] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] metrics.weight_of_unavailable = -10000.0 {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.569517] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] metrics.weight_setting = [] {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.569837] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] serial_console.base_url = ws://127.0.0.1:6083/ {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.570026] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] serial_console.enabled = False {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.570211] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] serial_console.port_range = 10000:20000 {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.570386] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] serial_console.proxyclient_address = 127.0.0.1 {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.570569] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] serial_console.serialproxy_host = 0.0.0.0 {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.570741] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] serial_console.serialproxy_port = 6083 {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.570914] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] service_user.auth_section = None {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.571104] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] service_user.auth_type = password {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.571270] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] service_user.cafile = None {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.571433] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] service_user.certfile = None {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.571600] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] service_user.collect_timing = False {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.571764] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] service_user.insecure = False {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.571924] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] service_user.keyfile = None {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.572109] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] service_user.send_service_user_token = True {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.572280] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] service_user.split_loggers = False {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.572441] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] service_user.timeout = None {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.572617] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] spice.agent_enabled = True {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.572781] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] spice.enabled = False {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.573128] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] spice.html5proxy_base_url = http://127.0.0.1:6082/spice_auto.html {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.573326] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] spice.html5proxy_host = 0.0.0.0 {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.573499] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] spice.html5proxy_port = 6082 {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.573672] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] spice.image_compression = None {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.573872] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] spice.jpeg_compression = None {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.574053] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] spice.playback_compression = None {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.574288] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] spice.require_secure = False {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.574526] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] spice.server_listen = 127.0.0.1 {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.574715] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] spice.server_proxyclient_address = 127.0.0.1 {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.574880] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] spice.streaming_mode = None {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.575052] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] spice.zlib_compression = None {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.575227] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] upgrade_levels.baseapi = None {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.575402] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] upgrade_levels.compute = auto {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.575569] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] upgrade_levels.conductor = None {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.575732] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] upgrade_levels.scheduler = None {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.575901] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] vendordata_dynamic_auth.auth_section = None {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.576078] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] vendordata_dynamic_auth.auth_type = None {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.576246] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] vendordata_dynamic_auth.cafile = None {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.576408] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] vendordata_dynamic_auth.certfile = None {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.576649] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] vendordata_dynamic_auth.collect_timing = False {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.576839] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] vendordata_dynamic_auth.insecure = False {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.577018] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] vendordata_dynamic_auth.keyfile = None {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.577258] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] vendordata_dynamic_auth.split_loggers = False {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.577446] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] vendordata_dynamic_auth.timeout = None {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.577630] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] vmware.api_retry_count = 10 {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.577800] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] vmware.ca_file = None {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.577976] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] vmware.cache_prefix = devstack-image-cache {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.578164] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] vmware.cluster_name = testcl1 {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.578351] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] vmware.connection_pool_size = 10 {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.578520] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] vmware.console_delay_seconds = None {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.578724] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] vmware.datastore_regex = ^datastore.* {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.578983] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] vmware.host_ip = vc1.osci.c.eu-de-1.cloud.sap {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.579185] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] vmware.host_password = **** {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.579364] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] vmware.host_port = 443 {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.579540] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] vmware.host_username = administrator@vsphere.local {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.579713] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] vmware.insecure = True {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.579878] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] vmware.integration_bridge = None {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.580057] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] vmware.maximum_objects = 100 {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.580224] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] vmware.pbm_default_policy = None {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.580390] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] vmware.pbm_enabled = False {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.580552] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] vmware.pbm_wsdl_location = None {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.580723] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] vmware.serial_log_dir = /opt/vmware/vspc {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.580887] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] vmware.serial_port_proxy_uri = None {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.581056] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] vmware.serial_port_service_uri = None {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.581229] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] vmware.task_poll_interval = 0.5 {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.581402] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] vmware.use_linked_clone = False {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.581573] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] vmware.vnc_keymap = en-us {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.581741] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] vmware.vnc_port = 5900 {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.581904] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] vmware.vnc_port_total = 10000 {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.582111] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] vnc.auth_schemes = ['none'] {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.582317] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] vnc.enabled = False {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.582640] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] vnc.novncproxy_base_url = http://127.0.0.1:6080/vnc_auto.html {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.582829] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] vnc.novncproxy_host = 0.0.0.0 {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.583013] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] vnc.novncproxy_port = 6080 {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.583201] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] vnc.server_listen = 127.0.0.1 {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.583373] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] vnc.server_proxyclient_address = 127.0.0.1 {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.583538] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] vnc.vencrypt_ca_certs = None {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.583696] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] vnc.vencrypt_client_cert = None {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.583853] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] vnc.vencrypt_client_key = None {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.584046] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] workarounds.disable_compute_service_check_for_ffu = False {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.584220] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] workarounds.disable_deep_image_inspection = False {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.584388] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] workarounds.disable_fallback_pcpu_query = False {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.584555] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] workarounds.disable_group_policy_check_upcall = False {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.584720] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] workarounds.disable_libvirt_livesnapshot = False {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.584884] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] workarounds.disable_rootwrap = False {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.585058] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] workarounds.enable_numa_live_migration = False {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.585227] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] workarounds.enable_qemu_monitor_announce_self = False {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.585394] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] workarounds.ensure_libvirt_rbd_instance_dir_cleanup = False {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.585561] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] workarounds.handle_virt_lifecycle_events = True {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.585774] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] workarounds.libvirt_disable_apic = False {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.585884] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] workarounds.never_download_image_if_on_rbd = False {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.586059] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] workarounds.qemu_monitor_announce_self_count = 3 {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.586239] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] workarounds.qemu_monitor_announce_self_interval = 1 {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.586719] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] workarounds.reserve_disk_resource_for_image_cache = False {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.586719] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] workarounds.skip_cpu_compare_at_startup = False {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.586817] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] workarounds.skip_cpu_compare_on_dest = False {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.586918] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] workarounds.skip_hypervisor_version_check_on_lm = False {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.587120] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] workarounds.skip_reserve_in_use_ironic_nodes = False {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.587256] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] workarounds.unified_limits_count_pcpu_as_vcpu = False {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.587425] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] workarounds.wait_for_vif_plugged_event_during_hard_reboot = [] {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.587614] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] wsgi.api_paste_config = /etc/nova/api-paste.ini {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.587788] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] wsgi.client_socket_timeout = 900 {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.587958] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] wsgi.default_pool_size = 1000 {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.588139] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] wsgi.keep_alive = True {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.588316] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] wsgi.max_header_line = 16384 {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.588503] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] wsgi.secure_proxy_ssl_header = None {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.588673] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] wsgi.ssl_ca_file = None {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.588836] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] wsgi.ssl_cert_file = None {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.589010] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] wsgi.ssl_key_file = None {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.589183] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] wsgi.tcp_keepidle = 600 {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.589367] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] wsgi.wsgi_log_format = %(client_ip)s "%(request_line)s" status: %(status_code)s len: %(body_length)s time: %(wall_seconds).7f {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.589539] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] zvm.ca_file = None {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.589703] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] zvm.cloud_connector_url = None {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.590025] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] zvm.image_tmp_path = /opt/stack/data/n-cpu-1/images {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.590206] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] zvm.reachable_timeout = 300 {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.590390] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] oslo_policy.enforce_new_defaults = True {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.590800] env[61806]: WARNING oslo_config.cfg [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] Deprecated: Option "enforce_scope" from group "oslo_policy" is deprecated for removal (This configuration was added temporarily to facilitate a smooth transition to the new RBAC. OpenStack will always enforce scope checks. This configuration option is deprecated and will be removed in the 2025.2 cycle.). Its value may be silently ignored in the future. [ 471.590988] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] oslo_policy.enforce_scope = True {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.591186] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] oslo_policy.policy_default_rule = default {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.591377] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] oslo_policy.policy_dirs = ['policy.d'] {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.591558] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] oslo_policy.policy_file = policy.yaml {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.591739] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] oslo_policy.remote_content_type = application/x-www-form-urlencoded {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.591905] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] oslo_policy.remote_ssl_ca_crt_file = None {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.592078] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] oslo_policy.remote_ssl_client_crt_file = None {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.592241] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] oslo_policy.remote_ssl_client_key_file = None {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.592409] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] oslo_policy.remote_ssl_verify_server_crt = False {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.592583] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] oslo_versionedobjects.fatal_exception_format_errors = False {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.592762] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] oslo_middleware.http_basic_auth_user_file = /etc/htpasswd {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.592940] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] profiler.connection_string = messaging:// {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.593124] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] profiler.enabled = False {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.593301] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] profiler.es_doc_type = notification {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.593467] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] profiler.es_scroll_size = 10000 {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.593641] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] profiler.es_scroll_time = 2m {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.593807] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] profiler.filter_error_trace = False {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.593978] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] profiler.hmac_keys = **** {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.594163] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] profiler.sentinel_service_name = mymaster {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.594335] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] profiler.socket_timeout = 0.1 {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.594502] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] profiler.trace_requests = False {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.594666] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] profiler.trace_sqlalchemy = False {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.594846] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] profiler_jaeger.process_tags = {} {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.595014] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] profiler_jaeger.service_name_prefix = None {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.595188] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] profiler_otlp.service_name_prefix = None {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.595358] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] remote_debug.host = None {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.595523] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] remote_debug.port = None {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.595706] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] oslo_messaging_rabbit.amqp_auto_delete = False {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.595885] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] oslo_messaging_rabbit.amqp_durable_queues = False {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.596044] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] oslo_messaging_rabbit.conn_pool_min_size = 2 {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.596218] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] oslo_messaging_rabbit.conn_pool_ttl = 1200 {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.596401] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] oslo_messaging_rabbit.direct_mandatory_flag = True {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.596569] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] oslo_messaging_rabbit.enable_cancel_on_failover = False {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.596734] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] oslo_messaging_rabbit.heartbeat_in_pthread = False {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.596899] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] oslo_messaging_rabbit.heartbeat_rate = 3 {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.597076] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] oslo_messaging_rabbit.heartbeat_timeout_threshold = 60 {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.597253] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] oslo_messaging_rabbit.hostname = devstack {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.597416] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] oslo_messaging_rabbit.kombu_compression = None {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.597592] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] oslo_messaging_rabbit.kombu_failover_strategy = round-robin {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.597762] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] oslo_messaging_rabbit.kombu_missing_consumer_retry_timeout = 60 {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.597934] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] oslo_messaging_rabbit.kombu_reconnect_delay = 1.0 {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.598119] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] oslo_messaging_rabbit.processname = nova-compute {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.598291] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] oslo_messaging_rabbit.rabbit_ha_queues = False {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.598504] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] oslo_messaging_rabbit.rabbit_interval_max = 30 {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.598712] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] oslo_messaging_rabbit.rabbit_login_method = AMQPLAIN {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.598886] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] oslo_messaging_rabbit.rabbit_qos_prefetch_count = 0 {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.599066] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] oslo_messaging_rabbit.rabbit_quorum_delivery_limit = 0 {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.599240] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] oslo_messaging_rabbit.rabbit_quorum_max_memory_bytes = 0 {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.599410] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] oslo_messaging_rabbit.rabbit_quorum_max_memory_length = 0 {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.599575] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] oslo_messaging_rabbit.rabbit_quorum_queue = False {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.599745] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] oslo_messaging_rabbit.rabbit_retry_backoff = 2 {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.599913] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] oslo_messaging_rabbit.rabbit_retry_interval = 1 {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.600089] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] oslo_messaging_rabbit.rabbit_stream_fanout = False {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.600260] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] oslo_messaging_rabbit.rabbit_transient_queues_ttl = 1800 {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.600426] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] oslo_messaging_rabbit.rabbit_transient_quorum_queue = False {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.600620] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] oslo_messaging_rabbit.rpc_conn_pool_size = 30 {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.600864] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] oslo_messaging_rabbit.ssl = False {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.601073] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] oslo_messaging_rabbit.ssl_ca_file = {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.601256] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] oslo_messaging_rabbit.ssl_cert_file = {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.601424] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] oslo_messaging_rabbit.ssl_enforce_fips_mode = False {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.601617] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] oslo_messaging_rabbit.ssl_key_file = {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.601814] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] oslo_messaging_rabbit.ssl_version = {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.601985] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] oslo_messaging_rabbit.use_queue_manager = False {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.602195] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] oslo_messaging_notifications.driver = ['messagingv2'] {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.602369] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] oslo_messaging_notifications.retry = -1 {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.602563] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] oslo_messaging_notifications.topics = ['notifications'] {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.602743] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] oslo_messaging_notifications.transport_url = **** {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.602919] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] oslo_limit.auth_section = None {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.603098] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] oslo_limit.auth_type = None {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.603265] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] oslo_limit.cafile = None {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.603429] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] oslo_limit.certfile = None {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.603595] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] oslo_limit.collect_timing = False {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.603757] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] oslo_limit.connect_retries = None {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.603919] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] oslo_limit.connect_retry_delay = None {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.604091] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] oslo_limit.endpoint_id = None {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.604253] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] oslo_limit.endpoint_override = None {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.604417] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] oslo_limit.insecure = False {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.604580] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] oslo_limit.keyfile = None {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.604813] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] oslo_limit.max_version = None {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.604992] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] oslo_limit.min_version = None {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.605170] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] oslo_limit.region_name = None {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.605337] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] oslo_limit.retriable_status_codes = None {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.605500] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] oslo_limit.service_name = None {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.605662] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] oslo_limit.service_type = None {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.605825] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] oslo_limit.split_loggers = False {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.605989] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] oslo_limit.status_code_retries = None {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.606158] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] oslo_limit.status_code_retry_delay = None {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.606319] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] oslo_limit.timeout = None {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.606482] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] oslo_limit.valid_interfaces = None {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.606640] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] oslo_limit.version = None {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.606807] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] oslo_reports.file_event_handler = None {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.606972] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] oslo_reports.file_event_handler_interval = 1 {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.607148] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] oslo_reports.log_dir = None {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.607325] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] vif_plug_linux_bridge_privileged.capabilities = [12] {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.607491] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] vif_plug_linux_bridge_privileged.group = None {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.607653] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] vif_plug_linux_bridge_privileged.helper_command = None {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.607847] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] vif_plug_linux_bridge_privileged.logger_name = oslo_privsep.daemon {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.608033] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] vif_plug_linux_bridge_privileged.thread_pool_size = 8 {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.608200] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] vif_plug_linux_bridge_privileged.user = None {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.608397] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] vif_plug_ovs_privileged.capabilities = [12, 1] {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.608574] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] vif_plug_ovs_privileged.group = None {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.608738] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] vif_plug_ovs_privileged.helper_command = None {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.608908] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] vif_plug_ovs_privileged.logger_name = oslo_privsep.daemon {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.609084] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] vif_plug_ovs_privileged.thread_pool_size = 8 {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.609248] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] vif_plug_ovs_privileged.user = None {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.609423] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] os_vif_linux_bridge.flat_interface = None {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.609609] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] os_vif_linux_bridge.forward_bridge_interface = ['all'] {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.609787] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] os_vif_linux_bridge.iptables_bottom_regex = {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.609963] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] os_vif_linux_bridge.iptables_drop_action = DROP {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.610151] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] os_vif_linux_bridge.iptables_top_regex = {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.610323] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] os_vif_linux_bridge.network_device_mtu = 1500 {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.610537] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] os_vif_linux_bridge.use_ipv6 = False {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.610718] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] os_vif_linux_bridge.vlan_interface = None {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.610934] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] os_vif_ovs.default_qos_type = linux-noop {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.611129] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] os_vif_ovs.isolate_vif = False {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.611309] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] os_vif_ovs.network_device_mtu = 1500 {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.611482] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] os_vif_ovs.ovs_vsctl_timeout = 120 {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.611688] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] os_vif_ovs.ovsdb_connection = tcp:127.0.0.1:6640 {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.611881] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] os_vif_ovs.ovsdb_interface = native {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.612064] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] os_vif_ovs.per_port_bridge = False {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.612238] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] os_brick.lock_path = None {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.612408] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] os_brick.wait_mpath_device_attempts = 4 {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.612582] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] os_brick.wait_mpath_device_interval = 1 {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.612800] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] privsep_osbrick.capabilities = [21] {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.612971] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] privsep_osbrick.group = None {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.613150] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] privsep_osbrick.helper_command = None {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.613322] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] privsep_osbrick.logger_name = os_brick.privileged {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.613493] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] privsep_osbrick.thread_pool_size = 8 {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.613660] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] privsep_osbrick.user = None {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.613870] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] nova_sys_admin.capabilities = [0, 1, 2, 3, 12, 21] {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.614050] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] nova_sys_admin.group = None {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.614220] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] nova_sys_admin.helper_command = None {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.614390] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] nova_sys_admin.logger_name = oslo_privsep.daemon {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.614560] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] nova_sys_admin.thread_pool_size = 8 {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.614741] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] nova_sys_admin.user = None {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 471.614901] env[61806]: DEBUG oslo_service.service [None req-2c088f27-04b4-4a76-8384-3a8ba57d4b3a None None] ******************************************************************************** {{(pid=61806) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2830}} [ 471.615413] env[61806]: INFO nova.service [-] Starting compute node (version 0.0.1) [ 472.118858] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-8b0a403f-8b06-4ccf-9017-b2d8765fa68a None None] Getting list of instances from cluster (obj){ [ 472.118858] env[61806]: value = "domain-c8" [ 472.118858] env[61806]: _type = "ClusterComputeResource" [ 472.118858] env[61806]: } {{(pid=61806) list_instances /opt/stack/nova/nova/virt/vmwareapi/vmops.py:2122}} [ 472.120232] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ba52b048-27ff-4088-a643-231a0dcd5396 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 472.129337] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-8b0a403f-8b06-4ccf-9017-b2d8765fa68a None None] Got total of 0 instances {{(pid=61806) list_instances /opt/stack/nova/nova/virt/vmwareapi/vmops.py:2131}} [ 472.129881] env[61806]: WARNING nova.virt.vmwareapi.driver [None req-8b0a403f-8b06-4ccf-9017-b2d8765fa68a None None] The vmwareapi driver is not tested by the OpenStack project nor does it have clear maintainer(s) and thus its quality can not be ensured. It should be considered experimental and may be removed in a future release. If you are using the driver in production please let us know via the openstack-discuss mailing list. [ 472.130352] env[61806]: INFO nova.virt.node [None req-8b0a403f-8b06-4ccf-9017-b2d8765fa68a None None] Generated node identity a2858be1-fd22-4e08-979e-87ad25293407 [ 472.130577] env[61806]: INFO nova.virt.node [None req-8b0a403f-8b06-4ccf-9017-b2d8765fa68a None None] Wrote node identity a2858be1-fd22-4e08-979e-87ad25293407 to /opt/stack/data/n-cpu-1/compute_id [ 472.633247] env[61806]: WARNING nova.compute.manager [None req-8b0a403f-8b06-4ccf-9017-b2d8765fa68a None None] Compute nodes ['a2858be1-fd22-4e08-979e-87ad25293407'] for host cpu-1 were not found in the database. If this is the first time this service is starting on this host, then you can ignore this warning. [ 473.639323] env[61806]: INFO nova.compute.manager [None req-8b0a403f-8b06-4ccf-9017-b2d8765fa68a None None] Looking for unclaimed instances stuck in BUILDING status for nodes managed by this host [ 474.645616] env[61806]: WARNING nova.compute.manager [None req-8b0a403f-8b06-4ccf-9017-b2d8765fa68a None None] No compute node record found for host cpu-1. If this is the first time this service is starting on this host, then you can ignore this warning.: nova.exception_Remote.ComputeHostNotFound_Remote: Compute host cpu-1 could not be found. [ 474.645974] env[61806]: DEBUG oslo_concurrency.lockutils [None req-8b0a403f-8b06-4ccf-9017-b2d8765fa68a None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 474.646083] env[61806]: DEBUG oslo_concurrency.lockutils [None req-8b0a403f-8b06-4ccf-9017-b2d8765fa68a None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 0.000s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 474.646181] env[61806]: DEBUG oslo_concurrency.lockutils [None req-8b0a403f-8b06-4ccf-9017-b2d8765fa68a None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 474.646361] env[61806]: DEBUG nova.compute.resource_tracker [None req-8b0a403f-8b06-4ccf-9017-b2d8765fa68a None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=61806) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 474.647279] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-07695b03-794f-495f-8822-b92af6a18758 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 474.655796] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-841ebf4b-7087-428e-b33d-962f01cc7b7b {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 474.669279] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bc50a47f-e7ca-47c8-bbdd-680cd4782fa7 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 474.675616] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-38154e51-820b-4ab8-9766-807a69b57bc2 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 474.704490] env[61806]: DEBUG nova.compute.resource_tracker [None req-8b0a403f-8b06-4ccf-9017-b2d8765fa68a None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=181464MB free_disk=139GB free_vcpus=48 pci_devices=None {{(pid=61806) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 474.704626] env[61806]: DEBUG oslo_concurrency.lockutils [None req-8b0a403f-8b06-4ccf-9017-b2d8765fa68a None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 474.704811] env[61806]: DEBUG oslo_concurrency.lockutils [None req-8b0a403f-8b06-4ccf-9017-b2d8765fa68a None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 0.000s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 475.206998] env[61806]: WARNING nova.compute.resource_tracker [None req-8b0a403f-8b06-4ccf-9017-b2d8765fa68a None None] No compute node record for cpu-1:a2858be1-fd22-4e08-979e-87ad25293407: nova.exception_Remote.ComputeHostNotFound_Remote: Compute host a2858be1-fd22-4e08-979e-87ad25293407 could not be found. [ 475.710299] env[61806]: INFO nova.compute.resource_tracker [None req-8b0a403f-8b06-4ccf-9017-b2d8765fa68a None None] Compute node record created for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 with uuid: a2858be1-fd22-4e08-979e-87ad25293407 [ 477.217927] env[61806]: DEBUG nova.compute.resource_tracker [None req-8b0a403f-8b06-4ccf-9017-b2d8765fa68a None None] Total usable vcpus: 48, total allocated vcpus: 0 {{(pid=61806) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 477.218321] env[61806]: DEBUG nova.compute.resource_tracker [None req-8b0a403f-8b06-4ccf-9017-b2d8765fa68a None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=512MB phys_disk=200GB used_disk=0GB total_vcpus=48 used_vcpus=0 pci_stats=[] {{(pid=61806) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 477.371618] env[61806]: INFO nova.scheduler.client.report [None req-8b0a403f-8b06-4ccf-9017-b2d8765fa68a None None] [req-6b421a78-8bc1-4203-9858-f17f74b49d7f] Created resource provider record via placement API for resource provider with UUID a2858be1-fd22-4e08-979e-87ad25293407 and name domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28. [ 477.388890] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-14377292-57f6-40a4-abfd-81bda9eccaed {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 477.396628] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-427858d9-995d-4bc4-86ba-005a7318704d {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 477.429186] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a25fee80-6d90-4967-aaaf-db2329928278 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 477.436904] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d70869cf-a89a-44d6-a63c-4f3e2e3638cf {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 477.449650] env[61806]: DEBUG nova.compute.provider_tree [None req-8b0a403f-8b06-4ccf-9017-b2d8765fa68a None None] Updating inventory in ProviderTree for provider a2858be1-fd22-4e08-979e-87ad25293407 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 139, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61806) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 477.990643] env[61806]: DEBUG nova.scheduler.client.report [None req-8b0a403f-8b06-4ccf-9017-b2d8765fa68a None None] Updated inventory for provider a2858be1-fd22-4e08-979e-87ad25293407 with generation 0 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 139, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61806) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:971}} [ 477.990898] env[61806]: DEBUG nova.compute.provider_tree [None req-8b0a403f-8b06-4ccf-9017-b2d8765fa68a None None] Updating resource provider a2858be1-fd22-4e08-979e-87ad25293407 generation from 0 to 1 during operation: update_inventory {{(pid=61806) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 477.991058] env[61806]: DEBUG nova.compute.provider_tree [None req-8b0a403f-8b06-4ccf-9017-b2d8765fa68a None None] Updating inventory in ProviderTree for provider a2858be1-fd22-4e08-979e-87ad25293407 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 139, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61806) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 478.044978] env[61806]: DEBUG nova.compute.provider_tree [None req-8b0a403f-8b06-4ccf-9017-b2d8765fa68a None None] Updating resource provider a2858be1-fd22-4e08-979e-87ad25293407 generation from 1 to 2 during operation: update_traits {{(pid=61806) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 478.550102] env[61806]: DEBUG nova.compute.resource_tracker [None req-8b0a403f-8b06-4ccf-9017-b2d8765fa68a None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=61806) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 478.550472] env[61806]: DEBUG oslo_concurrency.lockutils [None req-8b0a403f-8b06-4ccf-9017-b2d8765fa68a None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 3.845s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 478.550523] env[61806]: DEBUG nova.service [None req-8b0a403f-8b06-4ccf-9017-b2d8765fa68a None None] Creating RPC server for service compute {{(pid=61806) start /opt/stack/nova/nova/service.py:186}} [ 478.574012] env[61806]: DEBUG nova.service [None req-8b0a403f-8b06-4ccf-9017-b2d8765fa68a None None] Join ServiceGroup membership for this service compute {{(pid=61806) start /opt/stack/nova/nova/service.py:203}} [ 478.574272] env[61806]: DEBUG nova.servicegroup.drivers.db [None req-8b0a403f-8b06-4ccf-9017-b2d8765fa68a None None] DB_Driver: join new ServiceGroup member cpu-1 to the compute group, service = {{(pid=61806) join /opt/stack/nova/nova/servicegroup/drivers/db.py:44}} [ 509.802642] env[61806]: DEBUG oslo_concurrency.lockutils [None req-e67991d2-7862-4225-a618-8e964a5a624a tempest-ServerDiagnosticsNegativeTest-847099007 tempest-ServerDiagnosticsNegativeTest-847099007-project-member] Acquiring lock "e397a02e-26c6-4eb5-a242-5aa2b29fa29d" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 509.802919] env[61806]: DEBUG oslo_concurrency.lockutils [None req-e67991d2-7862-4225-a618-8e964a5a624a tempest-ServerDiagnosticsNegativeTest-847099007 tempest-ServerDiagnosticsNegativeTest-847099007-project-member] Lock "e397a02e-26c6-4eb5-a242-5aa2b29fa29d" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 510.307303] env[61806]: DEBUG nova.compute.manager [None req-e67991d2-7862-4225-a618-8e964a5a624a tempest-ServerDiagnosticsNegativeTest-847099007 tempest-ServerDiagnosticsNegativeTest-847099007-project-member] [instance: e397a02e-26c6-4eb5-a242-5aa2b29fa29d] Starting instance... {{(pid=61806) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 510.850124] env[61806]: DEBUG oslo_concurrency.lockutils [None req-e67991d2-7862-4225-a618-8e964a5a624a tempest-ServerDiagnosticsNegativeTest-847099007 tempest-ServerDiagnosticsNegativeTest-847099007-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 510.850388] env[61806]: DEBUG oslo_concurrency.lockutils [None req-e67991d2-7862-4225-a618-8e964a5a624a tempest-ServerDiagnosticsNegativeTest-847099007 tempest-ServerDiagnosticsNegativeTest-847099007-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.001s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 510.852681] env[61806]: INFO nova.compute.claims [None req-e67991d2-7862-4225-a618-8e964a5a624a tempest-ServerDiagnosticsNegativeTest-847099007 tempest-ServerDiagnosticsNegativeTest-847099007-project-member] [instance: e397a02e-26c6-4eb5-a242-5aa2b29fa29d] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 511.907058] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-578e0e9a-b060-47c9-8130-59c0e9056781 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 511.915977] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-79bccef7-579f-440f-81a9-a0054385171a {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 511.957105] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e754cc00-98a3-460a-89d0-2d0a7b661a48 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 511.968272] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-13f72cd0-1e1b-4fd3-9b13-6f83a17b46f5 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 511.982558] env[61806]: DEBUG nova.compute.provider_tree [None req-e67991d2-7862-4225-a618-8e964a5a624a tempest-ServerDiagnosticsNegativeTest-847099007 tempest-ServerDiagnosticsNegativeTest-847099007-project-member] Inventory has not changed in ProviderTree for provider: a2858be1-fd22-4e08-979e-87ad25293407 {{(pid=61806) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 512.487369] env[61806]: DEBUG nova.scheduler.client.report [None req-e67991d2-7862-4225-a618-8e964a5a624a tempest-ServerDiagnosticsNegativeTest-847099007 tempest-ServerDiagnosticsNegativeTest-847099007-project-member] Inventory has not changed for provider a2858be1-fd22-4e08-979e-87ad25293407 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 139, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61806) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 512.994012] env[61806]: DEBUG oslo_concurrency.lockutils [None req-e67991d2-7862-4225-a618-8e964a5a624a tempest-ServerDiagnosticsNegativeTest-847099007 tempest-ServerDiagnosticsNegativeTest-847099007-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.144s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 512.994629] env[61806]: DEBUG nova.compute.manager [None req-e67991d2-7862-4225-a618-8e964a5a624a tempest-ServerDiagnosticsNegativeTest-847099007 tempest-ServerDiagnosticsNegativeTest-847099007-project-member] [instance: e397a02e-26c6-4eb5-a242-5aa2b29fa29d] Start building networks asynchronously for instance. {{(pid=61806) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 513.500524] env[61806]: DEBUG nova.compute.utils [None req-e67991d2-7862-4225-a618-8e964a5a624a tempest-ServerDiagnosticsNegativeTest-847099007 tempest-ServerDiagnosticsNegativeTest-847099007-project-member] Using /dev/sd instead of None {{(pid=61806) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 513.501940] env[61806]: DEBUG nova.compute.manager [None req-e67991d2-7862-4225-a618-8e964a5a624a tempest-ServerDiagnosticsNegativeTest-847099007 tempest-ServerDiagnosticsNegativeTest-847099007-project-member] [instance: e397a02e-26c6-4eb5-a242-5aa2b29fa29d] Allocating IP information in the background. {{(pid=61806) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 513.505377] env[61806]: DEBUG nova.network.neutron [None req-e67991d2-7862-4225-a618-8e964a5a624a tempest-ServerDiagnosticsNegativeTest-847099007 tempest-ServerDiagnosticsNegativeTest-847099007-project-member] [instance: e397a02e-26c6-4eb5-a242-5aa2b29fa29d] allocate_for_instance() {{(pid=61806) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 514.017554] env[61806]: DEBUG nova.compute.manager [None req-e67991d2-7862-4225-a618-8e964a5a624a tempest-ServerDiagnosticsNegativeTest-847099007 tempest-ServerDiagnosticsNegativeTest-847099007-project-member] [instance: e397a02e-26c6-4eb5-a242-5aa2b29fa29d] Start building block device mappings for instance. {{(pid=61806) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 514.190910] env[61806]: DEBUG nova.policy [None req-e67991d2-7862-4225-a618-8e964a5a624a tempest-ServerDiagnosticsNegativeTest-847099007 tempest-ServerDiagnosticsNegativeTest-847099007-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'aad0853147624a7abf05e440bede769c', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '1fd106fa21494b21b41bac2d826f22d7', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61806) authorize /opt/stack/nova/nova/policy.py:201}} [ 515.033595] env[61806]: DEBUG nova.compute.manager [None req-e67991d2-7862-4225-a618-8e964a5a624a tempest-ServerDiagnosticsNegativeTest-847099007 tempest-ServerDiagnosticsNegativeTest-847099007-project-member] [instance: e397a02e-26c6-4eb5-a242-5aa2b29fa29d] Start spawning the instance on the hypervisor. {{(pid=61806) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 515.084559] env[61806]: DEBUG nova.virt.hardware [None req-e67991d2-7862-4225-a618-8e964a5a624a tempest-ServerDiagnosticsNegativeTest-847099007 tempest-ServerDiagnosticsNegativeTest-847099007-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-15T18:05:10Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-15T18:04:52Z,direct_url=,disk_format='vmdk',id=ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='84ba83fce2c349a6988173c5d6fc3b07',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-15T18:04:53Z,virtual_size=,visibility=), allow threads: False {{(pid=61806) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 515.084790] env[61806]: DEBUG nova.virt.hardware [None req-e67991d2-7862-4225-a618-8e964a5a624a tempest-ServerDiagnosticsNegativeTest-847099007 tempest-ServerDiagnosticsNegativeTest-847099007-project-member] Flavor limits 0:0:0 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 515.085178] env[61806]: DEBUG nova.virt.hardware [None req-e67991d2-7862-4225-a618-8e964a5a624a tempest-ServerDiagnosticsNegativeTest-847099007 tempest-ServerDiagnosticsNegativeTest-847099007-project-member] Image limits 0:0:0 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 515.085178] env[61806]: DEBUG nova.virt.hardware [None req-e67991d2-7862-4225-a618-8e964a5a624a tempest-ServerDiagnosticsNegativeTest-847099007 tempest-ServerDiagnosticsNegativeTest-847099007-project-member] Flavor pref 0:0:0 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 515.085294] env[61806]: DEBUG nova.virt.hardware [None req-e67991d2-7862-4225-a618-8e964a5a624a tempest-ServerDiagnosticsNegativeTest-847099007 tempest-ServerDiagnosticsNegativeTest-847099007-project-member] Image pref 0:0:0 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 515.085441] env[61806]: DEBUG nova.virt.hardware [None req-e67991d2-7862-4225-a618-8e964a5a624a tempest-ServerDiagnosticsNegativeTest-847099007 tempest-ServerDiagnosticsNegativeTest-847099007-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 515.085650] env[61806]: DEBUG nova.virt.hardware [None req-e67991d2-7862-4225-a618-8e964a5a624a tempest-ServerDiagnosticsNegativeTest-847099007 tempest-ServerDiagnosticsNegativeTest-847099007-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61806) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 515.085795] env[61806]: DEBUG nova.virt.hardware [None req-e67991d2-7862-4225-a618-8e964a5a624a tempest-ServerDiagnosticsNegativeTest-847099007 tempest-ServerDiagnosticsNegativeTest-847099007-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61806) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 515.086230] env[61806]: DEBUG nova.virt.hardware [None req-e67991d2-7862-4225-a618-8e964a5a624a tempest-ServerDiagnosticsNegativeTest-847099007 tempest-ServerDiagnosticsNegativeTest-847099007-project-member] Got 1 possible topologies {{(pid=61806) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 515.086533] env[61806]: DEBUG nova.virt.hardware [None req-e67991d2-7862-4225-a618-8e964a5a624a tempest-ServerDiagnosticsNegativeTest-847099007 tempest-ServerDiagnosticsNegativeTest-847099007-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61806) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 515.088314] env[61806]: DEBUG nova.virt.hardware [None req-e67991d2-7862-4225-a618-8e964a5a624a tempest-ServerDiagnosticsNegativeTest-847099007 tempest-ServerDiagnosticsNegativeTest-847099007-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61806) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 515.088960] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c8f9487e-804a-4ccc-9109-2acbd6bb0509 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 515.099913] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-82ac68e7-1b7c-4f02-9d7e-50ff2fc9cd1a {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 515.123457] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8778989a-9037-421d-8d9c-fe2bcea0f1b7 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 515.825901] env[61806]: DEBUG nova.network.neutron [None req-e67991d2-7862-4225-a618-8e964a5a624a tempest-ServerDiagnosticsNegativeTest-847099007 tempest-ServerDiagnosticsNegativeTest-847099007-project-member] [instance: e397a02e-26c6-4eb5-a242-5aa2b29fa29d] Successfully created port: 189034d5-836a-4442-b80c-60edb704f483 {{(pid=61806) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 516.825732] env[61806]: DEBUG oslo_concurrency.lockutils [None req-92c8b578-fe59-4328-af10-ad8c2b7cd69b tempest-DeleteServersAdminTestJSON-1553667494 tempest-DeleteServersAdminTestJSON-1553667494-project-member] Acquiring lock "22ce5e12-6c69-4316-b5f2-eed57a609369" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 516.828277] env[61806]: DEBUG oslo_concurrency.lockutils [None req-92c8b578-fe59-4328-af10-ad8c2b7cd69b tempest-DeleteServersAdminTestJSON-1553667494 tempest-DeleteServersAdminTestJSON-1553667494-project-member] Lock "22ce5e12-6c69-4316-b5f2-eed57a609369" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.003s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 517.333410] env[61806]: DEBUG nova.compute.manager [None req-92c8b578-fe59-4328-af10-ad8c2b7cd69b tempest-DeleteServersAdminTestJSON-1553667494 tempest-DeleteServersAdminTestJSON-1553667494-project-member] [instance: 22ce5e12-6c69-4316-b5f2-eed57a609369] Starting instance... {{(pid=61806) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 517.861842] env[61806]: DEBUG oslo_concurrency.lockutils [None req-92c8b578-fe59-4328-af10-ad8c2b7cd69b tempest-DeleteServersAdminTestJSON-1553667494 tempest-DeleteServersAdminTestJSON-1553667494-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 517.861842] env[61806]: DEBUG oslo_concurrency.lockutils [None req-92c8b578-fe59-4328-af10-ad8c2b7cd69b tempest-DeleteServersAdminTestJSON-1553667494 tempest-DeleteServersAdminTestJSON-1553667494-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.000s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 517.864400] env[61806]: INFO nova.compute.claims [None req-92c8b578-fe59-4328-af10-ad8c2b7cd69b tempest-DeleteServersAdminTestJSON-1553667494 tempest-DeleteServersAdminTestJSON-1553667494-project-member] [instance: 22ce5e12-6c69-4316-b5f2-eed57a609369] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 518.945141] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5a6eb35f-4ebb-414f-a1f5-d7c51f8e7fbc {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 518.955472] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f281b2b7-5969-48bb-89f1-2dcf5bce960d {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 519.000363] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-608a44d2-3136-4cf0-8a45-bc5d83e53fc3 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 519.008529] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-07a369df-e7ff-45dd-bb78-240cc90d26fd {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 519.026017] env[61806]: DEBUG nova.compute.provider_tree [None req-92c8b578-fe59-4328-af10-ad8c2b7cd69b tempest-DeleteServersAdminTestJSON-1553667494 tempest-DeleteServersAdminTestJSON-1553667494-project-member] Inventory has not changed in ProviderTree for provider: a2858be1-fd22-4e08-979e-87ad25293407 {{(pid=61806) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 519.218974] env[61806]: ERROR nova.compute.manager [None req-e67991d2-7862-4225-a618-8e964a5a624a tempest-ServerDiagnosticsNegativeTest-847099007 tempest-ServerDiagnosticsNegativeTest-847099007-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 189034d5-836a-4442-b80c-60edb704f483, please check neutron logs for more information. [ 519.218974] env[61806]: ERROR nova.compute.manager Traceback (most recent call last): [ 519.218974] env[61806]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 519.218974] env[61806]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 519.218974] env[61806]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 519.218974] env[61806]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 519.218974] env[61806]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 519.218974] env[61806]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 519.218974] env[61806]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 519.218974] env[61806]: ERROR nova.compute.manager self.force_reraise() [ 519.218974] env[61806]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 519.218974] env[61806]: ERROR nova.compute.manager raise self.value [ 519.218974] env[61806]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 519.218974] env[61806]: ERROR nova.compute.manager updated_port = self._update_port( [ 519.218974] env[61806]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 519.218974] env[61806]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 519.219452] env[61806]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 519.219452] env[61806]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 519.219452] env[61806]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 189034d5-836a-4442-b80c-60edb704f483, please check neutron logs for more information. [ 519.219452] env[61806]: ERROR nova.compute.manager [ 519.219557] env[61806]: Traceback (most recent call last): [ 519.219585] env[61806]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 519.219585] env[61806]: listener.cb(fileno) [ 519.219585] env[61806]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 519.219585] env[61806]: result = function(*args, **kwargs) [ 519.219585] env[61806]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 519.219585] env[61806]: return func(*args, **kwargs) [ 519.219585] env[61806]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 519.219585] env[61806]: raise e [ 519.219585] env[61806]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 519.219585] env[61806]: nwinfo = self.network_api.allocate_for_instance( [ 519.219585] env[61806]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 519.219585] env[61806]: created_port_ids = self._update_ports_for_instance( [ 519.219585] env[61806]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 519.219585] env[61806]: with excutils.save_and_reraise_exception(): [ 519.219585] env[61806]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 519.219585] env[61806]: self.force_reraise() [ 519.219981] env[61806]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 519.219981] env[61806]: raise self.value [ 519.219981] env[61806]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 519.219981] env[61806]: updated_port = self._update_port( [ 519.219981] env[61806]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 519.219981] env[61806]: _ensure_no_port_binding_failure(port) [ 519.219981] env[61806]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 519.219981] env[61806]: raise exception.PortBindingFailed(port_id=port['id']) [ 519.219981] env[61806]: nova.exception.PortBindingFailed: Binding failed for port 189034d5-836a-4442-b80c-60edb704f483, please check neutron logs for more information. [ 519.219981] env[61806]: Removing descriptor: 14 [ 519.221270] env[61806]: ERROR nova.compute.manager [None req-e67991d2-7862-4225-a618-8e964a5a624a tempest-ServerDiagnosticsNegativeTest-847099007 tempest-ServerDiagnosticsNegativeTest-847099007-project-member] [instance: e397a02e-26c6-4eb5-a242-5aa2b29fa29d] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 189034d5-836a-4442-b80c-60edb704f483, please check neutron logs for more information. [ 519.221270] env[61806]: ERROR nova.compute.manager [instance: e397a02e-26c6-4eb5-a242-5aa2b29fa29d] Traceback (most recent call last): [ 519.221270] env[61806]: ERROR nova.compute.manager [instance: e397a02e-26c6-4eb5-a242-5aa2b29fa29d] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 519.221270] env[61806]: ERROR nova.compute.manager [instance: e397a02e-26c6-4eb5-a242-5aa2b29fa29d] yield resources [ 519.221270] env[61806]: ERROR nova.compute.manager [instance: e397a02e-26c6-4eb5-a242-5aa2b29fa29d] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 519.221270] env[61806]: ERROR nova.compute.manager [instance: e397a02e-26c6-4eb5-a242-5aa2b29fa29d] self.driver.spawn(context, instance, image_meta, [ 519.221270] env[61806]: ERROR nova.compute.manager [instance: e397a02e-26c6-4eb5-a242-5aa2b29fa29d] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 519.221270] env[61806]: ERROR nova.compute.manager [instance: e397a02e-26c6-4eb5-a242-5aa2b29fa29d] self._vmops.spawn(context, instance, image_meta, injected_files, [ 519.221270] env[61806]: ERROR nova.compute.manager [instance: e397a02e-26c6-4eb5-a242-5aa2b29fa29d] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 519.221270] env[61806]: ERROR nova.compute.manager [instance: e397a02e-26c6-4eb5-a242-5aa2b29fa29d] vm_ref = self.build_virtual_machine(instance, [ 519.221270] env[61806]: ERROR nova.compute.manager [instance: e397a02e-26c6-4eb5-a242-5aa2b29fa29d] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 519.221605] env[61806]: ERROR nova.compute.manager [instance: e397a02e-26c6-4eb5-a242-5aa2b29fa29d] vif_infos = vmwarevif.get_vif_info(self._session, [ 519.221605] env[61806]: ERROR nova.compute.manager [instance: e397a02e-26c6-4eb5-a242-5aa2b29fa29d] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 519.221605] env[61806]: ERROR nova.compute.manager [instance: e397a02e-26c6-4eb5-a242-5aa2b29fa29d] for vif in network_info: [ 519.221605] env[61806]: ERROR nova.compute.manager [instance: e397a02e-26c6-4eb5-a242-5aa2b29fa29d] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 519.221605] env[61806]: ERROR nova.compute.manager [instance: e397a02e-26c6-4eb5-a242-5aa2b29fa29d] return self._sync_wrapper(fn, *args, **kwargs) [ 519.221605] env[61806]: ERROR nova.compute.manager [instance: e397a02e-26c6-4eb5-a242-5aa2b29fa29d] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 519.221605] env[61806]: ERROR nova.compute.manager [instance: e397a02e-26c6-4eb5-a242-5aa2b29fa29d] self.wait() [ 519.221605] env[61806]: ERROR nova.compute.manager [instance: e397a02e-26c6-4eb5-a242-5aa2b29fa29d] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 519.221605] env[61806]: ERROR nova.compute.manager [instance: e397a02e-26c6-4eb5-a242-5aa2b29fa29d] self[:] = self._gt.wait() [ 519.221605] env[61806]: ERROR nova.compute.manager [instance: e397a02e-26c6-4eb5-a242-5aa2b29fa29d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 519.221605] env[61806]: ERROR nova.compute.manager [instance: e397a02e-26c6-4eb5-a242-5aa2b29fa29d] return self._exit_event.wait() [ 519.221605] env[61806]: ERROR nova.compute.manager [instance: e397a02e-26c6-4eb5-a242-5aa2b29fa29d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 519.221605] env[61806]: ERROR nova.compute.manager [instance: e397a02e-26c6-4eb5-a242-5aa2b29fa29d] result = hub.switch() [ 519.221913] env[61806]: ERROR nova.compute.manager [instance: e397a02e-26c6-4eb5-a242-5aa2b29fa29d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 519.221913] env[61806]: ERROR nova.compute.manager [instance: e397a02e-26c6-4eb5-a242-5aa2b29fa29d] return self.greenlet.switch() [ 519.221913] env[61806]: ERROR nova.compute.manager [instance: e397a02e-26c6-4eb5-a242-5aa2b29fa29d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 519.221913] env[61806]: ERROR nova.compute.manager [instance: e397a02e-26c6-4eb5-a242-5aa2b29fa29d] result = function(*args, **kwargs) [ 519.221913] env[61806]: ERROR nova.compute.manager [instance: e397a02e-26c6-4eb5-a242-5aa2b29fa29d] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 519.221913] env[61806]: ERROR nova.compute.manager [instance: e397a02e-26c6-4eb5-a242-5aa2b29fa29d] return func(*args, **kwargs) [ 519.221913] env[61806]: ERROR nova.compute.manager [instance: e397a02e-26c6-4eb5-a242-5aa2b29fa29d] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 519.221913] env[61806]: ERROR nova.compute.manager [instance: e397a02e-26c6-4eb5-a242-5aa2b29fa29d] raise e [ 519.221913] env[61806]: ERROR nova.compute.manager [instance: e397a02e-26c6-4eb5-a242-5aa2b29fa29d] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 519.221913] env[61806]: ERROR nova.compute.manager [instance: e397a02e-26c6-4eb5-a242-5aa2b29fa29d] nwinfo = self.network_api.allocate_for_instance( [ 519.221913] env[61806]: ERROR nova.compute.manager [instance: e397a02e-26c6-4eb5-a242-5aa2b29fa29d] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 519.221913] env[61806]: ERROR nova.compute.manager [instance: e397a02e-26c6-4eb5-a242-5aa2b29fa29d] created_port_ids = self._update_ports_for_instance( [ 519.221913] env[61806]: ERROR nova.compute.manager [instance: e397a02e-26c6-4eb5-a242-5aa2b29fa29d] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 519.222238] env[61806]: ERROR nova.compute.manager [instance: e397a02e-26c6-4eb5-a242-5aa2b29fa29d] with excutils.save_and_reraise_exception(): [ 519.222238] env[61806]: ERROR nova.compute.manager [instance: e397a02e-26c6-4eb5-a242-5aa2b29fa29d] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 519.222238] env[61806]: ERROR nova.compute.manager [instance: e397a02e-26c6-4eb5-a242-5aa2b29fa29d] self.force_reraise() [ 519.222238] env[61806]: ERROR nova.compute.manager [instance: e397a02e-26c6-4eb5-a242-5aa2b29fa29d] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 519.222238] env[61806]: ERROR nova.compute.manager [instance: e397a02e-26c6-4eb5-a242-5aa2b29fa29d] raise self.value [ 519.222238] env[61806]: ERROR nova.compute.manager [instance: e397a02e-26c6-4eb5-a242-5aa2b29fa29d] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 519.222238] env[61806]: ERROR nova.compute.manager [instance: e397a02e-26c6-4eb5-a242-5aa2b29fa29d] updated_port = self._update_port( [ 519.222238] env[61806]: ERROR nova.compute.manager [instance: e397a02e-26c6-4eb5-a242-5aa2b29fa29d] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 519.222238] env[61806]: ERROR nova.compute.manager [instance: e397a02e-26c6-4eb5-a242-5aa2b29fa29d] _ensure_no_port_binding_failure(port) [ 519.222238] env[61806]: ERROR nova.compute.manager [instance: e397a02e-26c6-4eb5-a242-5aa2b29fa29d] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 519.222238] env[61806]: ERROR nova.compute.manager [instance: e397a02e-26c6-4eb5-a242-5aa2b29fa29d] raise exception.PortBindingFailed(port_id=port['id']) [ 519.222238] env[61806]: ERROR nova.compute.manager [instance: e397a02e-26c6-4eb5-a242-5aa2b29fa29d] nova.exception.PortBindingFailed: Binding failed for port 189034d5-836a-4442-b80c-60edb704f483, please check neutron logs for more information. [ 519.222238] env[61806]: ERROR nova.compute.manager [instance: e397a02e-26c6-4eb5-a242-5aa2b29fa29d] [ 519.222559] env[61806]: INFO nova.compute.manager [None req-e67991d2-7862-4225-a618-8e964a5a624a tempest-ServerDiagnosticsNegativeTest-847099007 tempest-ServerDiagnosticsNegativeTest-847099007-project-member] [instance: e397a02e-26c6-4eb5-a242-5aa2b29fa29d] Terminating instance [ 519.224780] env[61806]: DEBUG oslo_concurrency.lockutils [None req-e67991d2-7862-4225-a618-8e964a5a624a tempest-ServerDiagnosticsNegativeTest-847099007 tempest-ServerDiagnosticsNegativeTest-847099007-project-member] Acquiring lock "refresh_cache-e397a02e-26c6-4eb5-a242-5aa2b29fa29d" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 519.225114] env[61806]: DEBUG oslo_concurrency.lockutils [None req-e67991d2-7862-4225-a618-8e964a5a624a tempest-ServerDiagnosticsNegativeTest-847099007 tempest-ServerDiagnosticsNegativeTest-847099007-project-member] Acquired lock "refresh_cache-e397a02e-26c6-4eb5-a242-5aa2b29fa29d" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 519.225359] env[61806]: DEBUG nova.network.neutron [None req-e67991d2-7862-4225-a618-8e964a5a624a tempest-ServerDiagnosticsNegativeTest-847099007 tempest-ServerDiagnosticsNegativeTest-847099007-project-member] [instance: e397a02e-26c6-4eb5-a242-5aa2b29fa29d] Building network info cache for instance {{(pid=61806) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 519.529023] env[61806]: DEBUG nova.scheduler.client.report [None req-92c8b578-fe59-4328-af10-ad8c2b7cd69b tempest-DeleteServersAdminTestJSON-1553667494 tempest-DeleteServersAdminTestJSON-1553667494-project-member] Inventory has not changed for provider a2858be1-fd22-4e08-979e-87ad25293407 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 139, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61806) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 519.804013] env[61806]: DEBUG nova.network.neutron [None req-e67991d2-7862-4225-a618-8e964a5a624a tempest-ServerDiagnosticsNegativeTest-847099007 tempest-ServerDiagnosticsNegativeTest-847099007-project-member] [instance: e397a02e-26c6-4eb5-a242-5aa2b29fa29d] Instance cache missing network info. {{(pid=61806) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 519.884484] env[61806]: DEBUG oslo_concurrency.lockutils [None req-3568a2a1-f64b-42a9-bf53-df7d9370887d tempest-ImagesTestJSON-618852556 tempest-ImagesTestJSON-618852556-project-member] Acquiring lock "81538b78-9943-475d-aad3-f31607f2e493" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 519.884484] env[61806]: DEBUG oslo_concurrency.lockutils [None req-3568a2a1-f64b-42a9-bf53-df7d9370887d tempest-ImagesTestJSON-618852556 tempest-ImagesTestJSON-618852556-project-member] Lock "81538b78-9943-475d-aad3-f31607f2e493" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 520.038056] env[61806]: DEBUG oslo_concurrency.lockutils [None req-92c8b578-fe59-4328-af10-ad8c2b7cd69b tempest-DeleteServersAdminTestJSON-1553667494 tempest-DeleteServersAdminTestJSON-1553667494-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.174s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 520.038056] env[61806]: DEBUG nova.compute.manager [None req-92c8b578-fe59-4328-af10-ad8c2b7cd69b tempest-DeleteServersAdminTestJSON-1553667494 tempest-DeleteServersAdminTestJSON-1553667494-project-member] [instance: 22ce5e12-6c69-4316-b5f2-eed57a609369] Start building networks asynchronously for instance. {{(pid=61806) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 520.066826] env[61806]: DEBUG nova.network.neutron [None req-e67991d2-7862-4225-a618-8e964a5a624a tempest-ServerDiagnosticsNegativeTest-847099007 tempest-ServerDiagnosticsNegativeTest-847099007-project-member] [instance: e397a02e-26c6-4eb5-a242-5aa2b29fa29d] Updating instance_info_cache with network_info: [] {{(pid=61806) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 520.072823] env[61806]: DEBUG oslo_concurrency.lockutils [None req-2150b9c2-a011-40b9-950a-cddb0d2b8fb6 tempest-ImagesNegativeTestJSON-436875204 tempest-ImagesNegativeTestJSON-436875204-project-member] Acquiring lock "71380a36-8926-45d3-a70e-a7112bd1d06f" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 520.073067] env[61806]: DEBUG oslo_concurrency.lockutils [None req-2150b9c2-a011-40b9-950a-cddb0d2b8fb6 tempest-ImagesNegativeTestJSON-436875204 tempest-ImagesNegativeTestJSON-436875204-project-member] Lock "71380a36-8926-45d3-a70e-a7112bd1d06f" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 520.384986] env[61806]: DEBUG nova.compute.manager [None req-3568a2a1-f64b-42a9-bf53-df7d9370887d tempest-ImagesTestJSON-618852556 tempest-ImagesTestJSON-618852556-project-member] [instance: 81538b78-9943-475d-aad3-f31607f2e493] Starting instance... {{(pid=61806) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 520.543424] env[61806]: DEBUG nova.compute.utils [None req-92c8b578-fe59-4328-af10-ad8c2b7cd69b tempest-DeleteServersAdminTestJSON-1553667494 tempest-DeleteServersAdminTestJSON-1553667494-project-member] Using /dev/sd instead of None {{(pid=61806) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 520.544903] env[61806]: DEBUG nova.compute.manager [None req-92c8b578-fe59-4328-af10-ad8c2b7cd69b tempest-DeleteServersAdminTestJSON-1553667494 tempest-DeleteServersAdminTestJSON-1553667494-project-member] [instance: 22ce5e12-6c69-4316-b5f2-eed57a609369] Allocating IP information in the background. {{(pid=61806) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 520.545245] env[61806]: DEBUG nova.network.neutron [None req-92c8b578-fe59-4328-af10-ad8c2b7cd69b tempest-DeleteServersAdminTestJSON-1553667494 tempest-DeleteServersAdminTestJSON-1553667494-project-member] [instance: 22ce5e12-6c69-4316-b5f2-eed57a609369] allocate_for_instance() {{(pid=61806) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 520.577022] env[61806]: DEBUG oslo_concurrency.lockutils [None req-e67991d2-7862-4225-a618-8e964a5a624a tempest-ServerDiagnosticsNegativeTest-847099007 tempest-ServerDiagnosticsNegativeTest-847099007-project-member] Releasing lock "refresh_cache-e397a02e-26c6-4eb5-a242-5aa2b29fa29d" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 520.577022] env[61806]: DEBUG nova.compute.manager [None req-e67991d2-7862-4225-a618-8e964a5a624a tempest-ServerDiagnosticsNegativeTest-847099007 tempest-ServerDiagnosticsNegativeTest-847099007-project-member] [instance: e397a02e-26c6-4eb5-a242-5aa2b29fa29d] Start destroying the instance on the hypervisor. {{(pid=61806) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 520.577022] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-e67991d2-7862-4225-a618-8e964a5a624a tempest-ServerDiagnosticsNegativeTest-847099007 tempest-ServerDiagnosticsNegativeTest-847099007-project-member] [instance: e397a02e-26c6-4eb5-a242-5aa2b29fa29d] Destroying instance {{(pid=61806) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 520.577022] env[61806]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-d164f64f-e1bb-4241-9d62-2ae9f71d48df {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 520.578884] env[61806]: DEBUG oslo_service.periodic_task [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Running periodic task ComputeManager._sync_power_states {{(pid=61806) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 520.580837] env[61806]: DEBUG nova.compute.manager [None req-2150b9c2-a011-40b9-950a-cddb0d2b8fb6 tempest-ImagesNegativeTestJSON-436875204 tempest-ImagesNegativeTestJSON-436875204-project-member] [instance: 71380a36-8926-45d3-a70e-a7112bd1d06f] Starting instance... {{(pid=61806) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 520.596394] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8e9e915b-323a-4abb-8964-0b29b8fd7bc2 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 520.622776] env[61806]: WARNING nova.virt.vmwareapi.vmops [None req-e67991d2-7862-4225-a618-8e964a5a624a tempest-ServerDiagnosticsNegativeTest-847099007 tempest-ServerDiagnosticsNegativeTest-847099007-project-member] [instance: e397a02e-26c6-4eb5-a242-5aa2b29fa29d] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance e397a02e-26c6-4eb5-a242-5aa2b29fa29d could not be found. [ 520.623013] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-e67991d2-7862-4225-a618-8e964a5a624a tempest-ServerDiagnosticsNegativeTest-847099007 tempest-ServerDiagnosticsNegativeTest-847099007-project-member] [instance: e397a02e-26c6-4eb5-a242-5aa2b29fa29d] Instance destroyed {{(pid=61806) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 520.623417] env[61806]: INFO nova.compute.manager [None req-e67991d2-7862-4225-a618-8e964a5a624a tempest-ServerDiagnosticsNegativeTest-847099007 tempest-ServerDiagnosticsNegativeTest-847099007-project-member] [instance: e397a02e-26c6-4eb5-a242-5aa2b29fa29d] Took 0.05 seconds to destroy the instance on the hypervisor. [ 520.623669] env[61806]: DEBUG oslo.service.loopingcall [None req-e67991d2-7862-4225-a618-8e964a5a624a tempest-ServerDiagnosticsNegativeTest-847099007 tempest-ServerDiagnosticsNegativeTest-847099007-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61806) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 520.624082] env[61806]: DEBUG nova.compute.manager [-] [instance: e397a02e-26c6-4eb5-a242-5aa2b29fa29d] Deallocating network for instance {{(pid=61806) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 520.624193] env[61806]: DEBUG nova.network.neutron [-] [instance: e397a02e-26c6-4eb5-a242-5aa2b29fa29d] deallocate_for_instance() {{(pid=61806) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 520.664093] env[61806]: DEBUG nova.network.neutron [-] [instance: e397a02e-26c6-4eb5-a242-5aa2b29fa29d] Instance cache missing network info. {{(pid=61806) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 520.708989] env[61806]: DEBUG nova.policy [None req-92c8b578-fe59-4328-af10-ad8c2b7cd69b tempest-DeleteServersAdminTestJSON-1553667494 tempest-DeleteServersAdminTestJSON-1553667494-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '19ec5bc21d684c88a8004e53b9939ed1', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '45024eb0d0fe4ccdbb80b05fda894ce7', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61806) authorize /opt/stack/nova/nova/policy.py:201}} [ 520.923225] env[61806]: DEBUG oslo_concurrency.lockutils [None req-3568a2a1-f64b-42a9-bf53-df7d9370887d tempest-ImagesTestJSON-618852556 tempest-ImagesTestJSON-618852556-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 520.923225] env[61806]: DEBUG oslo_concurrency.lockutils [None req-3568a2a1-f64b-42a9-bf53-df7d9370887d tempest-ImagesTestJSON-618852556 tempest-ImagesTestJSON-618852556-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.000s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 520.924367] env[61806]: INFO nova.compute.claims [None req-3568a2a1-f64b-42a9-bf53-df7d9370887d tempest-ImagesTestJSON-618852556 tempest-ImagesTestJSON-618852556-project-member] [instance: 81538b78-9943-475d-aad3-f31607f2e493] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 521.049368] env[61806]: DEBUG nova.compute.manager [None req-92c8b578-fe59-4328-af10-ad8c2b7cd69b tempest-DeleteServersAdminTestJSON-1553667494 tempest-DeleteServersAdminTestJSON-1553667494-project-member] [instance: 22ce5e12-6c69-4316-b5f2-eed57a609369] Start building block device mappings for instance. {{(pid=61806) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 521.090092] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Getting list of instances from cluster (obj){ [ 521.090092] env[61806]: value = "domain-c8" [ 521.090092] env[61806]: _type = "ClusterComputeResource" [ 521.090092] env[61806]: } {{(pid=61806) list_instances /opt/stack/nova/nova/virt/vmwareapi/vmops.py:2122}} [ 521.091203] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-620a4eac-c63b-48a9-8a85-2021a00675be {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 521.107980] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Got total of 0 instances {{(pid=61806) list_instances /opt/stack/nova/nova/virt/vmwareapi/vmops.py:2131}} [ 521.108182] env[61806]: WARNING nova.compute.manager [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] While synchronizing instance power states, found 2 instances in the database and 0 instances on the hypervisor. [ 521.108293] env[61806]: DEBUG nova.compute.manager [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Triggering sync for uuid e397a02e-26c6-4eb5-a242-5aa2b29fa29d {{(pid=61806) _sync_power_states /opt/stack/nova/nova/compute/manager.py:10336}} [ 521.108504] env[61806]: DEBUG nova.compute.manager [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Triggering sync for uuid 22ce5e12-6c69-4316-b5f2-eed57a609369 {{(pid=61806) _sync_power_states /opt/stack/nova/nova/compute/manager.py:10336}} [ 521.109568] env[61806]: DEBUG oslo_concurrency.lockutils [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Acquiring lock "e397a02e-26c6-4eb5-a242-5aa2b29fa29d" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 521.109804] env[61806]: DEBUG oslo_concurrency.lockutils [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Acquiring lock "22ce5e12-6c69-4316-b5f2-eed57a609369" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 521.110053] env[61806]: DEBUG oslo_service.periodic_task [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Running periodic task ComputeManager._cleanup_running_deleted_instances {{(pid=61806) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 521.110408] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Getting list of instances from cluster (obj){ [ 521.110408] env[61806]: value = "domain-c8" [ 521.110408] env[61806]: _type = "ClusterComputeResource" [ 521.110408] env[61806]: } {{(pid=61806) list_instances /opt/stack/nova/nova/virt/vmwareapi/vmops.py:2122}} [ 521.111923] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1d30f54c-d4c9-43f8-81ba-1ea34e1c844d {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 521.121319] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Got total of 0 instances {{(pid=61806) list_instances /opt/stack/nova/nova/virt/vmwareapi/vmops.py:2131}} [ 521.148655] env[61806]: DEBUG oslo_concurrency.lockutils [None req-2150b9c2-a011-40b9-950a-cddb0d2b8fb6 tempest-ImagesNegativeTestJSON-436875204 tempest-ImagesNegativeTestJSON-436875204-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 521.169472] env[61806]: DEBUG nova.network.neutron [-] [instance: e397a02e-26c6-4eb5-a242-5aa2b29fa29d] Updating instance_info_cache with network_info: [] {{(pid=61806) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 521.672035] env[61806]: INFO nova.compute.manager [-] [instance: e397a02e-26c6-4eb5-a242-5aa2b29fa29d] Took 1.05 seconds to deallocate network for instance. [ 521.675204] env[61806]: DEBUG nova.compute.claims [None req-e67991d2-7862-4225-a618-8e964a5a624a tempest-ServerDiagnosticsNegativeTest-847099007 tempest-ServerDiagnosticsNegativeTest-847099007-project-member] [instance: e397a02e-26c6-4eb5-a242-5aa2b29fa29d] Aborting claim: {{(pid=61806) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 521.675382] env[61806]: DEBUG oslo_concurrency.lockutils [None req-e67991d2-7862-4225-a618-8e964a5a624a tempest-ServerDiagnosticsNegativeTest-847099007 tempest-ServerDiagnosticsNegativeTest-847099007-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 521.783740] env[61806]: DEBUG nova.compute.manager [req-637714c6-17ae-427e-b1c1-25644c845abe req-b119f738-7f7f-447b-bb7b-fa1b0bdb40c4 service nova] [instance: e397a02e-26c6-4eb5-a242-5aa2b29fa29d] Received event network-changed-189034d5-836a-4442-b80c-60edb704f483 {{(pid=61806) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 521.783931] env[61806]: DEBUG nova.compute.manager [req-637714c6-17ae-427e-b1c1-25644c845abe req-b119f738-7f7f-447b-bb7b-fa1b0bdb40c4 service nova] [instance: e397a02e-26c6-4eb5-a242-5aa2b29fa29d] Refreshing instance network info cache due to event network-changed-189034d5-836a-4442-b80c-60edb704f483. {{(pid=61806) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 521.784155] env[61806]: DEBUG oslo_concurrency.lockutils [req-637714c6-17ae-427e-b1c1-25644c845abe req-b119f738-7f7f-447b-bb7b-fa1b0bdb40c4 service nova] Acquiring lock "refresh_cache-e397a02e-26c6-4eb5-a242-5aa2b29fa29d" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 521.784293] env[61806]: DEBUG oslo_concurrency.lockutils [req-637714c6-17ae-427e-b1c1-25644c845abe req-b119f738-7f7f-447b-bb7b-fa1b0bdb40c4 service nova] Acquired lock "refresh_cache-e397a02e-26c6-4eb5-a242-5aa2b29fa29d" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 521.784448] env[61806]: DEBUG nova.network.neutron [req-637714c6-17ae-427e-b1c1-25644c845abe req-b119f738-7f7f-447b-bb7b-fa1b0bdb40c4 service nova] [instance: e397a02e-26c6-4eb5-a242-5aa2b29fa29d] Refreshing network info cache for port 189034d5-836a-4442-b80c-60edb704f483 {{(pid=61806) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 521.855705] env[61806]: DEBUG nova.network.neutron [None req-92c8b578-fe59-4328-af10-ad8c2b7cd69b tempest-DeleteServersAdminTestJSON-1553667494 tempest-DeleteServersAdminTestJSON-1553667494-project-member] [instance: 22ce5e12-6c69-4316-b5f2-eed57a609369] Successfully created port: 8fd4da63-a17c-4795-80db-6d4adad9e48c {{(pid=61806) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 522.038221] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f34a8457-c3e7-4314-a282-a66bee5f59ea {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 522.049748] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c7723bd0-d441-4dd5-810c-901fa64f8a2e {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 522.084453] env[61806]: DEBUG nova.compute.manager [None req-92c8b578-fe59-4328-af10-ad8c2b7cd69b tempest-DeleteServersAdminTestJSON-1553667494 tempest-DeleteServersAdminTestJSON-1553667494-project-member] [instance: 22ce5e12-6c69-4316-b5f2-eed57a609369] Start spawning the instance on the hypervisor. {{(pid=61806) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 522.087934] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-962c0135-d4f0-4de5-9851-03a18794722b {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 522.098830] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b29ed511-4abb-4068-94d3-7b968bfdb6e7 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 522.123077] env[61806]: DEBUG nova.compute.provider_tree [None req-3568a2a1-f64b-42a9-bf53-df7d9370887d tempest-ImagesTestJSON-618852556 tempest-ImagesTestJSON-618852556-project-member] Inventory has not changed in ProviderTree for provider: a2858be1-fd22-4e08-979e-87ad25293407 {{(pid=61806) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 522.137725] env[61806]: DEBUG nova.virt.hardware [None req-92c8b578-fe59-4328-af10-ad8c2b7cd69b tempest-DeleteServersAdminTestJSON-1553667494 tempest-DeleteServersAdminTestJSON-1553667494-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-15T18:05:10Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-15T18:04:52Z,direct_url=,disk_format='vmdk',id=ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='84ba83fce2c349a6988173c5d6fc3b07',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-15T18:04:53Z,virtual_size=,visibility=), allow threads: False {{(pid=61806) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 522.137725] env[61806]: DEBUG nova.virt.hardware [None req-92c8b578-fe59-4328-af10-ad8c2b7cd69b tempest-DeleteServersAdminTestJSON-1553667494 tempest-DeleteServersAdminTestJSON-1553667494-project-member] Flavor limits 0:0:0 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 522.137904] env[61806]: DEBUG nova.virt.hardware [None req-92c8b578-fe59-4328-af10-ad8c2b7cd69b tempest-DeleteServersAdminTestJSON-1553667494 tempest-DeleteServersAdminTestJSON-1553667494-project-member] Image limits 0:0:0 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 522.138077] env[61806]: DEBUG nova.virt.hardware [None req-92c8b578-fe59-4328-af10-ad8c2b7cd69b tempest-DeleteServersAdminTestJSON-1553667494 tempest-DeleteServersAdminTestJSON-1553667494-project-member] Flavor pref 0:0:0 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 522.138196] env[61806]: DEBUG nova.virt.hardware [None req-92c8b578-fe59-4328-af10-ad8c2b7cd69b tempest-DeleteServersAdminTestJSON-1553667494 tempest-DeleteServersAdminTestJSON-1553667494-project-member] Image pref 0:0:0 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 522.138348] env[61806]: DEBUG nova.virt.hardware [None req-92c8b578-fe59-4328-af10-ad8c2b7cd69b tempest-DeleteServersAdminTestJSON-1553667494 tempest-DeleteServersAdminTestJSON-1553667494-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 522.138646] env[61806]: DEBUG nova.virt.hardware [None req-92c8b578-fe59-4328-af10-ad8c2b7cd69b tempest-DeleteServersAdminTestJSON-1553667494 tempest-DeleteServersAdminTestJSON-1553667494-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61806) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 522.140488] env[61806]: DEBUG nova.virt.hardware [None req-92c8b578-fe59-4328-af10-ad8c2b7cd69b tempest-DeleteServersAdminTestJSON-1553667494 tempest-DeleteServersAdminTestJSON-1553667494-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61806) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 522.140488] env[61806]: DEBUG nova.virt.hardware [None req-92c8b578-fe59-4328-af10-ad8c2b7cd69b tempest-DeleteServersAdminTestJSON-1553667494 tempest-DeleteServersAdminTestJSON-1553667494-project-member] Got 1 possible topologies {{(pid=61806) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 522.140705] env[61806]: DEBUG nova.virt.hardware [None req-92c8b578-fe59-4328-af10-ad8c2b7cd69b tempest-DeleteServersAdminTestJSON-1553667494 tempest-DeleteServersAdminTestJSON-1553667494-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61806) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 522.142915] env[61806]: DEBUG nova.virt.hardware [None req-92c8b578-fe59-4328-af10-ad8c2b7cd69b tempest-DeleteServersAdminTestJSON-1553667494 tempest-DeleteServersAdminTestJSON-1553667494-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61806) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 522.143479] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8744d98f-cefc-44b4-a2d8-702cc2b71b3d {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 522.152165] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6315732a-ad11-46b7-a578-16a2faddc871 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 522.351488] env[61806]: DEBUG nova.network.neutron [req-637714c6-17ae-427e-b1c1-25644c845abe req-b119f738-7f7f-447b-bb7b-fa1b0bdb40c4 service nova] [instance: e397a02e-26c6-4eb5-a242-5aa2b29fa29d] Instance cache missing network info. {{(pid=61806) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 522.556013] env[61806]: DEBUG nova.network.neutron [req-637714c6-17ae-427e-b1c1-25644c845abe req-b119f738-7f7f-447b-bb7b-fa1b0bdb40c4 service nova] [instance: e397a02e-26c6-4eb5-a242-5aa2b29fa29d] Updating instance_info_cache with network_info: [] {{(pid=61806) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 522.631740] env[61806]: DEBUG nova.scheduler.client.report [None req-3568a2a1-f64b-42a9-bf53-df7d9370887d tempest-ImagesTestJSON-618852556 tempest-ImagesTestJSON-618852556-project-member] Inventory has not changed for provider a2858be1-fd22-4e08-979e-87ad25293407 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 139, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61806) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 523.060025] env[61806]: DEBUG oslo_concurrency.lockutils [req-637714c6-17ae-427e-b1c1-25644c845abe req-b119f738-7f7f-447b-bb7b-fa1b0bdb40c4 service nova] Releasing lock "refresh_cache-e397a02e-26c6-4eb5-a242-5aa2b29fa29d" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 523.145512] env[61806]: DEBUG oslo_concurrency.lockutils [None req-3568a2a1-f64b-42a9-bf53-df7d9370887d tempest-ImagesTestJSON-618852556 tempest-ImagesTestJSON-618852556-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.224s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 523.146132] env[61806]: DEBUG nova.compute.manager [None req-3568a2a1-f64b-42a9-bf53-df7d9370887d tempest-ImagesTestJSON-618852556 tempest-ImagesTestJSON-618852556-project-member] [instance: 81538b78-9943-475d-aad3-f31607f2e493] Start building networks asynchronously for instance. {{(pid=61806) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 523.154018] env[61806]: DEBUG oslo_concurrency.lockutils [None req-2150b9c2-a011-40b9-950a-cddb0d2b8fb6 tempest-ImagesNegativeTestJSON-436875204 tempest-ImagesNegativeTestJSON-436875204-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 2.002s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 523.154018] env[61806]: INFO nova.compute.claims [None req-2150b9c2-a011-40b9-950a-cddb0d2b8fb6 tempest-ImagesNegativeTestJSON-436875204 tempest-ImagesNegativeTestJSON-436875204-project-member] [instance: 71380a36-8926-45d3-a70e-a7112bd1d06f] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 523.658196] env[61806]: DEBUG nova.compute.utils [None req-3568a2a1-f64b-42a9-bf53-df7d9370887d tempest-ImagesTestJSON-618852556 tempest-ImagesTestJSON-618852556-project-member] Using /dev/sd instead of None {{(pid=61806) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 523.659827] env[61806]: DEBUG nova.compute.manager [None req-3568a2a1-f64b-42a9-bf53-df7d9370887d tempest-ImagesTestJSON-618852556 tempest-ImagesTestJSON-618852556-project-member] [instance: 81538b78-9943-475d-aad3-f31607f2e493] Allocating IP information in the background. {{(pid=61806) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 523.659827] env[61806]: DEBUG nova.network.neutron [None req-3568a2a1-f64b-42a9-bf53-df7d9370887d tempest-ImagesTestJSON-618852556 tempest-ImagesTestJSON-618852556-project-member] [instance: 81538b78-9943-475d-aad3-f31607f2e493] allocate_for_instance() {{(pid=61806) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 523.730857] env[61806]: DEBUG oslo_concurrency.lockutils [None req-0858196d-347c-4201-9cb8-c7f147f5d99d tempest-MigrationsAdminTest-470098049 tempest-MigrationsAdminTest-470098049-project-member] Acquiring lock "52b19347-a08f-44a7-8aad-22a3af5e291d" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 523.731306] env[61806]: DEBUG oslo_concurrency.lockutils [None req-0858196d-347c-4201-9cb8-c7f147f5d99d tempest-MigrationsAdminTest-470098049 tempest-MigrationsAdminTest-470098049-project-member] Lock "52b19347-a08f-44a7-8aad-22a3af5e291d" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 523.762979] env[61806]: DEBUG nova.policy [None req-3568a2a1-f64b-42a9-bf53-df7d9370887d tempest-ImagesTestJSON-618852556 tempest-ImagesTestJSON-618852556-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'e6fa11a9b63f4cd6b04baf3115431167', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '4b9aa91aff1d4008ac5096902b77f852', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61806) authorize /opt/stack/nova/nova/policy.py:201}} [ 524.176730] env[61806]: DEBUG nova.compute.manager [None req-3568a2a1-f64b-42a9-bf53-df7d9370887d tempest-ImagesTestJSON-618852556 tempest-ImagesTestJSON-618852556-project-member] [instance: 81538b78-9943-475d-aad3-f31607f2e493] Start building block device mappings for instance. {{(pid=61806) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 524.234651] env[61806]: DEBUG nova.compute.manager [None req-0858196d-347c-4201-9cb8-c7f147f5d99d tempest-MigrationsAdminTest-470098049 tempest-MigrationsAdminTest-470098049-project-member] [instance: 52b19347-a08f-44a7-8aad-22a3af5e291d] Starting instance... {{(pid=61806) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 524.295467] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-00a17310-4dec-445c-9935-cb9131c929db {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 524.304324] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5c587bb0-c199-45e1-be8d-910cb58995b0 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 524.341525] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-50862f6c-d042-4d0e-aa4b-1fadbd7d30e8 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 524.350332] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2910df54-29b7-49bc-9a80-7bca559b3653 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 524.366478] env[61806]: DEBUG nova.compute.provider_tree [None req-2150b9c2-a011-40b9-950a-cddb0d2b8fb6 tempest-ImagesNegativeTestJSON-436875204 tempest-ImagesNegativeTestJSON-436875204-project-member] Inventory has not changed in ProviderTree for provider: a2858be1-fd22-4e08-979e-87ad25293407 {{(pid=61806) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 524.679815] env[61806]: DEBUG nova.network.neutron [None req-3568a2a1-f64b-42a9-bf53-df7d9370887d tempest-ImagesTestJSON-618852556 tempest-ImagesTestJSON-618852556-project-member] [instance: 81538b78-9943-475d-aad3-f31607f2e493] Successfully created port: 91d5d4f4-3e0d-456f-98f8-483167092c82 {{(pid=61806) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 524.715383] env[61806]: DEBUG oslo_concurrency.lockutils [None req-3d02814b-d8ff-4079-980b-6c580616ce84 tempest-AttachInterfacesUnderV243Test-1683422373 tempest-AttachInterfacesUnderV243Test-1683422373-project-member] Acquiring lock "df2576da-dc49-4ced-9462-fa71710e2e2a" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 524.716011] env[61806]: DEBUG oslo_concurrency.lockutils [None req-3d02814b-d8ff-4079-980b-6c580616ce84 tempest-AttachInterfacesUnderV243Test-1683422373 tempest-AttachInterfacesUnderV243Test-1683422373-project-member] Lock "df2576da-dc49-4ced-9462-fa71710e2e2a" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 524.753809] env[61806]: DEBUG oslo_concurrency.lockutils [None req-0858196d-347c-4201-9cb8-c7f147f5d99d tempest-MigrationsAdminTest-470098049 tempest-MigrationsAdminTest-470098049-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 524.869501] env[61806]: DEBUG nova.scheduler.client.report [None req-2150b9c2-a011-40b9-950a-cddb0d2b8fb6 tempest-ImagesNegativeTestJSON-436875204 tempest-ImagesNegativeTestJSON-436875204-project-member] Inventory has not changed for provider a2858be1-fd22-4e08-979e-87ad25293407 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 139, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61806) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 525.025410] env[61806]: DEBUG oslo_concurrency.lockutils [None req-d8e4a121-0976-4eaa-b4c6-51deda86d081 tempest-InstanceActionsNegativeTestJSON-503825571 tempest-InstanceActionsNegativeTestJSON-503825571-project-member] Acquiring lock "a1aca638-e9c4-4e3b-9855-f92e7bd67204" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 525.025721] env[61806]: DEBUG oslo_concurrency.lockutils [None req-d8e4a121-0976-4eaa-b4c6-51deda86d081 tempest-InstanceActionsNegativeTestJSON-503825571 tempest-InstanceActionsNegativeTestJSON-503825571-project-member] Lock "a1aca638-e9c4-4e3b-9855-f92e7bd67204" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 525.124762] env[61806]: ERROR nova.compute.manager [None req-92c8b578-fe59-4328-af10-ad8c2b7cd69b tempest-DeleteServersAdminTestJSON-1553667494 tempest-DeleteServersAdminTestJSON-1553667494-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 8fd4da63-a17c-4795-80db-6d4adad9e48c, please check neutron logs for more information. [ 525.124762] env[61806]: ERROR nova.compute.manager Traceback (most recent call last): [ 525.124762] env[61806]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 525.124762] env[61806]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 525.124762] env[61806]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 525.124762] env[61806]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 525.124762] env[61806]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 525.124762] env[61806]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 525.124762] env[61806]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 525.124762] env[61806]: ERROR nova.compute.manager self.force_reraise() [ 525.124762] env[61806]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 525.124762] env[61806]: ERROR nova.compute.manager raise self.value [ 525.124762] env[61806]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 525.124762] env[61806]: ERROR nova.compute.manager updated_port = self._update_port( [ 525.124762] env[61806]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 525.124762] env[61806]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 525.125394] env[61806]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 525.125394] env[61806]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 525.125394] env[61806]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 8fd4da63-a17c-4795-80db-6d4adad9e48c, please check neutron logs for more information. [ 525.125394] env[61806]: ERROR nova.compute.manager [ 525.125394] env[61806]: Traceback (most recent call last): [ 525.125394] env[61806]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 525.125394] env[61806]: listener.cb(fileno) [ 525.125394] env[61806]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 525.125394] env[61806]: result = function(*args, **kwargs) [ 525.125394] env[61806]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 525.125394] env[61806]: return func(*args, **kwargs) [ 525.125394] env[61806]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 525.125394] env[61806]: raise e [ 525.125394] env[61806]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 525.125394] env[61806]: nwinfo = self.network_api.allocate_for_instance( [ 525.125394] env[61806]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 525.125394] env[61806]: created_port_ids = self._update_ports_for_instance( [ 525.125394] env[61806]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 525.125394] env[61806]: with excutils.save_and_reraise_exception(): [ 525.125394] env[61806]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 525.125394] env[61806]: self.force_reraise() [ 525.125394] env[61806]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 525.125394] env[61806]: raise self.value [ 525.125394] env[61806]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 525.125394] env[61806]: updated_port = self._update_port( [ 525.125394] env[61806]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 525.125394] env[61806]: _ensure_no_port_binding_failure(port) [ 525.125394] env[61806]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 525.125394] env[61806]: raise exception.PortBindingFailed(port_id=port['id']) [ 525.126268] env[61806]: nova.exception.PortBindingFailed: Binding failed for port 8fd4da63-a17c-4795-80db-6d4adad9e48c, please check neutron logs for more information. [ 525.126268] env[61806]: Removing descriptor: 14 [ 525.126268] env[61806]: ERROR nova.compute.manager [None req-92c8b578-fe59-4328-af10-ad8c2b7cd69b tempest-DeleteServersAdminTestJSON-1553667494 tempest-DeleteServersAdminTestJSON-1553667494-project-member] [instance: 22ce5e12-6c69-4316-b5f2-eed57a609369] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 8fd4da63-a17c-4795-80db-6d4adad9e48c, please check neutron logs for more information. [ 525.126268] env[61806]: ERROR nova.compute.manager [instance: 22ce5e12-6c69-4316-b5f2-eed57a609369] Traceback (most recent call last): [ 525.126268] env[61806]: ERROR nova.compute.manager [instance: 22ce5e12-6c69-4316-b5f2-eed57a609369] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 525.126268] env[61806]: ERROR nova.compute.manager [instance: 22ce5e12-6c69-4316-b5f2-eed57a609369] yield resources [ 525.126268] env[61806]: ERROR nova.compute.manager [instance: 22ce5e12-6c69-4316-b5f2-eed57a609369] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 525.126268] env[61806]: ERROR nova.compute.manager [instance: 22ce5e12-6c69-4316-b5f2-eed57a609369] self.driver.spawn(context, instance, image_meta, [ 525.126268] env[61806]: ERROR nova.compute.manager [instance: 22ce5e12-6c69-4316-b5f2-eed57a609369] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 525.126268] env[61806]: ERROR nova.compute.manager [instance: 22ce5e12-6c69-4316-b5f2-eed57a609369] self._vmops.spawn(context, instance, image_meta, injected_files, [ 525.126268] env[61806]: ERROR nova.compute.manager [instance: 22ce5e12-6c69-4316-b5f2-eed57a609369] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 525.126268] env[61806]: ERROR nova.compute.manager [instance: 22ce5e12-6c69-4316-b5f2-eed57a609369] vm_ref = self.build_virtual_machine(instance, [ 525.126670] env[61806]: ERROR nova.compute.manager [instance: 22ce5e12-6c69-4316-b5f2-eed57a609369] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 525.126670] env[61806]: ERROR nova.compute.manager [instance: 22ce5e12-6c69-4316-b5f2-eed57a609369] vif_infos = vmwarevif.get_vif_info(self._session, [ 525.126670] env[61806]: ERROR nova.compute.manager [instance: 22ce5e12-6c69-4316-b5f2-eed57a609369] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 525.126670] env[61806]: ERROR nova.compute.manager [instance: 22ce5e12-6c69-4316-b5f2-eed57a609369] for vif in network_info: [ 525.126670] env[61806]: ERROR nova.compute.manager [instance: 22ce5e12-6c69-4316-b5f2-eed57a609369] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 525.126670] env[61806]: ERROR nova.compute.manager [instance: 22ce5e12-6c69-4316-b5f2-eed57a609369] return self._sync_wrapper(fn, *args, **kwargs) [ 525.126670] env[61806]: ERROR nova.compute.manager [instance: 22ce5e12-6c69-4316-b5f2-eed57a609369] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 525.126670] env[61806]: ERROR nova.compute.manager [instance: 22ce5e12-6c69-4316-b5f2-eed57a609369] self.wait() [ 525.126670] env[61806]: ERROR nova.compute.manager [instance: 22ce5e12-6c69-4316-b5f2-eed57a609369] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 525.126670] env[61806]: ERROR nova.compute.manager [instance: 22ce5e12-6c69-4316-b5f2-eed57a609369] self[:] = self._gt.wait() [ 525.126670] env[61806]: ERROR nova.compute.manager [instance: 22ce5e12-6c69-4316-b5f2-eed57a609369] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 525.126670] env[61806]: ERROR nova.compute.manager [instance: 22ce5e12-6c69-4316-b5f2-eed57a609369] return self._exit_event.wait() [ 525.126670] env[61806]: ERROR nova.compute.manager [instance: 22ce5e12-6c69-4316-b5f2-eed57a609369] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 525.128587] env[61806]: ERROR nova.compute.manager [instance: 22ce5e12-6c69-4316-b5f2-eed57a609369] result = hub.switch() [ 525.128587] env[61806]: ERROR nova.compute.manager [instance: 22ce5e12-6c69-4316-b5f2-eed57a609369] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 525.128587] env[61806]: ERROR nova.compute.manager [instance: 22ce5e12-6c69-4316-b5f2-eed57a609369] return self.greenlet.switch() [ 525.128587] env[61806]: ERROR nova.compute.manager [instance: 22ce5e12-6c69-4316-b5f2-eed57a609369] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 525.128587] env[61806]: ERROR nova.compute.manager [instance: 22ce5e12-6c69-4316-b5f2-eed57a609369] result = function(*args, **kwargs) [ 525.128587] env[61806]: ERROR nova.compute.manager [instance: 22ce5e12-6c69-4316-b5f2-eed57a609369] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 525.128587] env[61806]: ERROR nova.compute.manager [instance: 22ce5e12-6c69-4316-b5f2-eed57a609369] return func(*args, **kwargs) [ 525.128587] env[61806]: ERROR nova.compute.manager [instance: 22ce5e12-6c69-4316-b5f2-eed57a609369] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 525.128587] env[61806]: ERROR nova.compute.manager [instance: 22ce5e12-6c69-4316-b5f2-eed57a609369] raise e [ 525.128587] env[61806]: ERROR nova.compute.manager [instance: 22ce5e12-6c69-4316-b5f2-eed57a609369] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 525.128587] env[61806]: ERROR nova.compute.manager [instance: 22ce5e12-6c69-4316-b5f2-eed57a609369] nwinfo = self.network_api.allocate_for_instance( [ 525.128587] env[61806]: ERROR nova.compute.manager [instance: 22ce5e12-6c69-4316-b5f2-eed57a609369] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 525.128587] env[61806]: ERROR nova.compute.manager [instance: 22ce5e12-6c69-4316-b5f2-eed57a609369] created_port_ids = self._update_ports_for_instance( [ 525.130242] env[61806]: ERROR nova.compute.manager [instance: 22ce5e12-6c69-4316-b5f2-eed57a609369] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 525.130242] env[61806]: ERROR nova.compute.manager [instance: 22ce5e12-6c69-4316-b5f2-eed57a609369] with excutils.save_and_reraise_exception(): [ 525.130242] env[61806]: ERROR nova.compute.manager [instance: 22ce5e12-6c69-4316-b5f2-eed57a609369] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 525.130242] env[61806]: ERROR nova.compute.manager [instance: 22ce5e12-6c69-4316-b5f2-eed57a609369] self.force_reraise() [ 525.130242] env[61806]: ERROR nova.compute.manager [instance: 22ce5e12-6c69-4316-b5f2-eed57a609369] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 525.130242] env[61806]: ERROR nova.compute.manager [instance: 22ce5e12-6c69-4316-b5f2-eed57a609369] raise self.value [ 525.130242] env[61806]: ERROR nova.compute.manager [instance: 22ce5e12-6c69-4316-b5f2-eed57a609369] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 525.130242] env[61806]: ERROR nova.compute.manager [instance: 22ce5e12-6c69-4316-b5f2-eed57a609369] updated_port = self._update_port( [ 525.130242] env[61806]: ERROR nova.compute.manager [instance: 22ce5e12-6c69-4316-b5f2-eed57a609369] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 525.130242] env[61806]: ERROR nova.compute.manager [instance: 22ce5e12-6c69-4316-b5f2-eed57a609369] _ensure_no_port_binding_failure(port) [ 525.130242] env[61806]: ERROR nova.compute.manager [instance: 22ce5e12-6c69-4316-b5f2-eed57a609369] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 525.130242] env[61806]: ERROR nova.compute.manager [instance: 22ce5e12-6c69-4316-b5f2-eed57a609369] raise exception.PortBindingFailed(port_id=port['id']) [ 525.133163] env[61806]: ERROR nova.compute.manager [instance: 22ce5e12-6c69-4316-b5f2-eed57a609369] nova.exception.PortBindingFailed: Binding failed for port 8fd4da63-a17c-4795-80db-6d4adad9e48c, please check neutron logs for more information. [ 525.133163] env[61806]: ERROR nova.compute.manager [instance: 22ce5e12-6c69-4316-b5f2-eed57a609369] [ 525.133163] env[61806]: INFO nova.compute.manager [None req-92c8b578-fe59-4328-af10-ad8c2b7cd69b tempest-DeleteServersAdminTestJSON-1553667494 tempest-DeleteServersAdminTestJSON-1553667494-project-member] [instance: 22ce5e12-6c69-4316-b5f2-eed57a609369] Terminating instance [ 525.133163] env[61806]: DEBUG oslo_concurrency.lockutils [None req-92c8b578-fe59-4328-af10-ad8c2b7cd69b tempest-DeleteServersAdminTestJSON-1553667494 tempest-DeleteServersAdminTestJSON-1553667494-project-member] Acquiring lock "refresh_cache-22ce5e12-6c69-4316-b5f2-eed57a609369" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 525.133163] env[61806]: DEBUG oslo_concurrency.lockutils [None req-92c8b578-fe59-4328-af10-ad8c2b7cd69b tempest-DeleteServersAdminTestJSON-1553667494 tempest-DeleteServersAdminTestJSON-1553667494-project-member] Acquired lock "refresh_cache-22ce5e12-6c69-4316-b5f2-eed57a609369" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 525.133163] env[61806]: DEBUG nova.network.neutron [None req-92c8b578-fe59-4328-af10-ad8c2b7cd69b tempest-DeleteServersAdminTestJSON-1553667494 tempest-DeleteServersAdminTestJSON-1553667494-project-member] [instance: 22ce5e12-6c69-4316-b5f2-eed57a609369] Building network info cache for instance {{(pid=61806) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 525.191950] env[61806]: DEBUG nova.compute.manager [None req-3568a2a1-f64b-42a9-bf53-df7d9370887d tempest-ImagesTestJSON-618852556 tempest-ImagesTestJSON-618852556-project-member] [instance: 81538b78-9943-475d-aad3-f31607f2e493] Start spawning the instance on the hypervisor. {{(pid=61806) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 525.217884] env[61806]: DEBUG nova.compute.manager [None req-3d02814b-d8ff-4079-980b-6c580616ce84 tempest-AttachInterfacesUnderV243Test-1683422373 tempest-AttachInterfacesUnderV243Test-1683422373-project-member] [instance: df2576da-dc49-4ced-9462-fa71710e2e2a] Starting instance... {{(pid=61806) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 525.237484] env[61806]: DEBUG nova.virt.hardware [None req-3568a2a1-f64b-42a9-bf53-df7d9370887d tempest-ImagesTestJSON-618852556 tempest-ImagesTestJSON-618852556-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-15T18:05:10Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-15T18:04:52Z,direct_url=,disk_format='vmdk',id=ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='84ba83fce2c349a6988173c5d6fc3b07',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-15T18:04:53Z,virtual_size=,visibility=), allow threads: False {{(pid=61806) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 525.237849] env[61806]: DEBUG nova.virt.hardware [None req-3568a2a1-f64b-42a9-bf53-df7d9370887d tempest-ImagesTestJSON-618852556 tempest-ImagesTestJSON-618852556-project-member] Flavor limits 0:0:0 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 525.240303] env[61806]: DEBUG nova.virt.hardware [None req-3568a2a1-f64b-42a9-bf53-df7d9370887d tempest-ImagesTestJSON-618852556 tempest-ImagesTestJSON-618852556-project-member] Image limits 0:0:0 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 525.240610] env[61806]: DEBUG nova.virt.hardware [None req-3568a2a1-f64b-42a9-bf53-df7d9370887d tempest-ImagesTestJSON-618852556 tempest-ImagesTestJSON-618852556-project-member] Flavor pref 0:0:0 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 525.240830] env[61806]: DEBUG nova.virt.hardware [None req-3568a2a1-f64b-42a9-bf53-df7d9370887d tempest-ImagesTestJSON-618852556 tempest-ImagesTestJSON-618852556-project-member] Image pref 0:0:0 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 525.241617] env[61806]: DEBUG nova.virt.hardware [None req-3568a2a1-f64b-42a9-bf53-df7d9370887d tempest-ImagesTestJSON-618852556 tempest-ImagesTestJSON-618852556-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 525.242261] env[61806]: DEBUG nova.virt.hardware [None req-3568a2a1-f64b-42a9-bf53-df7d9370887d tempest-ImagesTestJSON-618852556 tempest-ImagesTestJSON-618852556-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61806) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 525.242790] env[61806]: DEBUG nova.virt.hardware [None req-3568a2a1-f64b-42a9-bf53-df7d9370887d tempest-ImagesTestJSON-618852556 tempest-ImagesTestJSON-618852556-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61806) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 525.243619] env[61806]: DEBUG nova.virt.hardware [None req-3568a2a1-f64b-42a9-bf53-df7d9370887d tempest-ImagesTestJSON-618852556 tempest-ImagesTestJSON-618852556-project-member] Got 1 possible topologies {{(pid=61806) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 525.245311] env[61806]: DEBUG nova.virt.hardware [None req-3568a2a1-f64b-42a9-bf53-df7d9370887d tempest-ImagesTestJSON-618852556 tempest-ImagesTestJSON-618852556-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61806) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 525.245311] env[61806]: DEBUG nova.virt.hardware [None req-3568a2a1-f64b-42a9-bf53-df7d9370887d tempest-ImagesTestJSON-618852556 tempest-ImagesTestJSON-618852556-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61806) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 525.246821] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-34bb7d35-0e1d-4599-bdc8-26ca9563d2db {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 525.257265] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-08b71dc7-9bae-4046-9692-ba42fd3747fe {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 525.374195] env[61806]: DEBUG oslo_concurrency.lockutils [None req-2150b9c2-a011-40b9-950a-cddb0d2b8fb6 tempest-ImagesNegativeTestJSON-436875204 tempest-ImagesNegativeTestJSON-436875204-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.224s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 525.374720] env[61806]: DEBUG nova.compute.manager [None req-2150b9c2-a011-40b9-950a-cddb0d2b8fb6 tempest-ImagesNegativeTestJSON-436875204 tempest-ImagesNegativeTestJSON-436875204-project-member] [instance: 71380a36-8926-45d3-a70e-a7112bd1d06f] Start building networks asynchronously for instance. {{(pid=61806) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 525.377218] env[61806]: DEBUG oslo_concurrency.lockutils [None req-e67991d2-7862-4225-a618-8e964a5a624a tempest-ServerDiagnosticsNegativeTest-847099007 tempest-ServerDiagnosticsNegativeTest-847099007-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 3.702s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 525.529257] env[61806]: DEBUG nova.compute.manager [None req-d8e4a121-0976-4eaa-b4c6-51deda86d081 tempest-InstanceActionsNegativeTestJSON-503825571 tempest-InstanceActionsNegativeTestJSON-503825571-project-member] [instance: a1aca638-e9c4-4e3b-9855-f92e7bd67204] Starting instance... {{(pid=61806) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 525.665653] env[61806]: DEBUG nova.network.neutron [None req-92c8b578-fe59-4328-af10-ad8c2b7cd69b tempest-DeleteServersAdminTestJSON-1553667494 tempest-DeleteServersAdminTestJSON-1553667494-project-member] [instance: 22ce5e12-6c69-4316-b5f2-eed57a609369] Instance cache missing network info. {{(pid=61806) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 525.740422] env[61806]: DEBUG oslo_concurrency.lockutils [None req-3d02814b-d8ff-4079-980b-6c580616ce84 tempest-AttachInterfacesUnderV243Test-1683422373 tempest-AttachInterfacesUnderV243Test-1683422373-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 525.792400] env[61806]: DEBUG nova.network.neutron [None req-92c8b578-fe59-4328-af10-ad8c2b7cd69b tempest-DeleteServersAdminTestJSON-1553667494 tempest-DeleteServersAdminTestJSON-1553667494-project-member] [instance: 22ce5e12-6c69-4316-b5f2-eed57a609369] Updating instance_info_cache with network_info: [] {{(pid=61806) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 525.883934] env[61806]: DEBUG nova.compute.utils [None req-2150b9c2-a011-40b9-950a-cddb0d2b8fb6 tempest-ImagesNegativeTestJSON-436875204 tempest-ImagesNegativeTestJSON-436875204-project-member] Using /dev/sd instead of None {{(pid=61806) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 525.890172] env[61806]: DEBUG nova.compute.manager [None req-2150b9c2-a011-40b9-950a-cddb0d2b8fb6 tempest-ImagesNegativeTestJSON-436875204 tempest-ImagesNegativeTestJSON-436875204-project-member] [instance: 71380a36-8926-45d3-a70e-a7112bd1d06f] Allocating IP information in the background. {{(pid=61806) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 525.890172] env[61806]: DEBUG nova.network.neutron [None req-2150b9c2-a011-40b9-950a-cddb0d2b8fb6 tempest-ImagesNegativeTestJSON-436875204 tempest-ImagesNegativeTestJSON-436875204-project-member] [instance: 71380a36-8926-45d3-a70e-a7112bd1d06f] allocate_for_instance() {{(pid=61806) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 526.010671] env[61806]: DEBUG nova.policy [None req-2150b9c2-a011-40b9-950a-cddb0d2b8fb6 tempest-ImagesNegativeTestJSON-436875204 tempest-ImagesNegativeTestJSON-436875204-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '67f553cca9774acab789aa9fa0f16d5b', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '975702520a2c4628a4cc0defc0d5ce02', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61806) authorize /opt/stack/nova/nova/policy.py:201}} [ 526.037772] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9c832aff-c924-4a46-b136-f307f83ada3b {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 526.048435] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-86d0cf20-8f91-4962-98b9-b5eb3087514f {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 526.055022] env[61806]: DEBUG oslo_concurrency.lockutils [None req-d8e4a121-0976-4eaa-b4c6-51deda86d081 tempest-InstanceActionsNegativeTestJSON-503825571 tempest-InstanceActionsNegativeTestJSON-503825571-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 526.082732] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-97ac17e6-1961-48b3-a9ad-6d7f272f6fb5 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 526.092160] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1ac202fc-4dbf-4805-8dc1-4a0b428753b3 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 526.109657] env[61806]: DEBUG nova.compute.provider_tree [None req-e67991d2-7862-4225-a618-8e964a5a624a tempest-ServerDiagnosticsNegativeTest-847099007 tempest-ServerDiagnosticsNegativeTest-847099007-project-member] Inventory has not changed in ProviderTree for provider: a2858be1-fd22-4e08-979e-87ad25293407 {{(pid=61806) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 526.297577] env[61806]: DEBUG oslo_concurrency.lockutils [None req-92c8b578-fe59-4328-af10-ad8c2b7cd69b tempest-DeleteServersAdminTestJSON-1553667494 tempest-DeleteServersAdminTestJSON-1553667494-project-member] Releasing lock "refresh_cache-22ce5e12-6c69-4316-b5f2-eed57a609369" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 526.297577] env[61806]: DEBUG nova.compute.manager [None req-92c8b578-fe59-4328-af10-ad8c2b7cd69b tempest-DeleteServersAdminTestJSON-1553667494 tempest-DeleteServersAdminTestJSON-1553667494-project-member] [instance: 22ce5e12-6c69-4316-b5f2-eed57a609369] Start destroying the instance on the hypervisor. {{(pid=61806) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 526.297577] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-92c8b578-fe59-4328-af10-ad8c2b7cd69b tempest-DeleteServersAdminTestJSON-1553667494 tempest-DeleteServersAdminTestJSON-1553667494-project-member] [instance: 22ce5e12-6c69-4316-b5f2-eed57a609369] Destroying instance {{(pid=61806) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 526.298860] env[61806]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-5a5d7e19-0889-4f75-b123-deb895c2df7e {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 526.317248] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f9853f97-a81d-49ee-96de-99d3ecda2dcf {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 526.350375] env[61806]: WARNING nova.virt.vmwareapi.vmops [None req-92c8b578-fe59-4328-af10-ad8c2b7cd69b tempest-DeleteServersAdminTestJSON-1553667494 tempest-DeleteServersAdminTestJSON-1553667494-project-member] [instance: 22ce5e12-6c69-4316-b5f2-eed57a609369] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 22ce5e12-6c69-4316-b5f2-eed57a609369 could not be found. [ 526.350722] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-92c8b578-fe59-4328-af10-ad8c2b7cd69b tempest-DeleteServersAdminTestJSON-1553667494 tempest-DeleteServersAdminTestJSON-1553667494-project-member] [instance: 22ce5e12-6c69-4316-b5f2-eed57a609369] Instance destroyed {{(pid=61806) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 526.351022] env[61806]: INFO nova.compute.manager [None req-92c8b578-fe59-4328-af10-ad8c2b7cd69b tempest-DeleteServersAdminTestJSON-1553667494 tempest-DeleteServersAdminTestJSON-1553667494-project-member] [instance: 22ce5e12-6c69-4316-b5f2-eed57a609369] Took 0.05 seconds to destroy the instance on the hypervisor. [ 526.351373] env[61806]: DEBUG oslo.service.loopingcall [None req-92c8b578-fe59-4328-af10-ad8c2b7cd69b tempest-DeleteServersAdminTestJSON-1553667494 tempest-DeleteServersAdminTestJSON-1553667494-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61806) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 526.351693] env[61806]: DEBUG nova.compute.manager [-] [instance: 22ce5e12-6c69-4316-b5f2-eed57a609369] Deallocating network for instance {{(pid=61806) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 526.351892] env[61806]: DEBUG nova.network.neutron [-] [instance: 22ce5e12-6c69-4316-b5f2-eed57a609369] deallocate_for_instance() {{(pid=61806) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 526.390623] env[61806]: DEBUG nova.compute.manager [None req-2150b9c2-a011-40b9-950a-cddb0d2b8fb6 tempest-ImagesNegativeTestJSON-436875204 tempest-ImagesNegativeTestJSON-436875204-project-member] [instance: 71380a36-8926-45d3-a70e-a7112bd1d06f] Start building block device mappings for instance. {{(pid=61806) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 526.400127] env[61806]: DEBUG nova.network.neutron [-] [instance: 22ce5e12-6c69-4316-b5f2-eed57a609369] Instance cache missing network info. {{(pid=61806) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 526.480649] env[61806]: DEBUG oslo_concurrency.lockutils [None req-960d1abd-1b06-4b5a-8ddd-4b463d493a45 tempest-ListServerFiltersTestJSON-241228521 tempest-ListServerFiltersTestJSON-241228521-project-member] Acquiring lock "30e44576-dcd3-4036-9070-9032b775c90a" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 526.480784] env[61806]: DEBUG oslo_concurrency.lockutils [None req-960d1abd-1b06-4b5a-8ddd-4b463d493a45 tempest-ListServerFiltersTestJSON-241228521 tempest-ListServerFiltersTestJSON-241228521-project-member] Lock "30e44576-dcd3-4036-9070-9032b775c90a" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 526.612707] env[61806]: DEBUG nova.scheduler.client.report [None req-e67991d2-7862-4225-a618-8e964a5a624a tempest-ServerDiagnosticsNegativeTest-847099007 tempest-ServerDiagnosticsNegativeTest-847099007-project-member] Inventory has not changed for provider a2858be1-fd22-4e08-979e-87ad25293407 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 139, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61806) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 526.903778] env[61806]: DEBUG nova.network.neutron [-] [instance: 22ce5e12-6c69-4316-b5f2-eed57a609369] Updating instance_info_cache with network_info: [] {{(pid=61806) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 526.952345] env[61806]: DEBUG nova.network.neutron [None req-2150b9c2-a011-40b9-950a-cddb0d2b8fb6 tempest-ImagesNegativeTestJSON-436875204 tempest-ImagesNegativeTestJSON-436875204-project-member] [instance: 71380a36-8926-45d3-a70e-a7112bd1d06f] Successfully created port: 3f10bab7-3a39-46da-b6db-0d3525153d7b {{(pid=61806) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 526.987295] env[61806]: DEBUG nova.compute.manager [None req-960d1abd-1b06-4b5a-8ddd-4b463d493a45 tempest-ListServerFiltersTestJSON-241228521 tempest-ListServerFiltersTestJSON-241228521-project-member] [instance: 30e44576-dcd3-4036-9070-9032b775c90a] Starting instance... {{(pid=61806) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 527.035305] env[61806]: DEBUG oslo_service.periodic_task [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=61806) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 527.035305] env[61806]: DEBUG oslo_service.periodic_task [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=61806) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 527.035305] env[61806]: DEBUG nova.compute.manager [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Starting heal instance info cache {{(pid=61806) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9926}} [ 527.035407] env[61806]: DEBUG nova.compute.manager [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Rebuilding the list of instances to heal {{(pid=61806) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9930}} [ 527.123190] env[61806]: DEBUG oslo_concurrency.lockutils [None req-e67991d2-7862-4225-a618-8e964a5a624a tempest-ServerDiagnosticsNegativeTest-847099007 tempest-ServerDiagnosticsNegativeTest-847099007-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.746s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 527.124034] env[61806]: ERROR nova.compute.manager [None req-e67991d2-7862-4225-a618-8e964a5a624a tempest-ServerDiagnosticsNegativeTest-847099007 tempest-ServerDiagnosticsNegativeTest-847099007-project-member] [instance: e397a02e-26c6-4eb5-a242-5aa2b29fa29d] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 189034d5-836a-4442-b80c-60edb704f483, please check neutron logs for more information. [ 527.124034] env[61806]: ERROR nova.compute.manager [instance: e397a02e-26c6-4eb5-a242-5aa2b29fa29d] Traceback (most recent call last): [ 527.124034] env[61806]: ERROR nova.compute.manager [instance: e397a02e-26c6-4eb5-a242-5aa2b29fa29d] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 527.124034] env[61806]: ERROR nova.compute.manager [instance: e397a02e-26c6-4eb5-a242-5aa2b29fa29d] self.driver.spawn(context, instance, image_meta, [ 527.124034] env[61806]: ERROR nova.compute.manager [instance: e397a02e-26c6-4eb5-a242-5aa2b29fa29d] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 527.124034] env[61806]: ERROR nova.compute.manager [instance: e397a02e-26c6-4eb5-a242-5aa2b29fa29d] self._vmops.spawn(context, instance, image_meta, injected_files, [ 527.124034] env[61806]: ERROR nova.compute.manager [instance: e397a02e-26c6-4eb5-a242-5aa2b29fa29d] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 527.124034] env[61806]: ERROR nova.compute.manager [instance: e397a02e-26c6-4eb5-a242-5aa2b29fa29d] vm_ref = self.build_virtual_machine(instance, [ 527.124034] env[61806]: ERROR nova.compute.manager [instance: e397a02e-26c6-4eb5-a242-5aa2b29fa29d] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 527.124034] env[61806]: ERROR nova.compute.manager [instance: e397a02e-26c6-4eb5-a242-5aa2b29fa29d] vif_infos = vmwarevif.get_vif_info(self._session, [ 527.124034] env[61806]: ERROR nova.compute.manager [instance: e397a02e-26c6-4eb5-a242-5aa2b29fa29d] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 527.126759] env[61806]: ERROR nova.compute.manager [instance: e397a02e-26c6-4eb5-a242-5aa2b29fa29d] for vif in network_info: [ 527.126759] env[61806]: ERROR nova.compute.manager [instance: e397a02e-26c6-4eb5-a242-5aa2b29fa29d] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 527.126759] env[61806]: ERROR nova.compute.manager [instance: e397a02e-26c6-4eb5-a242-5aa2b29fa29d] return self._sync_wrapper(fn, *args, **kwargs) [ 527.126759] env[61806]: ERROR nova.compute.manager [instance: e397a02e-26c6-4eb5-a242-5aa2b29fa29d] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 527.126759] env[61806]: ERROR nova.compute.manager [instance: e397a02e-26c6-4eb5-a242-5aa2b29fa29d] self.wait() [ 527.126759] env[61806]: ERROR nova.compute.manager [instance: e397a02e-26c6-4eb5-a242-5aa2b29fa29d] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 527.126759] env[61806]: ERROR nova.compute.manager [instance: e397a02e-26c6-4eb5-a242-5aa2b29fa29d] self[:] = self._gt.wait() [ 527.126759] env[61806]: ERROR nova.compute.manager [instance: e397a02e-26c6-4eb5-a242-5aa2b29fa29d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 527.126759] env[61806]: ERROR nova.compute.manager [instance: e397a02e-26c6-4eb5-a242-5aa2b29fa29d] return self._exit_event.wait() [ 527.126759] env[61806]: ERROR nova.compute.manager [instance: e397a02e-26c6-4eb5-a242-5aa2b29fa29d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 527.126759] env[61806]: ERROR nova.compute.manager [instance: e397a02e-26c6-4eb5-a242-5aa2b29fa29d] result = hub.switch() [ 527.126759] env[61806]: ERROR nova.compute.manager [instance: e397a02e-26c6-4eb5-a242-5aa2b29fa29d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 527.126759] env[61806]: ERROR nova.compute.manager [instance: e397a02e-26c6-4eb5-a242-5aa2b29fa29d] return self.greenlet.switch() [ 527.127538] env[61806]: ERROR nova.compute.manager [instance: e397a02e-26c6-4eb5-a242-5aa2b29fa29d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 527.127538] env[61806]: ERROR nova.compute.manager [instance: e397a02e-26c6-4eb5-a242-5aa2b29fa29d] result = function(*args, **kwargs) [ 527.127538] env[61806]: ERROR nova.compute.manager [instance: e397a02e-26c6-4eb5-a242-5aa2b29fa29d] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 527.127538] env[61806]: ERROR nova.compute.manager [instance: e397a02e-26c6-4eb5-a242-5aa2b29fa29d] return func(*args, **kwargs) [ 527.127538] env[61806]: ERROR nova.compute.manager [instance: e397a02e-26c6-4eb5-a242-5aa2b29fa29d] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 527.127538] env[61806]: ERROR nova.compute.manager [instance: e397a02e-26c6-4eb5-a242-5aa2b29fa29d] raise e [ 527.127538] env[61806]: ERROR nova.compute.manager [instance: e397a02e-26c6-4eb5-a242-5aa2b29fa29d] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 527.127538] env[61806]: ERROR nova.compute.manager [instance: e397a02e-26c6-4eb5-a242-5aa2b29fa29d] nwinfo = self.network_api.allocate_for_instance( [ 527.127538] env[61806]: ERROR nova.compute.manager [instance: e397a02e-26c6-4eb5-a242-5aa2b29fa29d] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 527.127538] env[61806]: ERROR nova.compute.manager [instance: e397a02e-26c6-4eb5-a242-5aa2b29fa29d] created_port_ids = self._update_ports_for_instance( [ 527.127538] env[61806]: ERROR nova.compute.manager [instance: e397a02e-26c6-4eb5-a242-5aa2b29fa29d] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 527.127538] env[61806]: ERROR nova.compute.manager [instance: e397a02e-26c6-4eb5-a242-5aa2b29fa29d] with excutils.save_and_reraise_exception(): [ 527.127538] env[61806]: ERROR nova.compute.manager [instance: e397a02e-26c6-4eb5-a242-5aa2b29fa29d] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 527.127873] env[61806]: ERROR nova.compute.manager [instance: e397a02e-26c6-4eb5-a242-5aa2b29fa29d] self.force_reraise() [ 527.127873] env[61806]: ERROR nova.compute.manager [instance: e397a02e-26c6-4eb5-a242-5aa2b29fa29d] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 527.127873] env[61806]: ERROR nova.compute.manager [instance: e397a02e-26c6-4eb5-a242-5aa2b29fa29d] raise self.value [ 527.127873] env[61806]: ERROR nova.compute.manager [instance: e397a02e-26c6-4eb5-a242-5aa2b29fa29d] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 527.127873] env[61806]: ERROR nova.compute.manager [instance: e397a02e-26c6-4eb5-a242-5aa2b29fa29d] updated_port = self._update_port( [ 527.127873] env[61806]: ERROR nova.compute.manager [instance: e397a02e-26c6-4eb5-a242-5aa2b29fa29d] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 527.127873] env[61806]: ERROR nova.compute.manager [instance: e397a02e-26c6-4eb5-a242-5aa2b29fa29d] _ensure_no_port_binding_failure(port) [ 527.127873] env[61806]: ERROR nova.compute.manager [instance: e397a02e-26c6-4eb5-a242-5aa2b29fa29d] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 527.127873] env[61806]: ERROR nova.compute.manager [instance: e397a02e-26c6-4eb5-a242-5aa2b29fa29d] raise exception.PortBindingFailed(port_id=port['id']) [ 527.127873] env[61806]: ERROR nova.compute.manager [instance: e397a02e-26c6-4eb5-a242-5aa2b29fa29d] nova.exception.PortBindingFailed: Binding failed for port 189034d5-836a-4442-b80c-60edb704f483, please check neutron logs for more information. [ 527.127873] env[61806]: ERROR nova.compute.manager [instance: e397a02e-26c6-4eb5-a242-5aa2b29fa29d] [ 527.128216] env[61806]: DEBUG nova.compute.utils [None req-e67991d2-7862-4225-a618-8e964a5a624a tempest-ServerDiagnosticsNegativeTest-847099007 tempest-ServerDiagnosticsNegativeTest-847099007-project-member] [instance: e397a02e-26c6-4eb5-a242-5aa2b29fa29d] Binding failed for port 189034d5-836a-4442-b80c-60edb704f483, please check neutron logs for more information. {{(pid=61806) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 527.128216] env[61806]: DEBUG oslo_concurrency.lockutils [None req-0858196d-347c-4201-9cb8-c7f147f5d99d tempest-MigrationsAdminTest-470098049 tempest-MigrationsAdminTest-470098049-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 2.373s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 527.128216] env[61806]: INFO nova.compute.claims [None req-0858196d-347c-4201-9cb8-c7f147f5d99d tempest-MigrationsAdminTest-470098049 tempest-MigrationsAdminTest-470098049-project-member] [instance: 52b19347-a08f-44a7-8aad-22a3af5e291d] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 527.133564] env[61806]: DEBUG nova.compute.manager [None req-e67991d2-7862-4225-a618-8e964a5a624a tempest-ServerDiagnosticsNegativeTest-847099007 tempest-ServerDiagnosticsNegativeTest-847099007-project-member] [instance: e397a02e-26c6-4eb5-a242-5aa2b29fa29d] Build of instance e397a02e-26c6-4eb5-a242-5aa2b29fa29d was re-scheduled: Binding failed for port 189034d5-836a-4442-b80c-60edb704f483, please check neutron logs for more information. {{(pid=61806) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 527.134080] env[61806]: DEBUG nova.compute.manager [None req-e67991d2-7862-4225-a618-8e964a5a624a tempest-ServerDiagnosticsNegativeTest-847099007 tempest-ServerDiagnosticsNegativeTest-847099007-project-member] [instance: e397a02e-26c6-4eb5-a242-5aa2b29fa29d] Unplugging VIFs for instance {{(pid=61806) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 527.135064] env[61806]: DEBUG oslo_concurrency.lockutils [None req-e67991d2-7862-4225-a618-8e964a5a624a tempest-ServerDiagnosticsNegativeTest-847099007 tempest-ServerDiagnosticsNegativeTest-847099007-project-member] Acquiring lock "refresh_cache-e397a02e-26c6-4eb5-a242-5aa2b29fa29d" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 527.135064] env[61806]: DEBUG oslo_concurrency.lockutils [None req-e67991d2-7862-4225-a618-8e964a5a624a tempest-ServerDiagnosticsNegativeTest-847099007 tempest-ServerDiagnosticsNegativeTest-847099007-project-member] Acquired lock "refresh_cache-e397a02e-26c6-4eb5-a242-5aa2b29fa29d" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 527.135064] env[61806]: DEBUG nova.network.neutron [None req-e67991d2-7862-4225-a618-8e964a5a624a tempest-ServerDiagnosticsNegativeTest-847099007 tempest-ServerDiagnosticsNegativeTest-847099007-project-member] [instance: e397a02e-26c6-4eb5-a242-5aa2b29fa29d] Building network info cache for instance {{(pid=61806) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 527.406165] env[61806]: DEBUG nova.compute.manager [None req-2150b9c2-a011-40b9-950a-cddb0d2b8fb6 tempest-ImagesNegativeTestJSON-436875204 tempest-ImagesNegativeTestJSON-436875204-project-member] [instance: 71380a36-8926-45d3-a70e-a7112bd1d06f] Start spawning the instance on the hypervisor. {{(pid=61806) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 527.408964] env[61806]: INFO nova.compute.manager [-] [instance: 22ce5e12-6c69-4316-b5f2-eed57a609369] Took 1.06 seconds to deallocate network for instance. [ 527.412371] env[61806]: DEBUG nova.compute.claims [None req-92c8b578-fe59-4328-af10-ad8c2b7cd69b tempest-DeleteServersAdminTestJSON-1553667494 tempest-DeleteServersAdminTestJSON-1553667494-project-member] [instance: 22ce5e12-6c69-4316-b5f2-eed57a609369] Aborting claim: {{(pid=61806) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 527.412547] env[61806]: DEBUG oslo_concurrency.lockutils [None req-92c8b578-fe59-4328-af10-ad8c2b7cd69b tempest-DeleteServersAdminTestJSON-1553667494 tempest-DeleteServersAdminTestJSON-1553667494-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 527.445404] env[61806]: DEBUG nova.virt.hardware [None req-2150b9c2-a011-40b9-950a-cddb0d2b8fb6 tempest-ImagesNegativeTestJSON-436875204 tempest-ImagesNegativeTestJSON-436875204-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-15T18:05:10Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-15T18:04:52Z,direct_url=,disk_format='vmdk',id=ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='84ba83fce2c349a6988173c5d6fc3b07',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-15T18:04:53Z,virtual_size=,visibility=), allow threads: False {{(pid=61806) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 527.445559] env[61806]: DEBUG nova.virt.hardware [None req-2150b9c2-a011-40b9-950a-cddb0d2b8fb6 tempest-ImagesNegativeTestJSON-436875204 tempest-ImagesNegativeTestJSON-436875204-project-member] Flavor limits 0:0:0 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 527.445851] env[61806]: DEBUG nova.virt.hardware [None req-2150b9c2-a011-40b9-950a-cddb0d2b8fb6 tempest-ImagesNegativeTestJSON-436875204 tempest-ImagesNegativeTestJSON-436875204-project-member] Image limits 0:0:0 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 527.446268] env[61806]: DEBUG nova.virt.hardware [None req-2150b9c2-a011-40b9-950a-cddb0d2b8fb6 tempest-ImagesNegativeTestJSON-436875204 tempest-ImagesNegativeTestJSON-436875204-project-member] Flavor pref 0:0:0 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 527.446268] env[61806]: DEBUG nova.virt.hardware [None req-2150b9c2-a011-40b9-950a-cddb0d2b8fb6 tempest-ImagesNegativeTestJSON-436875204 tempest-ImagesNegativeTestJSON-436875204-project-member] Image pref 0:0:0 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 527.447390] env[61806]: DEBUG nova.virt.hardware [None req-2150b9c2-a011-40b9-950a-cddb0d2b8fb6 tempest-ImagesNegativeTestJSON-436875204 tempest-ImagesNegativeTestJSON-436875204-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 527.447390] env[61806]: DEBUG nova.virt.hardware [None req-2150b9c2-a011-40b9-950a-cddb0d2b8fb6 tempest-ImagesNegativeTestJSON-436875204 tempest-ImagesNegativeTestJSON-436875204-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61806) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 527.447390] env[61806]: DEBUG nova.virt.hardware [None req-2150b9c2-a011-40b9-950a-cddb0d2b8fb6 tempest-ImagesNegativeTestJSON-436875204 tempest-ImagesNegativeTestJSON-436875204-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61806) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 527.447390] env[61806]: DEBUG nova.virt.hardware [None req-2150b9c2-a011-40b9-950a-cddb0d2b8fb6 tempest-ImagesNegativeTestJSON-436875204 tempest-ImagesNegativeTestJSON-436875204-project-member] Got 1 possible topologies {{(pid=61806) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 527.447390] env[61806]: DEBUG nova.virt.hardware [None req-2150b9c2-a011-40b9-950a-cddb0d2b8fb6 tempest-ImagesNegativeTestJSON-436875204 tempest-ImagesNegativeTestJSON-436875204-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61806) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 527.448417] env[61806]: DEBUG nova.virt.hardware [None req-2150b9c2-a011-40b9-950a-cddb0d2b8fb6 tempest-ImagesNegativeTestJSON-436875204 tempest-ImagesNegativeTestJSON-436875204-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61806) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 527.449040] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-11bdb9f3-9d40-4e76-bb1b-87d0c47e5d5f {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 527.460858] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-29b068b4-529a-4dfd-9f33-bcb6feb47fcf {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 527.528657] env[61806]: DEBUG oslo_concurrency.lockutils [None req-960d1abd-1b06-4b5a-8ddd-4b463d493a45 tempest-ListServerFiltersTestJSON-241228521 tempest-ListServerFiltersTestJSON-241228521-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 527.541603] env[61806]: DEBUG nova.compute.manager [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] [instance: 22ce5e12-6c69-4316-b5f2-eed57a609369] Skipping network cache update for instance because it is Building. {{(pid=61806) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9939}} [ 527.541603] env[61806]: DEBUG nova.compute.manager [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] [instance: 81538b78-9943-475d-aad3-f31607f2e493] Skipping network cache update for instance because it is Building. {{(pid=61806) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9939}} [ 527.541603] env[61806]: DEBUG nova.compute.manager [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] [instance: 71380a36-8926-45d3-a70e-a7112bd1d06f] Skipping network cache update for instance because it is Building. {{(pid=61806) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9939}} [ 527.548646] env[61806]: DEBUG nova.compute.manager [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Didn't find any instances for network info cache update. {{(pid=61806) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10012}} [ 527.548924] env[61806]: DEBUG oslo_service.periodic_task [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=61806) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 527.549505] env[61806]: DEBUG oslo_service.periodic_task [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=61806) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 527.550288] env[61806]: DEBUG oslo_service.periodic_task [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=61806) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 527.550288] env[61806]: DEBUG oslo_service.periodic_task [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=61806) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 527.550288] env[61806]: DEBUG oslo_service.periodic_task [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=61806) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 527.550420] env[61806]: DEBUG oslo_service.periodic_task [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=61806) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 527.550511] env[61806]: DEBUG nova.compute.manager [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=61806) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10545}} [ 527.551541] env[61806]: DEBUG oslo_service.periodic_task [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Running periodic task ComputeManager.update_available_resource {{(pid=61806) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 527.666165] env[61806]: DEBUG nova.network.neutron [None req-e67991d2-7862-4225-a618-8e964a5a624a tempest-ServerDiagnosticsNegativeTest-847099007 tempest-ServerDiagnosticsNegativeTest-847099007-project-member] [instance: e397a02e-26c6-4eb5-a242-5aa2b29fa29d] Instance cache missing network info. {{(pid=61806) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 527.795744] env[61806]: DEBUG nova.network.neutron [None req-e67991d2-7862-4225-a618-8e964a5a624a tempest-ServerDiagnosticsNegativeTest-847099007 tempest-ServerDiagnosticsNegativeTest-847099007-project-member] [instance: e397a02e-26c6-4eb5-a242-5aa2b29fa29d] Updating instance_info_cache with network_info: [] {{(pid=61806) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 528.056525] env[61806]: DEBUG oslo_concurrency.lockutils [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 528.291100] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-588a5a2d-6825-4ca9-9202-e96afc8a0c62 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 528.301624] env[61806]: DEBUG oslo_concurrency.lockutils [None req-e67991d2-7862-4225-a618-8e964a5a624a tempest-ServerDiagnosticsNegativeTest-847099007 tempest-ServerDiagnosticsNegativeTest-847099007-project-member] Releasing lock "refresh_cache-e397a02e-26c6-4eb5-a242-5aa2b29fa29d" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 528.301951] env[61806]: DEBUG nova.compute.manager [None req-e67991d2-7862-4225-a618-8e964a5a624a tempest-ServerDiagnosticsNegativeTest-847099007 tempest-ServerDiagnosticsNegativeTest-847099007-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61806) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 528.301951] env[61806]: DEBUG nova.compute.manager [None req-e67991d2-7862-4225-a618-8e964a5a624a tempest-ServerDiagnosticsNegativeTest-847099007 tempest-ServerDiagnosticsNegativeTest-847099007-project-member] [instance: e397a02e-26c6-4eb5-a242-5aa2b29fa29d] Deallocating network for instance {{(pid=61806) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 528.302219] env[61806]: DEBUG nova.network.neutron [None req-e67991d2-7862-4225-a618-8e964a5a624a tempest-ServerDiagnosticsNegativeTest-847099007 tempest-ServerDiagnosticsNegativeTest-847099007-project-member] [instance: e397a02e-26c6-4eb5-a242-5aa2b29fa29d] deallocate_for_instance() {{(pid=61806) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 528.306488] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-53545d97-56d0-4a27-95f4-f585284412ba {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 528.352898] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0045776c-f8fa-4e98-933f-0d19510da42e {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 528.360617] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3c7dfdd3-200e-4e7a-a4f6-488c016d6186 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 528.375240] env[61806]: DEBUG nova.compute.provider_tree [None req-0858196d-347c-4201-9cb8-c7f147f5d99d tempest-MigrationsAdminTest-470098049 tempest-MigrationsAdminTest-470098049-project-member] Inventory has not changed in ProviderTree for provider: a2858be1-fd22-4e08-979e-87ad25293407 {{(pid=61806) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 528.498498] env[61806]: DEBUG nova.network.neutron [None req-e67991d2-7862-4225-a618-8e964a5a624a tempest-ServerDiagnosticsNegativeTest-847099007 tempest-ServerDiagnosticsNegativeTest-847099007-project-member] [instance: e397a02e-26c6-4eb5-a242-5aa2b29fa29d] Instance cache missing network info. {{(pid=61806) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 528.807695] env[61806]: DEBUG oslo_concurrency.lockutils [None req-7d4ae408-54ff-4248-838b-a36cbf45b0c5 tempest-ListServerFiltersTestJSON-241228521 tempest-ListServerFiltersTestJSON-241228521-project-member] Acquiring lock "9ddec22f-0fce-4d28-b97e-92b179f829f6" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 528.807964] env[61806]: DEBUG oslo_concurrency.lockutils [None req-7d4ae408-54ff-4248-838b-a36cbf45b0c5 tempest-ListServerFiltersTestJSON-241228521 tempest-ListServerFiltersTestJSON-241228521-project-member] Lock "9ddec22f-0fce-4d28-b97e-92b179f829f6" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 528.879548] env[61806]: DEBUG nova.scheduler.client.report [None req-0858196d-347c-4201-9cb8-c7f147f5d99d tempest-MigrationsAdminTest-470098049 tempest-MigrationsAdminTest-470098049-project-member] Inventory has not changed for provider a2858be1-fd22-4e08-979e-87ad25293407 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 139, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61806) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 528.964294] env[61806]: ERROR nova.compute.manager [None req-3568a2a1-f64b-42a9-bf53-df7d9370887d tempest-ImagesTestJSON-618852556 tempest-ImagesTestJSON-618852556-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 91d5d4f4-3e0d-456f-98f8-483167092c82, please check neutron logs for more information. [ 528.964294] env[61806]: ERROR nova.compute.manager Traceback (most recent call last): [ 528.964294] env[61806]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 528.964294] env[61806]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 528.964294] env[61806]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 528.964294] env[61806]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 528.964294] env[61806]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 528.964294] env[61806]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 528.964294] env[61806]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 528.964294] env[61806]: ERROR nova.compute.manager self.force_reraise() [ 528.964294] env[61806]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 528.964294] env[61806]: ERROR nova.compute.manager raise self.value [ 528.964294] env[61806]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 528.964294] env[61806]: ERROR nova.compute.manager updated_port = self._update_port( [ 528.964294] env[61806]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 528.964294] env[61806]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 528.964726] env[61806]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 528.964726] env[61806]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 528.964726] env[61806]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 91d5d4f4-3e0d-456f-98f8-483167092c82, please check neutron logs for more information. [ 528.964726] env[61806]: ERROR nova.compute.manager [ 528.964726] env[61806]: Traceback (most recent call last): [ 528.964726] env[61806]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 528.964726] env[61806]: listener.cb(fileno) [ 528.964726] env[61806]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 528.964726] env[61806]: result = function(*args, **kwargs) [ 528.964726] env[61806]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 528.964726] env[61806]: return func(*args, **kwargs) [ 528.964726] env[61806]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 528.964726] env[61806]: raise e [ 528.964726] env[61806]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 528.964726] env[61806]: nwinfo = self.network_api.allocate_for_instance( [ 528.964726] env[61806]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 528.964726] env[61806]: created_port_ids = self._update_ports_for_instance( [ 528.964726] env[61806]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 528.964726] env[61806]: with excutils.save_and_reraise_exception(): [ 528.964726] env[61806]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 528.964726] env[61806]: self.force_reraise() [ 528.964726] env[61806]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 528.964726] env[61806]: raise self.value [ 528.964726] env[61806]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 528.964726] env[61806]: updated_port = self._update_port( [ 528.964726] env[61806]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 528.964726] env[61806]: _ensure_no_port_binding_failure(port) [ 528.964726] env[61806]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 528.964726] env[61806]: raise exception.PortBindingFailed(port_id=port['id']) [ 528.965463] env[61806]: nova.exception.PortBindingFailed: Binding failed for port 91d5d4f4-3e0d-456f-98f8-483167092c82, please check neutron logs for more information. [ 528.965463] env[61806]: Removing descriptor: 16 [ 528.965463] env[61806]: ERROR nova.compute.manager [None req-3568a2a1-f64b-42a9-bf53-df7d9370887d tempest-ImagesTestJSON-618852556 tempest-ImagesTestJSON-618852556-project-member] [instance: 81538b78-9943-475d-aad3-f31607f2e493] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 91d5d4f4-3e0d-456f-98f8-483167092c82, please check neutron logs for more information. [ 528.965463] env[61806]: ERROR nova.compute.manager [instance: 81538b78-9943-475d-aad3-f31607f2e493] Traceback (most recent call last): [ 528.965463] env[61806]: ERROR nova.compute.manager [instance: 81538b78-9943-475d-aad3-f31607f2e493] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 528.965463] env[61806]: ERROR nova.compute.manager [instance: 81538b78-9943-475d-aad3-f31607f2e493] yield resources [ 528.965463] env[61806]: ERROR nova.compute.manager [instance: 81538b78-9943-475d-aad3-f31607f2e493] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 528.965463] env[61806]: ERROR nova.compute.manager [instance: 81538b78-9943-475d-aad3-f31607f2e493] self.driver.spawn(context, instance, image_meta, [ 528.965463] env[61806]: ERROR nova.compute.manager [instance: 81538b78-9943-475d-aad3-f31607f2e493] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 528.965463] env[61806]: ERROR nova.compute.manager [instance: 81538b78-9943-475d-aad3-f31607f2e493] self._vmops.spawn(context, instance, image_meta, injected_files, [ 528.965463] env[61806]: ERROR nova.compute.manager [instance: 81538b78-9943-475d-aad3-f31607f2e493] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 528.965463] env[61806]: ERROR nova.compute.manager [instance: 81538b78-9943-475d-aad3-f31607f2e493] vm_ref = self.build_virtual_machine(instance, [ 528.965780] env[61806]: ERROR nova.compute.manager [instance: 81538b78-9943-475d-aad3-f31607f2e493] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 528.965780] env[61806]: ERROR nova.compute.manager [instance: 81538b78-9943-475d-aad3-f31607f2e493] vif_infos = vmwarevif.get_vif_info(self._session, [ 528.965780] env[61806]: ERROR nova.compute.manager [instance: 81538b78-9943-475d-aad3-f31607f2e493] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 528.965780] env[61806]: ERROR nova.compute.manager [instance: 81538b78-9943-475d-aad3-f31607f2e493] for vif in network_info: [ 528.965780] env[61806]: ERROR nova.compute.manager [instance: 81538b78-9943-475d-aad3-f31607f2e493] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 528.965780] env[61806]: ERROR nova.compute.manager [instance: 81538b78-9943-475d-aad3-f31607f2e493] return self._sync_wrapper(fn, *args, **kwargs) [ 528.965780] env[61806]: ERROR nova.compute.manager [instance: 81538b78-9943-475d-aad3-f31607f2e493] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 528.965780] env[61806]: ERROR nova.compute.manager [instance: 81538b78-9943-475d-aad3-f31607f2e493] self.wait() [ 528.965780] env[61806]: ERROR nova.compute.manager [instance: 81538b78-9943-475d-aad3-f31607f2e493] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 528.965780] env[61806]: ERROR nova.compute.manager [instance: 81538b78-9943-475d-aad3-f31607f2e493] self[:] = self._gt.wait() [ 528.965780] env[61806]: ERROR nova.compute.manager [instance: 81538b78-9943-475d-aad3-f31607f2e493] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 528.965780] env[61806]: ERROR nova.compute.manager [instance: 81538b78-9943-475d-aad3-f31607f2e493] return self._exit_event.wait() [ 528.965780] env[61806]: ERROR nova.compute.manager [instance: 81538b78-9943-475d-aad3-f31607f2e493] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 528.966205] env[61806]: ERROR nova.compute.manager [instance: 81538b78-9943-475d-aad3-f31607f2e493] result = hub.switch() [ 528.966205] env[61806]: ERROR nova.compute.manager [instance: 81538b78-9943-475d-aad3-f31607f2e493] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 528.966205] env[61806]: ERROR nova.compute.manager [instance: 81538b78-9943-475d-aad3-f31607f2e493] return self.greenlet.switch() [ 528.966205] env[61806]: ERROR nova.compute.manager [instance: 81538b78-9943-475d-aad3-f31607f2e493] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 528.966205] env[61806]: ERROR nova.compute.manager [instance: 81538b78-9943-475d-aad3-f31607f2e493] result = function(*args, **kwargs) [ 528.966205] env[61806]: ERROR nova.compute.manager [instance: 81538b78-9943-475d-aad3-f31607f2e493] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 528.966205] env[61806]: ERROR nova.compute.manager [instance: 81538b78-9943-475d-aad3-f31607f2e493] return func(*args, **kwargs) [ 528.966205] env[61806]: ERROR nova.compute.manager [instance: 81538b78-9943-475d-aad3-f31607f2e493] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 528.966205] env[61806]: ERROR nova.compute.manager [instance: 81538b78-9943-475d-aad3-f31607f2e493] raise e [ 528.966205] env[61806]: ERROR nova.compute.manager [instance: 81538b78-9943-475d-aad3-f31607f2e493] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 528.966205] env[61806]: ERROR nova.compute.manager [instance: 81538b78-9943-475d-aad3-f31607f2e493] nwinfo = self.network_api.allocate_for_instance( [ 528.966205] env[61806]: ERROR nova.compute.manager [instance: 81538b78-9943-475d-aad3-f31607f2e493] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 528.966205] env[61806]: ERROR nova.compute.manager [instance: 81538b78-9943-475d-aad3-f31607f2e493] created_port_ids = self._update_ports_for_instance( [ 528.966561] env[61806]: ERROR nova.compute.manager [instance: 81538b78-9943-475d-aad3-f31607f2e493] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 528.966561] env[61806]: ERROR nova.compute.manager [instance: 81538b78-9943-475d-aad3-f31607f2e493] with excutils.save_and_reraise_exception(): [ 528.966561] env[61806]: ERROR nova.compute.manager [instance: 81538b78-9943-475d-aad3-f31607f2e493] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 528.966561] env[61806]: ERROR nova.compute.manager [instance: 81538b78-9943-475d-aad3-f31607f2e493] self.force_reraise() [ 528.966561] env[61806]: ERROR nova.compute.manager [instance: 81538b78-9943-475d-aad3-f31607f2e493] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 528.966561] env[61806]: ERROR nova.compute.manager [instance: 81538b78-9943-475d-aad3-f31607f2e493] raise self.value [ 528.966561] env[61806]: ERROR nova.compute.manager [instance: 81538b78-9943-475d-aad3-f31607f2e493] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 528.966561] env[61806]: ERROR nova.compute.manager [instance: 81538b78-9943-475d-aad3-f31607f2e493] updated_port = self._update_port( [ 528.966561] env[61806]: ERROR nova.compute.manager [instance: 81538b78-9943-475d-aad3-f31607f2e493] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 528.966561] env[61806]: ERROR nova.compute.manager [instance: 81538b78-9943-475d-aad3-f31607f2e493] _ensure_no_port_binding_failure(port) [ 528.966561] env[61806]: ERROR nova.compute.manager [instance: 81538b78-9943-475d-aad3-f31607f2e493] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 528.966561] env[61806]: ERROR nova.compute.manager [instance: 81538b78-9943-475d-aad3-f31607f2e493] raise exception.PortBindingFailed(port_id=port['id']) [ 528.968631] env[61806]: ERROR nova.compute.manager [instance: 81538b78-9943-475d-aad3-f31607f2e493] nova.exception.PortBindingFailed: Binding failed for port 91d5d4f4-3e0d-456f-98f8-483167092c82, please check neutron logs for more information. [ 528.968631] env[61806]: ERROR nova.compute.manager [instance: 81538b78-9943-475d-aad3-f31607f2e493] [ 528.968631] env[61806]: INFO nova.compute.manager [None req-3568a2a1-f64b-42a9-bf53-df7d9370887d tempest-ImagesTestJSON-618852556 tempest-ImagesTestJSON-618852556-project-member] [instance: 81538b78-9943-475d-aad3-f31607f2e493] Terminating instance [ 528.968631] env[61806]: DEBUG oslo_concurrency.lockutils [None req-3568a2a1-f64b-42a9-bf53-df7d9370887d tempest-ImagesTestJSON-618852556 tempest-ImagesTestJSON-618852556-project-member] Acquiring lock "refresh_cache-81538b78-9943-475d-aad3-f31607f2e493" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 528.969215] env[61806]: DEBUG oslo_concurrency.lockutils [None req-3568a2a1-f64b-42a9-bf53-df7d9370887d tempest-ImagesTestJSON-618852556 tempest-ImagesTestJSON-618852556-project-member] Acquired lock "refresh_cache-81538b78-9943-475d-aad3-f31607f2e493" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 528.969215] env[61806]: DEBUG nova.network.neutron [None req-3568a2a1-f64b-42a9-bf53-df7d9370887d tempest-ImagesTestJSON-618852556 tempest-ImagesTestJSON-618852556-project-member] [instance: 81538b78-9943-475d-aad3-f31607f2e493] Building network info cache for instance {{(pid=61806) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 529.001741] env[61806]: DEBUG nova.network.neutron [None req-e67991d2-7862-4225-a618-8e964a5a624a tempest-ServerDiagnosticsNegativeTest-847099007 tempest-ServerDiagnosticsNegativeTest-847099007-project-member] [instance: e397a02e-26c6-4eb5-a242-5aa2b29fa29d] Updating instance_info_cache with network_info: [] {{(pid=61806) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 529.312402] env[61806]: DEBUG nova.compute.manager [None req-7d4ae408-54ff-4248-838b-a36cbf45b0c5 tempest-ListServerFiltersTestJSON-241228521 tempest-ListServerFiltersTestJSON-241228521-project-member] [instance: 9ddec22f-0fce-4d28-b97e-92b179f829f6] Starting instance... {{(pid=61806) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 529.393060] env[61806]: DEBUG oslo_concurrency.lockutils [None req-0858196d-347c-4201-9cb8-c7f147f5d99d tempest-MigrationsAdminTest-470098049 tempest-MigrationsAdminTest-470098049-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.267s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 529.393608] env[61806]: DEBUG nova.compute.manager [None req-0858196d-347c-4201-9cb8-c7f147f5d99d tempest-MigrationsAdminTest-470098049 tempest-MigrationsAdminTest-470098049-project-member] [instance: 52b19347-a08f-44a7-8aad-22a3af5e291d] Start building networks asynchronously for instance. {{(pid=61806) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 529.397879] env[61806]: DEBUG oslo_concurrency.lockutils [None req-3d02814b-d8ff-4079-980b-6c580616ce84 tempest-AttachInterfacesUnderV243Test-1683422373 tempest-AttachInterfacesUnderV243Test-1683422373-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 3.656s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 529.397879] env[61806]: INFO nova.compute.claims [None req-3d02814b-d8ff-4079-980b-6c580616ce84 tempest-AttachInterfacesUnderV243Test-1683422373 tempest-AttachInterfacesUnderV243Test-1683422373-project-member] [instance: df2576da-dc49-4ced-9462-fa71710e2e2a] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 529.493955] env[61806]: ERROR nova.compute.manager [None req-2150b9c2-a011-40b9-950a-cddb0d2b8fb6 tempest-ImagesNegativeTestJSON-436875204 tempest-ImagesNegativeTestJSON-436875204-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 3f10bab7-3a39-46da-b6db-0d3525153d7b, please check neutron logs for more information. [ 529.493955] env[61806]: ERROR nova.compute.manager Traceback (most recent call last): [ 529.493955] env[61806]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 529.493955] env[61806]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 529.493955] env[61806]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 529.493955] env[61806]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 529.493955] env[61806]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 529.493955] env[61806]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 529.493955] env[61806]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 529.493955] env[61806]: ERROR nova.compute.manager self.force_reraise() [ 529.493955] env[61806]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 529.493955] env[61806]: ERROR nova.compute.manager raise self.value [ 529.493955] env[61806]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 529.493955] env[61806]: ERROR nova.compute.manager updated_port = self._update_port( [ 529.493955] env[61806]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 529.493955] env[61806]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 529.494427] env[61806]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 529.494427] env[61806]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 529.494427] env[61806]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 3f10bab7-3a39-46da-b6db-0d3525153d7b, please check neutron logs for more information. [ 529.494427] env[61806]: ERROR nova.compute.manager [ 529.495554] env[61806]: Traceback (most recent call last): [ 529.495621] env[61806]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 529.495621] env[61806]: listener.cb(fileno) [ 529.495621] env[61806]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 529.495621] env[61806]: result = function(*args, **kwargs) [ 529.495621] env[61806]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 529.495621] env[61806]: return func(*args, **kwargs) [ 529.495621] env[61806]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 529.495621] env[61806]: raise e [ 529.495621] env[61806]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 529.495621] env[61806]: nwinfo = self.network_api.allocate_for_instance( [ 529.495621] env[61806]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 529.495621] env[61806]: created_port_ids = self._update_ports_for_instance( [ 529.495621] env[61806]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 529.495621] env[61806]: with excutils.save_and_reraise_exception(): [ 529.495621] env[61806]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 529.495621] env[61806]: self.force_reraise() [ 529.495621] env[61806]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 529.495621] env[61806]: raise self.value [ 529.495621] env[61806]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 529.495621] env[61806]: updated_port = self._update_port( [ 529.495621] env[61806]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 529.495621] env[61806]: _ensure_no_port_binding_failure(port) [ 529.495621] env[61806]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 529.495621] env[61806]: raise exception.PortBindingFailed(port_id=port['id']) [ 529.495621] env[61806]: nova.exception.PortBindingFailed: Binding failed for port 3f10bab7-3a39-46da-b6db-0d3525153d7b, please check neutron logs for more information. [ 529.495621] env[61806]: Removing descriptor: 14 [ 529.496609] env[61806]: ERROR nova.compute.manager [None req-2150b9c2-a011-40b9-950a-cddb0d2b8fb6 tempest-ImagesNegativeTestJSON-436875204 tempest-ImagesNegativeTestJSON-436875204-project-member] [instance: 71380a36-8926-45d3-a70e-a7112bd1d06f] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 3f10bab7-3a39-46da-b6db-0d3525153d7b, please check neutron logs for more information. [ 529.496609] env[61806]: ERROR nova.compute.manager [instance: 71380a36-8926-45d3-a70e-a7112bd1d06f] Traceback (most recent call last): [ 529.496609] env[61806]: ERROR nova.compute.manager [instance: 71380a36-8926-45d3-a70e-a7112bd1d06f] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 529.496609] env[61806]: ERROR nova.compute.manager [instance: 71380a36-8926-45d3-a70e-a7112bd1d06f] yield resources [ 529.496609] env[61806]: ERROR nova.compute.manager [instance: 71380a36-8926-45d3-a70e-a7112bd1d06f] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 529.496609] env[61806]: ERROR nova.compute.manager [instance: 71380a36-8926-45d3-a70e-a7112bd1d06f] self.driver.spawn(context, instance, image_meta, [ 529.496609] env[61806]: ERROR nova.compute.manager [instance: 71380a36-8926-45d3-a70e-a7112bd1d06f] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 529.496609] env[61806]: ERROR nova.compute.manager [instance: 71380a36-8926-45d3-a70e-a7112bd1d06f] self._vmops.spawn(context, instance, image_meta, injected_files, [ 529.496609] env[61806]: ERROR nova.compute.manager [instance: 71380a36-8926-45d3-a70e-a7112bd1d06f] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 529.496609] env[61806]: ERROR nova.compute.manager [instance: 71380a36-8926-45d3-a70e-a7112bd1d06f] vm_ref = self.build_virtual_machine(instance, [ 529.496609] env[61806]: ERROR nova.compute.manager [instance: 71380a36-8926-45d3-a70e-a7112bd1d06f] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 529.496938] env[61806]: ERROR nova.compute.manager [instance: 71380a36-8926-45d3-a70e-a7112bd1d06f] vif_infos = vmwarevif.get_vif_info(self._session, [ 529.496938] env[61806]: ERROR nova.compute.manager [instance: 71380a36-8926-45d3-a70e-a7112bd1d06f] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 529.496938] env[61806]: ERROR nova.compute.manager [instance: 71380a36-8926-45d3-a70e-a7112bd1d06f] for vif in network_info: [ 529.496938] env[61806]: ERROR nova.compute.manager [instance: 71380a36-8926-45d3-a70e-a7112bd1d06f] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 529.496938] env[61806]: ERROR nova.compute.manager [instance: 71380a36-8926-45d3-a70e-a7112bd1d06f] return self._sync_wrapper(fn, *args, **kwargs) [ 529.496938] env[61806]: ERROR nova.compute.manager [instance: 71380a36-8926-45d3-a70e-a7112bd1d06f] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 529.496938] env[61806]: ERROR nova.compute.manager [instance: 71380a36-8926-45d3-a70e-a7112bd1d06f] self.wait() [ 529.496938] env[61806]: ERROR nova.compute.manager [instance: 71380a36-8926-45d3-a70e-a7112bd1d06f] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 529.496938] env[61806]: ERROR nova.compute.manager [instance: 71380a36-8926-45d3-a70e-a7112bd1d06f] self[:] = self._gt.wait() [ 529.496938] env[61806]: ERROR nova.compute.manager [instance: 71380a36-8926-45d3-a70e-a7112bd1d06f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 529.496938] env[61806]: ERROR nova.compute.manager [instance: 71380a36-8926-45d3-a70e-a7112bd1d06f] return self._exit_event.wait() [ 529.496938] env[61806]: ERROR nova.compute.manager [instance: 71380a36-8926-45d3-a70e-a7112bd1d06f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 529.496938] env[61806]: ERROR nova.compute.manager [instance: 71380a36-8926-45d3-a70e-a7112bd1d06f] result = hub.switch() [ 529.497301] env[61806]: ERROR nova.compute.manager [instance: 71380a36-8926-45d3-a70e-a7112bd1d06f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 529.497301] env[61806]: ERROR nova.compute.manager [instance: 71380a36-8926-45d3-a70e-a7112bd1d06f] return self.greenlet.switch() [ 529.497301] env[61806]: ERROR nova.compute.manager [instance: 71380a36-8926-45d3-a70e-a7112bd1d06f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 529.497301] env[61806]: ERROR nova.compute.manager [instance: 71380a36-8926-45d3-a70e-a7112bd1d06f] result = function(*args, **kwargs) [ 529.497301] env[61806]: ERROR nova.compute.manager [instance: 71380a36-8926-45d3-a70e-a7112bd1d06f] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 529.497301] env[61806]: ERROR nova.compute.manager [instance: 71380a36-8926-45d3-a70e-a7112bd1d06f] return func(*args, **kwargs) [ 529.497301] env[61806]: ERROR nova.compute.manager [instance: 71380a36-8926-45d3-a70e-a7112bd1d06f] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 529.497301] env[61806]: ERROR nova.compute.manager [instance: 71380a36-8926-45d3-a70e-a7112bd1d06f] raise e [ 529.497301] env[61806]: ERROR nova.compute.manager [instance: 71380a36-8926-45d3-a70e-a7112bd1d06f] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 529.497301] env[61806]: ERROR nova.compute.manager [instance: 71380a36-8926-45d3-a70e-a7112bd1d06f] nwinfo = self.network_api.allocate_for_instance( [ 529.497301] env[61806]: ERROR nova.compute.manager [instance: 71380a36-8926-45d3-a70e-a7112bd1d06f] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 529.497301] env[61806]: ERROR nova.compute.manager [instance: 71380a36-8926-45d3-a70e-a7112bd1d06f] created_port_ids = self._update_ports_for_instance( [ 529.497301] env[61806]: ERROR nova.compute.manager [instance: 71380a36-8926-45d3-a70e-a7112bd1d06f] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 529.497624] env[61806]: ERROR nova.compute.manager [instance: 71380a36-8926-45d3-a70e-a7112bd1d06f] with excutils.save_and_reraise_exception(): [ 529.497624] env[61806]: ERROR nova.compute.manager [instance: 71380a36-8926-45d3-a70e-a7112bd1d06f] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 529.497624] env[61806]: ERROR nova.compute.manager [instance: 71380a36-8926-45d3-a70e-a7112bd1d06f] self.force_reraise() [ 529.497624] env[61806]: ERROR nova.compute.manager [instance: 71380a36-8926-45d3-a70e-a7112bd1d06f] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 529.497624] env[61806]: ERROR nova.compute.manager [instance: 71380a36-8926-45d3-a70e-a7112bd1d06f] raise self.value [ 529.497624] env[61806]: ERROR nova.compute.manager [instance: 71380a36-8926-45d3-a70e-a7112bd1d06f] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 529.497624] env[61806]: ERROR nova.compute.manager [instance: 71380a36-8926-45d3-a70e-a7112bd1d06f] updated_port = self._update_port( [ 529.497624] env[61806]: ERROR nova.compute.manager [instance: 71380a36-8926-45d3-a70e-a7112bd1d06f] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 529.497624] env[61806]: ERROR nova.compute.manager [instance: 71380a36-8926-45d3-a70e-a7112bd1d06f] _ensure_no_port_binding_failure(port) [ 529.497624] env[61806]: ERROR nova.compute.manager [instance: 71380a36-8926-45d3-a70e-a7112bd1d06f] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 529.497624] env[61806]: ERROR nova.compute.manager [instance: 71380a36-8926-45d3-a70e-a7112bd1d06f] raise exception.PortBindingFailed(port_id=port['id']) [ 529.497624] env[61806]: ERROR nova.compute.manager [instance: 71380a36-8926-45d3-a70e-a7112bd1d06f] nova.exception.PortBindingFailed: Binding failed for port 3f10bab7-3a39-46da-b6db-0d3525153d7b, please check neutron logs for more information. [ 529.497624] env[61806]: ERROR nova.compute.manager [instance: 71380a36-8926-45d3-a70e-a7112bd1d06f] [ 529.497968] env[61806]: INFO nova.compute.manager [None req-2150b9c2-a011-40b9-950a-cddb0d2b8fb6 tempest-ImagesNegativeTestJSON-436875204 tempest-ImagesNegativeTestJSON-436875204-project-member] [instance: 71380a36-8926-45d3-a70e-a7112bd1d06f] Terminating instance [ 529.506099] env[61806]: INFO nova.compute.manager [None req-e67991d2-7862-4225-a618-8e964a5a624a tempest-ServerDiagnosticsNegativeTest-847099007 tempest-ServerDiagnosticsNegativeTest-847099007-project-member] [instance: e397a02e-26c6-4eb5-a242-5aa2b29fa29d] Took 1.20 seconds to deallocate network for instance. [ 529.510480] env[61806]: DEBUG oslo_concurrency.lockutils [None req-2150b9c2-a011-40b9-950a-cddb0d2b8fb6 tempest-ImagesNegativeTestJSON-436875204 tempest-ImagesNegativeTestJSON-436875204-project-member] Acquiring lock "refresh_cache-71380a36-8926-45d3-a70e-a7112bd1d06f" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 529.510480] env[61806]: DEBUG oslo_concurrency.lockutils [None req-2150b9c2-a011-40b9-950a-cddb0d2b8fb6 tempest-ImagesNegativeTestJSON-436875204 tempest-ImagesNegativeTestJSON-436875204-project-member] Acquired lock "refresh_cache-71380a36-8926-45d3-a70e-a7112bd1d06f" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 529.510480] env[61806]: DEBUG nova.network.neutron [None req-2150b9c2-a011-40b9-950a-cddb0d2b8fb6 tempest-ImagesNegativeTestJSON-436875204 tempest-ImagesNegativeTestJSON-436875204-project-member] [instance: 71380a36-8926-45d3-a70e-a7112bd1d06f] Building network info cache for instance {{(pid=61806) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 529.517619] env[61806]: DEBUG nova.network.neutron [None req-3568a2a1-f64b-42a9-bf53-df7d9370887d tempest-ImagesTestJSON-618852556 tempest-ImagesTestJSON-618852556-project-member] [instance: 81538b78-9943-475d-aad3-f31607f2e493] Instance cache missing network info. {{(pid=61806) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 529.724116] env[61806]: DEBUG nova.network.neutron [None req-3568a2a1-f64b-42a9-bf53-df7d9370887d tempest-ImagesTestJSON-618852556 tempest-ImagesTestJSON-618852556-project-member] [instance: 81538b78-9943-475d-aad3-f31607f2e493] Updating instance_info_cache with network_info: [] {{(pid=61806) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 529.840795] env[61806]: DEBUG oslo_concurrency.lockutils [None req-7d4ae408-54ff-4248-838b-a36cbf45b0c5 tempest-ListServerFiltersTestJSON-241228521 tempest-ListServerFiltersTestJSON-241228521-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 529.906246] env[61806]: DEBUG nova.compute.utils [None req-0858196d-347c-4201-9cb8-c7f147f5d99d tempest-MigrationsAdminTest-470098049 tempest-MigrationsAdminTest-470098049-project-member] Using /dev/sd instead of None {{(pid=61806) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 529.908818] env[61806]: DEBUG nova.compute.manager [None req-0858196d-347c-4201-9cb8-c7f147f5d99d tempest-MigrationsAdminTest-470098049 tempest-MigrationsAdminTest-470098049-project-member] [instance: 52b19347-a08f-44a7-8aad-22a3af5e291d] Allocating IP information in the background. {{(pid=61806) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 529.909271] env[61806]: DEBUG nova.network.neutron [None req-0858196d-347c-4201-9cb8-c7f147f5d99d tempest-MigrationsAdminTest-470098049 tempest-MigrationsAdminTest-470098049-project-member] [instance: 52b19347-a08f-44a7-8aad-22a3af5e291d] allocate_for_instance() {{(pid=61806) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 530.052815] env[61806]: DEBUG nova.network.neutron [None req-2150b9c2-a011-40b9-950a-cddb0d2b8fb6 tempest-ImagesNegativeTestJSON-436875204 tempest-ImagesNegativeTestJSON-436875204-project-member] [instance: 71380a36-8926-45d3-a70e-a7112bd1d06f] Instance cache missing network info. {{(pid=61806) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 530.089339] env[61806]: DEBUG nova.policy [None req-0858196d-347c-4201-9cb8-c7f147f5d99d tempest-MigrationsAdminTest-470098049 tempest-MigrationsAdminTest-470098049-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'b248b20100be4ecbaeaa9e7ccd1bcfc3', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'ca6e40a2aeee479388f3a098945dba6e', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61806) authorize /opt/stack/nova/nova/policy.py:201}} [ 530.175447] env[61806]: DEBUG nova.compute.manager [req-e414bafe-1e60-48ee-b44c-18fb6d600234 req-03f37150-7d0f-488a-915d-0c477d52fbc2 service nova] [instance: 81538b78-9943-475d-aad3-f31607f2e493] Received event network-changed-91d5d4f4-3e0d-456f-98f8-483167092c82 {{(pid=61806) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 530.176715] env[61806]: DEBUG nova.compute.manager [req-e414bafe-1e60-48ee-b44c-18fb6d600234 req-03f37150-7d0f-488a-915d-0c477d52fbc2 service nova] [instance: 81538b78-9943-475d-aad3-f31607f2e493] Refreshing instance network info cache due to event network-changed-91d5d4f4-3e0d-456f-98f8-483167092c82. {{(pid=61806) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 530.177689] env[61806]: DEBUG oslo_concurrency.lockutils [req-e414bafe-1e60-48ee-b44c-18fb6d600234 req-03f37150-7d0f-488a-915d-0c477d52fbc2 service nova] Acquiring lock "refresh_cache-81538b78-9943-475d-aad3-f31607f2e493" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 530.230235] env[61806]: DEBUG oslo_concurrency.lockutils [None req-3568a2a1-f64b-42a9-bf53-df7d9370887d tempest-ImagesTestJSON-618852556 tempest-ImagesTestJSON-618852556-project-member] Releasing lock "refresh_cache-81538b78-9943-475d-aad3-f31607f2e493" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 530.230235] env[61806]: DEBUG nova.compute.manager [None req-3568a2a1-f64b-42a9-bf53-df7d9370887d tempest-ImagesTestJSON-618852556 tempest-ImagesTestJSON-618852556-project-member] [instance: 81538b78-9943-475d-aad3-f31607f2e493] Start destroying the instance on the hypervisor. {{(pid=61806) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 530.230235] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-3568a2a1-f64b-42a9-bf53-df7d9370887d tempest-ImagesTestJSON-618852556 tempest-ImagesTestJSON-618852556-project-member] [instance: 81538b78-9943-475d-aad3-f31607f2e493] Destroying instance {{(pid=61806) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 530.230235] env[61806]: DEBUG oslo_concurrency.lockutils [req-e414bafe-1e60-48ee-b44c-18fb6d600234 req-03f37150-7d0f-488a-915d-0c477d52fbc2 service nova] Acquired lock "refresh_cache-81538b78-9943-475d-aad3-f31607f2e493" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 530.230235] env[61806]: DEBUG nova.network.neutron [req-e414bafe-1e60-48ee-b44c-18fb6d600234 req-03f37150-7d0f-488a-915d-0c477d52fbc2 service nova] [instance: 81538b78-9943-475d-aad3-f31607f2e493] Refreshing network info cache for port 91d5d4f4-3e0d-456f-98f8-483167092c82 {{(pid=61806) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 530.230415] env[61806]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-51e5bafe-8fda-44ba-9b0c-b7e83af201f4 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 530.243233] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a9f88087-9070-4b82-8005-13172e872930 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 530.265110] env[61806]: DEBUG nova.network.neutron [None req-2150b9c2-a011-40b9-950a-cddb0d2b8fb6 tempest-ImagesNegativeTestJSON-436875204 tempest-ImagesNegativeTestJSON-436875204-project-member] [instance: 71380a36-8926-45d3-a70e-a7112bd1d06f] Updating instance_info_cache with network_info: [] {{(pid=61806) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 530.282319] env[61806]: WARNING nova.virt.vmwareapi.vmops [None req-3568a2a1-f64b-42a9-bf53-df7d9370887d tempest-ImagesTestJSON-618852556 tempest-ImagesTestJSON-618852556-project-member] [instance: 81538b78-9943-475d-aad3-f31607f2e493] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 81538b78-9943-475d-aad3-f31607f2e493 could not be found. [ 530.282608] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-3568a2a1-f64b-42a9-bf53-df7d9370887d tempest-ImagesTestJSON-618852556 tempest-ImagesTestJSON-618852556-project-member] [instance: 81538b78-9943-475d-aad3-f31607f2e493] Instance destroyed {{(pid=61806) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 530.282819] env[61806]: INFO nova.compute.manager [None req-3568a2a1-f64b-42a9-bf53-df7d9370887d tempest-ImagesTestJSON-618852556 tempest-ImagesTestJSON-618852556-project-member] [instance: 81538b78-9943-475d-aad3-f31607f2e493] Took 0.06 seconds to destroy the instance on the hypervisor. [ 530.283677] env[61806]: DEBUG oslo.service.loopingcall [None req-3568a2a1-f64b-42a9-bf53-df7d9370887d tempest-ImagesTestJSON-618852556 tempest-ImagesTestJSON-618852556-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61806) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 530.284269] env[61806]: DEBUG nova.compute.manager [-] [instance: 81538b78-9943-475d-aad3-f31607f2e493] Deallocating network for instance {{(pid=61806) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 530.284379] env[61806]: DEBUG nova.network.neutron [-] [instance: 81538b78-9943-475d-aad3-f31607f2e493] deallocate_for_instance() {{(pid=61806) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 530.305987] env[61806]: DEBUG nova.network.neutron [-] [instance: 81538b78-9943-475d-aad3-f31607f2e493] Instance cache missing network info. {{(pid=61806) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 530.416811] env[61806]: DEBUG nova.compute.manager [None req-0858196d-347c-4201-9cb8-c7f147f5d99d tempest-MigrationsAdminTest-470098049 tempest-MigrationsAdminTest-470098049-project-member] [instance: 52b19347-a08f-44a7-8aad-22a3af5e291d] Start building block device mappings for instance. {{(pid=61806) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 530.455564] env[61806]: DEBUG nova.compute.manager [req-8c380068-98a6-4c14-9e25-57e90ce644f6 req-7e6be020-240c-4bd1-9884-1926f681edbf service nova] [instance: 22ce5e12-6c69-4316-b5f2-eed57a609369] Received event network-changed-8fd4da63-a17c-4795-80db-6d4adad9e48c {{(pid=61806) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 530.455564] env[61806]: DEBUG nova.compute.manager [req-8c380068-98a6-4c14-9e25-57e90ce644f6 req-7e6be020-240c-4bd1-9884-1926f681edbf service nova] [instance: 22ce5e12-6c69-4316-b5f2-eed57a609369] Refreshing instance network info cache due to event network-changed-8fd4da63-a17c-4795-80db-6d4adad9e48c. {{(pid=61806) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 530.455830] env[61806]: DEBUG oslo_concurrency.lockutils [req-8c380068-98a6-4c14-9e25-57e90ce644f6 req-7e6be020-240c-4bd1-9884-1926f681edbf service nova] Acquiring lock "refresh_cache-22ce5e12-6c69-4316-b5f2-eed57a609369" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 530.455952] env[61806]: DEBUG oslo_concurrency.lockutils [req-8c380068-98a6-4c14-9e25-57e90ce644f6 req-7e6be020-240c-4bd1-9884-1926f681edbf service nova] Acquired lock "refresh_cache-22ce5e12-6c69-4316-b5f2-eed57a609369" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 530.456363] env[61806]: DEBUG nova.network.neutron [req-8c380068-98a6-4c14-9e25-57e90ce644f6 req-7e6be020-240c-4bd1-9884-1926f681edbf service nova] [instance: 22ce5e12-6c69-4316-b5f2-eed57a609369] Refreshing network info cache for port 8fd4da63-a17c-4795-80db-6d4adad9e48c {{(pid=61806) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 530.567089] env[61806]: INFO nova.scheduler.client.report [None req-e67991d2-7862-4225-a618-8e964a5a624a tempest-ServerDiagnosticsNegativeTest-847099007 tempest-ServerDiagnosticsNegativeTest-847099007-project-member] Deleted allocations for instance e397a02e-26c6-4eb5-a242-5aa2b29fa29d [ 530.636363] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-32220112-00df-4686-ad1c-cd824a4a5361 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 530.651558] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-abc461fa-dd86-48bb-a89c-04b870956496 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 530.693178] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ca2a78e0-4649-4762-b171-1c4181eb5ad2 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 530.703060] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-492a020d-421a-41d2-a74b-96fb9df169e3 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 530.722617] env[61806]: DEBUG nova.compute.provider_tree [None req-3d02814b-d8ff-4079-980b-6c580616ce84 tempest-AttachInterfacesUnderV243Test-1683422373 tempest-AttachInterfacesUnderV243Test-1683422373-project-member] Inventory has not changed in ProviderTree for provider: a2858be1-fd22-4e08-979e-87ad25293407 {{(pid=61806) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 530.768599] env[61806]: DEBUG oslo_concurrency.lockutils [None req-2150b9c2-a011-40b9-950a-cddb0d2b8fb6 tempest-ImagesNegativeTestJSON-436875204 tempest-ImagesNegativeTestJSON-436875204-project-member] Releasing lock "refresh_cache-71380a36-8926-45d3-a70e-a7112bd1d06f" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 530.769064] env[61806]: DEBUG nova.compute.manager [None req-2150b9c2-a011-40b9-950a-cddb0d2b8fb6 tempest-ImagesNegativeTestJSON-436875204 tempest-ImagesNegativeTestJSON-436875204-project-member] [instance: 71380a36-8926-45d3-a70e-a7112bd1d06f] Start destroying the instance on the hypervisor. {{(pid=61806) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 530.769289] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-2150b9c2-a011-40b9-950a-cddb0d2b8fb6 tempest-ImagesNegativeTestJSON-436875204 tempest-ImagesNegativeTestJSON-436875204-project-member] [instance: 71380a36-8926-45d3-a70e-a7112bd1d06f] Destroying instance {{(pid=61806) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 530.769585] env[61806]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-59bce1bc-0c1d-4d9d-adcf-1929cb42d0ac {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 530.780439] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-91b0a29f-1fd5-40a8-99f1-3136c5bceae9 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 530.817135] env[61806]: DEBUG nova.network.neutron [-] [instance: 81538b78-9943-475d-aad3-f31607f2e493] Updating instance_info_cache with network_info: [] {{(pid=61806) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 530.818674] env[61806]: WARNING nova.virt.vmwareapi.vmops [None req-2150b9c2-a011-40b9-950a-cddb0d2b8fb6 tempest-ImagesNegativeTestJSON-436875204 tempest-ImagesNegativeTestJSON-436875204-project-member] [instance: 71380a36-8926-45d3-a70e-a7112bd1d06f] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 71380a36-8926-45d3-a70e-a7112bd1d06f could not be found. [ 530.818891] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-2150b9c2-a011-40b9-950a-cddb0d2b8fb6 tempest-ImagesNegativeTestJSON-436875204 tempest-ImagesNegativeTestJSON-436875204-project-member] [instance: 71380a36-8926-45d3-a70e-a7112bd1d06f] Instance destroyed {{(pid=61806) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 530.819128] env[61806]: INFO nova.compute.manager [None req-2150b9c2-a011-40b9-950a-cddb0d2b8fb6 tempest-ImagesNegativeTestJSON-436875204 tempest-ImagesNegativeTestJSON-436875204-project-member] [instance: 71380a36-8926-45d3-a70e-a7112bd1d06f] Took 0.05 seconds to destroy the instance on the hypervisor. [ 530.819481] env[61806]: DEBUG oslo.service.loopingcall [None req-2150b9c2-a011-40b9-950a-cddb0d2b8fb6 tempest-ImagesNegativeTestJSON-436875204 tempest-ImagesNegativeTestJSON-436875204-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61806) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 530.819915] env[61806]: DEBUG nova.compute.manager [-] [instance: 71380a36-8926-45d3-a70e-a7112bd1d06f] Deallocating network for instance {{(pid=61806) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 530.820032] env[61806]: DEBUG nova.network.neutron [-] [instance: 71380a36-8926-45d3-a70e-a7112bd1d06f] deallocate_for_instance() {{(pid=61806) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 530.841815] env[61806]: DEBUG nova.network.neutron [req-e414bafe-1e60-48ee-b44c-18fb6d600234 req-03f37150-7d0f-488a-915d-0c477d52fbc2 service nova] [instance: 81538b78-9943-475d-aad3-f31607f2e493] Instance cache missing network info. {{(pid=61806) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 530.886216] env[61806]: DEBUG nova.network.neutron [-] [instance: 71380a36-8926-45d3-a70e-a7112bd1d06f] Instance cache missing network info. {{(pid=61806) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 531.023603] env[61806]: DEBUG nova.network.neutron [req-8c380068-98a6-4c14-9e25-57e90ce644f6 req-7e6be020-240c-4bd1-9884-1926f681edbf service nova] [instance: 22ce5e12-6c69-4316-b5f2-eed57a609369] Instance cache missing network info. {{(pid=61806) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 531.081080] env[61806]: DEBUG oslo_concurrency.lockutils [None req-e67991d2-7862-4225-a618-8e964a5a624a tempest-ServerDiagnosticsNegativeTest-847099007 tempest-ServerDiagnosticsNegativeTest-847099007-project-member] Lock "e397a02e-26c6-4eb5-a242-5aa2b29fa29d" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 21.278s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 531.082257] env[61806]: DEBUG oslo_concurrency.lockutils [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Lock "e397a02e-26c6-4eb5-a242-5aa2b29fa29d" acquired by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: waited 9.972s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 531.082257] env[61806]: INFO nova.compute.manager [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] [instance: e397a02e-26c6-4eb5-a242-5aa2b29fa29d] During sync_power_state the instance has a pending task (spawning). Skip. [ 531.082257] env[61806]: DEBUG oslo_concurrency.lockutils [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Lock "e397a02e-26c6-4eb5-a242-5aa2b29fa29d" "released" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: held 0.000s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 531.194961] env[61806]: DEBUG nova.network.neutron [None req-0858196d-347c-4201-9cb8-c7f147f5d99d tempest-MigrationsAdminTest-470098049 tempest-MigrationsAdminTest-470098049-project-member] [instance: 52b19347-a08f-44a7-8aad-22a3af5e291d] Successfully created port: d93797ab-56f6-476d-b147-8051439da3f2 {{(pid=61806) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 531.226384] env[61806]: DEBUG nova.scheduler.client.report [None req-3d02814b-d8ff-4079-980b-6c580616ce84 tempest-AttachInterfacesUnderV243Test-1683422373 tempest-AttachInterfacesUnderV243Test-1683422373-project-member] Inventory has not changed for provider a2858be1-fd22-4e08-979e-87ad25293407 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 139, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61806) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 531.242150] env[61806]: DEBUG nova.network.neutron [req-e414bafe-1e60-48ee-b44c-18fb6d600234 req-03f37150-7d0f-488a-915d-0c477d52fbc2 service nova] [instance: 81538b78-9943-475d-aad3-f31607f2e493] Updating instance_info_cache with network_info: [] {{(pid=61806) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 531.323294] env[61806]: INFO nova.compute.manager [-] [instance: 81538b78-9943-475d-aad3-f31607f2e493] Took 1.04 seconds to deallocate network for instance. [ 531.326764] env[61806]: DEBUG nova.compute.claims [None req-3568a2a1-f64b-42a9-bf53-df7d9370887d tempest-ImagesTestJSON-618852556 tempest-ImagesTestJSON-618852556-project-member] [instance: 81538b78-9943-475d-aad3-f31607f2e493] Aborting claim: {{(pid=61806) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 531.326938] env[61806]: DEBUG oslo_concurrency.lockutils [None req-3568a2a1-f64b-42a9-bf53-df7d9370887d tempest-ImagesTestJSON-618852556 tempest-ImagesTestJSON-618852556-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 531.339840] env[61806]: DEBUG nova.network.neutron [req-8c380068-98a6-4c14-9e25-57e90ce644f6 req-7e6be020-240c-4bd1-9884-1926f681edbf service nova] [instance: 22ce5e12-6c69-4316-b5f2-eed57a609369] Updating instance_info_cache with network_info: [] {{(pid=61806) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 531.346790] env[61806]: DEBUG oslo_concurrency.lockutils [None req-aac795d3-c37f-4811-9be9-aa67ad520ab0 tempest-ListServerFiltersTestJSON-241228521 tempest-ListServerFiltersTestJSON-241228521-project-member] Acquiring lock "7fcee323-8ea6-4f69-9f1e-f2105af0fe96" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 531.347112] env[61806]: DEBUG oslo_concurrency.lockutils [None req-aac795d3-c37f-4811-9be9-aa67ad520ab0 tempest-ListServerFiltersTestJSON-241228521 tempest-ListServerFiltersTestJSON-241228521-project-member] Lock "7fcee323-8ea6-4f69-9f1e-f2105af0fe96" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 531.390989] env[61806]: DEBUG nova.network.neutron [-] [instance: 71380a36-8926-45d3-a70e-a7112bd1d06f] Updating instance_info_cache with network_info: [] {{(pid=61806) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 531.433046] env[61806]: DEBUG nova.compute.manager [None req-0858196d-347c-4201-9cb8-c7f147f5d99d tempest-MigrationsAdminTest-470098049 tempest-MigrationsAdminTest-470098049-project-member] [instance: 52b19347-a08f-44a7-8aad-22a3af5e291d] Start spawning the instance on the hypervisor. {{(pid=61806) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 531.471640] env[61806]: DEBUG nova.virt.hardware [None req-0858196d-347c-4201-9cb8-c7f147f5d99d tempest-MigrationsAdminTest-470098049 tempest-MigrationsAdminTest-470098049-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-15T18:05:10Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-15T18:04:52Z,direct_url=,disk_format='vmdk',id=ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='84ba83fce2c349a6988173c5d6fc3b07',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-15T18:04:53Z,virtual_size=,visibility=), allow threads: False {{(pid=61806) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 531.471900] env[61806]: DEBUG nova.virt.hardware [None req-0858196d-347c-4201-9cb8-c7f147f5d99d tempest-MigrationsAdminTest-470098049 tempest-MigrationsAdminTest-470098049-project-member] Flavor limits 0:0:0 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 531.472101] env[61806]: DEBUG nova.virt.hardware [None req-0858196d-347c-4201-9cb8-c7f147f5d99d tempest-MigrationsAdminTest-470098049 tempest-MigrationsAdminTest-470098049-project-member] Image limits 0:0:0 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 531.472291] env[61806]: DEBUG nova.virt.hardware [None req-0858196d-347c-4201-9cb8-c7f147f5d99d tempest-MigrationsAdminTest-470098049 tempest-MigrationsAdminTest-470098049-project-member] Flavor pref 0:0:0 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 531.472439] env[61806]: DEBUG nova.virt.hardware [None req-0858196d-347c-4201-9cb8-c7f147f5d99d tempest-MigrationsAdminTest-470098049 tempest-MigrationsAdminTest-470098049-project-member] Image pref 0:0:0 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 531.472580] env[61806]: DEBUG nova.virt.hardware [None req-0858196d-347c-4201-9cb8-c7f147f5d99d tempest-MigrationsAdminTest-470098049 tempest-MigrationsAdminTest-470098049-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 531.472781] env[61806]: DEBUG nova.virt.hardware [None req-0858196d-347c-4201-9cb8-c7f147f5d99d tempest-MigrationsAdminTest-470098049 tempest-MigrationsAdminTest-470098049-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61806) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 531.472931] env[61806]: DEBUG nova.virt.hardware [None req-0858196d-347c-4201-9cb8-c7f147f5d99d tempest-MigrationsAdminTest-470098049 tempest-MigrationsAdminTest-470098049-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61806) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 531.473104] env[61806]: DEBUG nova.virt.hardware [None req-0858196d-347c-4201-9cb8-c7f147f5d99d tempest-MigrationsAdminTest-470098049 tempest-MigrationsAdminTest-470098049-project-member] Got 1 possible topologies {{(pid=61806) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 531.473926] env[61806]: DEBUG nova.virt.hardware [None req-0858196d-347c-4201-9cb8-c7f147f5d99d tempest-MigrationsAdminTest-470098049 tempest-MigrationsAdminTest-470098049-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61806) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 531.473926] env[61806]: DEBUG nova.virt.hardware [None req-0858196d-347c-4201-9cb8-c7f147f5d99d tempest-MigrationsAdminTest-470098049 tempest-MigrationsAdminTest-470098049-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61806) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 531.474610] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-67f58a56-7687-4f6d-8db5-80571b9b7bc2 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 531.489868] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-46a66486-de34-4b4b-a9c9-6a4a6a443161 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 531.735217] env[61806]: DEBUG oslo_concurrency.lockutils [None req-3d02814b-d8ff-4079-980b-6c580616ce84 tempest-AttachInterfacesUnderV243Test-1683422373 tempest-AttachInterfacesUnderV243Test-1683422373-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.339s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 531.735452] env[61806]: DEBUG nova.compute.manager [None req-3d02814b-d8ff-4079-980b-6c580616ce84 tempest-AttachInterfacesUnderV243Test-1683422373 tempest-AttachInterfacesUnderV243Test-1683422373-project-member] [instance: df2576da-dc49-4ced-9462-fa71710e2e2a] Start building networks asynchronously for instance. {{(pid=61806) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 531.739605] env[61806]: DEBUG oslo_concurrency.lockutils [None req-d8e4a121-0976-4eaa-b4c6-51deda86d081 tempest-InstanceActionsNegativeTestJSON-503825571 tempest-InstanceActionsNegativeTestJSON-503825571-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 5.685s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 531.741146] env[61806]: INFO nova.compute.claims [None req-d8e4a121-0976-4eaa-b4c6-51deda86d081 tempest-InstanceActionsNegativeTestJSON-503825571 tempest-InstanceActionsNegativeTestJSON-503825571-project-member] [instance: a1aca638-e9c4-4e3b-9855-f92e7bd67204] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 531.747363] env[61806]: DEBUG oslo_concurrency.lockutils [req-e414bafe-1e60-48ee-b44c-18fb6d600234 req-03f37150-7d0f-488a-915d-0c477d52fbc2 service nova] Releasing lock "refresh_cache-81538b78-9943-475d-aad3-f31607f2e493" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 531.848131] env[61806]: DEBUG oslo_concurrency.lockutils [req-8c380068-98a6-4c14-9e25-57e90ce644f6 req-7e6be020-240c-4bd1-9884-1926f681edbf service nova] Releasing lock "refresh_cache-22ce5e12-6c69-4316-b5f2-eed57a609369" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 531.848362] env[61806]: DEBUG nova.compute.manager [req-8c380068-98a6-4c14-9e25-57e90ce644f6 req-7e6be020-240c-4bd1-9884-1926f681edbf service nova] [instance: 22ce5e12-6c69-4316-b5f2-eed57a609369] Received event network-vif-deleted-8fd4da63-a17c-4795-80db-6d4adad9e48c {{(pid=61806) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 531.851451] env[61806]: DEBUG nova.compute.manager [None req-aac795d3-c37f-4811-9be9-aa67ad520ab0 tempest-ListServerFiltersTestJSON-241228521 tempest-ListServerFiltersTestJSON-241228521-project-member] [instance: 7fcee323-8ea6-4f69-9f1e-f2105af0fe96] Starting instance... {{(pid=61806) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 531.895065] env[61806]: INFO nova.compute.manager [-] [instance: 71380a36-8926-45d3-a70e-a7112bd1d06f] Took 1.07 seconds to deallocate network for instance. [ 531.901553] env[61806]: DEBUG nova.compute.claims [None req-2150b9c2-a011-40b9-950a-cddb0d2b8fb6 tempest-ImagesNegativeTestJSON-436875204 tempest-ImagesNegativeTestJSON-436875204-project-member] [instance: 71380a36-8926-45d3-a70e-a7112bd1d06f] Aborting claim: {{(pid=61806) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 531.901747] env[61806]: DEBUG oslo_concurrency.lockutils [None req-2150b9c2-a011-40b9-950a-cddb0d2b8fb6 tempest-ImagesNegativeTestJSON-436875204 tempest-ImagesNegativeTestJSON-436875204-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 532.244435] env[61806]: DEBUG nova.compute.utils [None req-3d02814b-d8ff-4079-980b-6c580616ce84 tempest-AttachInterfacesUnderV243Test-1683422373 tempest-AttachInterfacesUnderV243Test-1683422373-project-member] Using /dev/sd instead of None {{(pid=61806) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 532.245363] env[61806]: DEBUG nova.compute.manager [None req-3d02814b-d8ff-4079-980b-6c580616ce84 tempest-AttachInterfacesUnderV243Test-1683422373 tempest-AttachInterfacesUnderV243Test-1683422373-project-member] [instance: df2576da-dc49-4ced-9462-fa71710e2e2a] Allocating IP information in the background. {{(pid=61806) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 532.245780] env[61806]: DEBUG nova.network.neutron [None req-3d02814b-d8ff-4079-980b-6c580616ce84 tempest-AttachInterfacesUnderV243Test-1683422373 tempest-AttachInterfacesUnderV243Test-1683422373-project-member] [instance: df2576da-dc49-4ced-9462-fa71710e2e2a] allocate_for_instance() {{(pid=61806) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 532.330199] env[61806]: DEBUG nova.policy [None req-3d02814b-d8ff-4079-980b-6c580616ce84 tempest-AttachInterfacesUnderV243Test-1683422373 tempest-AttachInterfacesUnderV243Test-1683422373-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'b113fcafe8024d78aee5584c8518e9dd', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '315ba755cf9d415eb9a18842ebed4f4e', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61806) authorize /opt/stack/nova/nova/policy.py:201}} [ 532.392538] env[61806]: DEBUG oslo_concurrency.lockutils [None req-aac795d3-c37f-4811-9be9-aa67ad520ab0 tempest-ListServerFiltersTestJSON-241228521 tempest-ListServerFiltersTestJSON-241228521-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 532.746286] env[61806]: DEBUG nova.compute.manager [None req-3d02814b-d8ff-4079-980b-6c580616ce84 tempest-AttachInterfacesUnderV243Test-1683422373 tempest-AttachInterfacesUnderV243Test-1683422373-project-member] [instance: df2576da-dc49-4ced-9462-fa71710e2e2a] Start building block device mappings for instance. {{(pid=61806) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 532.859815] env[61806]: DEBUG nova.network.neutron [None req-3d02814b-d8ff-4079-980b-6c580616ce84 tempest-AttachInterfacesUnderV243Test-1683422373 tempest-AttachInterfacesUnderV243Test-1683422373-project-member] [instance: df2576da-dc49-4ced-9462-fa71710e2e2a] Successfully created port: cab2110f-89a2-4c90-8780-f2deaabaf59a {{(pid=61806) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 532.958897] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ebcfd3d3-3234-48fc-a8e4-49cc4196c643 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 532.966758] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cc55c7ec-b32e-4ad5-baaa-fda5578d2bda {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 532.999677] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a252cc8c-e984-4f8a-a9dc-4dd143045b2a {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 533.007931] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-623e0b90-2db3-4159-9a97-f35620065712 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 533.022096] env[61806]: DEBUG nova.compute.provider_tree [None req-d8e4a121-0976-4eaa-b4c6-51deda86d081 tempest-InstanceActionsNegativeTestJSON-503825571 tempest-InstanceActionsNegativeTestJSON-503825571-project-member] Inventory has not changed in ProviderTree for provider: a2858be1-fd22-4e08-979e-87ad25293407 {{(pid=61806) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 533.035221] env[61806]: DEBUG oslo_concurrency.lockutils [None req-55a61261-0862-4d73-b971-6d11cbe88c5e tempest-ServersAdminTestJSON-14199920 tempest-ServersAdminTestJSON-14199920-project-member] Acquiring lock "26a0baf9-68bd-4380-87b3-53f50524bbbf" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 533.037482] env[61806]: DEBUG oslo_concurrency.lockutils [None req-55a61261-0862-4d73-b971-6d11cbe88c5e tempest-ServersAdminTestJSON-14199920 tempest-ServersAdminTestJSON-14199920-project-member] Lock "26a0baf9-68bd-4380-87b3-53f50524bbbf" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 533.041641] env[61806]: DEBUG oslo_concurrency.lockutils [None req-be9179b6-8d17-4f04-8823-d48174fcc5c1 tempest-InstanceActionsV221TestJSON-379107765 tempest-InstanceActionsV221TestJSON-379107765-project-member] Acquiring lock "123f15be-3b56-4999-8010-40d2e86a6813" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 533.041641] env[61806]: DEBUG oslo_concurrency.lockutils [None req-be9179b6-8d17-4f04-8823-d48174fcc5c1 tempest-InstanceActionsV221TestJSON-379107765 tempest-InstanceActionsV221TestJSON-379107765-project-member] Lock "123f15be-3b56-4999-8010-40d2e86a6813" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 533.501024] env[61806]: DEBUG nova.compute.manager [req-b573f315-c5a0-4c75-a06b-5a7a0668536e req-1b1adc2c-bced-418e-8ea2-b2945d211425 service nova] [instance: 81538b78-9943-475d-aad3-f31607f2e493] Received event network-vif-deleted-91d5d4f4-3e0d-456f-98f8-483167092c82 {{(pid=61806) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 533.524948] env[61806]: DEBUG nova.scheduler.client.report [None req-d8e4a121-0976-4eaa-b4c6-51deda86d081 tempest-InstanceActionsNegativeTestJSON-503825571 tempest-InstanceActionsNegativeTestJSON-503825571-project-member] Inventory has not changed for provider a2858be1-fd22-4e08-979e-87ad25293407 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 139, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61806) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 533.541362] env[61806]: DEBUG nova.compute.manager [None req-55a61261-0862-4d73-b971-6d11cbe88c5e tempest-ServersAdminTestJSON-14199920 tempest-ServersAdminTestJSON-14199920-project-member] [instance: 26a0baf9-68bd-4380-87b3-53f50524bbbf] Starting instance... {{(pid=61806) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 533.758865] env[61806]: DEBUG nova.compute.manager [None req-3d02814b-d8ff-4079-980b-6c580616ce84 tempest-AttachInterfacesUnderV243Test-1683422373 tempest-AttachInterfacesUnderV243Test-1683422373-project-member] [instance: df2576da-dc49-4ced-9462-fa71710e2e2a] Start spawning the instance on the hypervisor. {{(pid=61806) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 533.791490] env[61806]: DEBUG nova.virt.hardware [None req-3d02814b-d8ff-4079-980b-6c580616ce84 tempest-AttachInterfacesUnderV243Test-1683422373 tempest-AttachInterfacesUnderV243Test-1683422373-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-15T18:05:10Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-15T18:04:52Z,direct_url=,disk_format='vmdk',id=ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='84ba83fce2c349a6988173c5d6fc3b07',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-15T18:04:53Z,virtual_size=,visibility=), allow threads: False {{(pid=61806) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 533.793650] env[61806]: DEBUG nova.virt.hardware [None req-3d02814b-d8ff-4079-980b-6c580616ce84 tempest-AttachInterfacesUnderV243Test-1683422373 tempest-AttachInterfacesUnderV243Test-1683422373-project-member] Flavor limits 0:0:0 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 533.794686] env[61806]: DEBUG nova.virt.hardware [None req-3d02814b-d8ff-4079-980b-6c580616ce84 tempest-AttachInterfacesUnderV243Test-1683422373 tempest-AttachInterfacesUnderV243Test-1683422373-project-member] Image limits 0:0:0 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 533.795085] env[61806]: DEBUG nova.virt.hardware [None req-3d02814b-d8ff-4079-980b-6c580616ce84 tempest-AttachInterfacesUnderV243Test-1683422373 tempest-AttachInterfacesUnderV243Test-1683422373-project-member] Flavor pref 0:0:0 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 533.795085] env[61806]: DEBUG nova.virt.hardware [None req-3d02814b-d8ff-4079-980b-6c580616ce84 tempest-AttachInterfacesUnderV243Test-1683422373 tempest-AttachInterfacesUnderV243Test-1683422373-project-member] Image pref 0:0:0 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 533.795246] env[61806]: DEBUG nova.virt.hardware [None req-3d02814b-d8ff-4079-980b-6c580616ce84 tempest-AttachInterfacesUnderV243Test-1683422373 tempest-AttachInterfacesUnderV243Test-1683422373-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 533.795729] env[61806]: DEBUG nova.virt.hardware [None req-3d02814b-d8ff-4079-980b-6c580616ce84 tempest-AttachInterfacesUnderV243Test-1683422373 tempest-AttachInterfacesUnderV243Test-1683422373-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61806) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 533.796329] env[61806]: DEBUG nova.virt.hardware [None req-3d02814b-d8ff-4079-980b-6c580616ce84 tempest-AttachInterfacesUnderV243Test-1683422373 tempest-AttachInterfacesUnderV243Test-1683422373-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61806) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 533.796673] env[61806]: DEBUG nova.virt.hardware [None req-3d02814b-d8ff-4079-980b-6c580616ce84 tempest-AttachInterfacesUnderV243Test-1683422373 tempest-AttachInterfacesUnderV243Test-1683422373-project-member] Got 1 possible topologies {{(pid=61806) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 533.796760] env[61806]: DEBUG nova.virt.hardware [None req-3d02814b-d8ff-4079-980b-6c580616ce84 tempest-AttachInterfacesUnderV243Test-1683422373 tempest-AttachInterfacesUnderV243Test-1683422373-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61806) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 533.796948] env[61806]: DEBUG nova.virt.hardware [None req-3d02814b-d8ff-4079-980b-6c580616ce84 tempest-AttachInterfacesUnderV243Test-1683422373 tempest-AttachInterfacesUnderV243Test-1683422373-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61806) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 533.797905] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c54003f8-52fb-427b-b28d-c6df46c5df9c {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 533.808547] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2fa46835-55a4-4208-b653-4b44d9dcede3 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 533.835687] env[61806]: DEBUG nova.compute.manager [req-aa1a1b86-e088-46f5-98b9-fb3c16d08944 req-dd7ddf72-6ffb-4ea4-9aea-bd34a0794f39 service nova] [instance: 71380a36-8926-45d3-a70e-a7112bd1d06f] Received event network-changed-3f10bab7-3a39-46da-b6db-0d3525153d7b {{(pid=61806) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 533.835880] env[61806]: DEBUG nova.compute.manager [req-aa1a1b86-e088-46f5-98b9-fb3c16d08944 req-dd7ddf72-6ffb-4ea4-9aea-bd34a0794f39 service nova] [instance: 71380a36-8926-45d3-a70e-a7112bd1d06f] Refreshing instance network info cache due to event network-changed-3f10bab7-3a39-46da-b6db-0d3525153d7b. {{(pid=61806) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 533.835927] env[61806]: DEBUG oslo_concurrency.lockutils [req-aa1a1b86-e088-46f5-98b9-fb3c16d08944 req-dd7ddf72-6ffb-4ea4-9aea-bd34a0794f39 service nova] Acquiring lock "refresh_cache-71380a36-8926-45d3-a70e-a7112bd1d06f" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 533.836426] env[61806]: DEBUG oslo_concurrency.lockutils [req-aa1a1b86-e088-46f5-98b9-fb3c16d08944 req-dd7ddf72-6ffb-4ea4-9aea-bd34a0794f39 service nova] Acquired lock "refresh_cache-71380a36-8926-45d3-a70e-a7112bd1d06f" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 533.836426] env[61806]: DEBUG nova.network.neutron [req-aa1a1b86-e088-46f5-98b9-fb3c16d08944 req-dd7ddf72-6ffb-4ea4-9aea-bd34a0794f39 service nova] [instance: 71380a36-8926-45d3-a70e-a7112bd1d06f] Refreshing network info cache for port 3f10bab7-3a39-46da-b6db-0d3525153d7b {{(pid=61806) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 534.031580] env[61806]: DEBUG oslo_concurrency.lockutils [None req-d8e4a121-0976-4eaa-b4c6-51deda86d081 tempest-InstanceActionsNegativeTestJSON-503825571 tempest-InstanceActionsNegativeTestJSON-503825571-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.292s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 534.032529] env[61806]: DEBUG nova.compute.manager [None req-d8e4a121-0976-4eaa-b4c6-51deda86d081 tempest-InstanceActionsNegativeTestJSON-503825571 tempest-InstanceActionsNegativeTestJSON-503825571-project-member] [instance: a1aca638-e9c4-4e3b-9855-f92e7bd67204] Start building networks asynchronously for instance. {{(pid=61806) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 534.043502] env[61806]: DEBUG oslo_concurrency.lockutils [None req-92c8b578-fe59-4328-af10-ad8c2b7cd69b tempest-DeleteServersAdminTestJSON-1553667494 tempest-DeleteServersAdminTestJSON-1553667494-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 6.627s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 534.067922] env[61806]: DEBUG oslo_concurrency.lockutils [None req-55a61261-0862-4d73-b971-6d11cbe88c5e tempest-ServersAdminTestJSON-14199920 tempest-ServersAdminTestJSON-14199920-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 534.092257] env[61806]: ERROR nova.compute.manager [None req-0858196d-347c-4201-9cb8-c7f147f5d99d tempest-MigrationsAdminTest-470098049 tempest-MigrationsAdminTest-470098049-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port d93797ab-56f6-476d-b147-8051439da3f2, please check neutron logs for more information. [ 534.092257] env[61806]: ERROR nova.compute.manager Traceback (most recent call last): [ 534.092257] env[61806]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 534.092257] env[61806]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 534.092257] env[61806]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 534.092257] env[61806]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 534.092257] env[61806]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 534.092257] env[61806]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 534.092257] env[61806]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 534.092257] env[61806]: ERROR nova.compute.manager self.force_reraise() [ 534.092257] env[61806]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 534.092257] env[61806]: ERROR nova.compute.manager raise self.value [ 534.092257] env[61806]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 534.092257] env[61806]: ERROR nova.compute.manager updated_port = self._update_port( [ 534.092257] env[61806]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 534.092257] env[61806]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 534.092747] env[61806]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 534.092747] env[61806]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 534.092747] env[61806]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port d93797ab-56f6-476d-b147-8051439da3f2, please check neutron logs for more information. [ 534.092747] env[61806]: ERROR nova.compute.manager [ 534.092747] env[61806]: Traceback (most recent call last): [ 534.092747] env[61806]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 534.092747] env[61806]: listener.cb(fileno) [ 534.092747] env[61806]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 534.092747] env[61806]: result = function(*args, **kwargs) [ 534.092747] env[61806]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 534.092747] env[61806]: return func(*args, **kwargs) [ 534.092747] env[61806]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 534.092747] env[61806]: raise e [ 534.092747] env[61806]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 534.092747] env[61806]: nwinfo = self.network_api.allocate_for_instance( [ 534.092747] env[61806]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 534.092747] env[61806]: created_port_ids = self._update_ports_for_instance( [ 534.092747] env[61806]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 534.092747] env[61806]: with excutils.save_and_reraise_exception(): [ 534.092747] env[61806]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 534.092747] env[61806]: self.force_reraise() [ 534.092747] env[61806]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 534.092747] env[61806]: raise self.value [ 534.092747] env[61806]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 534.092747] env[61806]: updated_port = self._update_port( [ 534.092747] env[61806]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 534.092747] env[61806]: _ensure_no_port_binding_failure(port) [ 534.092747] env[61806]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 534.092747] env[61806]: raise exception.PortBindingFailed(port_id=port['id']) [ 534.093471] env[61806]: nova.exception.PortBindingFailed: Binding failed for port d93797ab-56f6-476d-b147-8051439da3f2, please check neutron logs for more information. [ 534.093471] env[61806]: Removing descriptor: 16 [ 534.093471] env[61806]: ERROR nova.compute.manager [None req-0858196d-347c-4201-9cb8-c7f147f5d99d tempest-MigrationsAdminTest-470098049 tempest-MigrationsAdminTest-470098049-project-member] [instance: 52b19347-a08f-44a7-8aad-22a3af5e291d] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port d93797ab-56f6-476d-b147-8051439da3f2, please check neutron logs for more information. [ 534.093471] env[61806]: ERROR nova.compute.manager [instance: 52b19347-a08f-44a7-8aad-22a3af5e291d] Traceback (most recent call last): [ 534.093471] env[61806]: ERROR nova.compute.manager [instance: 52b19347-a08f-44a7-8aad-22a3af5e291d] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 534.093471] env[61806]: ERROR nova.compute.manager [instance: 52b19347-a08f-44a7-8aad-22a3af5e291d] yield resources [ 534.093471] env[61806]: ERROR nova.compute.manager [instance: 52b19347-a08f-44a7-8aad-22a3af5e291d] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 534.093471] env[61806]: ERROR nova.compute.manager [instance: 52b19347-a08f-44a7-8aad-22a3af5e291d] self.driver.spawn(context, instance, image_meta, [ 534.093471] env[61806]: ERROR nova.compute.manager [instance: 52b19347-a08f-44a7-8aad-22a3af5e291d] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 534.093471] env[61806]: ERROR nova.compute.manager [instance: 52b19347-a08f-44a7-8aad-22a3af5e291d] self._vmops.spawn(context, instance, image_meta, injected_files, [ 534.093471] env[61806]: ERROR nova.compute.manager [instance: 52b19347-a08f-44a7-8aad-22a3af5e291d] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 534.093471] env[61806]: ERROR nova.compute.manager [instance: 52b19347-a08f-44a7-8aad-22a3af5e291d] vm_ref = self.build_virtual_machine(instance, [ 534.093787] env[61806]: ERROR nova.compute.manager [instance: 52b19347-a08f-44a7-8aad-22a3af5e291d] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 534.093787] env[61806]: ERROR nova.compute.manager [instance: 52b19347-a08f-44a7-8aad-22a3af5e291d] vif_infos = vmwarevif.get_vif_info(self._session, [ 534.093787] env[61806]: ERROR nova.compute.manager [instance: 52b19347-a08f-44a7-8aad-22a3af5e291d] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 534.093787] env[61806]: ERROR nova.compute.manager [instance: 52b19347-a08f-44a7-8aad-22a3af5e291d] for vif in network_info: [ 534.093787] env[61806]: ERROR nova.compute.manager [instance: 52b19347-a08f-44a7-8aad-22a3af5e291d] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 534.093787] env[61806]: ERROR nova.compute.manager [instance: 52b19347-a08f-44a7-8aad-22a3af5e291d] return self._sync_wrapper(fn, *args, **kwargs) [ 534.093787] env[61806]: ERROR nova.compute.manager [instance: 52b19347-a08f-44a7-8aad-22a3af5e291d] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 534.093787] env[61806]: ERROR nova.compute.manager [instance: 52b19347-a08f-44a7-8aad-22a3af5e291d] self.wait() [ 534.093787] env[61806]: ERROR nova.compute.manager [instance: 52b19347-a08f-44a7-8aad-22a3af5e291d] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 534.093787] env[61806]: ERROR nova.compute.manager [instance: 52b19347-a08f-44a7-8aad-22a3af5e291d] self[:] = self._gt.wait() [ 534.093787] env[61806]: ERROR nova.compute.manager [instance: 52b19347-a08f-44a7-8aad-22a3af5e291d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 534.093787] env[61806]: ERROR nova.compute.manager [instance: 52b19347-a08f-44a7-8aad-22a3af5e291d] return self._exit_event.wait() [ 534.093787] env[61806]: ERROR nova.compute.manager [instance: 52b19347-a08f-44a7-8aad-22a3af5e291d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 534.094133] env[61806]: ERROR nova.compute.manager [instance: 52b19347-a08f-44a7-8aad-22a3af5e291d] result = hub.switch() [ 534.094133] env[61806]: ERROR nova.compute.manager [instance: 52b19347-a08f-44a7-8aad-22a3af5e291d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 534.094133] env[61806]: ERROR nova.compute.manager [instance: 52b19347-a08f-44a7-8aad-22a3af5e291d] return self.greenlet.switch() [ 534.094133] env[61806]: ERROR nova.compute.manager [instance: 52b19347-a08f-44a7-8aad-22a3af5e291d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 534.094133] env[61806]: ERROR nova.compute.manager [instance: 52b19347-a08f-44a7-8aad-22a3af5e291d] result = function(*args, **kwargs) [ 534.094133] env[61806]: ERROR nova.compute.manager [instance: 52b19347-a08f-44a7-8aad-22a3af5e291d] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 534.094133] env[61806]: ERROR nova.compute.manager [instance: 52b19347-a08f-44a7-8aad-22a3af5e291d] return func(*args, **kwargs) [ 534.094133] env[61806]: ERROR nova.compute.manager [instance: 52b19347-a08f-44a7-8aad-22a3af5e291d] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 534.094133] env[61806]: ERROR nova.compute.manager [instance: 52b19347-a08f-44a7-8aad-22a3af5e291d] raise e [ 534.094133] env[61806]: ERROR nova.compute.manager [instance: 52b19347-a08f-44a7-8aad-22a3af5e291d] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 534.094133] env[61806]: ERROR nova.compute.manager [instance: 52b19347-a08f-44a7-8aad-22a3af5e291d] nwinfo = self.network_api.allocate_for_instance( [ 534.094133] env[61806]: ERROR nova.compute.manager [instance: 52b19347-a08f-44a7-8aad-22a3af5e291d] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 534.094133] env[61806]: ERROR nova.compute.manager [instance: 52b19347-a08f-44a7-8aad-22a3af5e291d] created_port_ids = self._update_ports_for_instance( [ 534.094687] env[61806]: ERROR nova.compute.manager [instance: 52b19347-a08f-44a7-8aad-22a3af5e291d] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 534.094687] env[61806]: ERROR nova.compute.manager [instance: 52b19347-a08f-44a7-8aad-22a3af5e291d] with excutils.save_and_reraise_exception(): [ 534.094687] env[61806]: ERROR nova.compute.manager [instance: 52b19347-a08f-44a7-8aad-22a3af5e291d] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 534.094687] env[61806]: ERROR nova.compute.manager [instance: 52b19347-a08f-44a7-8aad-22a3af5e291d] self.force_reraise() [ 534.094687] env[61806]: ERROR nova.compute.manager [instance: 52b19347-a08f-44a7-8aad-22a3af5e291d] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 534.094687] env[61806]: ERROR nova.compute.manager [instance: 52b19347-a08f-44a7-8aad-22a3af5e291d] raise self.value [ 534.094687] env[61806]: ERROR nova.compute.manager [instance: 52b19347-a08f-44a7-8aad-22a3af5e291d] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 534.094687] env[61806]: ERROR nova.compute.manager [instance: 52b19347-a08f-44a7-8aad-22a3af5e291d] updated_port = self._update_port( [ 534.094687] env[61806]: ERROR nova.compute.manager [instance: 52b19347-a08f-44a7-8aad-22a3af5e291d] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 534.094687] env[61806]: ERROR nova.compute.manager [instance: 52b19347-a08f-44a7-8aad-22a3af5e291d] _ensure_no_port_binding_failure(port) [ 534.094687] env[61806]: ERROR nova.compute.manager [instance: 52b19347-a08f-44a7-8aad-22a3af5e291d] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 534.094687] env[61806]: ERROR nova.compute.manager [instance: 52b19347-a08f-44a7-8aad-22a3af5e291d] raise exception.PortBindingFailed(port_id=port['id']) [ 534.095258] env[61806]: ERROR nova.compute.manager [instance: 52b19347-a08f-44a7-8aad-22a3af5e291d] nova.exception.PortBindingFailed: Binding failed for port d93797ab-56f6-476d-b147-8051439da3f2, please check neutron logs for more information. [ 534.095258] env[61806]: ERROR nova.compute.manager [instance: 52b19347-a08f-44a7-8aad-22a3af5e291d] [ 534.095258] env[61806]: INFO nova.compute.manager [None req-0858196d-347c-4201-9cb8-c7f147f5d99d tempest-MigrationsAdminTest-470098049 tempest-MigrationsAdminTest-470098049-project-member] [instance: 52b19347-a08f-44a7-8aad-22a3af5e291d] Terminating instance [ 534.098493] env[61806]: DEBUG oslo_concurrency.lockutils [None req-0858196d-347c-4201-9cb8-c7f147f5d99d tempest-MigrationsAdminTest-470098049 tempest-MigrationsAdminTest-470098049-project-member] Acquiring lock "refresh_cache-52b19347-a08f-44a7-8aad-22a3af5e291d" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 534.098825] env[61806]: DEBUG oslo_concurrency.lockutils [None req-0858196d-347c-4201-9cb8-c7f147f5d99d tempest-MigrationsAdminTest-470098049 tempest-MigrationsAdminTest-470098049-project-member] Acquired lock "refresh_cache-52b19347-a08f-44a7-8aad-22a3af5e291d" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 534.099098] env[61806]: DEBUG nova.network.neutron [None req-0858196d-347c-4201-9cb8-c7f147f5d99d tempest-MigrationsAdminTest-470098049 tempest-MigrationsAdminTest-470098049-project-member] [instance: 52b19347-a08f-44a7-8aad-22a3af5e291d] Building network info cache for instance {{(pid=61806) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 534.433270] env[61806]: DEBUG oslo_concurrency.lockutils [None req-d176ea08-f6ea-453a-9c54-dfba38bfade5 tempest-ServersListShow296Test-1633904279 tempest-ServersListShow296Test-1633904279-project-member] Acquiring lock "5a95d5a0-e5bf-4ca1-bbad-f16b5d10ba45" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 534.433470] env[61806]: DEBUG oslo_concurrency.lockutils [None req-d176ea08-f6ea-453a-9c54-dfba38bfade5 tempest-ServersListShow296Test-1633904279 tempest-ServersListShow296Test-1633904279-project-member] Lock "5a95d5a0-e5bf-4ca1-bbad-f16b5d10ba45" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 534.511124] env[61806]: ERROR nova.compute.manager [None req-3d02814b-d8ff-4079-980b-6c580616ce84 tempest-AttachInterfacesUnderV243Test-1683422373 tempest-AttachInterfacesUnderV243Test-1683422373-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port cab2110f-89a2-4c90-8780-f2deaabaf59a, please check neutron logs for more information. [ 534.511124] env[61806]: ERROR nova.compute.manager Traceback (most recent call last): [ 534.511124] env[61806]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 534.511124] env[61806]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 534.511124] env[61806]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 534.511124] env[61806]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 534.511124] env[61806]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 534.511124] env[61806]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 534.511124] env[61806]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 534.511124] env[61806]: ERROR nova.compute.manager self.force_reraise() [ 534.511124] env[61806]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 534.511124] env[61806]: ERROR nova.compute.manager raise self.value [ 534.511124] env[61806]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 534.511124] env[61806]: ERROR nova.compute.manager updated_port = self._update_port( [ 534.511124] env[61806]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 534.511124] env[61806]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 534.511701] env[61806]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 534.511701] env[61806]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 534.511701] env[61806]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port cab2110f-89a2-4c90-8780-f2deaabaf59a, please check neutron logs for more information. [ 534.511701] env[61806]: ERROR nova.compute.manager [ 534.511701] env[61806]: Traceback (most recent call last): [ 534.511701] env[61806]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 534.511701] env[61806]: listener.cb(fileno) [ 534.511701] env[61806]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 534.511701] env[61806]: result = function(*args, **kwargs) [ 534.511701] env[61806]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 534.511701] env[61806]: return func(*args, **kwargs) [ 534.511701] env[61806]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 534.511701] env[61806]: raise e [ 534.511701] env[61806]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 534.511701] env[61806]: nwinfo = self.network_api.allocate_for_instance( [ 534.511701] env[61806]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 534.511701] env[61806]: created_port_ids = self._update_ports_for_instance( [ 534.511701] env[61806]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 534.511701] env[61806]: with excutils.save_and_reraise_exception(): [ 534.511701] env[61806]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 534.511701] env[61806]: self.force_reraise() [ 534.511701] env[61806]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 534.511701] env[61806]: raise self.value [ 534.511701] env[61806]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 534.511701] env[61806]: updated_port = self._update_port( [ 534.511701] env[61806]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 534.511701] env[61806]: _ensure_no_port_binding_failure(port) [ 534.511701] env[61806]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 534.511701] env[61806]: raise exception.PortBindingFailed(port_id=port['id']) [ 534.512646] env[61806]: nova.exception.PortBindingFailed: Binding failed for port cab2110f-89a2-4c90-8780-f2deaabaf59a, please check neutron logs for more information. [ 534.512646] env[61806]: Removing descriptor: 17 [ 534.512646] env[61806]: ERROR nova.compute.manager [None req-3d02814b-d8ff-4079-980b-6c580616ce84 tempest-AttachInterfacesUnderV243Test-1683422373 tempest-AttachInterfacesUnderV243Test-1683422373-project-member] [instance: df2576da-dc49-4ced-9462-fa71710e2e2a] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port cab2110f-89a2-4c90-8780-f2deaabaf59a, please check neutron logs for more information. [ 534.512646] env[61806]: ERROR nova.compute.manager [instance: df2576da-dc49-4ced-9462-fa71710e2e2a] Traceback (most recent call last): [ 534.512646] env[61806]: ERROR nova.compute.manager [instance: df2576da-dc49-4ced-9462-fa71710e2e2a] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 534.512646] env[61806]: ERROR nova.compute.manager [instance: df2576da-dc49-4ced-9462-fa71710e2e2a] yield resources [ 534.512646] env[61806]: ERROR nova.compute.manager [instance: df2576da-dc49-4ced-9462-fa71710e2e2a] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 534.512646] env[61806]: ERROR nova.compute.manager [instance: df2576da-dc49-4ced-9462-fa71710e2e2a] self.driver.spawn(context, instance, image_meta, [ 534.512646] env[61806]: ERROR nova.compute.manager [instance: df2576da-dc49-4ced-9462-fa71710e2e2a] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 534.512646] env[61806]: ERROR nova.compute.manager [instance: df2576da-dc49-4ced-9462-fa71710e2e2a] self._vmops.spawn(context, instance, image_meta, injected_files, [ 534.512646] env[61806]: ERROR nova.compute.manager [instance: df2576da-dc49-4ced-9462-fa71710e2e2a] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 534.512646] env[61806]: ERROR nova.compute.manager [instance: df2576da-dc49-4ced-9462-fa71710e2e2a] vm_ref = self.build_virtual_machine(instance, [ 534.513061] env[61806]: ERROR nova.compute.manager [instance: df2576da-dc49-4ced-9462-fa71710e2e2a] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 534.513061] env[61806]: ERROR nova.compute.manager [instance: df2576da-dc49-4ced-9462-fa71710e2e2a] vif_infos = vmwarevif.get_vif_info(self._session, [ 534.513061] env[61806]: ERROR nova.compute.manager [instance: df2576da-dc49-4ced-9462-fa71710e2e2a] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 534.513061] env[61806]: ERROR nova.compute.manager [instance: df2576da-dc49-4ced-9462-fa71710e2e2a] for vif in network_info: [ 534.513061] env[61806]: ERROR nova.compute.manager [instance: df2576da-dc49-4ced-9462-fa71710e2e2a] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 534.513061] env[61806]: ERROR nova.compute.manager [instance: df2576da-dc49-4ced-9462-fa71710e2e2a] return self._sync_wrapper(fn, *args, **kwargs) [ 534.513061] env[61806]: ERROR nova.compute.manager [instance: df2576da-dc49-4ced-9462-fa71710e2e2a] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 534.513061] env[61806]: ERROR nova.compute.manager [instance: df2576da-dc49-4ced-9462-fa71710e2e2a] self.wait() [ 534.513061] env[61806]: ERROR nova.compute.manager [instance: df2576da-dc49-4ced-9462-fa71710e2e2a] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 534.513061] env[61806]: ERROR nova.compute.manager [instance: df2576da-dc49-4ced-9462-fa71710e2e2a] self[:] = self._gt.wait() [ 534.513061] env[61806]: ERROR nova.compute.manager [instance: df2576da-dc49-4ced-9462-fa71710e2e2a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 534.513061] env[61806]: ERROR nova.compute.manager [instance: df2576da-dc49-4ced-9462-fa71710e2e2a] return self._exit_event.wait() [ 534.513061] env[61806]: ERROR nova.compute.manager [instance: df2576da-dc49-4ced-9462-fa71710e2e2a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 534.513488] env[61806]: ERROR nova.compute.manager [instance: df2576da-dc49-4ced-9462-fa71710e2e2a] result = hub.switch() [ 534.513488] env[61806]: ERROR nova.compute.manager [instance: df2576da-dc49-4ced-9462-fa71710e2e2a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 534.513488] env[61806]: ERROR nova.compute.manager [instance: df2576da-dc49-4ced-9462-fa71710e2e2a] return self.greenlet.switch() [ 534.513488] env[61806]: ERROR nova.compute.manager [instance: df2576da-dc49-4ced-9462-fa71710e2e2a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 534.513488] env[61806]: ERROR nova.compute.manager [instance: df2576da-dc49-4ced-9462-fa71710e2e2a] result = function(*args, **kwargs) [ 534.513488] env[61806]: ERROR nova.compute.manager [instance: df2576da-dc49-4ced-9462-fa71710e2e2a] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 534.513488] env[61806]: ERROR nova.compute.manager [instance: df2576da-dc49-4ced-9462-fa71710e2e2a] return func(*args, **kwargs) [ 534.513488] env[61806]: ERROR nova.compute.manager [instance: df2576da-dc49-4ced-9462-fa71710e2e2a] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 534.513488] env[61806]: ERROR nova.compute.manager [instance: df2576da-dc49-4ced-9462-fa71710e2e2a] raise e [ 534.513488] env[61806]: ERROR nova.compute.manager [instance: df2576da-dc49-4ced-9462-fa71710e2e2a] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 534.513488] env[61806]: ERROR nova.compute.manager [instance: df2576da-dc49-4ced-9462-fa71710e2e2a] nwinfo = self.network_api.allocate_for_instance( [ 534.513488] env[61806]: ERROR nova.compute.manager [instance: df2576da-dc49-4ced-9462-fa71710e2e2a] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 534.513488] env[61806]: ERROR nova.compute.manager [instance: df2576da-dc49-4ced-9462-fa71710e2e2a] created_port_ids = self._update_ports_for_instance( [ 534.513912] env[61806]: ERROR nova.compute.manager [instance: df2576da-dc49-4ced-9462-fa71710e2e2a] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 534.513912] env[61806]: ERROR nova.compute.manager [instance: df2576da-dc49-4ced-9462-fa71710e2e2a] with excutils.save_and_reraise_exception(): [ 534.513912] env[61806]: ERROR nova.compute.manager [instance: df2576da-dc49-4ced-9462-fa71710e2e2a] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 534.513912] env[61806]: ERROR nova.compute.manager [instance: df2576da-dc49-4ced-9462-fa71710e2e2a] self.force_reraise() [ 534.513912] env[61806]: ERROR nova.compute.manager [instance: df2576da-dc49-4ced-9462-fa71710e2e2a] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 534.513912] env[61806]: ERROR nova.compute.manager [instance: df2576da-dc49-4ced-9462-fa71710e2e2a] raise self.value [ 534.513912] env[61806]: ERROR nova.compute.manager [instance: df2576da-dc49-4ced-9462-fa71710e2e2a] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 534.513912] env[61806]: ERROR nova.compute.manager [instance: df2576da-dc49-4ced-9462-fa71710e2e2a] updated_port = self._update_port( [ 534.513912] env[61806]: ERROR nova.compute.manager [instance: df2576da-dc49-4ced-9462-fa71710e2e2a] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 534.513912] env[61806]: ERROR nova.compute.manager [instance: df2576da-dc49-4ced-9462-fa71710e2e2a] _ensure_no_port_binding_failure(port) [ 534.513912] env[61806]: ERROR nova.compute.manager [instance: df2576da-dc49-4ced-9462-fa71710e2e2a] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 534.513912] env[61806]: ERROR nova.compute.manager [instance: df2576da-dc49-4ced-9462-fa71710e2e2a] raise exception.PortBindingFailed(port_id=port['id']) [ 534.514445] env[61806]: ERROR nova.compute.manager [instance: df2576da-dc49-4ced-9462-fa71710e2e2a] nova.exception.PortBindingFailed: Binding failed for port cab2110f-89a2-4c90-8780-f2deaabaf59a, please check neutron logs for more information. [ 534.514445] env[61806]: ERROR nova.compute.manager [instance: df2576da-dc49-4ced-9462-fa71710e2e2a] [ 534.514445] env[61806]: INFO nova.compute.manager [None req-3d02814b-d8ff-4079-980b-6c580616ce84 tempest-AttachInterfacesUnderV243Test-1683422373 tempest-AttachInterfacesUnderV243Test-1683422373-project-member] [instance: df2576da-dc49-4ced-9462-fa71710e2e2a] Terminating instance [ 534.516022] env[61806]: DEBUG oslo_concurrency.lockutils [None req-3d02814b-d8ff-4079-980b-6c580616ce84 tempest-AttachInterfacesUnderV243Test-1683422373 tempest-AttachInterfacesUnderV243Test-1683422373-project-member] Acquiring lock "refresh_cache-df2576da-dc49-4ced-9462-fa71710e2e2a" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 534.516022] env[61806]: DEBUG oslo_concurrency.lockutils [None req-3d02814b-d8ff-4079-980b-6c580616ce84 tempest-AttachInterfacesUnderV243Test-1683422373 tempest-AttachInterfacesUnderV243Test-1683422373-project-member] Acquired lock "refresh_cache-df2576da-dc49-4ced-9462-fa71710e2e2a" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 534.516022] env[61806]: DEBUG nova.network.neutron [None req-3d02814b-d8ff-4079-980b-6c580616ce84 tempest-AttachInterfacesUnderV243Test-1683422373 tempest-AttachInterfacesUnderV243Test-1683422373-project-member] [instance: df2576da-dc49-4ced-9462-fa71710e2e2a] Building network info cache for instance {{(pid=61806) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 534.532899] env[61806]: DEBUG nova.network.neutron [req-aa1a1b86-e088-46f5-98b9-fb3c16d08944 req-dd7ddf72-6ffb-4ea4-9aea-bd34a0794f39 service nova] [instance: 71380a36-8926-45d3-a70e-a7112bd1d06f] Instance cache missing network info. {{(pid=61806) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 534.546363] env[61806]: DEBUG nova.compute.utils [None req-d8e4a121-0976-4eaa-b4c6-51deda86d081 tempest-InstanceActionsNegativeTestJSON-503825571 tempest-InstanceActionsNegativeTestJSON-503825571-project-member] Using /dev/sd instead of None {{(pid=61806) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 534.550657] env[61806]: DEBUG nova.compute.manager [None req-d8e4a121-0976-4eaa-b4c6-51deda86d081 tempest-InstanceActionsNegativeTestJSON-503825571 tempest-InstanceActionsNegativeTestJSON-503825571-project-member] [instance: a1aca638-e9c4-4e3b-9855-f92e7bd67204] Allocating IP information in the background. {{(pid=61806) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 534.551272] env[61806]: DEBUG nova.network.neutron [None req-d8e4a121-0976-4eaa-b4c6-51deda86d081 tempest-InstanceActionsNegativeTestJSON-503825571 tempest-InstanceActionsNegativeTestJSON-503825571-project-member] [instance: a1aca638-e9c4-4e3b-9855-f92e7bd67204] allocate_for_instance() {{(pid=61806) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 534.624501] env[61806]: DEBUG nova.network.neutron [None req-0858196d-347c-4201-9cb8-c7f147f5d99d tempest-MigrationsAdminTest-470098049 tempest-MigrationsAdminTest-470098049-project-member] [instance: 52b19347-a08f-44a7-8aad-22a3af5e291d] Instance cache missing network info. {{(pid=61806) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 534.636172] env[61806]: DEBUG nova.policy [None req-d8e4a121-0976-4eaa-b4c6-51deda86d081 tempest-InstanceActionsNegativeTestJSON-503825571 tempest-InstanceActionsNegativeTestJSON-503825571-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '16d82fc5a2024983a812d058a58fe0a1', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'd7d5c8f62f5d4b7e8315da2e6f8ba8fc', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61806) authorize /opt/stack/nova/nova/policy.py:201}} [ 534.767630] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e05d3985-fff4-4f21-addb-4291276d5f63 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 534.778463] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b19f343c-46d5-4652-aeee-11b1aa3be70c {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 534.814179] env[61806]: DEBUG nova.network.neutron [None req-0858196d-347c-4201-9cb8-c7f147f5d99d tempest-MigrationsAdminTest-470098049 tempest-MigrationsAdminTest-470098049-project-member] [instance: 52b19347-a08f-44a7-8aad-22a3af5e291d] Updating instance_info_cache with network_info: [] {{(pid=61806) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 534.816138] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b9d45c10-2006-40f6-80b7-3b8c746ccee4 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 534.820819] env[61806]: DEBUG nova.network.neutron [req-aa1a1b86-e088-46f5-98b9-fb3c16d08944 req-dd7ddf72-6ffb-4ea4-9aea-bd34a0794f39 service nova] [instance: 71380a36-8926-45d3-a70e-a7112bd1d06f] Updating instance_info_cache with network_info: [] {{(pid=61806) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 534.828230] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-20b0e9ef-f9a0-4c6d-8be4-95f7f22687c8 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 534.846700] env[61806]: DEBUG nova.compute.provider_tree [None req-92c8b578-fe59-4328-af10-ad8c2b7cd69b tempest-DeleteServersAdminTestJSON-1553667494 tempest-DeleteServersAdminTestJSON-1553667494-project-member] Inventory has not changed in ProviderTree for provider: a2858be1-fd22-4e08-979e-87ad25293407 {{(pid=61806) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 535.058582] env[61806]: DEBUG nova.compute.manager [None req-d8e4a121-0976-4eaa-b4c6-51deda86d081 tempest-InstanceActionsNegativeTestJSON-503825571 tempest-InstanceActionsNegativeTestJSON-503825571-project-member] [instance: a1aca638-e9c4-4e3b-9855-f92e7bd67204] Start building block device mappings for instance. {{(pid=61806) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 535.065020] env[61806]: DEBUG nova.network.neutron [None req-3d02814b-d8ff-4079-980b-6c580616ce84 tempest-AttachInterfacesUnderV243Test-1683422373 tempest-AttachInterfacesUnderV243Test-1683422373-project-member] [instance: df2576da-dc49-4ced-9462-fa71710e2e2a] Instance cache missing network info. {{(pid=61806) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 535.164109] env[61806]: DEBUG oslo_concurrency.lockutils [None req-38c4eaec-22d8-4bf0-bd45-321ef43329a0 tempest-ServersAdminTestJSON-14199920 tempest-ServersAdminTestJSON-14199920-project-member] Acquiring lock "8905f20f-28a6-49f4-88bf-e52177c6a4b6" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 535.164356] env[61806]: DEBUG oslo_concurrency.lockutils [None req-38c4eaec-22d8-4bf0-bd45-321ef43329a0 tempest-ServersAdminTestJSON-14199920 tempest-ServersAdminTestJSON-14199920-project-member] Lock "8905f20f-28a6-49f4-88bf-e52177c6a4b6" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 535.325754] env[61806]: DEBUG oslo_concurrency.lockutils [None req-0858196d-347c-4201-9cb8-c7f147f5d99d tempest-MigrationsAdminTest-470098049 tempest-MigrationsAdminTest-470098049-project-member] Releasing lock "refresh_cache-52b19347-a08f-44a7-8aad-22a3af5e291d" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 535.326076] env[61806]: DEBUG nova.compute.manager [None req-0858196d-347c-4201-9cb8-c7f147f5d99d tempest-MigrationsAdminTest-470098049 tempest-MigrationsAdminTest-470098049-project-member] [instance: 52b19347-a08f-44a7-8aad-22a3af5e291d] Start destroying the instance on the hypervisor. {{(pid=61806) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 535.326307] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-0858196d-347c-4201-9cb8-c7f147f5d99d tempest-MigrationsAdminTest-470098049 tempest-MigrationsAdminTest-470098049-project-member] [instance: 52b19347-a08f-44a7-8aad-22a3af5e291d] Destroying instance {{(pid=61806) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 535.326836] env[61806]: DEBUG oslo_concurrency.lockutils [req-aa1a1b86-e088-46f5-98b9-fb3c16d08944 req-dd7ddf72-6ffb-4ea4-9aea-bd34a0794f39 service nova] Releasing lock "refresh_cache-71380a36-8926-45d3-a70e-a7112bd1d06f" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 535.327141] env[61806]: DEBUG nova.compute.manager [req-aa1a1b86-e088-46f5-98b9-fb3c16d08944 req-dd7ddf72-6ffb-4ea4-9aea-bd34a0794f39 service nova] [instance: 71380a36-8926-45d3-a70e-a7112bd1d06f] Received event network-vif-deleted-3f10bab7-3a39-46da-b6db-0d3525153d7b {{(pid=61806) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 535.328117] env[61806]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-ef54252c-f29a-444c-bb5d-ed08e8e1a4f5 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 535.342642] env[61806]: DEBUG nova.network.neutron [None req-3d02814b-d8ff-4079-980b-6c580616ce84 tempest-AttachInterfacesUnderV243Test-1683422373 tempest-AttachInterfacesUnderV243Test-1683422373-project-member] [instance: df2576da-dc49-4ced-9462-fa71710e2e2a] Updating instance_info_cache with network_info: [] {{(pid=61806) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 535.351480] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b65fec05-f75e-4747-8fd7-16f5a609c5ce {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 535.376937] env[61806]: DEBUG nova.scheduler.client.report [None req-92c8b578-fe59-4328-af10-ad8c2b7cd69b tempest-DeleteServersAdminTestJSON-1553667494 tempest-DeleteServersAdminTestJSON-1553667494-project-member] Inventory has not changed for provider a2858be1-fd22-4e08-979e-87ad25293407 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 139, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61806) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 535.395975] env[61806]: WARNING nova.virt.vmwareapi.vmops [None req-0858196d-347c-4201-9cb8-c7f147f5d99d tempest-MigrationsAdminTest-470098049 tempest-MigrationsAdminTest-470098049-project-member] [instance: 52b19347-a08f-44a7-8aad-22a3af5e291d] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 52b19347-a08f-44a7-8aad-22a3af5e291d could not be found. [ 535.395975] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-0858196d-347c-4201-9cb8-c7f147f5d99d tempest-MigrationsAdminTest-470098049 tempest-MigrationsAdminTest-470098049-project-member] [instance: 52b19347-a08f-44a7-8aad-22a3af5e291d] Instance destroyed {{(pid=61806) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 535.395975] env[61806]: INFO nova.compute.manager [None req-0858196d-347c-4201-9cb8-c7f147f5d99d tempest-MigrationsAdminTest-470098049 tempest-MigrationsAdminTest-470098049-project-member] [instance: 52b19347-a08f-44a7-8aad-22a3af5e291d] Took 0.07 seconds to destroy the instance on the hypervisor. [ 535.395975] env[61806]: DEBUG oslo.service.loopingcall [None req-0858196d-347c-4201-9cb8-c7f147f5d99d tempest-MigrationsAdminTest-470098049 tempest-MigrationsAdminTest-470098049-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61806) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 535.395975] env[61806]: DEBUG nova.compute.manager [-] [instance: 52b19347-a08f-44a7-8aad-22a3af5e291d] Deallocating network for instance {{(pid=61806) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 535.395975] env[61806]: DEBUG nova.network.neutron [-] [instance: 52b19347-a08f-44a7-8aad-22a3af5e291d] deallocate_for_instance() {{(pid=61806) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 535.428390] env[61806]: DEBUG nova.network.neutron [-] [instance: 52b19347-a08f-44a7-8aad-22a3af5e291d] Instance cache missing network info. {{(pid=61806) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 535.508506] env[61806]: DEBUG oslo_concurrency.lockutils [None req-37e71da7-cb5e-431d-ba3c-95daaf8af7f3 tempest-ServersAdminNegativeTestJSON-2128564044 tempest-ServersAdminNegativeTestJSON-2128564044-project-member] Acquiring lock "89ce0fd8-6a22-4b1c-bdff-d1584abd538a" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 535.508747] env[61806]: DEBUG oslo_concurrency.lockutils [None req-37e71da7-cb5e-431d-ba3c-95daaf8af7f3 tempest-ServersAdminNegativeTestJSON-2128564044 tempest-ServersAdminNegativeTestJSON-2128564044-project-member] Lock "89ce0fd8-6a22-4b1c-bdff-d1584abd538a" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 535.628142] env[61806]: DEBUG nova.network.neutron [None req-d8e4a121-0976-4eaa-b4c6-51deda86d081 tempest-InstanceActionsNegativeTestJSON-503825571 tempest-InstanceActionsNegativeTestJSON-503825571-project-member] [instance: a1aca638-e9c4-4e3b-9855-f92e7bd67204] Successfully created port: 6cc2dbc9-ee9c-42fa-b51e-2fd8a2c5b6d9 {{(pid=61806) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 535.877818] env[61806]: DEBUG oslo_concurrency.lockutils [None req-3d02814b-d8ff-4079-980b-6c580616ce84 tempest-AttachInterfacesUnderV243Test-1683422373 tempest-AttachInterfacesUnderV243Test-1683422373-project-member] Releasing lock "refresh_cache-df2576da-dc49-4ced-9462-fa71710e2e2a" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 535.878265] env[61806]: DEBUG nova.compute.manager [None req-3d02814b-d8ff-4079-980b-6c580616ce84 tempest-AttachInterfacesUnderV243Test-1683422373 tempest-AttachInterfacesUnderV243Test-1683422373-project-member] [instance: df2576da-dc49-4ced-9462-fa71710e2e2a] Start destroying the instance on the hypervisor. {{(pid=61806) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 535.878458] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-3d02814b-d8ff-4079-980b-6c580616ce84 tempest-AttachInterfacesUnderV243Test-1683422373 tempest-AttachInterfacesUnderV243Test-1683422373-project-member] [instance: df2576da-dc49-4ced-9462-fa71710e2e2a] Destroying instance {{(pid=61806) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 535.878756] env[61806]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-4e771d62-e8b4-43dc-8e10-1854c39bf2ef {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 535.884816] env[61806]: DEBUG oslo_concurrency.lockutils [None req-92c8b578-fe59-4328-af10-ad8c2b7cd69b tempest-DeleteServersAdminTestJSON-1553667494 tempest-DeleteServersAdminTestJSON-1553667494-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.845s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 535.885336] env[61806]: ERROR nova.compute.manager [None req-92c8b578-fe59-4328-af10-ad8c2b7cd69b tempest-DeleteServersAdminTestJSON-1553667494 tempest-DeleteServersAdminTestJSON-1553667494-project-member] [instance: 22ce5e12-6c69-4316-b5f2-eed57a609369] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 8fd4da63-a17c-4795-80db-6d4adad9e48c, please check neutron logs for more information. [ 535.885336] env[61806]: ERROR nova.compute.manager [instance: 22ce5e12-6c69-4316-b5f2-eed57a609369] Traceback (most recent call last): [ 535.885336] env[61806]: ERROR nova.compute.manager [instance: 22ce5e12-6c69-4316-b5f2-eed57a609369] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 535.885336] env[61806]: ERROR nova.compute.manager [instance: 22ce5e12-6c69-4316-b5f2-eed57a609369] self.driver.spawn(context, instance, image_meta, [ 535.885336] env[61806]: ERROR nova.compute.manager [instance: 22ce5e12-6c69-4316-b5f2-eed57a609369] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 535.885336] env[61806]: ERROR nova.compute.manager [instance: 22ce5e12-6c69-4316-b5f2-eed57a609369] self._vmops.spawn(context, instance, image_meta, injected_files, [ 535.885336] env[61806]: ERROR nova.compute.manager [instance: 22ce5e12-6c69-4316-b5f2-eed57a609369] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 535.885336] env[61806]: ERROR nova.compute.manager [instance: 22ce5e12-6c69-4316-b5f2-eed57a609369] vm_ref = self.build_virtual_machine(instance, [ 535.885336] env[61806]: ERROR nova.compute.manager [instance: 22ce5e12-6c69-4316-b5f2-eed57a609369] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 535.885336] env[61806]: ERROR nova.compute.manager [instance: 22ce5e12-6c69-4316-b5f2-eed57a609369] vif_infos = vmwarevif.get_vif_info(self._session, [ 535.885336] env[61806]: ERROR nova.compute.manager [instance: 22ce5e12-6c69-4316-b5f2-eed57a609369] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 535.885729] env[61806]: ERROR nova.compute.manager [instance: 22ce5e12-6c69-4316-b5f2-eed57a609369] for vif in network_info: [ 535.885729] env[61806]: ERROR nova.compute.manager [instance: 22ce5e12-6c69-4316-b5f2-eed57a609369] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 535.885729] env[61806]: ERROR nova.compute.manager [instance: 22ce5e12-6c69-4316-b5f2-eed57a609369] return self._sync_wrapper(fn, *args, **kwargs) [ 535.885729] env[61806]: ERROR nova.compute.manager [instance: 22ce5e12-6c69-4316-b5f2-eed57a609369] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 535.885729] env[61806]: ERROR nova.compute.manager [instance: 22ce5e12-6c69-4316-b5f2-eed57a609369] self.wait() [ 535.885729] env[61806]: ERROR nova.compute.manager [instance: 22ce5e12-6c69-4316-b5f2-eed57a609369] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 535.885729] env[61806]: ERROR nova.compute.manager [instance: 22ce5e12-6c69-4316-b5f2-eed57a609369] self[:] = self._gt.wait() [ 535.885729] env[61806]: ERROR nova.compute.manager [instance: 22ce5e12-6c69-4316-b5f2-eed57a609369] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 535.885729] env[61806]: ERROR nova.compute.manager [instance: 22ce5e12-6c69-4316-b5f2-eed57a609369] return self._exit_event.wait() [ 535.885729] env[61806]: ERROR nova.compute.manager [instance: 22ce5e12-6c69-4316-b5f2-eed57a609369] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 535.885729] env[61806]: ERROR nova.compute.manager [instance: 22ce5e12-6c69-4316-b5f2-eed57a609369] result = hub.switch() [ 535.885729] env[61806]: ERROR nova.compute.manager [instance: 22ce5e12-6c69-4316-b5f2-eed57a609369] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 535.885729] env[61806]: ERROR nova.compute.manager [instance: 22ce5e12-6c69-4316-b5f2-eed57a609369] return self.greenlet.switch() [ 535.886054] env[61806]: ERROR nova.compute.manager [instance: 22ce5e12-6c69-4316-b5f2-eed57a609369] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 535.886054] env[61806]: ERROR nova.compute.manager [instance: 22ce5e12-6c69-4316-b5f2-eed57a609369] result = function(*args, **kwargs) [ 535.886054] env[61806]: ERROR nova.compute.manager [instance: 22ce5e12-6c69-4316-b5f2-eed57a609369] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 535.886054] env[61806]: ERROR nova.compute.manager [instance: 22ce5e12-6c69-4316-b5f2-eed57a609369] return func(*args, **kwargs) [ 535.886054] env[61806]: ERROR nova.compute.manager [instance: 22ce5e12-6c69-4316-b5f2-eed57a609369] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 535.886054] env[61806]: ERROR nova.compute.manager [instance: 22ce5e12-6c69-4316-b5f2-eed57a609369] raise e [ 535.886054] env[61806]: ERROR nova.compute.manager [instance: 22ce5e12-6c69-4316-b5f2-eed57a609369] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 535.886054] env[61806]: ERROR nova.compute.manager [instance: 22ce5e12-6c69-4316-b5f2-eed57a609369] nwinfo = self.network_api.allocate_for_instance( [ 535.886054] env[61806]: ERROR nova.compute.manager [instance: 22ce5e12-6c69-4316-b5f2-eed57a609369] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 535.886054] env[61806]: ERROR nova.compute.manager [instance: 22ce5e12-6c69-4316-b5f2-eed57a609369] created_port_ids = self._update_ports_for_instance( [ 535.886054] env[61806]: ERROR nova.compute.manager [instance: 22ce5e12-6c69-4316-b5f2-eed57a609369] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 535.886054] env[61806]: ERROR nova.compute.manager [instance: 22ce5e12-6c69-4316-b5f2-eed57a609369] with excutils.save_and_reraise_exception(): [ 535.886054] env[61806]: ERROR nova.compute.manager [instance: 22ce5e12-6c69-4316-b5f2-eed57a609369] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 535.886421] env[61806]: ERROR nova.compute.manager [instance: 22ce5e12-6c69-4316-b5f2-eed57a609369] self.force_reraise() [ 535.886421] env[61806]: ERROR nova.compute.manager [instance: 22ce5e12-6c69-4316-b5f2-eed57a609369] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 535.886421] env[61806]: ERROR nova.compute.manager [instance: 22ce5e12-6c69-4316-b5f2-eed57a609369] raise self.value [ 535.886421] env[61806]: ERROR nova.compute.manager [instance: 22ce5e12-6c69-4316-b5f2-eed57a609369] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 535.886421] env[61806]: ERROR nova.compute.manager [instance: 22ce5e12-6c69-4316-b5f2-eed57a609369] updated_port = self._update_port( [ 535.886421] env[61806]: ERROR nova.compute.manager [instance: 22ce5e12-6c69-4316-b5f2-eed57a609369] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 535.886421] env[61806]: ERROR nova.compute.manager [instance: 22ce5e12-6c69-4316-b5f2-eed57a609369] _ensure_no_port_binding_failure(port) [ 535.886421] env[61806]: ERROR nova.compute.manager [instance: 22ce5e12-6c69-4316-b5f2-eed57a609369] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 535.886421] env[61806]: ERROR nova.compute.manager [instance: 22ce5e12-6c69-4316-b5f2-eed57a609369] raise exception.PortBindingFailed(port_id=port['id']) [ 535.886421] env[61806]: ERROR nova.compute.manager [instance: 22ce5e12-6c69-4316-b5f2-eed57a609369] nova.exception.PortBindingFailed: Binding failed for port 8fd4da63-a17c-4795-80db-6d4adad9e48c, please check neutron logs for more information. [ 535.886421] env[61806]: ERROR nova.compute.manager [instance: 22ce5e12-6c69-4316-b5f2-eed57a609369] [ 535.886695] env[61806]: DEBUG nova.compute.utils [None req-92c8b578-fe59-4328-af10-ad8c2b7cd69b tempest-DeleteServersAdminTestJSON-1553667494 tempest-DeleteServersAdminTestJSON-1553667494-project-member] [instance: 22ce5e12-6c69-4316-b5f2-eed57a609369] Binding failed for port 8fd4da63-a17c-4795-80db-6d4adad9e48c, please check neutron logs for more information. {{(pid=61806) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 535.889255] env[61806]: DEBUG oslo_concurrency.lockutils [None req-960d1abd-1b06-4b5a-8ddd-4b463d493a45 tempest-ListServerFiltersTestJSON-241228521 tempest-ListServerFiltersTestJSON-241228521-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 8.361s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 535.891357] env[61806]: INFO nova.compute.claims [None req-960d1abd-1b06-4b5a-8ddd-4b463d493a45 tempest-ListServerFiltersTestJSON-241228521 tempest-ListServerFiltersTestJSON-241228521-project-member] [instance: 30e44576-dcd3-4036-9070-9032b775c90a] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 535.898587] env[61806]: DEBUG nova.compute.manager [None req-92c8b578-fe59-4328-af10-ad8c2b7cd69b tempest-DeleteServersAdminTestJSON-1553667494 tempest-DeleteServersAdminTestJSON-1553667494-project-member] [instance: 22ce5e12-6c69-4316-b5f2-eed57a609369] Build of instance 22ce5e12-6c69-4316-b5f2-eed57a609369 was re-scheduled: Binding failed for port 8fd4da63-a17c-4795-80db-6d4adad9e48c, please check neutron logs for more information. {{(pid=61806) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 535.898972] env[61806]: DEBUG nova.compute.manager [None req-92c8b578-fe59-4328-af10-ad8c2b7cd69b tempest-DeleteServersAdminTestJSON-1553667494 tempest-DeleteServersAdminTestJSON-1553667494-project-member] [instance: 22ce5e12-6c69-4316-b5f2-eed57a609369] Unplugging VIFs for instance {{(pid=61806) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 535.899255] env[61806]: DEBUG oslo_concurrency.lockutils [None req-92c8b578-fe59-4328-af10-ad8c2b7cd69b tempest-DeleteServersAdminTestJSON-1553667494 tempest-DeleteServersAdminTestJSON-1553667494-project-member] Acquiring lock "refresh_cache-22ce5e12-6c69-4316-b5f2-eed57a609369" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 535.899401] env[61806]: DEBUG oslo_concurrency.lockutils [None req-92c8b578-fe59-4328-af10-ad8c2b7cd69b tempest-DeleteServersAdminTestJSON-1553667494 tempest-DeleteServersAdminTestJSON-1553667494-project-member] Acquired lock "refresh_cache-22ce5e12-6c69-4316-b5f2-eed57a609369" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 535.899559] env[61806]: DEBUG nova.network.neutron [None req-92c8b578-fe59-4328-af10-ad8c2b7cd69b tempest-DeleteServersAdminTestJSON-1553667494 tempest-DeleteServersAdminTestJSON-1553667494-project-member] [instance: 22ce5e12-6c69-4316-b5f2-eed57a609369] Building network info cache for instance {{(pid=61806) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 535.906485] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0c081096-389f-4e0c-aae2-56034e4a95fb {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 535.931850] env[61806]: DEBUG nova.network.neutron [-] [instance: 52b19347-a08f-44a7-8aad-22a3af5e291d] Updating instance_info_cache with network_info: [] {{(pid=61806) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 535.952974] env[61806]: WARNING nova.virt.vmwareapi.vmops [None req-3d02814b-d8ff-4079-980b-6c580616ce84 tempest-AttachInterfacesUnderV243Test-1683422373 tempest-AttachInterfacesUnderV243Test-1683422373-project-member] [instance: df2576da-dc49-4ced-9462-fa71710e2e2a] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance df2576da-dc49-4ced-9462-fa71710e2e2a could not be found. [ 535.953259] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-3d02814b-d8ff-4079-980b-6c580616ce84 tempest-AttachInterfacesUnderV243Test-1683422373 tempest-AttachInterfacesUnderV243Test-1683422373-project-member] [instance: df2576da-dc49-4ced-9462-fa71710e2e2a] Instance destroyed {{(pid=61806) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 535.953466] env[61806]: INFO nova.compute.manager [None req-3d02814b-d8ff-4079-980b-6c580616ce84 tempest-AttachInterfacesUnderV243Test-1683422373 tempest-AttachInterfacesUnderV243Test-1683422373-project-member] [instance: df2576da-dc49-4ced-9462-fa71710e2e2a] Took 0.08 seconds to destroy the instance on the hypervisor. [ 535.953708] env[61806]: DEBUG oslo.service.loopingcall [None req-3d02814b-d8ff-4079-980b-6c580616ce84 tempest-AttachInterfacesUnderV243Test-1683422373 tempest-AttachInterfacesUnderV243Test-1683422373-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61806) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 535.955162] env[61806]: DEBUG nova.compute.manager [-] [instance: df2576da-dc49-4ced-9462-fa71710e2e2a] Deallocating network for instance {{(pid=61806) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 535.955162] env[61806]: DEBUG nova.network.neutron [-] [instance: df2576da-dc49-4ced-9462-fa71710e2e2a] deallocate_for_instance() {{(pid=61806) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 535.979650] env[61806]: DEBUG nova.network.neutron [None req-92c8b578-fe59-4328-af10-ad8c2b7cd69b tempest-DeleteServersAdminTestJSON-1553667494 tempest-DeleteServersAdminTestJSON-1553667494-project-member] [instance: 22ce5e12-6c69-4316-b5f2-eed57a609369] Instance cache missing network info. {{(pid=61806) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 535.982273] env[61806]: DEBUG nova.network.neutron [-] [instance: df2576da-dc49-4ced-9462-fa71710e2e2a] Instance cache missing network info. {{(pid=61806) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 536.075112] env[61806]: DEBUG nova.compute.manager [None req-d8e4a121-0976-4eaa-b4c6-51deda86d081 tempest-InstanceActionsNegativeTestJSON-503825571 tempest-InstanceActionsNegativeTestJSON-503825571-project-member] [instance: a1aca638-e9c4-4e3b-9855-f92e7bd67204] Start spawning the instance on the hypervisor. {{(pid=61806) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 536.084364] env[61806]: DEBUG nova.network.neutron [None req-92c8b578-fe59-4328-af10-ad8c2b7cd69b tempest-DeleteServersAdminTestJSON-1553667494 tempest-DeleteServersAdminTestJSON-1553667494-project-member] [instance: 22ce5e12-6c69-4316-b5f2-eed57a609369] Updating instance_info_cache with network_info: [] {{(pid=61806) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 536.116017] env[61806]: DEBUG nova.virt.hardware [None req-d8e4a121-0976-4eaa-b4c6-51deda86d081 tempest-InstanceActionsNegativeTestJSON-503825571 tempest-InstanceActionsNegativeTestJSON-503825571-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-15T18:05:10Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-15T18:04:52Z,direct_url=,disk_format='vmdk',id=ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='84ba83fce2c349a6988173c5d6fc3b07',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-15T18:04:53Z,virtual_size=,visibility=), allow threads: False {{(pid=61806) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 536.116017] env[61806]: DEBUG nova.virt.hardware [None req-d8e4a121-0976-4eaa-b4c6-51deda86d081 tempest-InstanceActionsNegativeTestJSON-503825571 tempest-InstanceActionsNegativeTestJSON-503825571-project-member] Flavor limits 0:0:0 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 536.116017] env[61806]: DEBUG nova.virt.hardware [None req-d8e4a121-0976-4eaa-b4c6-51deda86d081 tempest-InstanceActionsNegativeTestJSON-503825571 tempest-InstanceActionsNegativeTestJSON-503825571-project-member] Image limits 0:0:0 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 536.116715] env[61806]: DEBUG nova.virt.hardware [None req-d8e4a121-0976-4eaa-b4c6-51deda86d081 tempest-InstanceActionsNegativeTestJSON-503825571 tempest-InstanceActionsNegativeTestJSON-503825571-project-member] Flavor pref 0:0:0 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 536.116895] env[61806]: DEBUG nova.virt.hardware [None req-d8e4a121-0976-4eaa-b4c6-51deda86d081 tempest-InstanceActionsNegativeTestJSON-503825571 tempest-InstanceActionsNegativeTestJSON-503825571-project-member] Image pref 0:0:0 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 536.119436] env[61806]: DEBUG nova.virt.hardware [None req-d8e4a121-0976-4eaa-b4c6-51deda86d081 tempest-InstanceActionsNegativeTestJSON-503825571 tempest-InstanceActionsNegativeTestJSON-503825571-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 536.119436] env[61806]: DEBUG nova.virt.hardware [None req-d8e4a121-0976-4eaa-b4c6-51deda86d081 tempest-InstanceActionsNegativeTestJSON-503825571 tempest-InstanceActionsNegativeTestJSON-503825571-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61806) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 536.119436] env[61806]: DEBUG nova.virt.hardware [None req-d8e4a121-0976-4eaa-b4c6-51deda86d081 tempest-InstanceActionsNegativeTestJSON-503825571 tempest-InstanceActionsNegativeTestJSON-503825571-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61806) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 536.119436] env[61806]: DEBUG nova.virt.hardware [None req-d8e4a121-0976-4eaa-b4c6-51deda86d081 tempest-InstanceActionsNegativeTestJSON-503825571 tempest-InstanceActionsNegativeTestJSON-503825571-project-member] Got 1 possible topologies {{(pid=61806) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 536.119436] env[61806]: DEBUG nova.virt.hardware [None req-d8e4a121-0976-4eaa-b4c6-51deda86d081 tempest-InstanceActionsNegativeTestJSON-503825571 tempest-InstanceActionsNegativeTestJSON-503825571-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61806) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 536.119702] env[61806]: DEBUG nova.virt.hardware [None req-d8e4a121-0976-4eaa-b4c6-51deda86d081 tempest-InstanceActionsNegativeTestJSON-503825571 tempest-InstanceActionsNegativeTestJSON-503825571-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61806) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 536.119702] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-51dafe3e-32d5-4efd-a8c0-41368e88e1a1 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 536.135391] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-61cdb727-c126-4f4a-a6fe-61ffbbb894ac {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 536.361348] env[61806]: DEBUG nova.compute.manager [req-ed1b7d9c-fddd-47ab-93e5-6e9059060081 req-f7d0b35e-41fe-4a53-92b7-e32e31dfa137 service nova] [instance: 52b19347-a08f-44a7-8aad-22a3af5e291d] Received event network-changed-d93797ab-56f6-476d-b147-8051439da3f2 {{(pid=61806) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 536.361348] env[61806]: DEBUG nova.compute.manager [req-ed1b7d9c-fddd-47ab-93e5-6e9059060081 req-f7d0b35e-41fe-4a53-92b7-e32e31dfa137 service nova] [instance: 52b19347-a08f-44a7-8aad-22a3af5e291d] Refreshing instance network info cache due to event network-changed-d93797ab-56f6-476d-b147-8051439da3f2. {{(pid=61806) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 536.361348] env[61806]: DEBUG oslo_concurrency.lockutils [req-ed1b7d9c-fddd-47ab-93e5-6e9059060081 req-f7d0b35e-41fe-4a53-92b7-e32e31dfa137 service nova] Acquiring lock "refresh_cache-52b19347-a08f-44a7-8aad-22a3af5e291d" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 536.361348] env[61806]: DEBUG oslo_concurrency.lockutils [req-ed1b7d9c-fddd-47ab-93e5-6e9059060081 req-f7d0b35e-41fe-4a53-92b7-e32e31dfa137 service nova] Acquired lock "refresh_cache-52b19347-a08f-44a7-8aad-22a3af5e291d" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 536.361348] env[61806]: DEBUG nova.network.neutron [req-ed1b7d9c-fddd-47ab-93e5-6e9059060081 req-f7d0b35e-41fe-4a53-92b7-e32e31dfa137 service nova] [instance: 52b19347-a08f-44a7-8aad-22a3af5e291d] Refreshing network info cache for port d93797ab-56f6-476d-b147-8051439da3f2 {{(pid=61806) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 536.438111] env[61806]: INFO nova.compute.manager [-] [instance: 52b19347-a08f-44a7-8aad-22a3af5e291d] Took 1.04 seconds to deallocate network for instance. [ 536.440244] env[61806]: DEBUG nova.compute.claims [None req-0858196d-347c-4201-9cb8-c7f147f5d99d tempest-MigrationsAdminTest-470098049 tempest-MigrationsAdminTest-470098049-project-member] [instance: 52b19347-a08f-44a7-8aad-22a3af5e291d] Aborting claim: {{(pid=61806) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 536.440468] env[61806]: DEBUG oslo_concurrency.lockutils [None req-0858196d-347c-4201-9cb8-c7f147f5d99d tempest-MigrationsAdminTest-470098049 tempest-MigrationsAdminTest-470098049-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 536.486860] env[61806]: DEBUG nova.network.neutron [-] [instance: df2576da-dc49-4ced-9462-fa71710e2e2a] Updating instance_info_cache with network_info: [] {{(pid=61806) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 536.587081] env[61806]: DEBUG oslo_concurrency.lockutils [None req-92c8b578-fe59-4328-af10-ad8c2b7cd69b tempest-DeleteServersAdminTestJSON-1553667494 tempest-DeleteServersAdminTestJSON-1553667494-project-member] Releasing lock "refresh_cache-22ce5e12-6c69-4316-b5f2-eed57a609369" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 536.590070] env[61806]: DEBUG nova.compute.manager [None req-92c8b578-fe59-4328-af10-ad8c2b7cd69b tempest-DeleteServersAdminTestJSON-1553667494 tempest-DeleteServersAdminTestJSON-1553667494-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61806) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 536.590070] env[61806]: DEBUG nova.compute.manager [None req-92c8b578-fe59-4328-af10-ad8c2b7cd69b tempest-DeleteServersAdminTestJSON-1553667494 tempest-DeleteServersAdminTestJSON-1553667494-project-member] [instance: 22ce5e12-6c69-4316-b5f2-eed57a609369] Deallocating network for instance {{(pid=61806) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 536.590070] env[61806]: DEBUG nova.network.neutron [None req-92c8b578-fe59-4328-af10-ad8c2b7cd69b tempest-DeleteServersAdminTestJSON-1553667494 tempest-DeleteServersAdminTestJSON-1553667494-project-member] [instance: 22ce5e12-6c69-4316-b5f2-eed57a609369] deallocate_for_instance() {{(pid=61806) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 536.612105] env[61806]: DEBUG nova.network.neutron [None req-92c8b578-fe59-4328-af10-ad8c2b7cd69b tempest-DeleteServersAdminTestJSON-1553667494 tempest-DeleteServersAdminTestJSON-1553667494-project-member] [instance: 22ce5e12-6c69-4316-b5f2-eed57a609369] Instance cache missing network info. {{(pid=61806) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 536.803753] env[61806]: DEBUG nova.compute.manager [req-8952b6ca-eb97-4a0a-8c69-13deb1d4ec2e req-707bf42e-d2cb-4d24-b9cb-a40347419c01 service nova] [instance: df2576da-dc49-4ced-9462-fa71710e2e2a] Received event network-changed-cab2110f-89a2-4c90-8780-f2deaabaf59a {{(pid=61806) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 536.804038] env[61806]: DEBUG nova.compute.manager [req-8952b6ca-eb97-4a0a-8c69-13deb1d4ec2e req-707bf42e-d2cb-4d24-b9cb-a40347419c01 service nova] [instance: df2576da-dc49-4ced-9462-fa71710e2e2a] Refreshing instance network info cache due to event network-changed-cab2110f-89a2-4c90-8780-f2deaabaf59a. {{(pid=61806) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 536.804484] env[61806]: DEBUG oslo_concurrency.lockutils [req-8952b6ca-eb97-4a0a-8c69-13deb1d4ec2e req-707bf42e-d2cb-4d24-b9cb-a40347419c01 service nova] Acquiring lock "refresh_cache-df2576da-dc49-4ced-9462-fa71710e2e2a" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 536.804764] env[61806]: DEBUG oslo_concurrency.lockutils [req-8952b6ca-eb97-4a0a-8c69-13deb1d4ec2e req-707bf42e-d2cb-4d24-b9cb-a40347419c01 service nova] Acquired lock "refresh_cache-df2576da-dc49-4ced-9462-fa71710e2e2a" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 536.805105] env[61806]: DEBUG nova.network.neutron [req-8952b6ca-eb97-4a0a-8c69-13deb1d4ec2e req-707bf42e-d2cb-4d24-b9cb-a40347419c01 service nova] [instance: df2576da-dc49-4ced-9462-fa71710e2e2a] Refreshing network info cache for port cab2110f-89a2-4c90-8780-f2deaabaf59a {{(pid=61806) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 536.892062] env[61806]: DEBUG nova.network.neutron [req-ed1b7d9c-fddd-47ab-93e5-6e9059060081 req-f7d0b35e-41fe-4a53-92b7-e32e31dfa137 service nova] [instance: 52b19347-a08f-44a7-8aad-22a3af5e291d] Instance cache missing network info. {{(pid=61806) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 536.989860] env[61806]: INFO nova.compute.manager [-] [instance: df2576da-dc49-4ced-9462-fa71710e2e2a] Took 1.04 seconds to deallocate network for instance. [ 536.993397] env[61806]: DEBUG nova.compute.claims [None req-3d02814b-d8ff-4079-980b-6c580616ce84 tempest-AttachInterfacesUnderV243Test-1683422373 tempest-AttachInterfacesUnderV243Test-1683422373-project-member] [instance: df2576da-dc49-4ced-9462-fa71710e2e2a] Aborting claim: {{(pid=61806) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 536.993397] env[61806]: DEBUG oslo_concurrency.lockutils [None req-3d02814b-d8ff-4079-980b-6c580616ce84 tempest-AttachInterfacesUnderV243Test-1683422373 tempest-AttachInterfacesUnderV243Test-1683422373-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 537.115490] env[61806]: DEBUG nova.network.neutron [None req-92c8b578-fe59-4328-af10-ad8c2b7cd69b tempest-DeleteServersAdminTestJSON-1553667494 tempest-DeleteServersAdminTestJSON-1553667494-project-member] [instance: 22ce5e12-6c69-4316-b5f2-eed57a609369] Updating instance_info_cache with network_info: [] {{(pid=61806) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 537.152069] env[61806]: DEBUG nova.network.neutron [req-ed1b7d9c-fddd-47ab-93e5-6e9059060081 req-f7d0b35e-41fe-4a53-92b7-e32e31dfa137 service nova] [instance: 52b19347-a08f-44a7-8aad-22a3af5e291d] Updating instance_info_cache with network_info: [] {{(pid=61806) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 537.205642] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-958b4cc1-9271-4082-8a19-949f860333b5 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 537.214988] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a28667bd-cdc3-4158-8a15-40ef5f869625 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 537.248692] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-628cecdb-f620-473b-bec2-d022c94cedcc {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 537.258447] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-29919222-a17c-4e60-adb3-404fa52a1c12 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 537.273400] env[61806]: DEBUG nova.compute.provider_tree [None req-960d1abd-1b06-4b5a-8ddd-4b463d493a45 tempest-ListServerFiltersTestJSON-241228521 tempest-ListServerFiltersTestJSON-241228521-project-member] Inventory has not changed in ProviderTree for provider: a2858be1-fd22-4e08-979e-87ad25293407 {{(pid=61806) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 537.359154] env[61806]: DEBUG nova.network.neutron [req-8952b6ca-eb97-4a0a-8c69-13deb1d4ec2e req-707bf42e-d2cb-4d24-b9cb-a40347419c01 service nova] [instance: df2576da-dc49-4ced-9462-fa71710e2e2a] Instance cache missing network info. {{(pid=61806) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 537.586846] env[61806]: DEBUG nova.network.neutron [req-8952b6ca-eb97-4a0a-8c69-13deb1d4ec2e req-707bf42e-d2cb-4d24-b9cb-a40347419c01 service nova] [instance: df2576da-dc49-4ced-9462-fa71710e2e2a] Updating instance_info_cache with network_info: [] {{(pid=61806) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 537.621406] env[61806]: INFO nova.compute.manager [None req-92c8b578-fe59-4328-af10-ad8c2b7cd69b tempest-DeleteServersAdminTestJSON-1553667494 tempest-DeleteServersAdminTestJSON-1553667494-project-member] [instance: 22ce5e12-6c69-4316-b5f2-eed57a609369] Took 1.03 seconds to deallocate network for instance. [ 537.655145] env[61806]: DEBUG oslo_concurrency.lockutils [req-ed1b7d9c-fddd-47ab-93e5-6e9059060081 req-f7d0b35e-41fe-4a53-92b7-e32e31dfa137 service nova] Releasing lock "refresh_cache-52b19347-a08f-44a7-8aad-22a3af5e291d" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 537.655814] env[61806]: DEBUG nova.compute.manager [req-ed1b7d9c-fddd-47ab-93e5-6e9059060081 req-f7d0b35e-41fe-4a53-92b7-e32e31dfa137 service nova] [instance: 52b19347-a08f-44a7-8aad-22a3af5e291d] Received event network-vif-deleted-d93797ab-56f6-476d-b147-8051439da3f2 {{(pid=61806) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 537.778241] env[61806]: DEBUG nova.scheduler.client.report [None req-960d1abd-1b06-4b5a-8ddd-4b463d493a45 tempest-ListServerFiltersTestJSON-241228521 tempest-ListServerFiltersTestJSON-241228521-project-member] Inventory has not changed for provider a2858be1-fd22-4e08-979e-87ad25293407 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 139, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61806) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 538.092019] env[61806]: DEBUG oslo_concurrency.lockutils [req-8952b6ca-eb97-4a0a-8c69-13deb1d4ec2e req-707bf42e-d2cb-4d24-b9cb-a40347419c01 service nova] Releasing lock "refresh_cache-df2576da-dc49-4ced-9462-fa71710e2e2a" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 538.092019] env[61806]: DEBUG nova.compute.manager [req-8952b6ca-eb97-4a0a-8c69-13deb1d4ec2e req-707bf42e-d2cb-4d24-b9cb-a40347419c01 service nova] [instance: df2576da-dc49-4ced-9462-fa71710e2e2a] Received event network-vif-deleted-cab2110f-89a2-4c90-8780-f2deaabaf59a {{(pid=61806) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 538.284377] env[61806]: DEBUG oslo_concurrency.lockutils [None req-960d1abd-1b06-4b5a-8ddd-4b463d493a45 tempest-ListServerFiltersTestJSON-241228521 tempest-ListServerFiltersTestJSON-241228521-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.395s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 538.288174] env[61806]: DEBUG nova.compute.manager [None req-960d1abd-1b06-4b5a-8ddd-4b463d493a45 tempest-ListServerFiltersTestJSON-241228521 tempest-ListServerFiltersTestJSON-241228521-project-member] [instance: 30e44576-dcd3-4036-9070-9032b775c90a] Start building networks asynchronously for instance. {{(pid=61806) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 538.292707] env[61806]: DEBUG oslo_concurrency.lockutils [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 10.236s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 538.292922] env[61806]: DEBUG oslo_concurrency.lockutils [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 538.293104] env[61806]: DEBUG nova.compute.resource_tracker [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=61806) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 538.293942] env[61806]: DEBUG oslo_concurrency.lockutils [None req-7d4ae408-54ff-4248-838b-a36cbf45b0c5 tempest-ListServerFiltersTestJSON-241228521 tempest-ListServerFiltersTestJSON-241228521-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 8.453s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 538.299159] env[61806]: INFO nova.compute.claims [None req-7d4ae408-54ff-4248-838b-a36cbf45b0c5 tempest-ListServerFiltersTestJSON-241228521 tempest-ListServerFiltersTestJSON-241228521-project-member] [instance: 9ddec22f-0fce-4d28-b97e-92b179f829f6] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 538.307219] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-08d7bb79-96fc-4036-a967-4193ccfb3e06 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 538.314913] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e1708def-2eb9-4454-a972-e5036b76cee3 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 538.336052] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-57f5ee22-2486-40a3-87a1-de8b7159df5f {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 538.345590] env[61806]: DEBUG oslo_concurrency.lockutils [None req-5c11e673-bd02-4846-9be6-b5679ed1117d tempest-VolumesAssistedSnapshotsTest-2037321669 tempest-VolumesAssistedSnapshotsTest-2037321669-project-member] Acquiring lock "76722306-a9ad-4ef9-b579-d958c36cf6bc" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 538.345840] env[61806]: DEBUG oslo_concurrency.lockutils [None req-5c11e673-bd02-4846-9be6-b5679ed1117d tempest-VolumesAssistedSnapshotsTest-2037321669 tempest-VolumesAssistedSnapshotsTest-2037321669-project-member] Lock "76722306-a9ad-4ef9-b579-d958c36cf6bc" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 538.346853] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0620be3d-ef42-4c59-8fbe-3a506b11f144 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 538.380284] env[61806]: DEBUG nova.compute.resource_tracker [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=181450MB free_disk=139GB free_vcpus=48 pci_devices=None {{(pid=61806) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 538.380523] env[61806]: DEBUG oslo_concurrency.lockutils [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 538.671739] env[61806]: INFO nova.scheduler.client.report [None req-92c8b578-fe59-4328-af10-ad8c2b7cd69b tempest-DeleteServersAdminTestJSON-1553667494 tempest-DeleteServersAdminTestJSON-1553667494-project-member] Deleted allocations for instance 22ce5e12-6c69-4316-b5f2-eed57a609369 [ 538.805486] env[61806]: DEBUG nova.compute.utils [None req-960d1abd-1b06-4b5a-8ddd-4b463d493a45 tempest-ListServerFiltersTestJSON-241228521 tempest-ListServerFiltersTestJSON-241228521-project-member] Using /dev/sd instead of None {{(pid=61806) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 538.811401] env[61806]: DEBUG nova.compute.manager [None req-960d1abd-1b06-4b5a-8ddd-4b463d493a45 tempest-ListServerFiltersTestJSON-241228521 tempest-ListServerFiltersTestJSON-241228521-project-member] [instance: 30e44576-dcd3-4036-9070-9032b775c90a] Allocating IP information in the background. {{(pid=61806) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 538.811914] env[61806]: DEBUG nova.network.neutron [None req-960d1abd-1b06-4b5a-8ddd-4b463d493a45 tempest-ListServerFiltersTestJSON-241228521 tempest-ListServerFiltersTestJSON-241228521-project-member] [instance: 30e44576-dcd3-4036-9070-9032b775c90a] allocate_for_instance() {{(pid=61806) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 538.928851] env[61806]: DEBUG nova.policy [None req-960d1abd-1b06-4b5a-8ddd-4b463d493a45 tempest-ListServerFiltersTestJSON-241228521 tempest-ListServerFiltersTestJSON-241228521-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'de9850b7a2714a5fa0eccc2dfe2af770', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '8e08874243144f6d871ac98314e24d3c', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61806) authorize /opt/stack/nova/nova/policy.py:201}} [ 539.010334] env[61806]: DEBUG oslo_concurrency.lockutils [None req-b50d0802-a025-4c5c-b8e9-95dd624e7ff5 tempest-ServersWithSpecificFlavorTestJSON-815343160 tempest-ServersWithSpecificFlavorTestJSON-815343160-project-member] Acquiring lock "b68b509e-96dc-489d-9cf1-93ba91095af1" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 539.011359] env[61806]: DEBUG oslo_concurrency.lockutils [None req-b50d0802-a025-4c5c-b8e9-95dd624e7ff5 tempest-ServersWithSpecificFlavorTestJSON-815343160 tempest-ServersWithSpecificFlavorTestJSON-815343160-project-member] Lock "b68b509e-96dc-489d-9cf1-93ba91095af1" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 539.183792] env[61806]: ERROR nova.compute.manager [None req-d8e4a121-0976-4eaa-b4c6-51deda86d081 tempest-InstanceActionsNegativeTestJSON-503825571 tempest-InstanceActionsNegativeTestJSON-503825571-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 6cc2dbc9-ee9c-42fa-b51e-2fd8a2c5b6d9, please check neutron logs for more information. [ 539.183792] env[61806]: ERROR nova.compute.manager Traceback (most recent call last): [ 539.183792] env[61806]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 539.183792] env[61806]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 539.183792] env[61806]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 539.183792] env[61806]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 539.183792] env[61806]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 539.183792] env[61806]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 539.183792] env[61806]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 539.183792] env[61806]: ERROR nova.compute.manager self.force_reraise() [ 539.183792] env[61806]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 539.183792] env[61806]: ERROR nova.compute.manager raise self.value [ 539.183792] env[61806]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 539.183792] env[61806]: ERROR nova.compute.manager updated_port = self._update_port( [ 539.183792] env[61806]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 539.183792] env[61806]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 539.184447] env[61806]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 539.184447] env[61806]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 539.184447] env[61806]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 6cc2dbc9-ee9c-42fa-b51e-2fd8a2c5b6d9, please check neutron logs for more information. [ 539.184447] env[61806]: ERROR nova.compute.manager [ 539.184447] env[61806]: Traceback (most recent call last): [ 539.184447] env[61806]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 539.184447] env[61806]: listener.cb(fileno) [ 539.184447] env[61806]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 539.184447] env[61806]: result = function(*args, **kwargs) [ 539.184447] env[61806]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 539.184447] env[61806]: return func(*args, **kwargs) [ 539.184447] env[61806]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 539.184447] env[61806]: raise e [ 539.184447] env[61806]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 539.184447] env[61806]: nwinfo = self.network_api.allocate_for_instance( [ 539.184447] env[61806]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 539.184447] env[61806]: created_port_ids = self._update_ports_for_instance( [ 539.184447] env[61806]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 539.184447] env[61806]: with excutils.save_and_reraise_exception(): [ 539.184447] env[61806]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 539.184447] env[61806]: self.force_reraise() [ 539.184447] env[61806]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 539.184447] env[61806]: raise self.value [ 539.184447] env[61806]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 539.184447] env[61806]: updated_port = self._update_port( [ 539.184447] env[61806]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 539.184447] env[61806]: _ensure_no_port_binding_failure(port) [ 539.184447] env[61806]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 539.184447] env[61806]: raise exception.PortBindingFailed(port_id=port['id']) [ 539.185207] env[61806]: nova.exception.PortBindingFailed: Binding failed for port 6cc2dbc9-ee9c-42fa-b51e-2fd8a2c5b6d9, please check neutron logs for more information. [ 539.185207] env[61806]: Removing descriptor: 17 [ 539.185207] env[61806]: DEBUG oslo_concurrency.lockutils [None req-92c8b578-fe59-4328-af10-ad8c2b7cd69b tempest-DeleteServersAdminTestJSON-1553667494 tempest-DeleteServersAdminTestJSON-1553667494-project-member] Lock "22ce5e12-6c69-4316-b5f2-eed57a609369" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 22.355s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 539.185207] env[61806]: ERROR nova.compute.manager [None req-d8e4a121-0976-4eaa-b4c6-51deda86d081 tempest-InstanceActionsNegativeTestJSON-503825571 tempest-InstanceActionsNegativeTestJSON-503825571-project-member] [instance: a1aca638-e9c4-4e3b-9855-f92e7bd67204] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 6cc2dbc9-ee9c-42fa-b51e-2fd8a2c5b6d9, please check neutron logs for more information. [ 539.185207] env[61806]: ERROR nova.compute.manager [instance: a1aca638-e9c4-4e3b-9855-f92e7bd67204] Traceback (most recent call last): [ 539.185207] env[61806]: ERROR nova.compute.manager [instance: a1aca638-e9c4-4e3b-9855-f92e7bd67204] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 539.185207] env[61806]: ERROR nova.compute.manager [instance: a1aca638-e9c4-4e3b-9855-f92e7bd67204] yield resources [ 539.185207] env[61806]: ERROR nova.compute.manager [instance: a1aca638-e9c4-4e3b-9855-f92e7bd67204] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 539.185207] env[61806]: ERROR nova.compute.manager [instance: a1aca638-e9c4-4e3b-9855-f92e7bd67204] self.driver.spawn(context, instance, image_meta, [ 539.185207] env[61806]: ERROR nova.compute.manager [instance: a1aca638-e9c4-4e3b-9855-f92e7bd67204] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 539.185504] env[61806]: ERROR nova.compute.manager [instance: a1aca638-e9c4-4e3b-9855-f92e7bd67204] self._vmops.spawn(context, instance, image_meta, injected_files, [ 539.185504] env[61806]: ERROR nova.compute.manager [instance: a1aca638-e9c4-4e3b-9855-f92e7bd67204] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 539.185504] env[61806]: ERROR nova.compute.manager [instance: a1aca638-e9c4-4e3b-9855-f92e7bd67204] vm_ref = self.build_virtual_machine(instance, [ 539.185504] env[61806]: ERROR nova.compute.manager [instance: a1aca638-e9c4-4e3b-9855-f92e7bd67204] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 539.185504] env[61806]: ERROR nova.compute.manager [instance: a1aca638-e9c4-4e3b-9855-f92e7bd67204] vif_infos = vmwarevif.get_vif_info(self._session, [ 539.185504] env[61806]: ERROR nova.compute.manager [instance: a1aca638-e9c4-4e3b-9855-f92e7bd67204] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 539.185504] env[61806]: ERROR nova.compute.manager [instance: a1aca638-e9c4-4e3b-9855-f92e7bd67204] for vif in network_info: [ 539.185504] env[61806]: ERROR nova.compute.manager [instance: a1aca638-e9c4-4e3b-9855-f92e7bd67204] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 539.185504] env[61806]: ERROR nova.compute.manager [instance: a1aca638-e9c4-4e3b-9855-f92e7bd67204] return self._sync_wrapper(fn, *args, **kwargs) [ 539.185504] env[61806]: ERROR nova.compute.manager [instance: a1aca638-e9c4-4e3b-9855-f92e7bd67204] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 539.185504] env[61806]: ERROR nova.compute.manager [instance: a1aca638-e9c4-4e3b-9855-f92e7bd67204] self.wait() [ 539.185504] env[61806]: ERROR nova.compute.manager [instance: a1aca638-e9c4-4e3b-9855-f92e7bd67204] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 539.185504] env[61806]: ERROR nova.compute.manager [instance: a1aca638-e9c4-4e3b-9855-f92e7bd67204] self[:] = self._gt.wait() [ 539.185829] env[61806]: ERROR nova.compute.manager [instance: a1aca638-e9c4-4e3b-9855-f92e7bd67204] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 539.185829] env[61806]: ERROR nova.compute.manager [instance: a1aca638-e9c4-4e3b-9855-f92e7bd67204] return self._exit_event.wait() [ 539.185829] env[61806]: ERROR nova.compute.manager [instance: a1aca638-e9c4-4e3b-9855-f92e7bd67204] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 539.185829] env[61806]: ERROR nova.compute.manager [instance: a1aca638-e9c4-4e3b-9855-f92e7bd67204] result = hub.switch() [ 539.185829] env[61806]: ERROR nova.compute.manager [instance: a1aca638-e9c4-4e3b-9855-f92e7bd67204] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 539.185829] env[61806]: ERROR nova.compute.manager [instance: a1aca638-e9c4-4e3b-9855-f92e7bd67204] return self.greenlet.switch() [ 539.185829] env[61806]: ERROR nova.compute.manager [instance: a1aca638-e9c4-4e3b-9855-f92e7bd67204] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 539.185829] env[61806]: ERROR nova.compute.manager [instance: a1aca638-e9c4-4e3b-9855-f92e7bd67204] result = function(*args, **kwargs) [ 539.185829] env[61806]: ERROR nova.compute.manager [instance: a1aca638-e9c4-4e3b-9855-f92e7bd67204] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 539.185829] env[61806]: ERROR nova.compute.manager [instance: a1aca638-e9c4-4e3b-9855-f92e7bd67204] return func(*args, **kwargs) [ 539.185829] env[61806]: ERROR nova.compute.manager [instance: a1aca638-e9c4-4e3b-9855-f92e7bd67204] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 539.185829] env[61806]: ERROR nova.compute.manager [instance: a1aca638-e9c4-4e3b-9855-f92e7bd67204] raise e [ 539.185829] env[61806]: ERROR nova.compute.manager [instance: a1aca638-e9c4-4e3b-9855-f92e7bd67204] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 539.186179] env[61806]: ERROR nova.compute.manager [instance: a1aca638-e9c4-4e3b-9855-f92e7bd67204] nwinfo = self.network_api.allocate_for_instance( [ 539.186179] env[61806]: ERROR nova.compute.manager [instance: a1aca638-e9c4-4e3b-9855-f92e7bd67204] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 539.186179] env[61806]: ERROR nova.compute.manager [instance: a1aca638-e9c4-4e3b-9855-f92e7bd67204] created_port_ids = self._update_ports_for_instance( [ 539.186179] env[61806]: ERROR nova.compute.manager [instance: a1aca638-e9c4-4e3b-9855-f92e7bd67204] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 539.186179] env[61806]: ERROR nova.compute.manager [instance: a1aca638-e9c4-4e3b-9855-f92e7bd67204] with excutils.save_and_reraise_exception(): [ 539.186179] env[61806]: ERROR nova.compute.manager [instance: a1aca638-e9c4-4e3b-9855-f92e7bd67204] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 539.186179] env[61806]: ERROR nova.compute.manager [instance: a1aca638-e9c4-4e3b-9855-f92e7bd67204] self.force_reraise() [ 539.186179] env[61806]: ERROR nova.compute.manager [instance: a1aca638-e9c4-4e3b-9855-f92e7bd67204] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 539.186179] env[61806]: ERROR nova.compute.manager [instance: a1aca638-e9c4-4e3b-9855-f92e7bd67204] raise self.value [ 539.186179] env[61806]: ERROR nova.compute.manager [instance: a1aca638-e9c4-4e3b-9855-f92e7bd67204] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 539.186179] env[61806]: ERROR nova.compute.manager [instance: a1aca638-e9c4-4e3b-9855-f92e7bd67204] updated_port = self._update_port( [ 539.186179] env[61806]: ERROR nova.compute.manager [instance: a1aca638-e9c4-4e3b-9855-f92e7bd67204] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 539.186179] env[61806]: ERROR nova.compute.manager [instance: a1aca638-e9c4-4e3b-9855-f92e7bd67204] _ensure_no_port_binding_failure(port) [ 539.186503] env[61806]: ERROR nova.compute.manager [instance: a1aca638-e9c4-4e3b-9855-f92e7bd67204] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 539.186503] env[61806]: ERROR nova.compute.manager [instance: a1aca638-e9c4-4e3b-9855-f92e7bd67204] raise exception.PortBindingFailed(port_id=port['id']) [ 539.186503] env[61806]: ERROR nova.compute.manager [instance: a1aca638-e9c4-4e3b-9855-f92e7bd67204] nova.exception.PortBindingFailed: Binding failed for port 6cc2dbc9-ee9c-42fa-b51e-2fd8a2c5b6d9, please check neutron logs for more information. [ 539.186503] env[61806]: ERROR nova.compute.manager [instance: a1aca638-e9c4-4e3b-9855-f92e7bd67204] [ 539.186503] env[61806]: INFO nova.compute.manager [None req-d8e4a121-0976-4eaa-b4c6-51deda86d081 tempest-InstanceActionsNegativeTestJSON-503825571 tempest-InstanceActionsNegativeTestJSON-503825571-project-member] [instance: a1aca638-e9c4-4e3b-9855-f92e7bd67204] Terminating instance [ 539.190095] env[61806]: DEBUG oslo_concurrency.lockutils [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Lock "22ce5e12-6c69-4316-b5f2-eed57a609369" acquired by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: waited 18.079s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 539.190095] env[61806]: INFO nova.compute.manager [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] [instance: 22ce5e12-6c69-4316-b5f2-eed57a609369] During sync_power_state the instance has a pending task (networking). Skip. [ 539.196153] env[61806]: DEBUG oslo_concurrency.lockutils [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Lock "22ce5e12-6c69-4316-b5f2-eed57a609369" "released" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: held 0.000s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 539.197552] env[61806]: DEBUG oslo_concurrency.lockutils [None req-d8e4a121-0976-4eaa-b4c6-51deda86d081 tempest-InstanceActionsNegativeTestJSON-503825571 tempest-InstanceActionsNegativeTestJSON-503825571-project-member] Acquiring lock "refresh_cache-a1aca638-e9c4-4e3b-9855-f92e7bd67204" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 539.197552] env[61806]: DEBUG oslo_concurrency.lockutils [None req-d8e4a121-0976-4eaa-b4c6-51deda86d081 tempest-InstanceActionsNegativeTestJSON-503825571 tempest-InstanceActionsNegativeTestJSON-503825571-project-member] Acquired lock "refresh_cache-a1aca638-e9c4-4e3b-9855-f92e7bd67204" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 539.197552] env[61806]: DEBUG nova.network.neutron [None req-d8e4a121-0976-4eaa-b4c6-51deda86d081 tempest-InstanceActionsNegativeTestJSON-503825571 tempest-InstanceActionsNegativeTestJSON-503825571-project-member] [instance: a1aca638-e9c4-4e3b-9855-f92e7bd67204] Building network info cache for instance {{(pid=61806) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 539.312536] env[61806]: DEBUG nova.compute.manager [None req-960d1abd-1b06-4b5a-8ddd-4b463d493a45 tempest-ListServerFiltersTestJSON-241228521 tempest-ListServerFiltersTestJSON-241228521-project-member] [instance: 30e44576-dcd3-4036-9070-9032b775c90a] Start building block device mappings for instance. {{(pid=61806) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 539.514286] env[61806]: DEBUG nova.network.neutron [None req-960d1abd-1b06-4b5a-8ddd-4b463d493a45 tempest-ListServerFiltersTestJSON-241228521 tempest-ListServerFiltersTestJSON-241228521-project-member] [instance: 30e44576-dcd3-4036-9070-9032b775c90a] Successfully created port: 5990bc22-0a46-437d-b518-9a9263b60655 {{(pid=61806) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 539.610427] env[61806]: DEBUG nova.compute.manager [req-b49ed356-2608-4fb5-bb84-c435e9800605 req-46c03413-b6ea-4b48-9d20-bfccf145606a service nova] [instance: a1aca638-e9c4-4e3b-9855-f92e7bd67204] Received event network-changed-6cc2dbc9-ee9c-42fa-b51e-2fd8a2c5b6d9 {{(pid=61806) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 539.610427] env[61806]: DEBUG nova.compute.manager [req-b49ed356-2608-4fb5-bb84-c435e9800605 req-46c03413-b6ea-4b48-9d20-bfccf145606a service nova] [instance: a1aca638-e9c4-4e3b-9855-f92e7bd67204] Refreshing instance network info cache due to event network-changed-6cc2dbc9-ee9c-42fa-b51e-2fd8a2c5b6d9. {{(pid=61806) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 539.611122] env[61806]: DEBUG oslo_concurrency.lockutils [req-b49ed356-2608-4fb5-bb84-c435e9800605 req-46c03413-b6ea-4b48-9d20-bfccf145606a service nova] Acquiring lock "refresh_cache-a1aca638-e9c4-4e3b-9855-f92e7bd67204" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 539.683328] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dc01e7f7-8a94-40f1-bd32-dceb13dcab92 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 539.693304] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-33e5a5d4-695e-4006-897c-68a257e09347 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 539.698609] env[61806]: DEBUG nova.compute.manager [None req-be9179b6-8d17-4f04-8823-d48174fcc5c1 tempest-InstanceActionsV221TestJSON-379107765 tempest-InstanceActionsV221TestJSON-379107765-project-member] [instance: 123f15be-3b56-4999-8010-40d2e86a6813] Starting instance... {{(pid=61806) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 539.736839] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8ff5fa4e-f0be-48b5-8bbe-ec25c8ab7be8 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 539.745404] env[61806]: DEBUG nova.network.neutron [None req-d8e4a121-0976-4eaa-b4c6-51deda86d081 tempest-InstanceActionsNegativeTestJSON-503825571 tempest-InstanceActionsNegativeTestJSON-503825571-project-member] [instance: a1aca638-e9c4-4e3b-9855-f92e7bd67204] Instance cache missing network info. {{(pid=61806) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 539.749148] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f57b647a-1a4b-44d9-9509-2f974e7d69f6 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 539.769397] env[61806]: DEBUG nova.compute.provider_tree [None req-7d4ae408-54ff-4248-838b-a36cbf45b0c5 tempest-ListServerFiltersTestJSON-241228521 tempest-ListServerFiltersTestJSON-241228521-project-member] Inventory has not changed in ProviderTree for provider: a2858be1-fd22-4e08-979e-87ad25293407 {{(pid=61806) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 539.930749] env[61806]: DEBUG nova.network.neutron [None req-d8e4a121-0976-4eaa-b4c6-51deda86d081 tempest-InstanceActionsNegativeTestJSON-503825571 tempest-InstanceActionsNegativeTestJSON-503825571-project-member] [instance: a1aca638-e9c4-4e3b-9855-f92e7bd67204] Updating instance_info_cache with network_info: [] {{(pid=61806) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 540.123795] env[61806]: DEBUG oslo_concurrency.lockutils [None req-9101cb9e-b191-4241-8bc1-9b126210ca50 tempest-ListImageFiltersTestJSON-680116525 tempest-ListImageFiltersTestJSON-680116525-project-member] Acquiring lock "7e730579-8d0c-4226-86bc-292db16c49ee" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 540.124331] env[61806]: DEBUG oslo_concurrency.lockutils [None req-9101cb9e-b191-4241-8bc1-9b126210ca50 tempest-ListImageFiltersTestJSON-680116525 tempest-ListImageFiltersTestJSON-680116525-project-member] Lock "7e730579-8d0c-4226-86bc-292db16c49ee" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 540.249445] env[61806]: DEBUG oslo_concurrency.lockutils [None req-be9179b6-8d17-4f04-8823-d48174fcc5c1 tempest-InstanceActionsV221TestJSON-379107765 tempest-InstanceActionsV221TestJSON-379107765-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 540.273543] env[61806]: DEBUG nova.scheduler.client.report [None req-7d4ae408-54ff-4248-838b-a36cbf45b0c5 tempest-ListServerFiltersTestJSON-241228521 tempest-ListServerFiltersTestJSON-241228521-project-member] Inventory has not changed for provider a2858be1-fd22-4e08-979e-87ad25293407 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 139, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61806) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 540.333293] env[61806]: DEBUG nova.compute.manager [None req-960d1abd-1b06-4b5a-8ddd-4b463d493a45 tempest-ListServerFiltersTestJSON-241228521 tempest-ListServerFiltersTestJSON-241228521-project-member] [instance: 30e44576-dcd3-4036-9070-9032b775c90a] Start spawning the instance on the hypervisor. {{(pid=61806) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 540.362551] env[61806]: DEBUG nova.virt.hardware [None req-960d1abd-1b06-4b5a-8ddd-4b463d493a45 tempest-ListServerFiltersTestJSON-241228521 tempest-ListServerFiltersTestJSON-241228521-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-15T18:05:10Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-15T18:04:52Z,direct_url=,disk_format='vmdk',id=ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='84ba83fce2c349a6988173c5d6fc3b07',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-15T18:04:53Z,virtual_size=,visibility=), allow threads: False {{(pid=61806) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 540.362799] env[61806]: DEBUG nova.virt.hardware [None req-960d1abd-1b06-4b5a-8ddd-4b463d493a45 tempest-ListServerFiltersTestJSON-241228521 tempest-ListServerFiltersTestJSON-241228521-project-member] Flavor limits 0:0:0 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 540.363098] env[61806]: DEBUG nova.virt.hardware [None req-960d1abd-1b06-4b5a-8ddd-4b463d493a45 tempest-ListServerFiltersTestJSON-241228521 tempest-ListServerFiltersTestJSON-241228521-project-member] Image limits 0:0:0 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 540.363173] env[61806]: DEBUG nova.virt.hardware [None req-960d1abd-1b06-4b5a-8ddd-4b463d493a45 tempest-ListServerFiltersTestJSON-241228521 tempest-ListServerFiltersTestJSON-241228521-project-member] Flavor pref 0:0:0 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 540.363320] env[61806]: DEBUG nova.virt.hardware [None req-960d1abd-1b06-4b5a-8ddd-4b463d493a45 tempest-ListServerFiltersTestJSON-241228521 tempest-ListServerFiltersTestJSON-241228521-project-member] Image pref 0:0:0 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 540.363426] env[61806]: DEBUG nova.virt.hardware [None req-960d1abd-1b06-4b5a-8ddd-4b463d493a45 tempest-ListServerFiltersTestJSON-241228521 tempest-ListServerFiltersTestJSON-241228521-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 540.363629] env[61806]: DEBUG nova.virt.hardware [None req-960d1abd-1b06-4b5a-8ddd-4b463d493a45 tempest-ListServerFiltersTestJSON-241228521 tempest-ListServerFiltersTestJSON-241228521-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61806) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 540.363769] env[61806]: DEBUG nova.virt.hardware [None req-960d1abd-1b06-4b5a-8ddd-4b463d493a45 tempest-ListServerFiltersTestJSON-241228521 tempest-ListServerFiltersTestJSON-241228521-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61806) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 540.363933] env[61806]: DEBUG nova.virt.hardware [None req-960d1abd-1b06-4b5a-8ddd-4b463d493a45 tempest-ListServerFiltersTestJSON-241228521 tempest-ListServerFiltersTestJSON-241228521-project-member] Got 1 possible topologies {{(pid=61806) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 540.364103] env[61806]: DEBUG nova.virt.hardware [None req-960d1abd-1b06-4b5a-8ddd-4b463d493a45 tempest-ListServerFiltersTestJSON-241228521 tempest-ListServerFiltersTestJSON-241228521-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61806) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 540.364302] env[61806]: DEBUG nova.virt.hardware [None req-960d1abd-1b06-4b5a-8ddd-4b463d493a45 tempest-ListServerFiltersTestJSON-241228521 tempest-ListServerFiltersTestJSON-241228521-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61806) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 540.365204] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-775c579c-a44e-4e67-ab19-d3cdadd2313e {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 540.373522] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8ed57156-3335-4341-8b39-9da5e234969f {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 540.434320] env[61806]: DEBUG oslo_concurrency.lockutils [None req-d8e4a121-0976-4eaa-b4c6-51deda86d081 tempest-InstanceActionsNegativeTestJSON-503825571 tempest-InstanceActionsNegativeTestJSON-503825571-project-member] Releasing lock "refresh_cache-a1aca638-e9c4-4e3b-9855-f92e7bd67204" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 540.435537] env[61806]: DEBUG nova.compute.manager [None req-d8e4a121-0976-4eaa-b4c6-51deda86d081 tempest-InstanceActionsNegativeTestJSON-503825571 tempest-InstanceActionsNegativeTestJSON-503825571-project-member] [instance: a1aca638-e9c4-4e3b-9855-f92e7bd67204] Start destroying the instance on the hypervisor. {{(pid=61806) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 540.437038] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-d8e4a121-0976-4eaa-b4c6-51deda86d081 tempest-InstanceActionsNegativeTestJSON-503825571 tempest-InstanceActionsNegativeTestJSON-503825571-project-member] [instance: a1aca638-e9c4-4e3b-9855-f92e7bd67204] Destroying instance {{(pid=61806) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 540.437038] env[61806]: DEBUG oslo_concurrency.lockutils [req-b49ed356-2608-4fb5-bb84-c435e9800605 req-46c03413-b6ea-4b48-9d20-bfccf145606a service nova] Acquired lock "refresh_cache-a1aca638-e9c4-4e3b-9855-f92e7bd67204" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 540.437172] env[61806]: DEBUG nova.network.neutron [req-b49ed356-2608-4fb5-bb84-c435e9800605 req-46c03413-b6ea-4b48-9d20-bfccf145606a service nova] [instance: a1aca638-e9c4-4e3b-9855-f92e7bd67204] Refreshing network info cache for port 6cc2dbc9-ee9c-42fa-b51e-2fd8a2c5b6d9 {{(pid=61806) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 540.439199] env[61806]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-f554720f-8db2-4b3f-9c7d-56f1091934b7 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 540.460459] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4d0b501a-611d-4b59-9251-b4c20b4363c1 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 540.491620] env[61806]: WARNING nova.virt.vmwareapi.vmops [None req-d8e4a121-0976-4eaa-b4c6-51deda86d081 tempest-InstanceActionsNegativeTestJSON-503825571 tempest-InstanceActionsNegativeTestJSON-503825571-project-member] [instance: a1aca638-e9c4-4e3b-9855-f92e7bd67204] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance a1aca638-e9c4-4e3b-9855-f92e7bd67204 could not be found. [ 540.492357] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-d8e4a121-0976-4eaa-b4c6-51deda86d081 tempest-InstanceActionsNegativeTestJSON-503825571 tempest-InstanceActionsNegativeTestJSON-503825571-project-member] [instance: a1aca638-e9c4-4e3b-9855-f92e7bd67204] Instance destroyed {{(pid=61806) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 540.492357] env[61806]: INFO nova.compute.manager [None req-d8e4a121-0976-4eaa-b4c6-51deda86d081 tempest-InstanceActionsNegativeTestJSON-503825571 tempest-InstanceActionsNegativeTestJSON-503825571-project-member] [instance: a1aca638-e9c4-4e3b-9855-f92e7bd67204] Took 0.06 seconds to destroy the instance on the hypervisor. [ 540.492357] env[61806]: DEBUG oslo.service.loopingcall [None req-d8e4a121-0976-4eaa-b4c6-51deda86d081 tempest-InstanceActionsNegativeTestJSON-503825571 tempest-InstanceActionsNegativeTestJSON-503825571-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61806) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 540.492584] env[61806]: DEBUG nova.compute.manager [-] [instance: a1aca638-e9c4-4e3b-9855-f92e7bd67204] Deallocating network for instance {{(pid=61806) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 540.492676] env[61806]: DEBUG nova.network.neutron [-] [instance: a1aca638-e9c4-4e3b-9855-f92e7bd67204] deallocate_for_instance() {{(pid=61806) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 540.536428] env[61806]: DEBUG nova.network.neutron [-] [instance: a1aca638-e9c4-4e3b-9855-f92e7bd67204] Instance cache missing network info. {{(pid=61806) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 540.783665] env[61806]: DEBUG oslo_concurrency.lockutils [None req-7d4ae408-54ff-4248-838b-a36cbf45b0c5 tempest-ListServerFiltersTestJSON-241228521 tempest-ListServerFiltersTestJSON-241228521-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.490s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 540.785016] env[61806]: DEBUG nova.compute.manager [None req-7d4ae408-54ff-4248-838b-a36cbf45b0c5 tempest-ListServerFiltersTestJSON-241228521 tempest-ListServerFiltersTestJSON-241228521-project-member] [instance: 9ddec22f-0fce-4d28-b97e-92b179f829f6] Start building networks asynchronously for instance. {{(pid=61806) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 540.788901] env[61806]: DEBUG oslo_concurrency.lockutils [None req-3568a2a1-f64b-42a9-bf53-df7d9370887d tempest-ImagesTestJSON-618852556 tempest-ImagesTestJSON-618852556-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 9.462s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 540.940386] env[61806]: ERROR nova.compute.manager [None req-960d1abd-1b06-4b5a-8ddd-4b463d493a45 tempest-ListServerFiltersTestJSON-241228521 tempest-ListServerFiltersTestJSON-241228521-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 5990bc22-0a46-437d-b518-9a9263b60655, please check neutron logs for more information. [ 540.940386] env[61806]: ERROR nova.compute.manager Traceback (most recent call last): [ 540.940386] env[61806]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 540.940386] env[61806]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 540.940386] env[61806]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 540.940386] env[61806]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 540.940386] env[61806]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 540.940386] env[61806]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 540.940386] env[61806]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 540.940386] env[61806]: ERROR nova.compute.manager self.force_reraise() [ 540.940386] env[61806]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 540.940386] env[61806]: ERROR nova.compute.manager raise self.value [ 540.940386] env[61806]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 540.940386] env[61806]: ERROR nova.compute.manager updated_port = self._update_port( [ 540.940386] env[61806]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 540.940386] env[61806]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 540.940843] env[61806]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 540.940843] env[61806]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 540.940843] env[61806]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 5990bc22-0a46-437d-b518-9a9263b60655, please check neutron logs for more information. [ 540.940843] env[61806]: ERROR nova.compute.manager [ 540.940843] env[61806]: Traceback (most recent call last): [ 540.940843] env[61806]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 540.940843] env[61806]: listener.cb(fileno) [ 540.940843] env[61806]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 540.940843] env[61806]: result = function(*args, **kwargs) [ 540.940843] env[61806]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 540.940843] env[61806]: return func(*args, **kwargs) [ 540.940843] env[61806]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 540.940843] env[61806]: raise e [ 540.940843] env[61806]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 540.940843] env[61806]: nwinfo = self.network_api.allocate_for_instance( [ 540.940843] env[61806]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 540.940843] env[61806]: created_port_ids = self._update_ports_for_instance( [ 540.940843] env[61806]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 540.940843] env[61806]: with excutils.save_and_reraise_exception(): [ 540.940843] env[61806]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 540.940843] env[61806]: self.force_reraise() [ 540.940843] env[61806]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 540.940843] env[61806]: raise self.value [ 540.940843] env[61806]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 540.940843] env[61806]: updated_port = self._update_port( [ 540.940843] env[61806]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 540.940843] env[61806]: _ensure_no_port_binding_failure(port) [ 540.940843] env[61806]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 540.940843] env[61806]: raise exception.PortBindingFailed(port_id=port['id']) [ 540.941671] env[61806]: nova.exception.PortBindingFailed: Binding failed for port 5990bc22-0a46-437d-b518-9a9263b60655, please check neutron logs for more information. [ 540.941671] env[61806]: Removing descriptor: 16 [ 540.941671] env[61806]: ERROR nova.compute.manager [None req-960d1abd-1b06-4b5a-8ddd-4b463d493a45 tempest-ListServerFiltersTestJSON-241228521 tempest-ListServerFiltersTestJSON-241228521-project-member] [instance: 30e44576-dcd3-4036-9070-9032b775c90a] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 5990bc22-0a46-437d-b518-9a9263b60655, please check neutron logs for more information. [ 540.941671] env[61806]: ERROR nova.compute.manager [instance: 30e44576-dcd3-4036-9070-9032b775c90a] Traceback (most recent call last): [ 540.941671] env[61806]: ERROR nova.compute.manager [instance: 30e44576-dcd3-4036-9070-9032b775c90a] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 540.941671] env[61806]: ERROR nova.compute.manager [instance: 30e44576-dcd3-4036-9070-9032b775c90a] yield resources [ 540.941671] env[61806]: ERROR nova.compute.manager [instance: 30e44576-dcd3-4036-9070-9032b775c90a] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 540.941671] env[61806]: ERROR nova.compute.manager [instance: 30e44576-dcd3-4036-9070-9032b775c90a] self.driver.spawn(context, instance, image_meta, [ 540.941671] env[61806]: ERROR nova.compute.manager [instance: 30e44576-dcd3-4036-9070-9032b775c90a] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 540.941671] env[61806]: ERROR nova.compute.manager [instance: 30e44576-dcd3-4036-9070-9032b775c90a] self._vmops.spawn(context, instance, image_meta, injected_files, [ 540.941671] env[61806]: ERROR nova.compute.manager [instance: 30e44576-dcd3-4036-9070-9032b775c90a] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 540.941671] env[61806]: ERROR nova.compute.manager [instance: 30e44576-dcd3-4036-9070-9032b775c90a] vm_ref = self.build_virtual_machine(instance, [ 540.941987] env[61806]: ERROR nova.compute.manager [instance: 30e44576-dcd3-4036-9070-9032b775c90a] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 540.941987] env[61806]: ERROR nova.compute.manager [instance: 30e44576-dcd3-4036-9070-9032b775c90a] vif_infos = vmwarevif.get_vif_info(self._session, [ 540.941987] env[61806]: ERROR nova.compute.manager [instance: 30e44576-dcd3-4036-9070-9032b775c90a] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 540.941987] env[61806]: ERROR nova.compute.manager [instance: 30e44576-dcd3-4036-9070-9032b775c90a] for vif in network_info: [ 540.941987] env[61806]: ERROR nova.compute.manager [instance: 30e44576-dcd3-4036-9070-9032b775c90a] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 540.941987] env[61806]: ERROR nova.compute.manager [instance: 30e44576-dcd3-4036-9070-9032b775c90a] return self._sync_wrapper(fn, *args, **kwargs) [ 540.941987] env[61806]: ERROR nova.compute.manager [instance: 30e44576-dcd3-4036-9070-9032b775c90a] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 540.941987] env[61806]: ERROR nova.compute.manager [instance: 30e44576-dcd3-4036-9070-9032b775c90a] self.wait() [ 540.941987] env[61806]: ERROR nova.compute.manager [instance: 30e44576-dcd3-4036-9070-9032b775c90a] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 540.941987] env[61806]: ERROR nova.compute.manager [instance: 30e44576-dcd3-4036-9070-9032b775c90a] self[:] = self._gt.wait() [ 540.941987] env[61806]: ERROR nova.compute.manager [instance: 30e44576-dcd3-4036-9070-9032b775c90a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 540.941987] env[61806]: ERROR nova.compute.manager [instance: 30e44576-dcd3-4036-9070-9032b775c90a] return self._exit_event.wait() [ 540.941987] env[61806]: ERROR nova.compute.manager [instance: 30e44576-dcd3-4036-9070-9032b775c90a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 540.942333] env[61806]: ERROR nova.compute.manager [instance: 30e44576-dcd3-4036-9070-9032b775c90a] result = hub.switch() [ 540.942333] env[61806]: ERROR nova.compute.manager [instance: 30e44576-dcd3-4036-9070-9032b775c90a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 540.942333] env[61806]: ERROR nova.compute.manager [instance: 30e44576-dcd3-4036-9070-9032b775c90a] return self.greenlet.switch() [ 540.942333] env[61806]: ERROR nova.compute.manager [instance: 30e44576-dcd3-4036-9070-9032b775c90a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 540.942333] env[61806]: ERROR nova.compute.manager [instance: 30e44576-dcd3-4036-9070-9032b775c90a] result = function(*args, **kwargs) [ 540.942333] env[61806]: ERROR nova.compute.manager [instance: 30e44576-dcd3-4036-9070-9032b775c90a] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 540.942333] env[61806]: ERROR nova.compute.manager [instance: 30e44576-dcd3-4036-9070-9032b775c90a] return func(*args, **kwargs) [ 540.942333] env[61806]: ERROR nova.compute.manager [instance: 30e44576-dcd3-4036-9070-9032b775c90a] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 540.942333] env[61806]: ERROR nova.compute.manager [instance: 30e44576-dcd3-4036-9070-9032b775c90a] raise e [ 540.942333] env[61806]: ERROR nova.compute.manager [instance: 30e44576-dcd3-4036-9070-9032b775c90a] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 540.942333] env[61806]: ERROR nova.compute.manager [instance: 30e44576-dcd3-4036-9070-9032b775c90a] nwinfo = self.network_api.allocate_for_instance( [ 540.942333] env[61806]: ERROR nova.compute.manager [instance: 30e44576-dcd3-4036-9070-9032b775c90a] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 540.942333] env[61806]: ERROR nova.compute.manager [instance: 30e44576-dcd3-4036-9070-9032b775c90a] created_port_ids = self._update_ports_for_instance( [ 540.942667] env[61806]: ERROR nova.compute.manager [instance: 30e44576-dcd3-4036-9070-9032b775c90a] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 540.942667] env[61806]: ERROR nova.compute.manager [instance: 30e44576-dcd3-4036-9070-9032b775c90a] with excutils.save_and_reraise_exception(): [ 540.942667] env[61806]: ERROR nova.compute.manager [instance: 30e44576-dcd3-4036-9070-9032b775c90a] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 540.942667] env[61806]: ERROR nova.compute.manager [instance: 30e44576-dcd3-4036-9070-9032b775c90a] self.force_reraise() [ 540.942667] env[61806]: ERROR nova.compute.manager [instance: 30e44576-dcd3-4036-9070-9032b775c90a] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 540.942667] env[61806]: ERROR nova.compute.manager [instance: 30e44576-dcd3-4036-9070-9032b775c90a] raise self.value [ 540.942667] env[61806]: ERROR nova.compute.manager [instance: 30e44576-dcd3-4036-9070-9032b775c90a] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 540.942667] env[61806]: ERROR nova.compute.manager [instance: 30e44576-dcd3-4036-9070-9032b775c90a] updated_port = self._update_port( [ 540.942667] env[61806]: ERROR nova.compute.manager [instance: 30e44576-dcd3-4036-9070-9032b775c90a] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 540.942667] env[61806]: ERROR nova.compute.manager [instance: 30e44576-dcd3-4036-9070-9032b775c90a] _ensure_no_port_binding_failure(port) [ 540.942667] env[61806]: ERROR nova.compute.manager [instance: 30e44576-dcd3-4036-9070-9032b775c90a] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 540.942667] env[61806]: ERROR nova.compute.manager [instance: 30e44576-dcd3-4036-9070-9032b775c90a] raise exception.PortBindingFailed(port_id=port['id']) [ 540.943087] env[61806]: ERROR nova.compute.manager [instance: 30e44576-dcd3-4036-9070-9032b775c90a] nova.exception.PortBindingFailed: Binding failed for port 5990bc22-0a46-437d-b518-9a9263b60655, please check neutron logs for more information. [ 540.943087] env[61806]: ERROR nova.compute.manager [instance: 30e44576-dcd3-4036-9070-9032b775c90a] [ 540.943087] env[61806]: INFO nova.compute.manager [None req-960d1abd-1b06-4b5a-8ddd-4b463d493a45 tempest-ListServerFiltersTestJSON-241228521 tempest-ListServerFiltersTestJSON-241228521-project-member] [instance: 30e44576-dcd3-4036-9070-9032b775c90a] Terminating instance [ 540.949050] env[61806]: DEBUG oslo_concurrency.lockutils [None req-960d1abd-1b06-4b5a-8ddd-4b463d493a45 tempest-ListServerFiltersTestJSON-241228521 tempest-ListServerFiltersTestJSON-241228521-project-member] Acquiring lock "refresh_cache-30e44576-dcd3-4036-9070-9032b775c90a" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 540.949050] env[61806]: DEBUG oslo_concurrency.lockutils [None req-960d1abd-1b06-4b5a-8ddd-4b463d493a45 tempest-ListServerFiltersTestJSON-241228521 tempest-ListServerFiltersTestJSON-241228521-project-member] Acquired lock "refresh_cache-30e44576-dcd3-4036-9070-9032b775c90a" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 540.949050] env[61806]: DEBUG nova.network.neutron [None req-960d1abd-1b06-4b5a-8ddd-4b463d493a45 tempest-ListServerFiltersTestJSON-241228521 tempest-ListServerFiltersTestJSON-241228521-project-member] [instance: 30e44576-dcd3-4036-9070-9032b775c90a] Building network info cache for instance {{(pid=61806) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 541.001142] env[61806]: DEBUG nova.network.neutron [req-b49ed356-2608-4fb5-bb84-c435e9800605 req-46c03413-b6ea-4b48-9d20-bfccf145606a service nova] [instance: a1aca638-e9c4-4e3b-9855-f92e7bd67204] Instance cache missing network info. {{(pid=61806) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 541.044101] env[61806]: DEBUG nova.network.neutron [-] [instance: a1aca638-e9c4-4e3b-9855-f92e7bd67204] Updating instance_info_cache with network_info: [] {{(pid=61806) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 541.301619] env[61806]: DEBUG nova.compute.utils [None req-7d4ae408-54ff-4248-838b-a36cbf45b0c5 tempest-ListServerFiltersTestJSON-241228521 tempest-ListServerFiltersTestJSON-241228521-project-member] Using /dev/sd instead of None {{(pid=61806) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 541.304512] env[61806]: DEBUG nova.compute.manager [None req-7d4ae408-54ff-4248-838b-a36cbf45b0c5 tempest-ListServerFiltersTestJSON-241228521 tempest-ListServerFiltersTestJSON-241228521-project-member] [instance: 9ddec22f-0fce-4d28-b97e-92b179f829f6] Allocating IP information in the background. {{(pid=61806) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 541.304512] env[61806]: DEBUG nova.network.neutron [None req-7d4ae408-54ff-4248-838b-a36cbf45b0c5 tempest-ListServerFiltersTestJSON-241228521 tempest-ListServerFiltersTestJSON-241228521-project-member] [instance: 9ddec22f-0fce-4d28-b97e-92b179f829f6] allocate_for_instance() {{(pid=61806) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 541.382961] env[61806]: DEBUG nova.policy [None req-7d4ae408-54ff-4248-838b-a36cbf45b0c5 tempest-ListServerFiltersTestJSON-241228521 tempest-ListServerFiltersTestJSON-241228521-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'de9850b7a2714a5fa0eccc2dfe2af770', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '8e08874243144f6d871ac98314e24d3c', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61806) authorize /opt/stack/nova/nova/policy.py:201}} [ 541.482838] env[61806]: DEBUG nova.network.neutron [None req-960d1abd-1b06-4b5a-8ddd-4b463d493a45 tempest-ListServerFiltersTestJSON-241228521 tempest-ListServerFiltersTestJSON-241228521-project-member] [instance: 30e44576-dcd3-4036-9070-9032b775c90a] Instance cache missing network info. {{(pid=61806) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 541.487577] env[61806]: DEBUG nova.network.neutron [req-b49ed356-2608-4fb5-bb84-c435e9800605 req-46c03413-b6ea-4b48-9d20-bfccf145606a service nova] [instance: a1aca638-e9c4-4e3b-9855-f92e7bd67204] Updating instance_info_cache with network_info: [] {{(pid=61806) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 541.546861] env[61806]: INFO nova.compute.manager [-] [instance: a1aca638-e9c4-4e3b-9855-f92e7bd67204] Took 1.05 seconds to deallocate network for instance. [ 541.548399] env[61806]: DEBUG nova.compute.manager [req-21a52603-630d-4c44-a737-1e529dafd7b2 req-f323930e-b5be-4c38-b5b6-9a33416721db service nova] [instance: 30e44576-dcd3-4036-9070-9032b775c90a] Received event network-changed-5990bc22-0a46-437d-b518-9a9263b60655 {{(pid=61806) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 541.548399] env[61806]: DEBUG nova.compute.manager [req-21a52603-630d-4c44-a737-1e529dafd7b2 req-f323930e-b5be-4c38-b5b6-9a33416721db service nova] [instance: 30e44576-dcd3-4036-9070-9032b775c90a] Refreshing instance network info cache due to event network-changed-5990bc22-0a46-437d-b518-9a9263b60655. {{(pid=61806) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 541.548698] env[61806]: DEBUG oslo_concurrency.lockutils [req-21a52603-630d-4c44-a737-1e529dafd7b2 req-f323930e-b5be-4c38-b5b6-9a33416721db service nova] Acquiring lock "refresh_cache-30e44576-dcd3-4036-9070-9032b775c90a" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 541.551039] env[61806]: DEBUG nova.compute.claims [None req-d8e4a121-0976-4eaa-b4c6-51deda86d081 tempest-InstanceActionsNegativeTestJSON-503825571 tempest-InstanceActionsNegativeTestJSON-503825571-project-member] [instance: a1aca638-e9c4-4e3b-9855-f92e7bd67204] Aborting claim: {{(pid=61806) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 541.551039] env[61806]: DEBUG oslo_concurrency.lockutils [None req-d8e4a121-0976-4eaa-b4c6-51deda86d081 tempest-InstanceActionsNegativeTestJSON-503825571 tempest-InstanceActionsNegativeTestJSON-503825571-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 541.641031] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c67befba-834d-45e2-9780-f62bbd0ac503 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 541.650112] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-73fd777e-a10f-4b62-a69b-e282c6b11847 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 541.691925] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d8ca43e8-d01f-4e1d-9751-4e0bc2deb43f {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 541.701962] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ecfff7a9-e308-43e7-a5ab-f2c2939cdbd2 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 541.716759] env[61806]: DEBUG nova.compute.provider_tree [None req-3568a2a1-f64b-42a9-bf53-df7d9370887d tempest-ImagesTestJSON-618852556 tempest-ImagesTestJSON-618852556-project-member] Inventory has not changed in ProviderTree for provider: a2858be1-fd22-4e08-979e-87ad25293407 {{(pid=61806) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 541.811934] env[61806]: DEBUG nova.compute.manager [None req-7d4ae408-54ff-4248-838b-a36cbf45b0c5 tempest-ListServerFiltersTestJSON-241228521 tempest-ListServerFiltersTestJSON-241228521-project-member] [instance: 9ddec22f-0fce-4d28-b97e-92b179f829f6] Start building block device mappings for instance. {{(pid=61806) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 541.943163] env[61806]: DEBUG nova.network.neutron [None req-960d1abd-1b06-4b5a-8ddd-4b463d493a45 tempest-ListServerFiltersTestJSON-241228521 tempest-ListServerFiltersTestJSON-241228521-project-member] [instance: 30e44576-dcd3-4036-9070-9032b775c90a] Updating instance_info_cache with network_info: [] {{(pid=61806) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 541.992975] env[61806]: DEBUG oslo_concurrency.lockutils [req-b49ed356-2608-4fb5-bb84-c435e9800605 req-46c03413-b6ea-4b48-9d20-bfccf145606a service nova] Releasing lock "refresh_cache-a1aca638-e9c4-4e3b-9855-f92e7bd67204" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 541.993763] env[61806]: DEBUG oslo_concurrency.lockutils [None req-962d5603-e71b-4a26-895d-6dce1320b70e tempest-ListImageFiltersTestJSON-680116525 tempest-ListImageFiltersTestJSON-680116525-project-member] Acquiring lock "53d3e2f2-f63d-45c7-a5d1-1fa9a9f0ac0f" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 541.993763] env[61806]: DEBUG oslo_concurrency.lockutils [None req-962d5603-e71b-4a26-895d-6dce1320b70e tempest-ListImageFiltersTestJSON-680116525 tempest-ListImageFiltersTestJSON-680116525-project-member] Lock "53d3e2f2-f63d-45c7-a5d1-1fa9a9f0ac0f" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 542.220286] env[61806]: DEBUG nova.scheduler.client.report [None req-3568a2a1-f64b-42a9-bf53-df7d9370887d tempest-ImagesTestJSON-618852556 tempest-ImagesTestJSON-618852556-project-member] Inventory has not changed for provider a2858be1-fd22-4e08-979e-87ad25293407 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 139, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61806) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 542.329121] env[61806]: DEBUG nova.network.neutron [None req-7d4ae408-54ff-4248-838b-a36cbf45b0c5 tempest-ListServerFiltersTestJSON-241228521 tempest-ListServerFiltersTestJSON-241228521-project-member] [instance: 9ddec22f-0fce-4d28-b97e-92b179f829f6] Successfully created port: c8f9a917-21ab-46bb-bbd3-1717da5a257b {{(pid=61806) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 542.397668] env[61806]: DEBUG nova.compute.manager [req-d52e5fb5-cf85-45b7-a445-87a38c88757c req-06cfd4d7-7fb1-4d87-b3fd-305c7f4d350f service nova] [instance: a1aca638-e9c4-4e3b-9855-f92e7bd67204] Received event network-vif-deleted-6cc2dbc9-ee9c-42fa-b51e-2fd8a2c5b6d9 {{(pid=61806) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 542.446459] env[61806]: DEBUG oslo_concurrency.lockutils [None req-960d1abd-1b06-4b5a-8ddd-4b463d493a45 tempest-ListServerFiltersTestJSON-241228521 tempest-ListServerFiltersTestJSON-241228521-project-member] Releasing lock "refresh_cache-30e44576-dcd3-4036-9070-9032b775c90a" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 542.446905] env[61806]: DEBUG nova.compute.manager [None req-960d1abd-1b06-4b5a-8ddd-4b463d493a45 tempest-ListServerFiltersTestJSON-241228521 tempest-ListServerFiltersTestJSON-241228521-project-member] [instance: 30e44576-dcd3-4036-9070-9032b775c90a] Start destroying the instance on the hypervisor. {{(pid=61806) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 542.446905] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-960d1abd-1b06-4b5a-8ddd-4b463d493a45 tempest-ListServerFiltersTestJSON-241228521 tempest-ListServerFiltersTestJSON-241228521-project-member] [instance: 30e44576-dcd3-4036-9070-9032b775c90a] Destroying instance {{(pid=61806) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 542.447512] env[61806]: DEBUG oslo_concurrency.lockutils [req-21a52603-630d-4c44-a737-1e529dafd7b2 req-f323930e-b5be-4c38-b5b6-9a33416721db service nova] Acquired lock "refresh_cache-30e44576-dcd3-4036-9070-9032b775c90a" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 542.447751] env[61806]: DEBUG nova.network.neutron [req-21a52603-630d-4c44-a737-1e529dafd7b2 req-f323930e-b5be-4c38-b5b6-9a33416721db service nova] [instance: 30e44576-dcd3-4036-9070-9032b775c90a] Refreshing network info cache for port 5990bc22-0a46-437d-b518-9a9263b60655 {{(pid=61806) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 542.449025] env[61806]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-34fd7329-1c95-49c1-be15-acbf9a5231f0 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 542.459959] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c9e3c482-d9d1-4780-a8ab-d99ed984a6e4 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 542.485527] env[61806]: WARNING nova.virt.vmwareapi.vmops [None req-960d1abd-1b06-4b5a-8ddd-4b463d493a45 tempest-ListServerFiltersTestJSON-241228521 tempest-ListServerFiltersTestJSON-241228521-project-member] [instance: 30e44576-dcd3-4036-9070-9032b775c90a] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 30e44576-dcd3-4036-9070-9032b775c90a could not be found. [ 542.485936] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-960d1abd-1b06-4b5a-8ddd-4b463d493a45 tempest-ListServerFiltersTestJSON-241228521 tempest-ListServerFiltersTestJSON-241228521-project-member] [instance: 30e44576-dcd3-4036-9070-9032b775c90a] Instance destroyed {{(pid=61806) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 542.486041] env[61806]: INFO nova.compute.manager [None req-960d1abd-1b06-4b5a-8ddd-4b463d493a45 tempest-ListServerFiltersTestJSON-241228521 tempest-ListServerFiltersTestJSON-241228521-project-member] [instance: 30e44576-dcd3-4036-9070-9032b775c90a] Took 0.04 seconds to destroy the instance on the hypervisor. [ 542.486343] env[61806]: DEBUG oslo.service.loopingcall [None req-960d1abd-1b06-4b5a-8ddd-4b463d493a45 tempest-ListServerFiltersTestJSON-241228521 tempest-ListServerFiltersTestJSON-241228521-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61806) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 542.486786] env[61806]: DEBUG nova.compute.manager [-] [instance: 30e44576-dcd3-4036-9070-9032b775c90a] Deallocating network for instance {{(pid=61806) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 542.487229] env[61806]: DEBUG nova.network.neutron [-] [instance: 30e44576-dcd3-4036-9070-9032b775c90a] deallocate_for_instance() {{(pid=61806) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 542.550785] env[61806]: DEBUG nova.network.neutron [-] [instance: 30e44576-dcd3-4036-9070-9032b775c90a] Instance cache missing network info. {{(pid=61806) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 542.726375] env[61806]: DEBUG oslo_concurrency.lockutils [None req-3568a2a1-f64b-42a9-bf53-df7d9370887d tempest-ImagesTestJSON-618852556 tempest-ImagesTestJSON-618852556-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.937s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 542.727639] env[61806]: ERROR nova.compute.manager [None req-3568a2a1-f64b-42a9-bf53-df7d9370887d tempest-ImagesTestJSON-618852556 tempest-ImagesTestJSON-618852556-project-member] [instance: 81538b78-9943-475d-aad3-f31607f2e493] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 91d5d4f4-3e0d-456f-98f8-483167092c82, please check neutron logs for more information. [ 542.727639] env[61806]: ERROR nova.compute.manager [instance: 81538b78-9943-475d-aad3-f31607f2e493] Traceback (most recent call last): [ 542.727639] env[61806]: ERROR nova.compute.manager [instance: 81538b78-9943-475d-aad3-f31607f2e493] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 542.727639] env[61806]: ERROR nova.compute.manager [instance: 81538b78-9943-475d-aad3-f31607f2e493] self.driver.spawn(context, instance, image_meta, [ 542.727639] env[61806]: ERROR nova.compute.manager [instance: 81538b78-9943-475d-aad3-f31607f2e493] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 542.727639] env[61806]: ERROR nova.compute.manager [instance: 81538b78-9943-475d-aad3-f31607f2e493] self._vmops.spawn(context, instance, image_meta, injected_files, [ 542.727639] env[61806]: ERROR nova.compute.manager [instance: 81538b78-9943-475d-aad3-f31607f2e493] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 542.727639] env[61806]: ERROR nova.compute.manager [instance: 81538b78-9943-475d-aad3-f31607f2e493] vm_ref = self.build_virtual_machine(instance, [ 542.727639] env[61806]: ERROR nova.compute.manager [instance: 81538b78-9943-475d-aad3-f31607f2e493] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 542.727639] env[61806]: ERROR nova.compute.manager [instance: 81538b78-9943-475d-aad3-f31607f2e493] vif_infos = vmwarevif.get_vif_info(self._session, [ 542.727639] env[61806]: ERROR nova.compute.manager [instance: 81538b78-9943-475d-aad3-f31607f2e493] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 542.728096] env[61806]: ERROR nova.compute.manager [instance: 81538b78-9943-475d-aad3-f31607f2e493] for vif in network_info: [ 542.728096] env[61806]: ERROR nova.compute.manager [instance: 81538b78-9943-475d-aad3-f31607f2e493] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 542.728096] env[61806]: ERROR nova.compute.manager [instance: 81538b78-9943-475d-aad3-f31607f2e493] return self._sync_wrapper(fn, *args, **kwargs) [ 542.728096] env[61806]: ERROR nova.compute.manager [instance: 81538b78-9943-475d-aad3-f31607f2e493] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 542.728096] env[61806]: ERROR nova.compute.manager [instance: 81538b78-9943-475d-aad3-f31607f2e493] self.wait() [ 542.728096] env[61806]: ERROR nova.compute.manager [instance: 81538b78-9943-475d-aad3-f31607f2e493] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 542.728096] env[61806]: ERROR nova.compute.manager [instance: 81538b78-9943-475d-aad3-f31607f2e493] self[:] = self._gt.wait() [ 542.728096] env[61806]: ERROR nova.compute.manager [instance: 81538b78-9943-475d-aad3-f31607f2e493] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 542.728096] env[61806]: ERROR nova.compute.manager [instance: 81538b78-9943-475d-aad3-f31607f2e493] return self._exit_event.wait() [ 542.728096] env[61806]: ERROR nova.compute.manager [instance: 81538b78-9943-475d-aad3-f31607f2e493] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 542.728096] env[61806]: ERROR nova.compute.manager [instance: 81538b78-9943-475d-aad3-f31607f2e493] result = hub.switch() [ 542.728096] env[61806]: ERROR nova.compute.manager [instance: 81538b78-9943-475d-aad3-f31607f2e493] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 542.728096] env[61806]: ERROR nova.compute.manager [instance: 81538b78-9943-475d-aad3-f31607f2e493] return self.greenlet.switch() [ 542.728460] env[61806]: ERROR nova.compute.manager [instance: 81538b78-9943-475d-aad3-f31607f2e493] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 542.728460] env[61806]: ERROR nova.compute.manager [instance: 81538b78-9943-475d-aad3-f31607f2e493] result = function(*args, **kwargs) [ 542.728460] env[61806]: ERROR nova.compute.manager [instance: 81538b78-9943-475d-aad3-f31607f2e493] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 542.728460] env[61806]: ERROR nova.compute.manager [instance: 81538b78-9943-475d-aad3-f31607f2e493] return func(*args, **kwargs) [ 542.728460] env[61806]: ERROR nova.compute.manager [instance: 81538b78-9943-475d-aad3-f31607f2e493] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 542.728460] env[61806]: ERROR nova.compute.manager [instance: 81538b78-9943-475d-aad3-f31607f2e493] raise e [ 542.728460] env[61806]: ERROR nova.compute.manager [instance: 81538b78-9943-475d-aad3-f31607f2e493] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 542.728460] env[61806]: ERROR nova.compute.manager [instance: 81538b78-9943-475d-aad3-f31607f2e493] nwinfo = self.network_api.allocate_for_instance( [ 542.728460] env[61806]: ERROR nova.compute.manager [instance: 81538b78-9943-475d-aad3-f31607f2e493] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 542.728460] env[61806]: ERROR nova.compute.manager [instance: 81538b78-9943-475d-aad3-f31607f2e493] created_port_ids = self._update_ports_for_instance( [ 542.728460] env[61806]: ERROR nova.compute.manager [instance: 81538b78-9943-475d-aad3-f31607f2e493] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 542.728460] env[61806]: ERROR nova.compute.manager [instance: 81538b78-9943-475d-aad3-f31607f2e493] with excutils.save_and_reraise_exception(): [ 542.728460] env[61806]: ERROR nova.compute.manager [instance: 81538b78-9943-475d-aad3-f31607f2e493] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 542.728813] env[61806]: ERROR nova.compute.manager [instance: 81538b78-9943-475d-aad3-f31607f2e493] self.force_reraise() [ 542.728813] env[61806]: ERROR nova.compute.manager [instance: 81538b78-9943-475d-aad3-f31607f2e493] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 542.728813] env[61806]: ERROR nova.compute.manager [instance: 81538b78-9943-475d-aad3-f31607f2e493] raise self.value [ 542.728813] env[61806]: ERROR nova.compute.manager [instance: 81538b78-9943-475d-aad3-f31607f2e493] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 542.728813] env[61806]: ERROR nova.compute.manager [instance: 81538b78-9943-475d-aad3-f31607f2e493] updated_port = self._update_port( [ 542.728813] env[61806]: ERROR nova.compute.manager [instance: 81538b78-9943-475d-aad3-f31607f2e493] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 542.728813] env[61806]: ERROR nova.compute.manager [instance: 81538b78-9943-475d-aad3-f31607f2e493] _ensure_no_port_binding_failure(port) [ 542.728813] env[61806]: ERROR nova.compute.manager [instance: 81538b78-9943-475d-aad3-f31607f2e493] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 542.728813] env[61806]: ERROR nova.compute.manager [instance: 81538b78-9943-475d-aad3-f31607f2e493] raise exception.PortBindingFailed(port_id=port['id']) [ 542.728813] env[61806]: ERROR nova.compute.manager [instance: 81538b78-9943-475d-aad3-f31607f2e493] nova.exception.PortBindingFailed: Binding failed for port 91d5d4f4-3e0d-456f-98f8-483167092c82, please check neutron logs for more information. [ 542.728813] env[61806]: ERROR nova.compute.manager [instance: 81538b78-9943-475d-aad3-f31607f2e493] [ 542.729127] env[61806]: DEBUG nova.compute.utils [None req-3568a2a1-f64b-42a9-bf53-df7d9370887d tempest-ImagesTestJSON-618852556 tempest-ImagesTestJSON-618852556-project-member] [instance: 81538b78-9943-475d-aad3-f31607f2e493] Binding failed for port 91d5d4f4-3e0d-456f-98f8-483167092c82, please check neutron logs for more information. {{(pid=61806) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 542.730879] env[61806]: DEBUG oslo_concurrency.lockutils [None req-2150b9c2-a011-40b9-950a-cddb0d2b8fb6 tempest-ImagesNegativeTestJSON-436875204 tempest-ImagesNegativeTestJSON-436875204-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 10.829s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 542.736679] env[61806]: DEBUG nova.compute.manager [None req-3568a2a1-f64b-42a9-bf53-df7d9370887d tempest-ImagesTestJSON-618852556 tempest-ImagesTestJSON-618852556-project-member] [instance: 81538b78-9943-475d-aad3-f31607f2e493] Build of instance 81538b78-9943-475d-aad3-f31607f2e493 was re-scheduled: Binding failed for port 91d5d4f4-3e0d-456f-98f8-483167092c82, please check neutron logs for more information. {{(pid=61806) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 542.737356] env[61806]: DEBUG nova.compute.manager [None req-3568a2a1-f64b-42a9-bf53-df7d9370887d tempest-ImagesTestJSON-618852556 tempest-ImagesTestJSON-618852556-project-member] [instance: 81538b78-9943-475d-aad3-f31607f2e493] Unplugging VIFs for instance {{(pid=61806) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 542.738066] env[61806]: DEBUG oslo_concurrency.lockutils [None req-3568a2a1-f64b-42a9-bf53-df7d9370887d tempest-ImagesTestJSON-618852556 tempest-ImagesTestJSON-618852556-project-member] Acquiring lock "refresh_cache-81538b78-9943-475d-aad3-f31607f2e493" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 542.738278] env[61806]: DEBUG oslo_concurrency.lockutils [None req-3568a2a1-f64b-42a9-bf53-df7d9370887d tempest-ImagesTestJSON-618852556 tempest-ImagesTestJSON-618852556-project-member] Acquired lock "refresh_cache-81538b78-9943-475d-aad3-f31607f2e493" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 542.738480] env[61806]: DEBUG nova.network.neutron [None req-3568a2a1-f64b-42a9-bf53-df7d9370887d tempest-ImagesTestJSON-618852556 tempest-ImagesTestJSON-618852556-project-member] [instance: 81538b78-9943-475d-aad3-f31607f2e493] Building network info cache for instance {{(pid=61806) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 542.821464] env[61806]: DEBUG nova.compute.manager [None req-7d4ae408-54ff-4248-838b-a36cbf45b0c5 tempest-ListServerFiltersTestJSON-241228521 tempest-ListServerFiltersTestJSON-241228521-project-member] [instance: 9ddec22f-0fce-4d28-b97e-92b179f829f6] Start spawning the instance on the hypervisor. {{(pid=61806) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 542.837463] env[61806]: DEBUG oslo_concurrency.lockutils [None req-11b5d1a0-518c-42d9-940e-4bf652a6d358 tempest-DeleteServersAdminTestJSON-1553667494 tempest-DeleteServersAdminTestJSON-1553667494-project-member] Acquiring lock "ddc801a0-0782-4c6d-bd72-95d0afd01339" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 542.837740] env[61806]: DEBUG oslo_concurrency.lockutils [None req-11b5d1a0-518c-42d9-940e-4bf652a6d358 tempest-DeleteServersAdminTestJSON-1553667494 tempest-DeleteServersAdminTestJSON-1553667494-project-member] Lock "ddc801a0-0782-4c6d-bd72-95d0afd01339" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 542.863087] env[61806]: DEBUG nova.virt.hardware [None req-7d4ae408-54ff-4248-838b-a36cbf45b0c5 tempest-ListServerFiltersTestJSON-241228521 tempest-ListServerFiltersTestJSON-241228521-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-15T18:05:10Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-15T18:04:52Z,direct_url=,disk_format='vmdk',id=ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='84ba83fce2c349a6988173c5d6fc3b07',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-15T18:04:53Z,virtual_size=,visibility=), allow threads: False {{(pid=61806) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 542.863340] env[61806]: DEBUG nova.virt.hardware [None req-7d4ae408-54ff-4248-838b-a36cbf45b0c5 tempest-ListServerFiltersTestJSON-241228521 tempest-ListServerFiltersTestJSON-241228521-project-member] Flavor limits 0:0:0 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 542.863492] env[61806]: DEBUG nova.virt.hardware [None req-7d4ae408-54ff-4248-838b-a36cbf45b0c5 tempest-ListServerFiltersTestJSON-241228521 tempest-ListServerFiltersTestJSON-241228521-project-member] Image limits 0:0:0 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 542.863669] env[61806]: DEBUG nova.virt.hardware [None req-7d4ae408-54ff-4248-838b-a36cbf45b0c5 tempest-ListServerFiltersTestJSON-241228521 tempest-ListServerFiltersTestJSON-241228521-project-member] Flavor pref 0:0:0 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 542.863812] env[61806]: DEBUG nova.virt.hardware [None req-7d4ae408-54ff-4248-838b-a36cbf45b0c5 tempest-ListServerFiltersTestJSON-241228521 tempest-ListServerFiltersTestJSON-241228521-project-member] Image pref 0:0:0 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 542.863955] env[61806]: DEBUG nova.virt.hardware [None req-7d4ae408-54ff-4248-838b-a36cbf45b0c5 tempest-ListServerFiltersTestJSON-241228521 tempest-ListServerFiltersTestJSON-241228521-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 542.864649] env[61806]: DEBUG nova.virt.hardware [None req-7d4ae408-54ff-4248-838b-a36cbf45b0c5 tempest-ListServerFiltersTestJSON-241228521 tempest-ListServerFiltersTestJSON-241228521-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61806) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 542.864818] env[61806]: DEBUG nova.virt.hardware [None req-7d4ae408-54ff-4248-838b-a36cbf45b0c5 tempest-ListServerFiltersTestJSON-241228521 tempest-ListServerFiltersTestJSON-241228521-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61806) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 542.864996] env[61806]: DEBUG nova.virt.hardware [None req-7d4ae408-54ff-4248-838b-a36cbf45b0c5 tempest-ListServerFiltersTestJSON-241228521 tempest-ListServerFiltersTestJSON-241228521-project-member] Got 1 possible topologies {{(pid=61806) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 542.865183] env[61806]: DEBUG nova.virt.hardware [None req-7d4ae408-54ff-4248-838b-a36cbf45b0c5 tempest-ListServerFiltersTestJSON-241228521 tempest-ListServerFiltersTestJSON-241228521-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61806) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 542.865360] env[61806]: DEBUG nova.virt.hardware [None req-7d4ae408-54ff-4248-838b-a36cbf45b0c5 tempest-ListServerFiltersTestJSON-241228521 tempest-ListServerFiltersTestJSON-241228521-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61806) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 542.866327] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-290e9791-bab6-4acf-8a2e-fe62ef33e59b {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 542.876304] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d754604f-73be-4ffd-b35f-10f760ecd6f3 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 542.991486] env[61806]: DEBUG nova.network.neutron [req-21a52603-630d-4c44-a737-1e529dafd7b2 req-f323930e-b5be-4c38-b5b6-9a33416721db service nova] [instance: 30e44576-dcd3-4036-9070-9032b775c90a] Instance cache missing network info. {{(pid=61806) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 543.057671] env[61806]: DEBUG nova.network.neutron [-] [instance: 30e44576-dcd3-4036-9070-9032b775c90a] Updating instance_info_cache with network_info: [] {{(pid=61806) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 543.247047] env[61806]: DEBUG nova.network.neutron [req-21a52603-630d-4c44-a737-1e529dafd7b2 req-f323930e-b5be-4c38-b5b6-9a33416721db service nova] [instance: 30e44576-dcd3-4036-9070-9032b775c90a] Updating instance_info_cache with network_info: [] {{(pid=61806) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 543.293247] env[61806]: DEBUG nova.network.neutron [None req-3568a2a1-f64b-42a9-bf53-df7d9370887d tempest-ImagesTestJSON-618852556 tempest-ImagesTestJSON-618852556-project-member] [instance: 81538b78-9943-475d-aad3-f31607f2e493] Instance cache missing network info. {{(pid=61806) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 543.565530] env[61806]: INFO nova.compute.manager [-] [instance: 30e44576-dcd3-4036-9070-9032b775c90a] Took 1.08 seconds to deallocate network for instance. [ 543.571354] env[61806]: DEBUG nova.compute.claims [None req-960d1abd-1b06-4b5a-8ddd-4b463d493a45 tempest-ListServerFiltersTestJSON-241228521 tempest-ListServerFiltersTestJSON-241228521-project-member] [instance: 30e44576-dcd3-4036-9070-9032b775c90a] Aborting claim: {{(pid=61806) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 543.571554] env[61806]: DEBUG oslo_concurrency.lockutils [None req-960d1abd-1b06-4b5a-8ddd-4b463d493a45 tempest-ListServerFiltersTestJSON-241228521 tempest-ListServerFiltersTestJSON-241228521-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 543.581966] env[61806]: DEBUG nova.network.neutron [None req-3568a2a1-f64b-42a9-bf53-df7d9370887d tempest-ImagesTestJSON-618852556 tempest-ImagesTestJSON-618852556-project-member] [instance: 81538b78-9943-475d-aad3-f31607f2e493] Updating instance_info_cache with network_info: [] {{(pid=61806) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 543.683769] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-64c9f8f8-ae06-4d6b-afa3-e8ace3cbaa42 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 543.693162] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e64903f8-adff-4881-9ac2-e0412daf22a4 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 543.733859] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-abc02887-d01b-489c-b5e4-69863605da27 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 543.742254] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3887ba53-4391-4ec2-b465-2cbd7bb668a0 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 543.755300] env[61806]: DEBUG oslo_concurrency.lockutils [req-21a52603-630d-4c44-a737-1e529dafd7b2 req-f323930e-b5be-4c38-b5b6-9a33416721db service nova] Releasing lock "refresh_cache-30e44576-dcd3-4036-9070-9032b775c90a" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 543.755875] env[61806]: DEBUG nova.compute.provider_tree [None req-2150b9c2-a011-40b9-950a-cddb0d2b8fb6 tempest-ImagesNegativeTestJSON-436875204 tempest-ImagesNegativeTestJSON-436875204-project-member] Inventory has not changed in ProviderTree for provider: a2858be1-fd22-4e08-979e-87ad25293407 {{(pid=61806) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 543.788618] env[61806]: DEBUG nova.compute.manager [req-178f7f02-aff9-4780-ab03-8881a5c076b7 req-55011b51-2bda-4912-9c0d-6b0766ef677d service nova] [instance: 30e44576-dcd3-4036-9070-9032b775c90a] Received event network-vif-deleted-5990bc22-0a46-437d-b518-9a9263b60655 {{(pid=61806) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 544.087025] env[61806]: DEBUG oslo_concurrency.lockutils [None req-3568a2a1-f64b-42a9-bf53-df7d9370887d tempest-ImagesTestJSON-618852556 tempest-ImagesTestJSON-618852556-project-member] Releasing lock "refresh_cache-81538b78-9943-475d-aad3-f31607f2e493" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 544.087025] env[61806]: DEBUG nova.compute.manager [None req-3568a2a1-f64b-42a9-bf53-df7d9370887d tempest-ImagesTestJSON-618852556 tempest-ImagesTestJSON-618852556-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61806) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 544.087025] env[61806]: DEBUG nova.compute.manager [None req-3568a2a1-f64b-42a9-bf53-df7d9370887d tempest-ImagesTestJSON-618852556 tempest-ImagesTestJSON-618852556-project-member] [instance: 81538b78-9943-475d-aad3-f31607f2e493] Deallocating network for instance {{(pid=61806) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 544.087025] env[61806]: DEBUG nova.network.neutron [None req-3568a2a1-f64b-42a9-bf53-df7d9370887d tempest-ImagesTestJSON-618852556 tempest-ImagesTestJSON-618852556-project-member] [instance: 81538b78-9943-475d-aad3-f31607f2e493] deallocate_for_instance() {{(pid=61806) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 544.141040] env[61806]: DEBUG nova.network.neutron [None req-3568a2a1-f64b-42a9-bf53-df7d9370887d tempest-ImagesTestJSON-618852556 tempest-ImagesTestJSON-618852556-project-member] [instance: 81538b78-9943-475d-aad3-f31607f2e493] Instance cache missing network info. {{(pid=61806) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 544.259435] env[61806]: DEBUG nova.scheduler.client.report [None req-2150b9c2-a011-40b9-950a-cddb0d2b8fb6 tempest-ImagesNegativeTestJSON-436875204 tempest-ImagesNegativeTestJSON-436875204-project-member] Inventory has not changed for provider a2858be1-fd22-4e08-979e-87ad25293407 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 139, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61806) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 544.643667] env[61806]: DEBUG nova.network.neutron [None req-3568a2a1-f64b-42a9-bf53-df7d9370887d tempest-ImagesTestJSON-618852556 tempest-ImagesTestJSON-618852556-project-member] [instance: 81538b78-9943-475d-aad3-f31607f2e493] Updating instance_info_cache with network_info: [] {{(pid=61806) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 544.765264] env[61806]: DEBUG oslo_concurrency.lockutils [None req-2150b9c2-a011-40b9-950a-cddb0d2b8fb6 tempest-ImagesNegativeTestJSON-436875204 tempest-ImagesNegativeTestJSON-436875204-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.034s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 544.765912] env[61806]: ERROR nova.compute.manager [None req-2150b9c2-a011-40b9-950a-cddb0d2b8fb6 tempest-ImagesNegativeTestJSON-436875204 tempest-ImagesNegativeTestJSON-436875204-project-member] [instance: 71380a36-8926-45d3-a70e-a7112bd1d06f] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 3f10bab7-3a39-46da-b6db-0d3525153d7b, please check neutron logs for more information. [ 544.765912] env[61806]: ERROR nova.compute.manager [instance: 71380a36-8926-45d3-a70e-a7112bd1d06f] Traceback (most recent call last): [ 544.765912] env[61806]: ERROR nova.compute.manager [instance: 71380a36-8926-45d3-a70e-a7112bd1d06f] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 544.765912] env[61806]: ERROR nova.compute.manager [instance: 71380a36-8926-45d3-a70e-a7112bd1d06f] self.driver.spawn(context, instance, image_meta, [ 544.765912] env[61806]: ERROR nova.compute.manager [instance: 71380a36-8926-45d3-a70e-a7112bd1d06f] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 544.765912] env[61806]: ERROR nova.compute.manager [instance: 71380a36-8926-45d3-a70e-a7112bd1d06f] self._vmops.spawn(context, instance, image_meta, injected_files, [ 544.765912] env[61806]: ERROR nova.compute.manager [instance: 71380a36-8926-45d3-a70e-a7112bd1d06f] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 544.765912] env[61806]: ERROR nova.compute.manager [instance: 71380a36-8926-45d3-a70e-a7112bd1d06f] vm_ref = self.build_virtual_machine(instance, [ 544.765912] env[61806]: ERROR nova.compute.manager [instance: 71380a36-8926-45d3-a70e-a7112bd1d06f] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 544.765912] env[61806]: ERROR nova.compute.manager [instance: 71380a36-8926-45d3-a70e-a7112bd1d06f] vif_infos = vmwarevif.get_vif_info(self._session, [ 544.765912] env[61806]: ERROR nova.compute.manager [instance: 71380a36-8926-45d3-a70e-a7112bd1d06f] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 544.766406] env[61806]: ERROR nova.compute.manager [instance: 71380a36-8926-45d3-a70e-a7112bd1d06f] for vif in network_info: [ 544.766406] env[61806]: ERROR nova.compute.manager [instance: 71380a36-8926-45d3-a70e-a7112bd1d06f] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 544.766406] env[61806]: ERROR nova.compute.manager [instance: 71380a36-8926-45d3-a70e-a7112bd1d06f] return self._sync_wrapper(fn, *args, **kwargs) [ 544.766406] env[61806]: ERROR nova.compute.manager [instance: 71380a36-8926-45d3-a70e-a7112bd1d06f] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 544.766406] env[61806]: ERROR nova.compute.manager [instance: 71380a36-8926-45d3-a70e-a7112bd1d06f] self.wait() [ 544.766406] env[61806]: ERROR nova.compute.manager [instance: 71380a36-8926-45d3-a70e-a7112bd1d06f] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 544.766406] env[61806]: ERROR nova.compute.manager [instance: 71380a36-8926-45d3-a70e-a7112bd1d06f] self[:] = self._gt.wait() [ 544.766406] env[61806]: ERROR nova.compute.manager [instance: 71380a36-8926-45d3-a70e-a7112bd1d06f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 544.766406] env[61806]: ERROR nova.compute.manager [instance: 71380a36-8926-45d3-a70e-a7112bd1d06f] return self._exit_event.wait() [ 544.766406] env[61806]: ERROR nova.compute.manager [instance: 71380a36-8926-45d3-a70e-a7112bd1d06f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 544.766406] env[61806]: ERROR nova.compute.manager [instance: 71380a36-8926-45d3-a70e-a7112bd1d06f] result = hub.switch() [ 544.766406] env[61806]: ERROR nova.compute.manager [instance: 71380a36-8926-45d3-a70e-a7112bd1d06f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 544.766406] env[61806]: ERROR nova.compute.manager [instance: 71380a36-8926-45d3-a70e-a7112bd1d06f] return self.greenlet.switch() [ 544.766802] env[61806]: ERROR nova.compute.manager [instance: 71380a36-8926-45d3-a70e-a7112bd1d06f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 544.766802] env[61806]: ERROR nova.compute.manager [instance: 71380a36-8926-45d3-a70e-a7112bd1d06f] result = function(*args, **kwargs) [ 544.766802] env[61806]: ERROR nova.compute.manager [instance: 71380a36-8926-45d3-a70e-a7112bd1d06f] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 544.766802] env[61806]: ERROR nova.compute.manager [instance: 71380a36-8926-45d3-a70e-a7112bd1d06f] return func(*args, **kwargs) [ 544.766802] env[61806]: ERROR nova.compute.manager [instance: 71380a36-8926-45d3-a70e-a7112bd1d06f] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 544.766802] env[61806]: ERROR nova.compute.manager [instance: 71380a36-8926-45d3-a70e-a7112bd1d06f] raise e [ 544.766802] env[61806]: ERROR nova.compute.manager [instance: 71380a36-8926-45d3-a70e-a7112bd1d06f] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 544.766802] env[61806]: ERROR nova.compute.manager [instance: 71380a36-8926-45d3-a70e-a7112bd1d06f] nwinfo = self.network_api.allocate_for_instance( [ 544.766802] env[61806]: ERROR nova.compute.manager [instance: 71380a36-8926-45d3-a70e-a7112bd1d06f] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 544.766802] env[61806]: ERROR nova.compute.manager [instance: 71380a36-8926-45d3-a70e-a7112bd1d06f] created_port_ids = self._update_ports_for_instance( [ 544.766802] env[61806]: ERROR nova.compute.manager [instance: 71380a36-8926-45d3-a70e-a7112bd1d06f] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 544.766802] env[61806]: ERROR nova.compute.manager [instance: 71380a36-8926-45d3-a70e-a7112bd1d06f] with excutils.save_and_reraise_exception(): [ 544.766802] env[61806]: ERROR nova.compute.manager [instance: 71380a36-8926-45d3-a70e-a7112bd1d06f] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 544.767231] env[61806]: ERROR nova.compute.manager [instance: 71380a36-8926-45d3-a70e-a7112bd1d06f] self.force_reraise() [ 544.767231] env[61806]: ERROR nova.compute.manager [instance: 71380a36-8926-45d3-a70e-a7112bd1d06f] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 544.767231] env[61806]: ERROR nova.compute.manager [instance: 71380a36-8926-45d3-a70e-a7112bd1d06f] raise self.value [ 544.767231] env[61806]: ERROR nova.compute.manager [instance: 71380a36-8926-45d3-a70e-a7112bd1d06f] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 544.767231] env[61806]: ERROR nova.compute.manager [instance: 71380a36-8926-45d3-a70e-a7112bd1d06f] updated_port = self._update_port( [ 544.767231] env[61806]: ERROR nova.compute.manager [instance: 71380a36-8926-45d3-a70e-a7112bd1d06f] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 544.767231] env[61806]: ERROR nova.compute.manager [instance: 71380a36-8926-45d3-a70e-a7112bd1d06f] _ensure_no_port_binding_failure(port) [ 544.767231] env[61806]: ERROR nova.compute.manager [instance: 71380a36-8926-45d3-a70e-a7112bd1d06f] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 544.767231] env[61806]: ERROR nova.compute.manager [instance: 71380a36-8926-45d3-a70e-a7112bd1d06f] raise exception.PortBindingFailed(port_id=port['id']) [ 544.767231] env[61806]: ERROR nova.compute.manager [instance: 71380a36-8926-45d3-a70e-a7112bd1d06f] nova.exception.PortBindingFailed: Binding failed for port 3f10bab7-3a39-46da-b6db-0d3525153d7b, please check neutron logs for more information. [ 544.767231] env[61806]: ERROR nova.compute.manager [instance: 71380a36-8926-45d3-a70e-a7112bd1d06f] [ 544.767562] env[61806]: DEBUG nova.compute.utils [None req-2150b9c2-a011-40b9-950a-cddb0d2b8fb6 tempest-ImagesNegativeTestJSON-436875204 tempest-ImagesNegativeTestJSON-436875204-project-member] [instance: 71380a36-8926-45d3-a70e-a7112bd1d06f] Binding failed for port 3f10bab7-3a39-46da-b6db-0d3525153d7b, please check neutron logs for more information. {{(pid=61806) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 544.768759] env[61806]: DEBUG oslo_concurrency.lockutils [None req-aac795d3-c37f-4811-9be9-aa67ad520ab0 tempest-ListServerFiltersTestJSON-241228521 tempest-ListServerFiltersTestJSON-241228521-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 12.376s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 544.773357] env[61806]: INFO nova.compute.claims [None req-aac795d3-c37f-4811-9be9-aa67ad520ab0 tempest-ListServerFiltersTestJSON-241228521 tempest-ListServerFiltersTestJSON-241228521-project-member] [instance: 7fcee323-8ea6-4f69-9f1e-f2105af0fe96] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 544.777537] env[61806]: DEBUG nova.compute.manager [None req-2150b9c2-a011-40b9-950a-cddb0d2b8fb6 tempest-ImagesNegativeTestJSON-436875204 tempest-ImagesNegativeTestJSON-436875204-project-member] [instance: 71380a36-8926-45d3-a70e-a7112bd1d06f] Build of instance 71380a36-8926-45d3-a70e-a7112bd1d06f was re-scheduled: Binding failed for port 3f10bab7-3a39-46da-b6db-0d3525153d7b, please check neutron logs for more information. {{(pid=61806) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 544.777847] env[61806]: DEBUG nova.compute.manager [None req-2150b9c2-a011-40b9-950a-cddb0d2b8fb6 tempest-ImagesNegativeTestJSON-436875204 tempest-ImagesNegativeTestJSON-436875204-project-member] [instance: 71380a36-8926-45d3-a70e-a7112bd1d06f] Unplugging VIFs for instance {{(pid=61806) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 544.779038] env[61806]: DEBUG oslo_concurrency.lockutils [None req-2150b9c2-a011-40b9-950a-cddb0d2b8fb6 tempest-ImagesNegativeTestJSON-436875204 tempest-ImagesNegativeTestJSON-436875204-project-member] Acquiring lock "refresh_cache-71380a36-8926-45d3-a70e-a7112bd1d06f" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 544.779267] env[61806]: DEBUG oslo_concurrency.lockutils [None req-2150b9c2-a011-40b9-950a-cddb0d2b8fb6 tempest-ImagesNegativeTestJSON-436875204 tempest-ImagesNegativeTestJSON-436875204-project-member] Acquired lock "refresh_cache-71380a36-8926-45d3-a70e-a7112bd1d06f" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 544.779483] env[61806]: DEBUG nova.network.neutron [None req-2150b9c2-a011-40b9-950a-cddb0d2b8fb6 tempest-ImagesNegativeTestJSON-436875204 tempest-ImagesNegativeTestJSON-436875204-project-member] [instance: 71380a36-8926-45d3-a70e-a7112bd1d06f] Building network info cache for instance {{(pid=61806) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 545.061836] env[61806]: DEBUG nova.compute.manager [req-dcdf3b75-e6be-4eb8-84ce-e623c69e158e req-85ffaf28-da84-4310-81a2-b1852d80ea22 service nova] [instance: 9ddec22f-0fce-4d28-b97e-92b179f829f6] Received event network-changed-c8f9a917-21ab-46bb-bbd3-1717da5a257b {{(pid=61806) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 545.061836] env[61806]: DEBUG nova.compute.manager [req-dcdf3b75-e6be-4eb8-84ce-e623c69e158e req-85ffaf28-da84-4310-81a2-b1852d80ea22 service nova] [instance: 9ddec22f-0fce-4d28-b97e-92b179f829f6] Refreshing instance network info cache due to event network-changed-c8f9a917-21ab-46bb-bbd3-1717da5a257b. {{(pid=61806) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 545.061836] env[61806]: DEBUG oslo_concurrency.lockutils [req-dcdf3b75-e6be-4eb8-84ce-e623c69e158e req-85ffaf28-da84-4310-81a2-b1852d80ea22 service nova] Acquiring lock "refresh_cache-9ddec22f-0fce-4d28-b97e-92b179f829f6" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 545.061836] env[61806]: DEBUG oslo_concurrency.lockutils [req-dcdf3b75-e6be-4eb8-84ce-e623c69e158e req-85ffaf28-da84-4310-81a2-b1852d80ea22 service nova] Acquired lock "refresh_cache-9ddec22f-0fce-4d28-b97e-92b179f829f6" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 545.062101] env[61806]: DEBUG nova.network.neutron [req-dcdf3b75-e6be-4eb8-84ce-e623c69e158e req-85ffaf28-da84-4310-81a2-b1852d80ea22 service nova] [instance: 9ddec22f-0fce-4d28-b97e-92b179f829f6] Refreshing network info cache for port c8f9a917-21ab-46bb-bbd3-1717da5a257b {{(pid=61806) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 545.147028] env[61806]: INFO nova.compute.manager [None req-3568a2a1-f64b-42a9-bf53-df7d9370887d tempest-ImagesTestJSON-618852556 tempest-ImagesTestJSON-618852556-project-member] [instance: 81538b78-9943-475d-aad3-f31607f2e493] Took 1.06 seconds to deallocate network for instance. [ 545.314264] env[61806]: DEBUG nova.network.neutron [None req-2150b9c2-a011-40b9-950a-cddb0d2b8fb6 tempest-ImagesNegativeTestJSON-436875204 tempest-ImagesNegativeTestJSON-436875204-project-member] [instance: 71380a36-8926-45d3-a70e-a7112bd1d06f] Instance cache missing network info. {{(pid=61806) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 545.455440] env[61806]: ERROR nova.compute.manager [None req-7d4ae408-54ff-4248-838b-a36cbf45b0c5 tempest-ListServerFiltersTestJSON-241228521 tempest-ListServerFiltersTestJSON-241228521-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port c8f9a917-21ab-46bb-bbd3-1717da5a257b, please check neutron logs for more information. [ 545.455440] env[61806]: ERROR nova.compute.manager Traceback (most recent call last): [ 545.455440] env[61806]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 545.455440] env[61806]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 545.455440] env[61806]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 545.455440] env[61806]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 545.455440] env[61806]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 545.455440] env[61806]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 545.455440] env[61806]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 545.455440] env[61806]: ERROR nova.compute.manager self.force_reraise() [ 545.455440] env[61806]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 545.455440] env[61806]: ERROR nova.compute.manager raise self.value [ 545.455440] env[61806]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 545.455440] env[61806]: ERROR nova.compute.manager updated_port = self._update_port( [ 545.455440] env[61806]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 545.455440] env[61806]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 545.455955] env[61806]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 545.455955] env[61806]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 545.455955] env[61806]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port c8f9a917-21ab-46bb-bbd3-1717da5a257b, please check neutron logs for more information. [ 545.455955] env[61806]: ERROR nova.compute.manager [ 545.455955] env[61806]: Traceback (most recent call last): [ 545.455955] env[61806]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 545.455955] env[61806]: listener.cb(fileno) [ 545.455955] env[61806]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 545.455955] env[61806]: result = function(*args, **kwargs) [ 545.455955] env[61806]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 545.455955] env[61806]: return func(*args, **kwargs) [ 545.455955] env[61806]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 545.455955] env[61806]: raise e [ 545.455955] env[61806]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 545.455955] env[61806]: nwinfo = self.network_api.allocate_for_instance( [ 545.455955] env[61806]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 545.455955] env[61806]: created_port_ids = self._update_ports_for_instance( [ 545.455955] env[61806]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 545.455955] env[61806]: with excutils.save_and_reraise_exception(): [ 545.455955] env[61806]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 545.455955] env[61806]: self.force_reraise() [ 545.455955] env[61806]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 545.455955] env[61806]: raise self.value [ 545.455955] env[61806]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 545.455955] env[61806]: updated_port = self._update_port( [ 545.455955] env[61806]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 545.455955] env[61806]: _ensure_no_port_binding_failure(port) [ 545.455955] env[61806]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 545.455955] env[61806]: raise exception.PortBindingFailed(port_id=port['id']) [ 545.456742] env[61806]: nova.exception.PortBindingFailed: Binding failed for port c8f9a917-21ab-46bb-bbd3-1717da5a257b, please check neutron logs for more information. [ 545.456742] env[61806]: Removing descriptor: 17 [ 545.456742] env[61806]: ERROR nova.compute.manager [None req-7d4ae408-54ff-4248-838b-a36cbf45b0c5 tempest-ListServerFiltersTestJSON-241228521 tempest-ListServerFiltersTestJSON-241228521-project-member] [instance: 9ddec22f-0fce-4d28-b97e-92b179f829f6] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port c8f9a917-21ab-46bb-bbd3-1717da5a257b, please check neutron logs for more information. [ 545.456742] env[61806]: ERROR nova.compute.manager [instance: 9ddec22f-0fce-4d28-b97e-92b179f829f6] Traceback (most recent call last): [ 545.456742] env[61806]: ERROR nova.compute.manager [instance: 9ddec22f-0fce-4d28-b97e-92b179f829f6] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 545.456742] env[61806]: ERROR nova.compute.manager [instance: 9ddec22f-0fce-4d28-b97e-92b179f829f6] yield resources [ 545.456742] env[61806]: ERROR nova.compute.manager [instance: 9ddec22f-0fce-4d28-b97e-92b179f829f6] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 545.456742] env[61806]: ERROR nova.compute.manager [instance: 9ddec22f-0fce-4d28-b97e-92b179f829f6] self.driver.spawn(context, instance, image_meta, [ 545.456742] env[61806]: ERROR nova.compute.manager [instance: 9ddec22f-0fce-4d28-b97e-92b179f829f6] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 545.456742] env[61806]: ERROR nova.compute.manager [instance: 9ddec22f-0fce-4d28-b97e-92b179f829f6] self._vmops.spawn(context, instance, image_meta, injected_files, [ 545.456742] env[61806]: ERROR nova.compute.manager [instance: 9ddec22f-0fce-4d28-b97e-92b179f829f6] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 545.456742] env[61806]: ERROR nova.compute.manager [instance: 9ddec22f-0fce-4d28-b97e-92b179f829f6] vm_ref = self.build_virtual_machine(instance, [ 545.457203] env[61806]: ERROR nova.compute.manager [instance: 9ddec22f-0fce-4d28-b97e-92b179f829f6] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 545.457203] env[61806]: ERROR nova.compute.manager [instance: 9ddec22f-0fce-4d28-b97e-92b179f829f6] vif_infos = vmwarevif.get_vif_info(self._session, [ 545.457203] env[61806]: ERROR nova.compute.manager [instance: 9ddec22f-0fce-4d28-b97e-92b179f829f6] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 545.457203] env[61806]: ERROR nova.compute.manager [instance: 9ddec22f-0fce-4d28-b97e-92b179f829f6] for vif in network_info: [ 545.457203] env[61806]: ERROR nova.compute.manager [instance: 9ddec22f-0fce-4d28-b97e-92b179f829f6] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 545.457203] env[61806]: ERROR nova.compute.manager [instance: 9ddec22f-0fce-4d28-b97e-92b179f829f6] return self._sync_wrapper(fn, *args, **kwargs) [ 545.457203] env[61806]: ERROR nova.compute.manager [instance: 9ddec22f-0fce-4d28-b97e-92b179f829f6] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 545.457203] env[61806]: ERROR nova.compute.manager [instance: 9ddec22f-0fce-4d28-b97e-92b179f829f6] self.wait() [ 545.457203] env[61806]: ERROR nova.compute.manager [instance: 9ddec22f-0fce-4d28-b97e-92b179f829f6] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 545.457203] env[61806]: ERROR nova.compute.manager [instance: 9ddec22f-0fce-4d28-b97e-92b179f829f6] self[:] = self._gt.wait() [ 545.457203] env[61806]: ERROR nova.compute.manager [instance: 9ddec22f-0fce-4d28-b97e-92b179f829f6] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 545.457203] env[61806]: ERROR nova.compute.manager [instance: 9ddec22f-0fce-4d28-b97e-92b179f829f6] return self._exit_event.wait() [ 545.457203] env[61806]: ERROR nova.compute.manager [instance: 9ddec22f-0fce-4d28-b97e-92b179f829f6] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 545.457928] env[61806]: ERROR nova.compute.manager [instance: 9ddec22f-0fce-4d28-b97e-92b179f829f6] result = hub.switch() [ 545.457928] env[61806]: ERROR nova.compute.manager [instance: 9ddec22f-0fce-4d28-b97e-92b179f829f6] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 545.457928] env[61806]: ERROR nova.compute.manager [instance: 9ddec22f-0fce-4d28-b97e-92b179f829f6] return self.greenlet.switch() [ 545.457928] env[61806]: ERROR nova.compute.manager [instance: 9ddec22f-0fce-4d28-b97e-92b179f829f6] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 545.457928] env[61806]: ERROR nova.compute.manager [instance: 9ddec22f-0fce-4d28-b97e-92b179f829f6] result = function(*args, **kwargs) [ 545.457928] env[61806]: ERROR nova.compute.manager [instance: 9ddec22f-0fce-4d28-b97e-92b179f829f6] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 545.457928] env[61806]: ERROR nova.compute.manager [instance: 9ddec22f-0fce-4d28-b97e-92b179f829f6] return func(*args, **kwargs) [ 545.457928] env[61806]: ERROR nova.compute.manager [instance: 9ddec22f-0fce-4d28-b97e-92b179f829f6] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 545.457928] env[61806]: ERROR nova.compute.manager [instance: 9ddec22f-0fce-4d28-b97e-92b179f829f6] raise e [ 545.457928] env[61806]: ERROR nova.compute.manager [instance: 9ddec22f-0fce-4d28-b97e-92b179f829f6] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 545.457928] env[61806]: ERROR nova.compute.manager [instance: 9ddec22f-0fce-4d28-b97e-92b179f829f6] nwinfo = self.network_api.allocate_for_instance( [ 545.457928] env[61806]: ERROR nova.compute.manager [instance: 9ddec22f-0fce-4d28-b97e-92b179f829f6] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 545.457928] env[61806]: ERROR nova.compute.manager [instance: 9ddec22f-0fce-4d28-b97e-92b179f829f6] created_port_ids = self._update_ports_for_instance( [ 545.458326] env[61806]: ERROR nova.compute.manager [instance: 9ddec22f-0fce-4d28-b97e-92b179f829f6] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 545.458326] env[61806]: ERROR nova.compute.manager [instance: 9ddec22f-0fce-4d28-b97e-92b179f829f6] with excutils.save_and_reraise_exception(): [ 545.458326] env[61806]: ERROR nova.compute.manager [instance: 9ddec22f-0fce-4d28-b97e-92b179f829f6] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 545.458326] env[61806]: ERROR nova.compute.manager [instance: 9ddec22f-0fce-4d28-b97e-92b179f829f6] self.force_reraise() [ 545.458326] env[61806]: ERROR nova.compute.manager [instance: 9ddec22f-0fce-4d28-b97e-92b179f829f6] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 545.458326] env[61806]: ERROR nova.compute.manager [instance: 9ddec22f-0fce-4d28-b97e-92b179f829f6] raise self.value [ 545.458326] env[61806]: ERROR nova.compute.manager [instance: 9ddec22f-0fce-4d28-b97e-92b179f829f6] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 545.458326] env[61806]: ERROR nova.compute.manager [instance: 9ddec22f-0fce-4d28-b97e-92b179f829f6] updated_port = self._update_port( [ 545.458326] env[61806]: ERROR nova.compute.manager [instance: 9ddec22f-0fce-4d28-b97e-92b179f829f6] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 545.458326] env[61806]: ERROR nova.compute.manager [instance: 9ddec22f-0fce-4d28-b97e-92b179f829f6] _ensure_no_port_binding_failure(port) [ 545.458326] env[61806]: ERROR nova.compute.manager [instance: 9ddec22f-0fce-4d28-b97e-92b179f829f6] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 545.458326] env[61806]: ERROR nova.compute.manager [instance: 9ddec22f-0fce-4d28-b97e-92b179f829f6] raise exception.PortBindingFailed(port_id=port['id']) [ 545.458711] env[61806]: ERROR nova.compute.manager [instance: 9ddec22f-0fce-4d28-b97e-92b179f829f6] nova.exception.PortBindingFailed: Binding failed for port c8f9a917-21ab-46bb-bbd3-1717da5a257b, please check neutron logs for more information. [ 545.458711] env[61806]: ERROR nova.compute.manager [instance: 9ddec22f-0fce-4d28-b97e-92b179f829f6] [ 545.458711] env[61806]: INFO nova.compute.manager [None req-7d4ae408-54ff-4248-838b-a36cbf45b0c5 tempest-ListServerFiltersTestJSON-241228521 tempest-ListServerFiltersTestJSON-241228521-project-member] [instance: 9ddec22f-0fce-4d28-b97e-92b179f829f6] Terminating instance [ 545.458711] env[61806]: DEBUG oslo_concurrency.lockutils [None req-7d4ae408-54ff-4248-838b-a36cbf45b0c5 tempest-ListServerFiltersTestJSON-241228521 tempest-ListServerFiltersTestJSON-241228521-project-member] Acquiring lock "refresh_cache-9ddec22f-0fce-4d28-b97e-92b179f829f6" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 545.495334] env[61806]: DEBUG nova.network.neutron [None req-2150b9c2-a011-40b9-950a-cddb0d2b8fb6 tempest-ImagesNegativeTestJSON-436875204 tempest-ImagesNegativeTestJSON-436875204-project-member] [instance: 71380a36-8926-45d3-a70e-a7112bd1d06f] Updating instance_info_cache with network_info: [] {{(pid=61806) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 545.597979] env[61806]: DEBUG nova.network.neutron [req-dcdf3b75-e6be-4eb8-84ce-e623c69e158e req-85ffaf28-da84-4310-81a2-b1852d80ea22 service nova] [instance: 9ddec22f-0fce-4d28-b97e-92b179f829f6] Instance cache missing network info. {{(pid=61806) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 545.793603] env[61806]: DEBUG nova.network.neutron [req-dcdf3b75-e6be-4eb8-84ce-e623c69e158e req-85ffaf28-da84-4310-81a2-b1852d80ea22 service nova] [instance: 9ddec22f-0fce-4d28-b97e-92b179f829f6] Updating instance_info_cache with network_info: [] {{(pid=61806) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 546.000310] env[61806]: DEBUG oslo_concurrency.lockutils [None req-2150b9c2-a011-40b9-950a-cddb0d2b8fb6 tempest-ImagesNegativeTestJSON-436875204 tempest-ImagesNegativeTestJSON-436875204-project-member] Releasing lock "refresh_cache-71380a36-8926-45d3-a70e-a7112bd1d06f" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 546.000563] env[61806]: DEBUG nova.compute.manager [None req-2150b9c2-a011-40b9-950a-cddb0d2b8fb6 tempest-ImagesNegativeTestJSON-436875204 tempest-ImagesNegativeTestJSON-436875204-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61806) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 546.000748] env[61806]: DEBUG nova.compute.manager [None req-2150b9c2-a011-40b9-950a-cddb0d2b8fb6 tempest-ImagesNegativeTestJSON-436875204 tempest-ImagesNegativeTestJSON-436875204-project-member] [instance: 71380a36-8926-45d3-a70e-a7112bd1d06f] Deallocating network for instance {{(pid=61806) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 546.000917] env[61806]: DEBUG nova.network.neutron [None req-2150b9c2-a011-40b9-950a-cddb0d2b8fb6 tempest-ImagesNegativeTestJSON-436875204 tempest-ImagesNegativeTestJSON-436875204-project-member] [instance: 71380a36-8926-45d3-a70e-a7112bd1d06f] deallocate_for_instance() {{(pid=61806) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 546.056090] env[61806]: DEBUG nova.network.neutron [None req-2150b9c2-a011-40b9-950a-cddb0d2b8fb6 tempest-ImagesNegativeTestJSON-436875204 tempest-ImagesNegativeTestJSON-436875204-project-member] [instance: 71380a36-8926-45d3-a70e-a7112bd1d06f] Instance cache missing network info. {{(pid=61806) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 546.144232] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6f23fb3d-fabd-4ee7-bee6-70e99c6fcdce {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 546.153266] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dbe6e071-a13f-47e0-a868-702719723bc9 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 546.195894] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-220a84db-d637-4cad-9546-852aeced436c {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 546.201665] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d39119ac-fe2c-4660-89aa-212c7c1aeee8 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 546.221770] env[61806]: DEBUG nova.compute.provider_tree [None req-aac795d3-c37f-4811-9be9-aa67ad520ab0 tempest-ListServerFiltersTestJSON-241228521 tempest-ListServerFiltersTestJSON-241228521-project-member] Inventory has not changed in ProviderTree for provider: a2858be1-fd22-4e08-979e-87ad25293407 {{(pid=61806) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 546.235514] env[61806]: INFO nova.scheduler.client.report [None req-3568a2a1-f64b-42a9-bf53-df7d9370887d tempest-ImagesTestJSON-618852556 tempest-ImagesTestJSON-618852556-project-member] Deleted allocations for instance 81538b78-9943-475d-aad3-f31607f2e493 [ 546.296493] env[61806]: DEBUG oslo_concurrency.lockutils [req-dcdf3b75-e6be-4eb8-84ce-e623c69e158e req-85ffaf28-da84-4310-81a2-b1852d80ea22 service nova] Releasing lock "refresh_cache-9ddec22f-0fce-4d28-b97e-92b179f829f6" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 546.296895] env[61806]: DEBUG oslo_concurrency.lockutils [None req-7d4ae408-54ff-4248-838b-a36cbf45b0c5 tempest-ListServerFiltersTestJSON-241228521 tempest-ListServerFiltersTestJSON-241228521-project-member] Acquired lock "refresh_cache-9ddec22f-0fce-4d28-b97e-92b179f829f6" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 546.297093] env[61806]: DEBUG nova.network.neutron [None req-7d4ae408-54ff-4248-838b-a36cbf45b0c5 tempest-ListServerFiltersTestJSON-241228521 tempest-ListServerFiltersTestJSON-241228521-project-member] [instance: 9ddec22f-0fce-4d28-b97e-92b179f829f6] Building network info cache for instance {{(pid=61806) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 546.563584] env[61806]: DEBUG nova.network.neutron [None req-2150b9c2-a011-40b9-950a-cddb0d2b8fb6 tempest-ImagesNegativeTestJSON-436875204 tempest-ImagesNegativeTestJSON-436875204-project-member] [instance: 71380a36-8926-45d3-a70e-a7112bd1d06f] Updating instance_info_cache with network_info: [] {{(pid=61806) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 546.728394] env[61806]: DEBUG nova.scheduler.client.report [None req-aac795d3-c37f-4811-9be9-aa67ad520ab0 tempest-ListServerFiltersTestJSON-241228521 tempest-ListServerFiltersTestJSON-241228521-project-member] Inventory has not changed for provider a2858be1-fd22-4e08-979e-87ad25293407 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 139, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61806) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 546.746811] env[61806]: DEBUG oslo_concurrency.lockutils [None req-3568a2a1-f64b-42a9-bf53-df7d9370887d tempest-ImagesTestJSON-618852556 tempest-ImagesTestJSON-618852556-project-member] Lock "81538b78-9943-475d-aad3-f31607f2e493" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 26.864s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 546.851541] env[61806]: DEBUG nova.network.neutron [None req-7d4ae408-54ff-4248-838b-a36cbf45b0c5 tempest-ListServerFiltersTestJSON-241228521 tempest-ListServerFiltersTestJSON-241228521-project-member] [instance: 9ddec22f-0fce-4d28-b97e-92b179f829f6] Instance cache missing network info. {{(pid=61806) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 546.927856] env[61806]: DEBUG oslo_concurrency.lockutils [None req-ebff8349-d5ea-4a95-aee1-e4c9e8b6d3a4 tempest-ServerDiagnosticsTest-195484958 tempest-ServerDiagnosticsTest-195484958-project-member] Acquiring lock "79fc9ea2-c427-47ce-acdb-1002ee5d78c9" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 546.928044] env[61806]: DEBUG oslo_concurrency.lockutils [None req-ebff8349-d5ea-4a95-aee1-e4c9e8b6d3a4 tempest-ServerDiagnosticsTest-195484958 tempest-ServerDiagnosticsTest-195484958-project-member] Lock "79fc9ea2-c427-47ce-acdb-1002ee5d78c9" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 547.066619] env[61806]: INFO nova.compute.manager [None req-2150b9c2-a011-40b9-950a-cddb0d2b8fb6 tempest-ImagesNegativeTestJSON-436875204 tempest-ImagesNegativeTestJSON-436875204-project-member] [instance: 71380a36-8926-45d3-a70e-a7112bd1d06f] Took 1.07 seconds to deallocate network for instance. [ 547.237308] env[61806]: DEBUG oslo_concurrency.lockutils [None req-aac795d3-c37f-4811-9be9-aa67ad520ab0 tempest-ListServerFiltersTestJSON-241228521 tempest-ListServerFiltersTestJSON-241228521-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.468s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 547.238538] env[61806]: DEBUG nova.compute.manager [None req-aac795d3-c37f-4811-9be9-aa67ad520ab0 tempest-ListServerFiltersTestJSON-241228521 tempest-ListServerFiltersTestJSON-241228521-project-member] [instance: 7fcee323-8ea6-4f69-9f1e-f2105af0fe96] Start building networks asynchronously for instance. {{(pid=61806) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 547.241642] env[61806]: DEBUG oslo_concurrency.lockutils [None req-55a61261-0862-4d73-b971-6d11cbe88c5e tempest-ServersAdminTestJSON-14199920 tempest-ServersAdminTestJSON-14199920-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 13.174s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 547.243203] env[61806]: INFO nova.compute.claims [None req-55a61261-0862-4d73-b971-6d11cbe88c5e tempest-ServersAdminTestJSON-14199920 tempest-ServersAdminTestJSON-14199920-project-member] [instance: 26a0baf9-68bd-4380-87b3-53f50524bbbf] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 547.255529] env[61806]: DEBUG nova.compute.manager [None req-d176ea08-f6ea-453a-9c54-dfba38bfade5 tempest-ServersListShow296Test-1633904279 tempest-ServersListShow296Test-1633904279-project-member] [instance: 5a95d5a0-e5bf-4ca1-bbad-f16b5d10ba45] Starting instance... {{(pid=61806) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 547.325907] env[61806]: DEBUG nova.network.neutron [None req-7d4ae408-54ff-4248-838b-a36cbf45b0c5 tempest-ListServerFiltersTestJSON-241228521 tempest-ListServerFiltersTestJSON-241228521-project-member] [instance: 9ddec22f-0fce-4d28-b97e-92b179f829f6] Updating instance_info_cache with network_info: [] {{(pid=61806) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 547.751419] env[61806]: DEBUG nova.compute.utils [None req-aac795d3-c37f-4811-9be9-aa67ad520ab0 tempest-ListServerFiltersTestJSON-241228521 tempest-ListServerFiltersTestJSON-241228521-project-member] Using /dev/sd instead of None {{(pid=61806) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 547.757257] env[61806]: DEBUG nova.compute.manager [None req-aac795d3-c37f-4811-9be9-aa67ad520ab0 tempest-ListServerFiltersTestJSON-241228521 tempest-ListServerFiltersTestJSON-241228521-project-member] [instance: 7fcee323-8ea6-4f69-9f1e-f2105af0fe96] Allocating IP information in the background. {{(pid=61806) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 547.757423] env[61806]: DEBUG nova.network.neutron [None req-aac795d3-c37f-4811-9be9-aa67ad520ab0 tempest-ListServerFiltersTestJSON-241228521 tempest-ListServerFiltersTestJSON-241228521-project-member] [instance: 7fcee323-8ea6-4f69-9f1e-f2105af0fe96] allocate_for_instance() {{(pid=61806) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 547.761445] env[61806]: DEBUG nova.compute.manager [None req-d176ea08-f6ea-453a-9c54-dfba38bfade5 tempest-ServersListShow296Test-1633904279 tempest-ServersListShow296Test-1633904279-project-member] [instance: 5a95d5a0-e5bf-4ca1-bbad-f16b5d10ba45] Instance disappeared before build. {{(pid=61806) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2439}} [ 547.833306] env[61806]: DEBUG oslo_concurrency.lockutils [None req-7d4ae408-54ff-4248-838b-a36cbf45b0c5 tempest-ListServerFiltersTestJSON-241228521 tempest-ListServerFiltersTestJSON-241228521-project-member] Releasing lock "refresh_cache-9ddec22f-0fce-4d28-b97e-92b179f829f6" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 547.833306] env[61806]: DEBUG nova.compute.manager [None req-7d4ae408-54ff-4248-838b-a36cbf45b0c5 tempest-ListServerFiltersTestJSON-241228521 tempest-ListServerFiltersTestJSON-241228521-project-member] [instance: 9ddec22f-0fce-4d28-b97e-92b179f829f6] Start destroying the instance on the hypervisor. {{(pid=61806) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 547.833306] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-7d4ae408-54ff-4248-838b-a36cbf45b0c5 tempest-ListServerFiltersTestJSON-241228521 tempest-ListServerFiltersTestJSON-241228521-project-member] [instance: 9ddec22f-0fce-4d28-b97e-92b179f829f6] Destroying instance {{(pid=61806) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 547.833306] env[61806]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-025f2d47-9e99-419f-8a0f-d3d6a5233784 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 547.843111] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bf0f2f91-a244-4ae8-862e-a799c7756087 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 547.873880] env[61806]: WARNING nova.virt.vmwareapi.vmops [None req-7d4ae408-54ff-4248-838b-a36cbf45b0c5 tempest-ListServerFiltersTestJSON-241228521 tempest-ListServerFiltersTestJSON-241228521-project-member] [instance: 9ddec22f-0fce-4d28-b97e-92b179f829f6] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 9ddec22f-0fce-4d28-b97e-92b179f829f6 could not be found. [ 547.875233] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-7d4ae408-54ff-4248-838b-a36cbf45b0c5 tempest-ListServerFiltersTestJSON-241228521 tempest-ListServerFiltersTestJSON-241228521-project-member] [instance: 9ddec22f-0fce-4d28-b97e-92b179f829f6] Instance destroyed {{(pid=61806) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 547.875737] env[61806]: INFO nova.compute.manager [None req-7d4ae408-54ff-4248-838b-a36cbf45b0c5 tempest-ListServerFiltersTestJSON-241228521 tempest-ListServerFiltersTestJSON-241228521-project-member] [instance: 9ddec22f-0fce-4d28-b97e-92b179f829f6] Took 0.05 seconds to destroy the instance on the hypervisor. [ 547.876125] env[61806]: DEBUG oslo.service.loopingcall [None req-7d4ae408-54ff-4248-838b-a36cbf45b0c5 tempest-ListServerFiltersTestJSON-241228521 tempest-ListServerFiltersTestJSON-241228521-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61806) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 547.876458] env[61806]: DEBUG nova.compute.manager [-] [instance: 9ddec22f-0fce-4d28-b97e-92b179f829f6] Deallocating network for instance {{(pid=61806) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 547.877321] env[61806]: DEBUG nova.network.neutron [-] [instance: 9ddec22f-0fce-4d28-b97e-92b179f829f6] deallocate_for_instance() {{(pid=61806) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 547.920606] env[61806]: DEBUG nova.network.neutron [-] [instance: 9ddec22f-0fce-4d28-b97e-92b179f829f6] Instance cache missing network info. {{(pid=61806) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 547.929142] env[61806]: DEBUG nova.policy [None req-aac795d3-c37f-4811-9be9-aa67ad520ab0 tempest-ListServerFiltersTestJSON-241228521 tempest-ListServerFiltersTestJSON-241228521-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'de9850b7a2714a5fa0eccc2dfe2af770', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '8e08874243144f6d871ac98314e24d3c', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61806) authorize /opt/stack/nova/nova/policy.py:201}} [ 548.126394] env[61806]: INFO nova.scheduler.client.report [None req-2150b9c2-a011-40b9-950a-cddb0d2b8fb6 tempest-ImagesNegativeTestJSON-436875204 tempest-ImagesNegativeTestJSON-436875204-project-member] Deleted allocations for instance 71380a36-8926-45d3-a70e-a7112bd1d06f [ 548.259317] env[61806]: DEBUG nova.compute.manager [None req-aac795d3-c37f-4811-9be9-aa67ad520ab0 tempest-ListServerFiltersTestJSON-241228521 tempest-ListServerFiltersTestJSON-241228521-project-member] [instance: 7fcee323-8ea6-4f69-9f1e-f2105af0fe96] Start building block device mappings for instance. {{(pid=61806) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 548.286473] env[61806]: DEBUG oslo_concurrency.lockutils [None req-d176ea08-f6ea-453a-9c54-dfba38bfade5 tempest-ServersListShow296Test-1633904279 tempest-ServersListShow296Test-1633904279-project-member] Lock "5a95d5a0-e5bf-4ca1-bbad-f16b5d10ba45" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 13.853s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 548.302770] env[61806]: DEBUG nova.compute.manager [req-634d3e6a-6f19-45a6-a5d3-9a0f18ed6dae req-17b698cd-6a18-48a0-97b5-1a3a31e40741 service nova] [instance: 9ddec22f-0fce-4d28-b97e-92b179f829f6] Received event network-vif-deleted-c8f9a917-21ab-46bb-bbd3-1717da5a257b {{(pid=61806) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 548.423121] env[61806]: DEBUG nova.network.neutron [-] [instance: 9ddec22f-0fce-4d28-b97e-92b179f829f6] Updating instance_info_cache with network_info: [] {{(pid=61806) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 548.637681] env[61806]: DEBUG oslo_concurrency.lockutils [None req-2150b9c2-a011-40b9-950a-cddb0d2b8fb6 tempest-ImagesNegativeTestJSON-436875204 tempest-ImagesNegativeTestJSON-436875204-project-member] Lock "71380a36-8926-45d3-a70e-a7112bd1d06f" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 28.564s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 548.674786] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-354b931d-3ebc-4ec1-a622-a8e616708e87 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 548.684927] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c488732a-a578-48cc-8f50-6337a4e58173 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 548.738777] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-02c35b28-e612-4c85-b725-15dc0f2d70ac {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 548.750348] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-578fbb12-e1a4-4303-bd7a-ed5738de9085 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 548.775267] env[61806]: DEBUG nova.compute.provider_tree [None req-55a61261-0862-4d73-b971-6d11cbe88c5e tempest-ServersAdminTestJSON-14199920 tempest-ServersAdminTestJSON-14199920-project-member] Inventory has not changed in ProviderTree for provider: a2858be1-fd22-4e08-979e-87ad25293407 {{(pid=61806) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 548.791285] env[61806]: DEBUG nova.compute.manager [None req-38c4eaec-22d8-4bf0-bd45-321ef43329a0 tempest-ServersAdminTestJSON-14199920 tempest-ServersAdminTestJSON-14199920-project-member] [instance: 8905f20f-28a6-49f4-88bf-e52177c6a4b6] Starting instance... {{(pid=61806) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 548.932296] env[61806]: INFO nova.compute.manager [-] [instance: 9ddec22f-0fce-4d28-b97e-92b179f829f6] Took 1.05 seconds to deallocate network for instance. [ 548.934052] env[61806]: DEBUG nova.compute.claims [None req-7d4ae408-54ff-4248-838b-a36cbf45b0c5 tempest-ListServerFiltersTestJSON-241228521 tempest-ListServerFiltersTestJSON-241228521-project-member] [instance: 9ddec22f-0fce-4d28-b97e-92b179f829f6] Aborting claim: {{(pid=61806) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 548.934800] env[61806]: DEBUG oslo_concurrency.lockutils [None req-7d4ae408-54ff-4248-838b-a36cbf45b0c5 tempest-ListServerFiltersTestJSON-241228521 tempest-ListServerFiltersTestJSON-241228521-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 548.953335] env[61806]: DEBUG nova.network.neutron [None req-aac795d3-c37f-4811-9be9-aa67ad520ab0 tempest-ListServerFiltersTestJSON-241228521 tempest-ListServerFiltersTestJSON-241228521-project-member] [instance: 7fcee323-8ea6-4f69-9f1e-f2105af0fe96] Successfully created port: eed1949d-29dd-4127-a2a4-d7dc2369df55 {{(pid=61806) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 549.103877] env[61806]: DEBUG oslo_concurrency.lockutils [None req-24088d08-d3d7-4f62-ad63-91029bd9b22d tempest-ServersTestMultiNic-1480543082 tempest-ServersTestMultiNic-1480543082-project-member] Acquiring lock "c2d1c9f6-f458-4843-982c-afdb2a90edae" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 549.104013] env[61806]: DEBUG oslo_concurrency.lockutils [None req-24088d08-d3d7-4f62-ad63-91029bd9b22d tempest-ServersTestMultiNic-1480543082 tempest-ServersTestMultiNic-1480543082-project-member] Lock "c2d1c9f6-f458-4843-982c-afdb2a90edae" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 549.141831] env[61806]: DEBUG nova.compute.manager [None req-37e71da7-cb5e-431d-ba3c-95daaf8af7f3 tempest-ServersAdminNegativeTestJSON-2128564044 tempest-ServersAdminNegativeTestJSON-2128564044-project-member] [instance: 89ce0fd8-6a22-4b1c-bdff-d1584abd538a] Starting instance... {{(pid=61806) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 549.280999] env[61806]: DEBUG nova.compute.manager [None req-aac795d3-c37f-4811-9be9-aa67ad520ab0 tempest-ListServerFiltersTestJSON-241228521 tempest-ListServerFiltersTestJSON-241228521-project-member] [instance: 7fcee323-8ea6-4f69-9f1e-f2105af0fe96] Start spawning the instance on the hypervisor. {{(pid=61806) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 549.285583] env[61806]: DEBUG nova.scheduler.client.report [None req-55a61261-0862-4d73-b971-6d11cbe88c5e tempest-ServersAdminTestJSON-14199920 tempest-ServersAdminTestJSON-14199920-project-member] Inventory has not changed for provider a2858be1-fd22-4e08-979e-87ad25293407 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 139, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61806) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 549.320649] env[61806]: DEBUG oslo_concurrency.lockutils [None req-38c4eaec-22d8-4bf0-bd45-321ef43329a0 tempest-ServersAdminTestJSON-14199920 tempest-ServersAdminTestJSON-14199920-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 549.328193] env[61806]: DEBUG nova.virt.hardware [None req-aac795d3-c37f-4811-9be9-aa67ad520ab0 tempest-ListServerFiltersTestJSON-241228521 tempest-ListServerFiltersTestJSON-241228521-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-15T18:05:12Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='84',id=12,is_public=True,memory_mb=256,name='m1.micro',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-15T18:04:52Z,direct_url=,disk_format='vmdk',id=ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='84ba83fce2c349a6988173c5d6fc3b07',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-15T18:04:53Z,virtual_size=,visibility=), allow threads: False {{(pid=61806) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 549.329373] env[61806]: DEBUG nova.virt.hardware [None req-aac795d3-c37f-4811-9be9-aa67ad520ab0 tempest-ListServerFiltersTestJSON-241228521 tempest-ListServerFiltersTestJSON-241228521-project-member] Flavor limits 0:0:0 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 549.330432] env[61806]: DEBUG nova.virt.hardware [None req-aac795d3-c37f-4811-9be9-aa67ad520ab0 tempest-ListServerFiltersTestJSON-241228521 tempest-ListServerFiltersTestJSON-241228521-project-member] Image limits 0:0:0 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 549.330432] env[61806]: DEBUG nova.virt.hardware [None req-aac795d3-c37f-4811-9be9-aa67ad520ab0 tempest-ListServerFiltersTestJSON-241228521 tempest-ListServerFiltersTestJSON-241228521-project-member] Flavor pref 0:0:0 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 549.330432] env[61806]: DEBUG nova.virt.hardware [None req-aac795d3-c37f-4811-9be9-aa67ad520ab0 tempest-ListServerFiltersTestJSON-241228521 tempest-ListServerFiltersTestJSON-241228521-project-member] Image pref 0:0:0 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 549.330432] env[61806]: DEBUG nova.virt.hardware [None req-aac795d3-c37f-4811-9be9-aa67ad520ab0 tempest-ListServerFiltersTestJSON-241228521 tempest-ListServerFiltersTestJSON-241228521-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 549.332264] env[61806]: DEBUG nova.virt.hardware [None req-aac795d3-c37f-4811-9be9-aa67ad520ab0 tempest-ListServerFiltersTestJSON-241228521 tempest-ListServerFiltersTestJSON-241228521-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61806) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 549.332348] env[61806]: DEBUG nova.virt.hardware [None req-aac795d3-c37f-4811-9be9-aa67ad520ab0 tempest-ListServerFiltersTestJSON-241228521 tempest-ListServerFiltersTestJSON-241228521-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61806) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 549.333310] env[61806]: DEBUG nova.virt.hardware [None req-aac795d3-c37f-4811-9be9-aa67ad520ab0 tempest-ListServerFiltersTestJSON-241228521 tempest-ListServerFiltersTestJSON-241228521-project-member] Got 1 possible topologies {{(pid=61806) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 549.333441] env[61806]: DEBUG nova.virt.hardware [None req-aac795d3-c37f-4811-9be9-aa67ad520ab0 tempest-ListServerFiltersTestJSON-241228521 tempest-ListServerFiltersTestJSON-241228521-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61806) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 549.333754] env[61806]: DEBUG nova.virt.hardware [None req-aac795d3-c37f-4811-9be9-aa67ad520ab0 tempest-ListServerFiltersTestJSON-241228521 tempest-ListServerFiltersTestJSON-241228521-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61806) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 549.335048] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d485d546-67d5-4310-9598-cdb4c5b77027 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 549.348666] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-29ca6a80-f425-4b3a-9e6f-b02d41c68260 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 549.683366] env[61806]: DEBUG oslo_concurrency.lockutils [None req-37e71da7-cb5e-431d-ba3c-95daaf8af7f3 tempest-ServersAdminNegativeTestJSON-2128564044 tempest-ServersAdminNegativeTestJSON-2128564044-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 549.792043] env[61806]: DEBUG oslo_concurrency.lockutils [None req-55a61261-0862-4d73-b971-6d11cbe88c5e tempest-ServersAdminTestJSON-14199920 tempest-ServersAdminTestJSON-14199920-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.550s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 549.793072] env[61806]: DEBUG nova.compute.manager [None req-55a61261-0862-4d73-b971-6d11cbe88c5e tempest-ServersAdminTestJSON-14199920 tempest-ServersAdminTestJSON-14199920-project-member] [instance: 26a0baf9-68bd-4380-87b3-53f50524bbbf] Start building networks asynchronously for instance. {{(pid=61806) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 549.799165] env[61806]: DEBUG oslo_concurrency.lockutils [None req-0858196d-347c-4201-9cb8-c7f147f5d99d tempest-MigrationsAdminTest-470098049 tempest-MigrationsAdminTest-470098049-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 13.358s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 550.308140] env[61806]: DEBUG nova.compute.utils [None req-55a61261-0862-4d73-b971-6d11cbe88c5e tempest-ServersAdminTestJSON-14199920 tempest-ServersAdminTestJSON-14199920-project-member] Using /dev/sd instead of None {{(pid=61806) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 550.316219] env[61806]: DEBUG nova.compute.manager [None req-55a61261-0862-4d73-b971-6d11cbe88c5e tempest-ServersAdminTestJSON-14199920 tempest-ServersAdminTestJSON-14199920-project-member] [instance: 26a0baf9-68bd-4380-87b3-53f50524bbbf] Allocating IP information in the background. {{(pid=61806) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 550.317653] env[61806]: DEBUG nova.network.neutron [None req-55a61261-0862-4d73-b971-6d11cbe88c5e tempest-ServersAdminTestJSON-14199920 tempest-ServersAdminTestJSON-14199920-project-member] [instance: 26a0baf9-68bd-4380-87b3-53f50524bbbf] allocate_for_instance() {{(pid=61806) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 550.436738] env[61806]: DEBUG nova.policy [None req-55a61261-0862-4d73-b971-6d11cbe88c5e tempest-ServersAdminTestJSON-14199920 tempest-ServersAdminTestJSON-14199920-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '4147cb03f0904970bf0784dccb850d0b', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '738e028f263e4026ba95a5d491905d22', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61806) authorize /opt/stack/nova/nova/policy.py:201}} [ 550.627292] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f139ad78-029a-41b7-bf09-57f6b507df01 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 550.641768] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f044efd1-f751-41d7-9ebc-5d8c2ceac903 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 550.683093] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f773ca40-cbb8-4048-b13e-06e939a4748b {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 550.692595] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-470f4a38-4e56-4f11-885e-ff703465dd5c {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 550.708232] env[61806]: DEBUG nova.compute.provider_tree [None req-0858196d-347c-4201-9cb8-c7f147f5d99d tempest-MigrationsAdminTest-470098049 tempest-MigrationsAdminTest-470098049-project-member] Inventory has not changed in ProviderTree for provider: a2858be1-fd22-4e08-979e-87ad25293407 {{(pid=61806) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 550.813632] env[61806]: DEBUG nova.compute.manager [None req-55a61261-0862-4d73-b971-6d11cbe88c5e tempest-ServersAdminTestJSON-14199920 tempest-ServersAdminTestJSON-14199920-project-member] [instance: 26a0baf9-68bd-4380-87b3-53f50524bbbf] Start building block device mappings for instance. {{(pid=61806) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 551.212158] env[61806]: DEBUG nova.scheduler.client.report [None req-0858196d-347c-4201-9cb8-c7f147f5d99d tempest-MigrationsAdminTest-470098049 tempest-MigrationsAdminTest-470098049-project-member] Inventory has not changed for provider a2858be1-fd22-4e08-979e-87ad25293407 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 139, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61806) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 551.485924] env[61806]: DEBUG oslo_concurrency.lockutils [None req-e34d2bcf-266f-4faa-8f31-6bd93f6ab176 tempest-ImagesTestJSON-618852556 tempest-ImagesTestJSON-618852556-project-member] Acquiring lock "65f3b2f3-6d25-44f0-95cb-f9a7e75f8e82" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 551.485924] env[61806]: DEBUG oslo_concurrency.lockutils [None req-e34d2bcf-266f-4faa-8f31-6bd93f6ab176 tempest-ImagesTestJSON-618852556 tempest-ImagesTestJSON-618852556-project-member] Lock "65f3b2f3-6d25-44f0-95cb-f9a7e75f8e82" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 551.586534] env[61806]: DEBUG nova.network.neutron [None req-55a61261-0862-4d73-b971-6d11cbe88c5e tempest-ServersAdminTestJSON-14199920 tempest-ServersAdminTestJSON-14199920-project-member] [instance: 26a0baf9-68bd-4380-87b3-53f50524bbbf] Successfully created port: 31cd9389-2251-4222-885e-268c99301a23 {{(pid=61806) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 551.723459] env[61806]: DEBUG oslo_concurrency.lockutils [None req-0858196d-347c-4201-9cb8-c7f147f5d99d tempest-MigrationsAdminTest-470098049 tempest-MigrationsAdminTest-470098049-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.922s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 551.723459] env[61806]: ERROR nova.compute.manager [None req-0858196d-347c-4201-9cb8-c7f147f5d99d tempest-MigrationsAdminTest-470098049 tempest-MigrationsAdminTest-470098049-project-member] [instance: 52b19347-a08f-44a7-8aad-22a3af5e291d] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port d93797ab-56f6-476d-b147-8051439da3f2, please check neutron logs for more information. [ 551.723459] env[61806]: ERROR nova.compute.manager [instance: 52b19347-a08f-44a7-8aad-22a3af5e291d] Traceback (most recent call last): [ 551.723459] env[61806]: ERROR nova.compute.manager [instance: 52b19347-a08f-44a7-8aad-22a3af5e291d] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 551.723459] env[61806]: ERROR nova.compute.manager [instance: 52b19347-a08f-44a7-8aad-22a3af5e291d] self.driver.spawn(context, instance, image_meta, [ 551.723459] env[61806]: ERROR nova.compute.manager [instance: 52b19347-a08f-44a7-8aad-22a3af5e291d] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 551.723459] env[61806]: ERROR nova.compute.manager [instance: 52b19347-a08f-44a7-8aad-22a3af5e291d] self._vmops.spawn(context, instance, image_meta, injected_files, [ 551.723459] env[61806]: ERROR nova.compute.manager [instance: 52b19347-a08f-44a7-8aad-22a3af5e291d] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 551.723459] env[61806]: ERROR nova.compute.manager [instance: 52b19347-a08f-44a7-8aad-22a3af5e291d] vm_ref = self.build_virtual_machine(instance, [ 551.724708] env[61806]: ERROR nova.compute.manager [instance: 52b19347-a08f-44a7-8aad-22a3af5e291d] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 551.724708] env[61806]: ERROR nova.compute.manager [instance: 52b19347-a08f-44a7-8aad-22a3af5e291d] vif_infos = vmwarevif.get_vif_info(self._session, [ 551.724708] env[61806]: ERROR nova.compute.manager [instance: 52b19347-a08f-44a7-8aad-22a3af5e291d] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 551.724708] env[61806]: ERROR nova.compute.manager [instance: 52b19347-a08f-44a7-8aad-22a3af5e291d] for vif in network_info: [ 551.724708] env[61806]: ERROR nova.compute.manager [instance: 52b19347-a08f-44a7-8aad-22a3af5e291d] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 551.724708] env[61806]: ERROR nova.compute.manager [instance: 52b19347-a08f-44a7-8aad-22a3af5e291d] return self._sync_wrapper(fn, *args, **kwargs) [ 551.724708] env[61806]: ERROR nova.compute.manager [instance: 52b19347-a08f-44a7-8aad-22a3af5e291d] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 551.724708] env[61806]: ERROR nova.compute.manager [instance: 52b19347-a08f-44a7-8aad-22a3af5e291d] self.wait() [ 551.724708] env[61806]: ERROR nova.compute.manager [instance: 52b19347-a08f-44a7-8aad-22a3af5e291d] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 551.724708] env[61806]: ERROR nova.compute.manager [instance: 52b19347-a08f-44a7-8aad-22a3af5e291d] self[:] = self._gt.wait() [ 551.724708] env[61806]: ERROR nova.compute.manager [instance: 52b19347-a08f-44a7-8aad-22a3af5e291d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 551.724708] env[61806]: ERROR nova.compute.manager [instance: 52b19347-a08f-44a7-8aad-22a3af5e291d] return self._exit_event.wait() [ 551.724708] env[61806]: ERROR nova.compute.manager [instance: 52b19347-a08f-44a7-8aad-22a3af5e291d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 551.726203] env[61806]: ERROR nova.compute.manager [instance: 52b19347-a08f-44a7-8aad-22a3af5e291d] result = hub.switch() [ 551.726203] env[61806]: ERROR nova.compute.manager [instance: 52b19347-a08f-44a7-8aad-22a3af5e291d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 551.726203] env[61806]: ERROR nova.compute.manager [instance: 52b19347-a08f-44a7-8aad-22a3af5e291d] return self.greenlet.switch() [ 551.726203] env[61806]: ERROR nova.compute.manager [instance: 52b19347-a08f-44a7-8aad-22a3af5e291d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 551.726203] env[61806]: ERROR nova.compute.manager [instance: 52b19347-a08f-44a7-8aad-22a3af5e291d] result = function(*args, **kwargs) [ 551.726203] env[61806]: ERROR nova.compute.manager [instance: 52b19347-a08f-44a7-8aad-22a3af5e291d] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 551.726203] env[61806]: ERROR nova.compute.manager [instance: 52b19347-a08f-44a7-8aad-22a3af5e291d] return func(*args, **kwargs) [ 551.726203] env[61806]: ERROR nova.compute.manager [instance: 52b19347-a08f-44a7-8aad-22a3af5e291d] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 551.726203] env[61806]: ERROR nova.compute.manager [instance: 52b19347-a08f-44a7-8aad-22a3af5e291d] raise e [ 551.726203] env[61806]: ERROR nova.compute.manager [instance: 52b19347-a08f-44a7-8aad-22a3af5e291d] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 551.726203] env[61806]: ERROR nova.compute.manager [instance: 52b19347-a08f-44a7-8aad-22a3af5e291d] nwinfo = self.network_api.allocate_for_instance( [ 551.726203] env[61806]: ERROR nova.compute.manager [instance: 52b19347-a08f-44a7-8aad-22a3af5e291d] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 551.726203] env[61806]: ERROR nova.compute.manager [instance: 52b19347-a08f-44a7-8aad-22a3af5e291d] created_port_ids = self._update_ports_for_instance( [ 551.726656] env[61806]: ERROR nova.compute.manager [instance: 52b19347-a08f-44a7-8aad-22a3af5e291d] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 551.726656] env[61806]: ERROR nova.compute.manager [instance: 52b19347-a08f-44a7-8aad-22a3af5e291d] with excutils.save_and_reraise_exception(): [ 551.726656] env[61806]: ERROR nova.compute.manager [instance: 52b19347-a08f-44a7-8aad-22a3af5e291d] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 551.726656] env[61806]: ERROR nova.compute.manager [instance: 52b19347-a08f-44a7-8aad-22a3af5e291d] self.force_reraise() [ 551.726656] env[61806]: ERROR nova.compute.manager [instance: 52b19347-a08f-44a7-8aad-22a3af5e291d] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 551.726656] env[61806]: ERROR nova.compute.manager [instance: 52b19347-a08f-44a7-8aad-22a3af5e291d] raise self.value [ 551.726656] env[61806]: ERROR nova.compute.manager [instance: 52b19347-a08f-44a7-8aad-22a3af5e291d] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 551.726656] env[61806]: ERROR nova.compute.manager [instance: 52b19347-a08f-44a7-8aad-22a3af5e291d] updated_port = self._update_port( [ 551.726656] env[61806]: ERROR nova.compute.manager [instance: 52b19347-a08f-44a7-8aad-22a3af5e291d] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 551.726656] env[61806]: ERROR nova.compute.manager [instance: 52b19347-a08f-44a7-8aad-22a3af5e291d] _ensure_no_port_binding_failure(port) [ 551.726656] env[61806]: ERROR nova.compute.manager [instance: 52b19347-a08f-44a7-8aad-22a3af5e291d] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 551.726656] env[61806]: ERROR nova.compute.manager [instance: 52b19347-a08f-44a7-8aad-22a3af5e291d] raise exception.PortBindingFailed(port_id=port['id']) [ 551.732536] env[61806]: ERROR nova.compute.manager [instance: 52b19347-a08f-44a7-8aad-22a3af5e291d] nova.exception.PortBindingFailed: Binding failed for port d93797ab-56f6-476d-b147-8051439da3f2, please check neutron logs for more information. [ 551.732536] env[61806]: ERROR nova.compute.manager [instance: 52b19347-a08f-44a7-8aad-22a3af5e291d] [ 551.732536] env[61806]: DEBUG nova.compute.utils [None req-0858196d-347c-4201-9cb8-c7f147f5d99d tempest-MigrationsAdminTest-470098049 tempest-MigrationsAdminTest-470098049-project-member] [instance: 52b19347-a08f-44a7-8aad-22a3af5e291d] Binding failed for port d93797ab-56f6-476d-b147-8051439da3f2, please check neutron logs for more information. {{(pid=61806) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 551.732536] env[61806]: DEBUG oslo_concurrency.lockutils [None req-3d02814b-d8ff-4079-980b-6c580616ce84 tempest-AttachInterfacesUnderV243Test-1683422373 tempest-AttachInterfacesUnderV243Test-1683422373-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 14.736s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 551.732536] env[61806]: DEBUG nova.compute.manager [None req-0858196d-347c-4201-9cb8-c7f147f5d99d tempest-MigrationsAdminTest-470098049 tempest-MigrationsAdminTest-470098049-project-member] [instance: 52b19347-a08f-44a7-8aad-22a3af5e291d] Build of instance 52b19347-a08f-44a7-8aad-22a3af5e291d was re-scheduled: Binding failed for port d93797ab-56f6-476d-b147-8051439da3f2, please check neutron logs for more information. {{(pid=61806) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 551.732825] env[61806]: DEBUG nova.compute.manager [None req-0858196d-347c-4201-9cb8-c7f147f5d99d tempest-MigrationsAdminTest-470098049 tempest-MigrationsAdminTest-470098049-project-member] [instance: 52b19347-a08f-44a7-8aad-22a3af5e291d] Unplugging VIFs for instance {{(pid=61806) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 551.733048] env[61806]: DEBUG oslo_concurrency.lockutils [None req-0858196d-347c-4201-9cb8-c7f147f5d99d tempest-MigrationsAdminTest-470098049 tempest-MigrationsAdminTest-470098049-project-member] Acquiring lock "refresh_cache-52b19347-a08f-44a7-8aad-22a3af5e291d" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 551.733145] env[61806]: DEBUG oslo_concurrency.lockutils [None req-0858196d-347c-4201-9cb8-c7f147f5d99d tempest-MigrationsAdminTest-470098049 tempest-MigrationsAdminTest-470098049-project-member] Acquired lock "refresh_cache-52b19347-a08f-44a7-8aad-22a3af5e291d" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 551.733240] env[61806]: DEBUG nova.network.neutron [None req-0858196d-347c-4201-9cb8-c7f147f5d99d tempest-MigrationsAdminTest-470098049 tempest-MigrationsAdminTest-470098049-project-member] [instance: 52b19347-a08f-44a7-8aad-22a3af5e291d] Building network info cache for instance {{(pid=61806) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 551.830026] env[61806]: DEBUG nova.compute.manager [None req-55a61261-0862-4d73-b971-6d11cbe88c5e tempest-ServersAdminTestJSON-14199920 tempest-ServersAdminTestJSON-14199920-project-member] [instance: 26a0baf9-68bd-4380-87b3-53f50524bbbf] Start spawning the instance on the hypervisor. {{(pid=61806) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 551.867396] env[61806]: DEBUG nova.virt.hardware [None req-55a61261-0862-4d73-b971-6d11cbe88c5e tempest-ServersAdminTestJSON-14199920 tempest-ServersAdminTestJSON-14199920-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-15T18:05:10Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-15T18:04:52Z,direct_url=,disk_format='vmdk',id=ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='84ba83fce2c349a6988173c5d6fc3b07',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-15T18:04:53Z,virtual_size=,visibility=), allow threads: False {{(pid=61806) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 551.867711] env[61806]: DEBUG nova.virt.hardware [None req-55a61261-0862-4d73-b971-6d11cbe88c5e tempest-ServersAdminTestJSON-14199920 tempest-ServersAdminTestJSON-14199920-project-member] Flavor limits 0:0:0 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 551.867931] env[61806]: DEBUG nova.virt.hardware [None req-55a61261-0862-4d73-b971-6d11cbe88c5e tempest-ServersAdminTestJSON-14199920 tempest-ServersAdminTestJSON-14199920-project-member] Image limits 0:0:0 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 551.868228] env[61806]: DEBUG nova.virt.hardware [None req-55a61261-0862-4d73-b971-6d11cbe88c5e tempest-ServersAdminTestJSON-14199920 tempest-ServersAdminTestJSON-14199920-project-member] Flavor pref 0:0:0 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 551.868390] env[61806]: DEBUG nova.virt.hardware [None req-55a61261-0862-4d73-b971-6d11cbe88c5e tempest-ServersAdminTestJSON-14199920 tempest-ServersAdminTestJSON-14199920-project-member] Image pref 0:0:0 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 551.868844] env[61806]: DEBUG nova.virt.hardware [None req-55a61261-0862-4d73-b971-6d11cbe88c5e tempest-ServersAdminTestJSON-14199920 tempest-ServersAdminTestJSON-14199920-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 551.868844] env[61806]: DEBUG nova.virt.hardware [None req-55a61261-0862-4d73-b971-6d11cbe88c5e tempest-ServersAdminTestJSON-14199920 tempest-ServersAdminTestJSON-14199920-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61806) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 551.869045] env[61806]: DEBUG nova.virt.hardware [None req-55a61261-0862-4d73-b971-6d11cbe88c5e tempest-ServersAdminTestJSON-14199920 tempest-ServersAdminTestJSON-14199920-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61806) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 551.869086] env[61806]: DEBUG nova.virt.hardware [None req-55a61261-0862-4d73-b971-6d11cbe88c5e tempest-ServersAdminTestJSON-14199920 tempest-ServersAdminTestJSON-14199920-project-member] Got 1 possible topologies {{(pid=61806) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 551.869317] env[61806]: DEBUG nova.virt.hardware [None req-55a61261-0862-4d73-b971-6d11cbe88c5e tempest-ServersAdminTestJSON-14199920 tempest-ServersAdminTestJSON-14199920-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61806) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 551.869485] env[61806]: DEBUG nova.virt.hardware [None req-55a61261-0862-4d73-b971-6d11cbe88c5e tempest-ServersAdminTestJSON-14199920 tempest-ServersAdminTestJSON-14199920-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61806) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 551.870396] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d130b440-e3df-4892-8d00-47aeff167ac9 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 551.881990] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7b75072d-0d22-4f18-bd2b-116c93739bbb {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 552.283703] env[61806]: DEBUG nova.network.neutron [None req-0858196d-347c-4201-9cb8-c7f147f5d99d tempest-MigrationsAdminTest-470098049 tempest-MigrationsAdminTest-470098049-project-member] [instance: 52b19347-a08f-44a7-8aad-22a3af5e291d] Instance cache missing network info. {{(pid=61806) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 552.465537] env[61806]: DEBUG nova.network.neutron [None req-0858196d-347c-4201-9cb8-c7f147f5d99d tempest-MigrationsAdminTest-470098049 tempest-MigrationsAdminTest-470098049-project-member] [instance: 52b19347-a08f-44a7-8aad-22a3af5e291d] Updating instance_info_cache with network_info: [] {{(pid=61806) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 552.628341] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-63db0fbd-e907-4fae-a232-37cc23227b6f {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 552.641644] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-20d842f7-0465-4bde-8c7d-f10dcfb77113 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 552.690433] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-52531fc5-2e3f-447a-a99c-6efbe71165fb {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 552.697774] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ce247b1a-ebee-4406-b236-45834149ffb3 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 552.712817] env[61806]: DEBUG nova.compute.provider_tree [None req-3d02814b-d8ff-4079-980b-6c580616ce84 tempest-AttachInterfacesUnderV243Test-1683422373 tempest-AttachInterfacesUnderV243Test-1683422373-project-member] Inventory has not changed in ProviderTree for provider: a2858be1-fd22-4e08-979e-87ad25293407 {{(pid=61806) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 552.770652] env[61806]: ERROR nova.compute.manager [None req-aac795d3-c37f-4811-9be9-aa67ad520ab0 tempest-ListServerFiltersTestJSON-241228521 tempest-ListServerFiltersTestJSON-241228521-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port eed1949d-29dd-4127-a2a4-d7dc2369df55, please check neutron logs for more information. [ 552.770652] env[61806]: ERROR nova.compute.manager Traceback (most recent call last): [ 552.770652] env[61806]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 552.770652] env[61806]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 552.770652] env[61806]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 552.770652] env[61806]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 552.770652] env[61806]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 552.770652] env[61806]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 552.770652] env[61806]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 552.770652] env[61806]: ERROR nova.compute.manager self.force_reraise() [ 552.770652] env[61806]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 552.770652] env[61806]: ERROR nova.compute.manager raise self.value [ 552.770652] env[61806]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 552.770652] env[61806]: ERROR nova.compute.manager updated_port = self._update_port( [ 552.770652] env[61806]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 552.770652] env[61806]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 552.771199] env[61806]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 552.771199] env[61806]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 552.771199] env[61806]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port eed1949d-29dd-4127-a2a4-d7dc2369df55, please check neutron logs for more information. [ 552.771199] env[61806]: ERROR nova.compute.manager [ 552.771199] env[61806]: Traceback (most recent call last): [ 552.771199] env[61806]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 552.771199] env[61806]: listener.cb(fileno) [ 552.771199] env[61806]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 552.771199] env[61806]: result = function(*args, **kwargs) [ 552.771199] env[61806]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 552.771199] env[61806]: return func(*args, **kwargs) [ 552.771199] env[61806]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 552.771199] env[61806]: raise e [ 552.771199] env[61806]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 552.771199] env[61806]: nwinfo = self.network_api.allocate_for_instance( [ 552.771199] env[61806]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 552.771199] env[61806]: created_port_ids = self._update_ports_for_instance( [ 552.771199] env[61806]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 552.771199] env[61806]: with excutils.save_and_reraise_exception(): [ 552.771199] env[61806]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 552.771199] env[61806]: self.force_reraise() [ 552.771199] env[61806]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 552.771199] env[61806]: raise self.value [ 552.771199] env[61806]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 552.771199] env[61806]: updated_port = self._update_port( [ 552.771199] env[61806]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 552.771199] env[61806]: _ensure_no_port_binding_failure(port) [ 552.771199] env[61806]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 552.771199] env[61806]: raise exception.PortBindingFailed(port_id=port['id']) [ 552.772073] env[61806]: nova.exception.PortBindingFailed: Binding failed for port eed1949d-29dd-4127-a2a4-d7dc2369df55, please check neutron logs for more information. [ 552.772073] env[61806]: Removing descriptor: 16 [ 552.772073] env[61806]: ERROR nova.compute.manager [None req-aac795d3-c37f-4811-9be9-aa67ad520ab0 tempest-ListServerFiltersTestJSON-241228521 tempest-ListServerFiltersTestJSON-241228521-project-member] [instance: 7fcee323-8ea6-4f69-9f1e-f2105af0fe96] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port eed1949d-29dd-4127-a2a4-d7dc2369df55, please check neutron logs for more information. [ 552.772073] env[61806]: ERROR nova.compute.manager [instance: 7fcee323-8ea6-4f69-9f1e-f2105af0fe96] Traceback (most recent call last): [ 552.772073] env[61806]: ERROR nova.compute.manager [instance: 7fcee323-8ea6-4f69-9f1e-f2105af0fe96] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 552.772073] env[61806]: ERROR nova.compute.manager [instance: 7fcee323-8ea6-4f69-9f1e-f2105af0fe96] yield resources [ 552.772073] env[61806]: ERROR nova.compute.manager [instance: 7fcee323-8ea6-4f69-9f1e-f2105af0fe96] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 552.772073] env[61806]: ERROR nova.compute.manager [instance: 7fcee323-8ea6-4f69-9f1e-f2105af0fe96] self.driver.spawn(context, instance, image_meta, [ 552.772073] env[61806]: ERROR nova.compute.manager [instance: 7fcee323-8ea6-4f69-9f1e-f2105af0fe96] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 552.772073] env[61806]: ERROR nova.compute.manager [instance: 7fcee323-8ea6-4f69-9f1e-f2105af0fe96] self._vmops.spawn(context, instance, image_meta, injected_files, [ 552.772073] env[61806]: ERROR nova.compute.manager [instance: 7fcee323-8ea6-4f69-9f1e-f2105af0fe96] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 552.772073] env[61806]: ERROR nova.compute.manager [instance: 7fcee323-8ea6-4f69-9f1e-f2105af0fe96] vm_ref = self.build_virtual_machine(instance, [ 552.772495] env[61806]: ERROR nova.compute.manager [instance: 7fcee323-8ea6-4f69-9f1e-f2105af0fe96] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 552.772495] env[61806]: ERROR nova.compute.manager [instance: 7fcee323-8ea6-4f69-9f1e-f2105af0fe96] vif_infos = vmwarevif.get_vif_info(self._session, [ 552.772495] env[61806]: ERROR nova.compute.manager [instance: 7fcee323-8ea6-4f69-9f1e-f2105af0fe96] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 552.772495] env[61806]: ERROR nova.compute.manager [instance: 7fcee323-8ea6-4f69-9f1e-f2105af0fe96] for vif in network_info: [ 552.772495] env[61806]: ERROR nova.compute.manager [instance: 7fcee323-8ea6-4f69-9f1e-f2105af0fe96] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 552.772495] env[61806]: ERROR nova.compute.manager [instance: 7fcee323-8ea6-4f69-9f1e-f2105af0fe96] return self._sync_wrapper(fn, *args, **kwargs) [ 552.772495] env[61806]: ERROR nova.compute.manager [instance: 7fcee323-8ea6-4f69-9f1e-f2105af0fe96] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 552.772495] env[61806]: ERROR nova.compute.manager [instance: 7fcee323-8ea6-4f69-9f1e-f2105af0fe96] self.wait() [ 552.772495] env[61806]: ERROR nova.compute.manager [instance: 7fcee323-8ea6-4f69-9f1e-f2105af0fe96] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 552.772495] env[61806]: ERROR nova.compute.manager [instance: 7fcee323-8ea6-4f69-9f1e-f2105af0fe96] self[:] = self._gt.wait() [ 552.772495] env[61806]: ERROR nova.compute.manager [instance: 7fcee323-8ea6-4f69-9f1e-f2105af0fe96] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 552.772495] env[61806]: ERROR nova.compute.manager [instance: 7fcee323-8ea6-4f69-9f1e-f2105af0fe96] return self._exit_event.wait() [ 552.772495] env[61806]: ERROR nova.compute.manager [instance: 7fcee323-8ea6-4f69-9f1e-f2105af0fe96] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 552.772904] env[61806]: ERROR nova.compute.manager [instance: 7fcee323-8ea6-4f69-9f1e-f2105af0fe96] result = hub.switch() [ 552.772904] env[61806]: ERROR nova.compute.manager [instance: 7fcee323-8ea6-4f69-9f1e-f2105af0fe96] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 552.772904] env[61806]: ERROR nova.compute.manager [instance: 7fcee323-8ea6-4f69-9f1e-f2105af0fe96] return self.greenlet.switch() [ 552.772904] env[61806]: ERROR nova.compute.manager [instance: 7fcee323-8ea6-4f69-9f1e-f2105af0fe96] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 552.772904] env[61806]: ERROR nova.compute.manager [instance: 7fcee323-8ea6-4f69-9f1e-f2105af0fe96] result = function(*args, **kwargs) [ 552.772904] env[61806]: ERROR nova.compute.manager [instance: 7fcee323-8ea6-4f69-9f1e-f2105af0fe96] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 552.772904] env[61806]: ERROR nova.compute.manager [instance: 7fcee323-8ea6-4f69-9f1e-f2105af0fe96] return func(*args, **kwargs) [ 552.772904] env[61806]: ERROR nova.compute.manager [instance: 7fcee323-8ea6-4f69-9f1e-f2105af0fe96] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 552.772904] env[61806]: ERROR nova.compute.manager [instance: 7fcee323-8ea6-4f69-9f1e-f2105af0fe96] raise e [ 552.772904] env[61806]: ERROR nova.compute.manager [instance: 7fcee323-8ea6-4f69-9f1e-f2105af0fe96] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 552.772904] env[61806]: ERROR nova.compute.manager [instance: 7fcee323-8ea6-4f69-9f1e-f2105af0fe96] nwinfo = self.network_api.allocate_for_instance( [ 552.772904] env[61806]: ERROR nova.compute.manager [instance: 7fcee323-8ea6-4f69-9f1e-f2105af0fe96] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 552.772904] env[61806]: ERROR nova.compute.manager [instance: 7fcee323-8ea6-4f69-9f1e-f2105af0fe96] created_port_ids = self._update_ports_for_instance( [ 552.773326] env[61806]: ERROR nova.compute.manager [instance: 7fcee323-8ea6-4f69-9f1e-f2105af0fe96] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 552.773326] env[61806]: ERROR nova.compute.manager [instance: 7fcee323-8ea6-4f69-9f1e-f2105af0fe96] with excutils.save_and_reraise_exception(): [ 552.773326] env[61806]: ERROR nova.compute.manager [instance: 7fcee323-8ea6-4f69-9f1e-f2105af0fe96] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 552.773326] env[61806]: ERROR nova.compute.manager [instance: 7fcee323-8ea6-4f69-9f1e-f2105af0fe96] self.force_reraise() [ 552.773326] env[61806]: ERROR nova.compute.manager [instance: 7fcee323-8ea6-4f69-9f1e-f2105af0fe96] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 552.773326] env[61806]: ERROR nova.compute.manager [instance: 7fcee323-8ea6-4f69-9f1e-f2105af0fe96] raise self.value [ 552.773326] env[61806]: ERROR nova.compute.manager [instance: 7fcee323-8ea6-4f69-9f1e-f2105af0fe96] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 552.773326] env[61806]: ERROR nova.compute.manager [instance: 7fcee323-8ea6-4f69-9f1e-f2105af0fe96] updated_port = self._update_port( [ 552.773326] env[61806]: ERROR nova.compute.manager [instance: 7fcee323-8ea6-4f69-9f1e-f2105af0fe96] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 552.773326] env[61806]: ERROR nova.compute.manager [instance: 7fcee323-8ea6-4f69-9f1e-f2105af0fe96] _ensure_no_port_binding_failure(port) [ 552.773326] env[61806]: ERROR nova.compute.manager [instance: 7fcee323-8ea6-4f69-9f1e-f2105af0fe96] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 552.773326] env[61806]: ERROR nova.compute.manager [instance: 7fcee323-8ea6-4f69-9f1e-f2105af0fe96] raise exception.PortBindingFailed(port_id=port['id']) [ 552.773881] env[61806]: ERROR nova.compute.manager [instance: 7fcee323-8ea6-4f69-9f1e-f2105af0fe96] nova.exception.PortBindingFailed: Binding failed for port eed1949d-29dd-4127-a2a4-d7dc2369df55, please check neutron logs for more information. [ 552.773881] env[61806]: ERROR nova.compute.manager [instance: 7fcee323-8ea6-4f69-9f1e-f2105af0fe96] [ 552.773881] env[61806]: INFO nova.compute.manager [None req-aac795d3-c37f-4811-9be9-aa67ad520ab0 tempest-ListServerFiltersTestJSON-241228521 tempest-ListServerFiltersTestJSON-241228521-project-member] [instance: 7fcee323-8ea6-4f69-9f1e-f2105af0fe96] Terminating instance [ 552.773881] env[61806]: DEBUG oslo_concurrency.lockutils [None req-aac795d3-c37f-4811-9be9-aa67ad520ab0 tempest-ListServerFiltersTestJSON-241228521 tempest-ListServerFiltersTestJSON-241228521-project-member] Acquiring lock "refresh_cache-7fcee323-8ea6-4f69-9f1e-f2105af0fe96" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 552.773881] env[61806]: DEBUG oslo_concurrency.lockutils [None req-aac795d3-c37f-4811-9be9-aa67ad520ab0 tempest-ListServerFiltersTestJSON-241228521 tempest-ListServerFiltersTestJSON-241228521-project-member] Acquired lock "refresh_cache-7fcee323-8ea6-4f69-9f1e-f2105af0fe96" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 552.773881] env[61806]: DEBUG nova.network.neutron [None req-aac795d3-c37f-4811-9be9-aa67ad520ab0 tempest-ListServerFiltersTestJSON-241228521 tempest-ListServerFiltersTestJSON-241228521-project-member] [instance: 7fcee323-8ea6-4f69-9f1e-f2105af0fe96] Building network info cache for instance {{(pid=61806) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 552.972712] env[61806]: DEBUG oslo_concurrency.lockutils [None req-0858196d-347c-4201-9cb8-c7f147f5d99d tempest-MigrationsAdminTest-470098049 tempest-MigrationsAdminTest-470098049-project-member] Releasing lock "refresh_cache-52b19347-a08f-44a7-8aad-22a3af5e291d" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 552.973249] env[61806]: DEBUG nova.compute.manager [None req-0858196d-347c-4201-9cb8-c7f147f5d99d tempest-MigrationsAdminTest-470098049 tempest-MigrationsAdminTest-470098049-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61806) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 552.973467] env[61806]: DEBUG nova.compute.manager [None req-0858196d-347c-4201-9cb8-c7f147f5d99d tempest-MigrationsAdminTest-470098049 tempest-MigrationsAdminTest-470098049-project-member] [instance: 52b19347-a08f-44a7-8aad-22a3af5e291d] Deallocating network for instance {{(pid=61806) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 552.973704] env[61806]: DEBUG nova.network.neutron [None req-0858196d-347c-4201-9cb8-c7f147f5d99d tempest-MigrationsAdminTest-470098049 tempest-MigrationsAdminTest-470098049-project-member] [instance: 52b19347-a08f-44a7-8aad-22a3af5e291d] deallocate_for_instance() {{(pid=61806) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 553.011873] env[61806]: DEBUG nova.network.neutron [None req-0858196d-347c-4201-9cb8-c7f147f5d99d tempest-MigrationsAdminTest-470098049 tempest-MigrationsAdminTest-470098049-project-member] [instance: 52b19347-a08f-44a7-8aad-22a3af5e291d] Instance cache missing network info. {{(pid=61806) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 553.076942] env[61806]: DEBUG nova.compute.manager [req-b670eeef-4a14-48c0-bd92-75fa8a1927ae req-5cfc66b1-e883-4135-b2fc-ddc16a8254b6 service nova] [instance: 7fcee323-8ea6-4f69-9f1e-f2105af0fe96] Received event network-changed-eed1949d-29dd-4127-a2a4-d7dc2369df55 {{(pid=61806) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 553.077236] env[61806]: DEBUG nova.compute.manager [req-b670eeef-4a14-48c0-bd92-75fa8a1927ae req-5cfc66b1-e883-4135-b2fc-ddc16a8254b6 service nova] [instance: 7fcee323-8ea6-4f69-9f1e-f2105af0fe96] Refreshing instance network info cache due to event network-changed-eed1949d-29dd-4127-a2a4-d7dc2369df55. {{(pid=61806) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 553.077357] env[61806]: DEBUG oslo_concurrency.lockutils [req-b670eeef-4a14-48c0-bd92-75fa8a1927ae req-5cfc66b1-e883-4135-b2fc-ddc16a8254b6 service nova] Acquiring lock "refresh_cache-7fcee323-8ea6-4f69-9f1e-f2105af0fe96" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 553.216991] env[61806]: DEBUG nova.scheduler.client.report [None req-3d02814b-d8ff-4079-980b-6c580616ce84 tempest-AttachInterfacesUnderV243Test-1683422373 tempest-AttachInterfacesUnderV243Test-1683422373-project-member] Inventory has not changed for provider a2858be1-fd22-4e08-979e-87ad25293407 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 139, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61806) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 553.304886] env[61806]: DEBUG nova.network.neutron [None req-aac795d3-c37f-4811-9be9-aa67ad520ab0 tempest-ListServerFiltersTestJSON-241228521 tempest-ListServerFiltersTestJSON-241228521-project-member] [instance: 7fcee323-8ea6-4f69-9f1e-f2105af0fe96] Instance cache missing network info. {{(pid=61806) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 553.485984] env[61806]: DEBUG nova.network.neutron [None req-aac795d3-c37f-4811-9be9-aa67ad520ab0 tempest-ListServerFiltersTestJSON-241228521 tempest-ListServerFiltersTestJSON-241228521-project-member] [instance: 7fcee323-8ea6-4f69-9f1e-f2105af0fe96] Updating instance_info_cache with network_info: [] {{(pid=61806) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 553.514984] env[61806]: DEBUG nova.network.neutron [None req-0858196d-347c-4201-9cb8-c7f147f5d99d tempest-MigrationsAdminTest-470098049 tempest-MigrationsAdminTest-470098049-project-member] [instance: 52b19347-a08f-44a7-8aad-22a3af5e291d] Updating instance_info_cache with network_info: [] {{(pid=61806) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 553.723201] env[61806]: DEBUG oslo_concurrency.lockutils [None req-3d02814b-d8ff-4079-980b-6c580616ce84 tempest-AttachInterfacesUnderV243Test-1683422373 tempest-AttachInterfacesUnderV243Test-1683422373-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.995s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 553.723474] env[61806]: ERROR nova.compute.manager [None req-3d02814b-d8ff-4079-980b-6c580616ce84 tempest-AttachInterfacesUnderV243Test-1683422373 tempest-AttachInterfacesUnderV243Test-1683422373-project-member] [instance: df2576da-dc49-4ced-9462-fa71710e2e2a] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port cab2110f-89a2-4c90-8780-f2deaabaf59a, please check neutron logs for more information. [ 553.723474] env[61806]: ERROR nova.compute.manager [instance: df2576da-dc49-4ced-9462-fa71710e2e2a] Traceback (most recent call last): [ 553.723474] env[61806]: ERROR nova.compute.manager [instance: df2576da-dc49-4ced-9462-fa71710e2e2a] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 553.723474] env[61806]: ERROR nova.compute.manager [instance: df2576da-dc49-4ced-9462-fa71710e2e2a] self.driver.spawn(context, instance, image_meta, [ 553.723474] env[61806]: ERROR nova.compute.manager [instance: df2576da-dc49-4ced-9462-fa71710e2e2a] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 553.723474] env[61806]: ERROR nova.compute.manager [instance: df2576da-dc49-4ced-9462-fa71710e2e2a] self._vmops.spawn(context, instance, image_meta, injected_files, [ 553.723474] env[61806]: ERROR nova.compute.manager [instance: df2576da-dc49-4ced-9462-fa71710e2e2a] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 553.723474] env[61806]: ERROR nova.compute.manager [instance: df2576da-dc49-4ced-9462-fa71710e2e2a] vm_ref = self.build_virtual_machine(instance, [ 553.723474] env[61806]: ERROR nova.compute.manager [instance: df2576da-dc49-4ced-9462-fa71710e2e2a] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 553.723474] env[61806]: ERROR nova.compute.manager [instance: df2576da-dc49-4ced-9462-fa71710e2e2a] vif_infos = vmwarevif.get_vif_info(self._session, [ 553.723474] env[61806]: ERROR nova.compute.manager [instance: df2576da-dc49-4ced-9462-fa71710e2e2a] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 553.724265] env[61806]: ERROR nova.compute.manager [instance: df2576da-dc49-4ced-9462-fa71710e2e2a] for vif in network_info: [ 553.724265] env[61806]: ERROR nova.compute.manager [instance: df2576da-dc49-4ced-9462-fa71710e2e2a] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 553.724265] env[61806]: ERROR nova.compute.manager [instance: df2576da-dc49-4ced-9462-fa71710e2e2a] return self._sync_wrapper(fn, *args, **kwargs) [ 553.724265] env[61806]: ERROR nova.compute.manager [instance: df2576da-dc49-4ced-9462-fa71710e2e2a] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 553.724265] env[61806]: ERROR nova.compute.manager [instance: df2576da-dc49-4ced-9462-fa71710e2e2a] self.wait() [ 553.724265] env[61806]: ERROR nova.compute.manager [instance: df2576da-dc49-4ced-9462-fa71710e2e2a] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 553.724265] env[61806]: ERROR nova.compute.manager [instance: df2576da-dc49-4ced-9462-fa71710e2e2a] self[:] = self._gt.wait() [ 553.724265] env[61806]: ERROR nova.compute.manager [instance: df2576da-dc49-4ced-9462-fa71710e2e2a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 553.724265] env[61806]: ERROR nova.compute.manager [instance: df2576da-dc49-4ced-9462-fa71710e2e2a] return self._exit_event.wait() [ 553.724265] env[61806]: ERROR nova.compute.manager [instance: df2576da-dc49-4ced-9462-fa71710e2e2a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 553.724265] env[61806]: ERROR nova.compute.manager [instance: df2576da-dc49-4ced-9462-fa71710e2e2a] result = hub.switch() [ 553.724265] env[61806]: ERROR nova.compute.manager [instance: df2576da-dc49-4ced-9462-fa71710e2e2a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 553.724265] env[61806]: ERROR nova.compute.manager [instance: df2576da-dc49-4ced-9462-fa71710e2e2a] return self.greenlet.switch() [ 553.724823] env[61806]: ERROR nova.compute.manager [instance: df2576da-dc49-4ced-9462-fa71710e2e2a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 553.724823] env[61806]: ERROR nova.compute.manager [instance: df2576da-dc49-4ced-9462-fa71710e2e2a] result = function(*args, **kwargs) [ 553.724823] env[61806]: ERROR nova.compute.manager [instance: df2576da-dc49-4ced-9462-fa71710e2e2a] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 553.724823] env[61806]: ERROR nova.compute.manager [instance: df2576da-dc49-4ced-9462-fa71710e2e2a] return func(*args, **kwargs) [ 553.724823] env[61806]: ERROR nova.compute.manager [instance: df2576da-dc49-4ced-9462-fa71710e2e2a] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 553.724823] env[61806]: ERROR nova.compute.manager [instance: df2576da-dc49-4ced-9462-fa71710e2e2a] raise e [ 553.724823] env[61806]: ERROR nova.compute.manager [instance: df2576da-dc49-4ced-9462-fa71710e2e2a] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 553.724823] env[61806]: ERROR nova.compute.manager [instance: df2576da-dc49-4ced-9462-fa71710e2e2a] nwinfo = self.network_api.allocate_for_instance( [ 553.724823] env[61806]: ERROR nova.compute.manager [instance: df2576da-dc49-4ced-9462-fa71710e2e2a] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 553.724823] env[61806]: ERROR nova.compute.manager [instance: df2576da-dc49-4ced-9462-fa71710e2e2a] created_port_ids = self._update_ports_for_instance( [ 553.724823] env[61806]: ERROR nova.compute.manager [instance: df2576da-dc49-4ced-9462-fa71710e2e2a] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 553.724823] env[61806]: ERROR nova.compute.manager [instance: df2576da-dc49-4ced-9462-fa71710e2e2a] with excutils.save_and_reraise_exception(): [ 553.724823] env[61806]: ERROR nova.compute.manager [instance: df2576da-dc49-4ced-9462-fa71710e2e2a] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 553.726437] env[61806]: ERROR nova.compute.manager [instance: df2576da-dc49-4ced-9462-fa71710e2e2a] self.force_reraise() [ 553.726437] env[61806]: ERROR nova.compute.manager [instance: df2576da-dc49-4ced-9462-fa71710e2e2a] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 553.726437] env[61806]: ERROR nova.compute.manager [instance: df2576da-dc49-4ced-9462-fa71710e2e2a] raise self.value [ 553.726437] env[61806]: ERROR nova.compute.manager [instance: df2576da-dc49-4ced-9462-fa71710e2e2a] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 553.726437] env[61806]: ERROR nova.compute.manager [instance: df2576da-dc49-4ced-9462-fa71710e2e2a] updated_port = self._update_port( [ 553.726437] env[61806]: ERROR nova.compute.manager [instance: df2576da-dc49-4ced-9462-fa71710e2e2a] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 553.726437] env[61806]: ERROR nova.compute.manager [instance: df2576da-dc49-4ced-9462-fa71710e2e2a] _ensure_no_port_binding_failure(port) [ 553.726437] env[61806]: ERROR nova.compute.manager [instance: df2576da-dc49-4ced-9462-fa71710e2e2a] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 553.726437] env[61806]: ERROR nova.compute.manager [instance: df2576da-dc49-4ced-9462-fa71710e2e2a] raise exception.PortBindingFailed(port_id=port['id']) [ 553.726437] env[61806]: ERROR nova.compute.manager [instance: df2576da-dc49-4ced-9462-fa71710e2e2a] nova.exception.PortBindingFailed: Binding failed for port cab2110f-89a2-4c90-8780-f2deaabaf59a, please check neutron logs for more information. [ 553.726437] env[61806]: ERROR nova.compute.manager [instance: df2576da-dc49-4ced-9462-fa71710e2e2a] [ 553.727156] env[61806]: DEBUG nova.compute.utils [None req-3d02814b-d8ff-4079-980b-6c580616ce84 tempest-AttachInterfacesUnderV243Test-1683422373 tempest-AttachInterfacesUnderV243Test-1683422373-project-member] [instance: df2576da-dc49-4ced-9462-fa71710e2e2a] Binding failed for port cab2110f-89a2-4c90-8780-f2deaabaf59a, please check neutron logs for more information. {{(pid=61806) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 553.728441] env[61806]: DEBUG nova.compute.manager [None req-3d02814b-d8ff-4079-980b-6c580616ce84 tempest-AttachInterfacesUnderV243Test-1683422373 tempest-AttachInterfacesUnderV243Test-1683422373-project-member] [instance: df2576da-dc49-4ced-9462-fa71710e2e2a] Build of instance df2576da-dc49-4ced-9462-fa71710e2e2a was re-scheduled: Binding failed for port cab2110f-89a2-4c90-8780-f2deaabaf59a, please check neutron logs for more information. {{(pid=61806) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 553.728882] env[61806]: DEBUG nova.compute.manager [None req-3d02814b-d8ff-4079-980b-6c580616ce84 tempest-AttachInterfacesUnderV243Test-1683422373 tempest-AttachInterfacesUnderV243Test-1683422373-project-member] [instance: df2576da-dc49-4ced-9462-fa71710e2e2a] Unplugging VIFs for instance {{(pid=61806) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 553.729128] env[61806]: DEBUG oslo_concurrency.lockutils [None req-3d02814b-d8ff-4079-980b-6c580616ce84 tempest-AttachInterfacesUnderV243Test-1683422373 tempest-AttachInterfacesUnderV243Test-1683422373-project-member] Acquiring lock "refresh_cache-df2576da-dc49-4ced-9462-fa71710e2e2a" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 553.729308] env[61806]: DEBUG oslo_concurrency.lockutils [None req-3d02814b-d8ff-4079-980b-6c580616ce84 tempest-AttachInterfacesUnderV243Test-1683422373 tempest-AttachInterfacesUnderV243Test-1683422373-project-member] Acquired lock "refresh_cache-df2576da-dc49-4ced-9462-fa71710e2e2a" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 553.729481] env[61806]: DEBUG nova.network.neutron [None req-3d02814b-d8ff-4079-980b-6c580616ce84 tempest-AttachInterfacesUnderV243Test-1683422373 tempest-AttachInterfacesUnderV243Test-1683422373-project-member] [instance: df2576da-dc49-4ced-9462-fa71710e2e2a] Building network info cache for instance {{(pid=61806) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 553.730566] env[61806]: DEBUG oslo_concurrency.lockutils [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 15.350s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 553.887800] env[61806]: ERROR nova.compute.manager [None req-55a61261-0862-4d73-b971-6d11cbe88c5e tempest-ServersAdminTestJSON-14199920 tempest-ServersAdminTestJSON-14199920-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 31cd9389-2251-4222-885e-268c99301a23, please check neutron logs for more information. [ 553.887800] env[61806]: ERROR nova.compute.manager Traceback (most recent call last): [ 553.887800] env[61806]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 553.887800] env[61806]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 553.887800] env[61806]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 553.887800] env[61806]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 553.887800] env[61806]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 553.887800] env[61806]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 553.887800] env[61806]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 553.887800] env[61806]: ERROR nova.compute.manager self.force_reraise() [ 553.887800] env[61806]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 553.887800] env[61806]: ERROR nova.compute.manager raise self.value [ 553.887800] env[61806]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 553.887800] env[61806]: ERROR nova.compute.manager updated_port = self._update_port( [ 553.887800] env[61806]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 553.887800] env[61806]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 553.888389] env[61806]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 553.888389] env[61806]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 553.888389] env[61806]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 31cd9389-2251-4222-885e-268c99301a23, please check neutron logs for more information. [ 553.888389] env[61806]: ERROR nova.compute.manager [ 553.888389] env[61806]: Traceback (most recent call last): [ 553.888389] env[61806]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 553.888389] env[61806]: listener.cb(fileno) [ 553.888389] env[61806]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 553.888389] env[61806]: result = function(*args, **kwargs) [ 553.888389] env[61806]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 553.888389] env[61806]: return func(*args, **kwargs) [ 553.888389] env[61806]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 553.888389] env[61806]: raise e [ 553.888389] env[61806]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 553.888389] env[61806]: nwinfo = self.network_api.allocate_for_instance( [ 553.888389] env[61806]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 553.888389] env[61806]: created_port_ids = self._update_ports_for_instance( [ 553.888389] env[61806]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 553.888389] env[61806]: with excutils.save_and_reraise_exception(): [ 553.888389] env[61806]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 553.888389] env[61806]: self.force_reraise() [ 553.888389] env[61806]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 553.888389] env[61806]: raise self.value [ 553.888389] env[61806]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 553.888389] env[61806]: updated_port = self._update_port( [ 553.888389] env[61806]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 553.888389] env[61806]: _ensure_no_port_binding_failure(port) [ 553.888389] env[61806]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 553.888389] env[61806]: raise exception.PortBindingFailed(port_id=port['id']) [ 553.889332] env[61806]: nova.exception.PortBindingFailed: Binding failed for port 31cd9389-2251-4222-885e-268c99301a23, please check neutron logs for more information. [ 553.889332] env[61806]: Removing descriptor: 17 [ 553.889332] env[61806]: ERROR nova.compute.manager [None req-55a61261-0862-4d73-b971-6d11cbe88c5e tempest-ServersAdminTestJSON-14199920 tempest-ServersAdminTestJSON-14199920-project-member] [instance: 26a0baf9-68bd-4380-87b3-53f50524bbbf] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 31cd9389-2251-4222-885e-268c99301a23, please check neutron logs for more information. [ 553.889332] env[61806]: ERROR nova.compute.manager [instance: 26a0baf9-68bd-4380-87b3-53f50524bbbf] Traceback (most recent call last): [ 553.889332] env[61806]: ERROR nova.compute.manager [instance: 26a0baf9-68bd-4380-87b3-53f50524bbbf] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 553.889332] env[61806]: ERROR nova.compute.manager [instance: 26a0baf9-68bd-4380-87b3-53f50524bbbf] yield resources [ 553.889332] env[61806]: ERROR nova.compute.manager [instance: 26a0baf9-68bd-4380-87b3-53f50524bbbf] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 553.889332] env[61806]: ERROR nova.compute.manager [instance: 26a0baf9-68bd-4380-87b3-53f50524bbbf] self.driver.spawn(context, instance, image_meta, [ 553.889332] env[61806]: ERROR nova.compute.manager [instance: 26a0baf9-68bd-4380-87b3-53f50524bbbf] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 553.889332] env[61806]: ERROR nova.compute.manager [instance: 26a0baf9-68bd-4380-87b3-53f50524bbbf] self._vmops.spawn(context, instance, image_meta, injected_files, [ 553.889332] env[61806]: ERROR nova.compute.manager [instance: 26a0baf9-68bd-4380-87b3-53f50524bbbf] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 553.889332] env[61806]: ERROR nova.compute.manager [instance: 26a0baf9-68bd-4380-87b3-53f50524bbbf] vm_ref = self.build_virtual_machine(instance, [ 553.889712] env[61806]: ERROR nova.compute.manager [instance: 26a0baf9-68bd-4380-87b3-53f50524bbbf] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 553.889712] env[61806]: ERROR nova.compute.manager [instance: 26a0baf9-68bd-4380-87b3-53f50524bbbf] vif_infos = vmwarevif.get_vif_info(self._session, [ 553.889712] env[61806]: ERROR nova.compute.manager [instance: 26a0baf9-68bd-4380-87b3-53f50524bbbf] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 553.889712] env[61806]: ERROR nova.compute.manager [instance: 26a0baf9-68bd-4380-87b3-53f50524bbbf] for vif in network_info: [ 553.889712] env[61806]: ERROR nova.compute.manager [instance: 26a0baf9-68bd-4380-87b3-53f50524bbbf] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 553.889712] env[61806]: ERROR nova.compute.manager [instance: 26a0baf9-68bd-4380-87b3-53f50524bbbf] return self._sync_wrapper(fn, *args, **kwargs) [ 553.889712] env[61806]: ERROR nova.compute.manager [instance: 26a0baf9-68bd-4380-87b3-53f50524bbbf] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 553.889712] env[61806]: ERROR nova.compute.manager [instance: 26a0baf9-68bd-4380-87b3-53f50524bbbf] self.wait() [ 553.889712] env[61806]: ERROR nova.compute.manager [instance: 26a0baf9-68bd-4380-87b3-53f50524bbbf] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 553.889712] env[61806]: ERROR nova.compute.manager [instance: 26a0baf9-68bd-4380-87b3-53f50524bbbf] self[:] = self._gt.wait() [ 553.889712] env[61806]: ERROR nova.compute.manager [instance: 26a0baf9-68bd-4380-87b3-53f50524bbbf] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 553.889712] env[61806]: ERROR nova.compute.manager [instance: 26a0baf9-68bd-4380-87b3-53f50524bbbf] return self._exit_event.wait() [ 553.889712] env[61806]: ERROR nova.compute.manager [instance: 26a0baf9-68bd-4380-87b3-53f50524bbbf] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 553.890142] env[61806]: ERROR nova.compute.manager [instance: 26a0baf9-68bd-4380-87b3-53f50524bbbf] result = hub.switch() [ 553.890142] env[61806]: ERROR nova.compute.manager [instance: 26a0baf9-68bd-4380-87b3-53f50524bbbf] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 553.890142] env[61806]: ERROR nova.compute.manager [instance: 26a0baf9-68bd-4380-87b3-53f50524bbbf] return self.greenlet.switch() [ 553.890142] env[61806]: ERROR nova.compute.manager [instance: 26a0baf9-68bd-4380-87b3-53f50524bbbf] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 553.890142] env[61806]: ERROR nova.compute.manager [instance: 26a0baf9-68bd-4380-87b3-53f50524bbbf] result = function(*args, **kwargs) [ 553.890142] env[61806]: ERROR nova.compute.manager [instance: 26a0baf9-68bd-4380-87b3-53f50524bbbf] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 553.890142] env[61806]: ERROR nova.compute.manager [instance: 26a0baf9-68bd-4380-87b3-53f50524bbbf] return func(*args, **kwargs) [ 553.890142] env[61806]: ERROR nova.compute.manager [instance: 26a0baf9-68bd-4380-87b3-53f50524bbbf] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 553.890142] env[61806]: ERROR nova.compute.manager [instance: 26a0baf9-68bd-4380-87b3-53f50524bbbf] raise e [ 553.890142] env[61806]: ERROR nova.compute.manager [instance: 26a0baf9-68bd-4380-87b3-53f50524bbbf] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 553.890142] env[61806]: ERROR nova.compute.manager [instance: 26a0baf9-68bd-4380-87b3-53f50524bbbf] nwinfo = self.network_api.allocate_for_instance( [ 553.890142] env[61806]: ERROR nova.compute.manager [instance: 26a0baf9-68bd-4380-87b3-53f50524bbbf] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 553.890142] env[61806]: ERROR nova.compute.manager [instance: 26a0baf9-68bd-4380-87b3-53f50524bbbf] created_port_ids = self._update_ports_for_instance( [ 553.890563] env[61806]: ERROR nova.compute.manager [instance: 26a0baf9-68bd-4380-87b3-53f50524bbbf] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 553.890563] env[61806]: ERROR nova.compute.manager [instance: 26a0baf9-68bd-4380-87b3-53f50524bbbf] with excutils.save_and_reraise_exception(): [ 553.890563] env[61806]: ERROR nova.compute.manager [instance: 26a0baf9-68bd-4380-87b3-53f50524bbbf] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 553.890563] env[61806]: ERROR nova.compute.manager [instance: 26a0baf9-68bd-4380-87b3-53f50524bbbf] self.force_reraise() [ 553.890563] env[61806]: ERROR nova.compute.manager [instance: 26a0baf9-68bd-4380-87b3-53f50524bbbf] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 553.890563] env[61806]: ERROR nova.compute.manager [instance: 26a0baf9-68bd-4380-87b3-53f50524bbbf] raise self.value [ 553.890563] env[61806]: ERROR nova.compute.manager [instance: 26a0baf9-68bd-4380-87b3-53f50524bbbf] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 553.890563] env[61806]: ERROR nova.compute.manager [instance: 26a0baf9-68bd-4380-87b3-53f50524bbbf] updated_port = self._update_port( [ 553.890563] env[61806]: ERROR nova.compute.manager [instance: 26a0baf9-68bd-4380-87b3-53f50524bbbf] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 553.890563] env[61806]: ERROR nova.compute.manager [instance: 26a0baf9-68bd-4380-87b3-53f50524bbbf] _ensure_no_port_binding_failure(port) [ 553.890563] env[61806]: ERROR nova.compute.manager [instance: 26a0baf9-68bd-4380-87b3-53f50524bbbf] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 553.890563] env[61806]: ERROR nova.compute.manager [instance: 26a0baf9-68bd-4380-87b3-53f50524bbbf] raise exception.PortBindingFailed(port_id=port['id']) [ 553.890961] env[61806]: ERROR nova.compute.manager [instance: 26a0baf9-68bd-4380-87b3-53f50524bbbf] nova.exception.PortBindingFailed: Binding failed for port 31cd9389-2251-4222-885e-268c99301a23, please check neutron logs for more information. [ 553.890961] env[61806]: ERROR nova.compute.manager [instance: 26a0baf9-68bd-4380-87b3-53f50524bbbf] [ 553.890961] env[61806]: INFO nova.compute.manager [None req-55a61261-0862-4d73-b971-6d11cbe88c5e tempest-ServersAdminTestJSON-14199920 tempest-ServersAdminTestJSON-14199920-project-member] [instance: 26a0baf9-68bd-4380-87b3-53f50524bbbf] Terminating instance [ 553.892937] env[61806]: DEBUG oslo_concurrency.lockutils [None req-55a61261-0862-4d73-b971-6d11cbe88c5e tempest-ServersAdminTestJSON-14199920 tempest-ServersAdminTestJSON-14199920-project-member] Acquiring lock "refresh_cache-26a0baf9-68bd-4380-87b3-53f50524bbbf" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 553.892937] env[61806]: DEBUG oslo_concurrency.lockutils [None req-55a61261-0862-4d73-b971-6d11cbe88c5e tempest-ServersAdminTestJSON-14199920 tempest-ServersAdminTestJSON-14199920-project-member] Acquired lock "refresh_cache-26a0baf9-68bd-4380-87b3-53f50524bbbf" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 553.892937] env[61806]: DEBUG nova.network.neutron [None req-55a61261-0862-4d73-b971-6d11cbe88c5e tempest-ServersAdminTestJSON-14199920 tempest-ServersAdminTestJSON-14199920-project-member] [instance: 26a0baf9-68bd-4380-87b3-53f50524bbbf] Building network info cache for instance {{(pid=61806) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 553.957808] env[61806]: DEBUG oslo_concurrency.lockutils [None req-a6b8da3a-7829-486a-884d-0f28be02e38c tempest-ServersTestManualDisk-151630353 tempest-ServersTestManualDisk-151630353-project-member] Acquiring lock "38a79453-939d-4649-9419-a2799636180d" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 553.958032] env[61806]: DEBUG oslo_concurrency.lockutils [None req-a6b8da3a-7829-486a-884d-0f28be02e38c tempest-ServersTestManualDisk-151630353 tempest-ServersTestManualDisk-151630353-project-member] Lock "38a79453-939d-4649-9419-a2799636180d" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 553.989383] env[61806]: DEBUG oslo_concurrency.lockutils [None req-aac795d3-c37f-4811-9be9-aa67ad520ab0 tempest-ListServerFiltersTestJSON-241228521 tempest-ListServerFiltersTestJSON-241228521-project-member] Releasing lock "refresh_cache-7fcee323-8ea6-4f69-9f1e-f2105af0fe96" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 553.989826] env[61806]: DEBUG nova.compute.manager [None req-aac795d3-c37f-4811-9be9-aa67ad520ab0 tempest-ListServerFiltersTestJSON-241228521 tempest-ListServerFiltersTestJSON-241228521-project-member] [instance: 7fcee323-8ea6-4f69-9f1e-f2105af0fe96] Start destroying the instance on the hypervisor. {{(pid=61806) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 553.990778] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-aac795d3-c37f-4811-9be9-aa67ad520ab0 tempest-ListServerFiltersTestJSON-241228521 tempest-ListServerFiltersTestJSON-241228521-project-member] [instance: 7fcee323-8ea6-4f69-9f1e-f2105af0fe96] Destroying instance {{(pid=61806) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 553.990778] env[61806]: DEBUG oslo_concurrency.lockutils [req-b670eeef-4a14-48c0-bd92-75fa8a1927ae req-5cfc66b1-e883-4135-b2fc-ddc16a8254b6 service nova] Acquired lock "refresh_cache-7fcee323-8ea6-4f69-9f1e-f2105af0fe96" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 553.990910] env[61806]: DEBUG nova.network.neutron [req-b670eeef-4a14-48c0-bd92-75fa8a1927ae req-5cfc66b1-e883-4135-b2fc-ddc16a8254b6 service nova] [instance: 7fcee323-8ea6-4f69-9f1e-f2105af0fe96] Refreshing network info cache for port eed1949d-29dd-4127-a2a4-d7dc2369df55 {{(pid=61806) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 553.994586] env[61806]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-249ed225-32df-4a71-ac33-c8d5ec4eab33 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 554.009405] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-80b740e3-ddf6-4911-b3c3-2df2f883df4a {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 554.019476] env[61806]: INFO nova.compute.manager [None req-0858196d-347c-4201-9cb8-c7f147f5d99d tempest-MigrationsAdminTest-470098049 tempest-MigrationsAdminTest-470098049-project-member] [instance: 52b19347-a08f-44a7-8aad-22a3af5e291d] Took 1.05 seconds to deallocate network for instance. [ 554.036707] env[61806]: WARNING nova.virt.vmwareapi.vmops [None req-aac795d3-c37f-4811-9be9-aa67ad520ab0 tempest-ListServerFiltersTestJSON-241228521 tempest-ListServerFiltersTestJSON-241228521-project-member] [instance: 7fcee323-8ea6-4f69-9f1e-f2105af0fe96] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 7fcee323-8ea6-4f69-9f1e-f2105af0fe96 could not be found. [ 554.036942] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-aac795d3-c37f-4811-9be9-aa67ad520ab0 tempest-ListServerFiltersTestJSON-241228521 tempest-ListServerFiltersTestJSON-241228521-project-member] [instance: 7fcee323-8ea6-4f69-9f1e-f2105af0fe96] Instance destroyed {{(pid=61806) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 554.037135] env[61806]: INFO nova.compute.manager [None req-aac795d3-c37f-4811-9be9-aa67ad520ab0 tempest-ListServerFiltersTestJSON-241228521 tempest-ListServerFiltersTestJSON-241228521-project-member] [instance: 7fcee323-8ea6-4f69-9f1e-f2105af0fe96] Took 0.05 seconds to destroy the instance on the hypervisor. [ 554.037722] env[61806]: DEBUG oslo.service.loopingcall [None req-aac795d3-c37f-4811-9be9-aa67ad520ab0 tempest-ListServerFiltersTestJSON-241228521 tempest-ListServerFiltersTestJSON-241228521-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61806) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 554.038010] env[61806]: DEBUG nova.compute.manager [-] [instance: 7fcee323-8ea6-4f69-9f1e-f2105af0fe96] Deallocating network for instance {{(pid=61806) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 554.038284] env[61806]: DEBUG nova.network.neutron [-] [instance: 7fcee323-8ea6-4f69-9f1e-f2105af0fe96] deallocate_for_instance() {{(pid=61806) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 554.067143] env[61806]: DEBUG nova.network.neutron [-] [instance: 7fcee323-8ea6-4f69-9f1e-f2105af0fe96] Instance cache missing network info. {{(pid=61806) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 554.265367] env[61806]: DEBUG nova.network.neutron [None req-3d02814b-d8ff-4079-980b-6c580616ce84 tempest-AttachInterfacesUnderV243Test-1683422373 tempest-AttachInterfacesUnderV243Test-1683422373-project-member] [instance: df2576da-dc49-4ced-9462-fa71710e2e2a] Instance cache missing network info. {{(pid=61806) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 554.358831] env[61806]: DEBUG oslo_concurrency.lockutils [None req-98d962a0-2632-4a9a-8d12-70ce4acc382b tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] Acquiring lock "b8249dc5-5561-454a-aa01-477f9f1fc0e1" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 554.359116] env[61806]: DEBUG oslo_concurrency.lockutils [None req-98d962a0-2632-4a9a-8d12-70ce4acc382b tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] Lock "b8249dc5-5561-454a-aa01-477f9f1fc0e1" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 554.396209] env[61806]: DEBUG nova.network.neutron [None req-3d02814b-d8ff-4079-980b-6c580616ce84 tempest-AttachInterfacesUnderV243Test-1683422373 tempest-AttachInterfacesUnderV243Test-1683422373-project-member] [instance: df2576da-dc49-4ced-9462-fa71710e2e2a] Updating instance_info_cache with network_info: [] {{(pid=61806) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 554.411549] env[61806]: DEBUG nova.network.neutron [None req-55a61261-0862-4d73-b971-6d11cbe88c5e tempest-ServersAdminTestJSON-14199920 tempest-ServersAdminTestJSON-14199920-project-member] [instance: 26a0baf9-68bd-4380-87b3-53f50524bbbf] Instance cache missing network info. {{(pid=61806) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 554.427434] env[61806]: DEBUG nova.compute.manager [req-0dee6e58-ecae-4166-be6f-633d7bc307e3 req-c810f074-cc7f-4f62-a40c-3c11e6000901 service nova] [instance: 26a0baf9-68bd-4380-87b3-53f50524bbbf] Received event network-changed-31cd9389-2251-4222-885e-268c99301a23 {{(pid=61806) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 554.427434] env[61806]: DEBUG nova.compute.manager [req-0dee6e58-ecae-4166-be6f-633d7bc307e3 req-c810f074-cc7f-4f62-a40c-3c11e6000901 service nova] [instance: 26a0baf9-68bd-4380-87b3-53f50524bbbf] Refreshing instance network info cache due to event network-changed-31cd9389-2251-4222-885e-268c99301a23. {{(pid=61806) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 554.427434] env[61806]: DEBUG oslo_concurrency.lockutils [req-0dee6e58-ecae-4166-be6f-633d7bc307e3 req-c810f074-cc7f-4f62-a40c-3c11e6000901 service nova] Acquiring lock "refresh_cache-26a0baf9-68bd-4380-87b3-53f50524bbbf" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 554.501513] env[61806]: DEBUG nova.network.neutron [None req-55a61261-0862-4d73-b971-6d11cbe88c5e tempest-ServersAdminTestJSON-14199920 tempest-ServersAdminTestJSON-14199920-project-member] [instance: 26a0baf9-68bd-4380-87b3-53f50524bbbf] Updating instance_info_cache with network_info: [] {{(pid=61806) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 554.519567] env[61806]: DEBUG nova.network.neutron [req-b670eeef-4a14-48c0-bd92-75fa8a1927ae req-5cfc66b1-e883-4135-b2fc-ddc16a8254b6 service nova] [instance: 7fcee323-8ea6-4f69-9f1e-f2105af0fe96] Instance cache missing network info. {{(pid=61806) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 554.569639] env[61806]: DEBUG nova.network.neutron [-] [instance: 7fcee323-8ea6-4f69-9f1e-f2105af0fe96] Updating instance_info_cache with network_info: [] {{(pid=61806) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 554.677391] env[61806]: DEBUG nova.network.neutron [req-b670eeef-4a14-48c0-bd92-75fa8a1927ae req-5cfc66b1-e883-4135-b2fc-ddc16a8254b6 service nova] [instance: 7fcee323-8ea6-4f69-9f1e-f2105af0fe96] Updating instance_info_cache with network_info: [] {{(pid=61806) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 554.901025] env[61806]: DEBUG oslo_concurrency.lockutils [None req-3d02814b-d8ff-4079-980b-6c580616ce84 tempest-AttachInterfacesUnderV243Test-1683422373 tempest-AttachInterfacesUnderV243Test-1683422373-project-member] Releasing lock "refresh_cache-df2576da-dc49-4ced-9462-fa71710e2e2a" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 554.901025] env[61806]: DEBUG nova.compute.manager [None req-3d02814b-d8ff-4079-980b-6c580616ce84 tempest-AttachInterfacesUnderV243Test-1683422373 tempest-AttachInterfacesUnderV243Test-1683422373-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61806) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 554.901025] env[61806]: DEBUG nova.compute.manager [None req-3d02814b-d8ff-4079-980b-6c580616ce84 tempest-AttachInterfacesUnderV243Test-1683422373 tempest-AttachInterfacesUnderV243Test-1683422373-project-member] [instance: df2576da-dc49-4ced-9462-fa71710e2e2a] Deallocating network for instance {{(pid=61806) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 554.901025] env[61806]: DEBUG nova.network.neutron [None req-3d02814b-d8ff-4079-980b-6c580616ce84 tempest-AttachInterfacesUnderV243Test-1683422373 tempest-AttachInterfacesUnderV243Test-1683422373-project-member] [instance: df2576da-dc49-4ced-9462-fa71710e2e2a] deallocate_for_instance() {{(pid=61806) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 554.929419] env[61806]: DEBUG nova.network.neutron [None req-3d02814b-d8ff-4079-980b-6c580616ce84 tempest-AttachInterfacesUnderV243Test-1683422373 tempest-AttachInterfacesUnderV243Test-1683422373-project-member] [instance: df2576da-dc49-4ced-9462-fa71710e2e2a] Instance cache missing network info. {{(pid=61806) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 555.004759] env[61806]: DEBUG oslo_concurrency.lockutils [None req-55a61261-0862-4d73-b971-6d11cbe88c5e tempest-ServersAdminTestJSON-14199920 tempest-ServersAdminTestJSON-14199920-project-member] Releasing lock "refresh_cache-26a0baf9-68bd-4380-87b3-53f50524bbbf" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 555.005207] env[61806]: DEBUG nova.compute.manager [None req-55a61261-0862-4d73-b971-6d11cbe88c5e tempest-ServersAdminTestJSON-14199920 tempest-ServersAdminTestJSON-14199920-project-member] [instance: 26a0baf9-68bd-4380-87b3-53f50524bbbf] Start destroying the instance on the hypervisor. {{(pid=61806) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 555.005408] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-55a61261-0862-4d73-b971-6d11cbe88c5e tempest-ServersAdminTestJSON-14199920 tempest-ServersAdminTestJSON-14199920-project-member] [instance: 26a0baf9-68bd-4380-87b3-53f50524bbbf] Destroying instance {{(pid=61806) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 555.005860] env[61806]: DEBUG oslo_concurrency.lockutils [req-0dee6e58-ecae-4166-be6f-633d7bc307e3 req-c810f074-cc7f-4f62-a40c-3c11e6000901 service nova] Acquired lock "refresh_cache-26a0baf9-68bd-4380-87b3-53f50524bbbf" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 555.005983] env[61806]: DEBUG nova.network.neutron [req-0dee6e58-ecae-4166-be6f-633d7bc307e3 req-c810f074-cc7f-4f62-a40c-3c11e6000901 service nova] [instance: 26a0baf9-68bd-4380-87b3-53f50524bbbf] Refreshing network info cache for port 31cd9389-2251-4222-885e-268c99301a23 {{(pid=61806) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 555.009012] env[61806]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-0f2e72e8-0d9d-4772-ae4e-f776be62e67e {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 555.020662] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-906e9f36-9a04-4bd5-9252-35b35c33eec9 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 555.056837] env[61806]: WARNING nova.virt.vmwareapi.vmops [None req-55a61261-0862-4d73-b971-6d11cbe88c5e tempest-ServersAdminTestJSON-14199920 tempest-ServersAdminTestJSON-14199920-project-member] [instance: 26a0baf9-68bd-4380-87b3-53f50524bbbf] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 26a0baf9-68bd-4380-87b3-53f50524bbbf could not be found. [ 555.057096] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-55a61261-0862-4d73-b971-6d11cbe88c5e tempest-ServersAdminTestJSON-14199920 tempest-ServersAdminTestJSON-14199920-project-member] [instance: 26a0baf9-68bd-4380-87b3-53f50524bbbf] Instance destroyed {{(pid=61806) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 555.057296] env[61806]: INFO nova.compute.manager [None req-55a61261-0862-4d73-b971-6d11cbe88c5e tempest-ServersAdminTestJSON-14199920 tempest-ServersAdminTestJSON-14199920-project-member] [instance: 26a0baf9-68bd-4380-87b3-53f50524bbbf] Took 0.05 seconds to destroy the instance on the hypervisor. [ 555.057543] env[61806]: DEBUG oslo.service.loopingcall [None req-55a61261-0862-4d73-b971-6d11cbe88c5e tempest-ServersAdminTestJSON-14199920 tempest-ServersAdminTestJSON-14199920-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61806) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 555.057760] env[61806]: DEBUG nova.compute.manager [-] [instance: 26a0baf9-68bd-4380-87b3-53f50524bbbf] Deallocating network for instance {{(pid=61806) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 555.057848] env[61806]: DEBUG nova.network.neutron [-] [instance: 26a0baf9-68bd-4380-87b3-53f50524bbbf] deallocate_for_instance() {{(pid=61806) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 555.064159] env[61806]: INFO nova.scheduler.client.report [None req-0858196d-347c-4201-9cb8-c7f147f5d99d tempest-MigrationsAdminTest-470098049 tempest-MigrationsAdminTest-470098049-project-member] Deleted allocations for instance 52b19347-a08f-44a7-8aad-22a3af5e291d [ 555.075285] env[61806]: INFO nova.compute.manager [-] [instance: 7fcee323-8ea6-4f69-9f1e-f2105af0fe96] Took 1.04 seconds to deallocate network for instance. [ 555.078752] env[61806]: DEBUG nova.compute.claims [None req-aac795d3-c37f-4811-9be9-aa67ad520ab0 tempest-ListServerFiltersTestJSON-241228521 tempest-ListServerFiltersTestJSON-241228521-project-member] [instance: 7fcee323-8ea6-4f69-9f1e-f2105af0fe96] Aborting claim: {{(pid=61806) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 555.078752] env[61806]: DEBUG oslo_concurrency.lockutils [None req-aac795d3-c37f-4811-9be9-aa67ad520ab0 tempest-ListServerFiltersTestJSON-241228521 tempest-ListServerFiltersTestJSON-241228521-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 555.098399] env[61806]: DEBUG nova.network.neutron [-] [instance: 26a0baf9-68bd-4380-87b3-53f50524bbbf] Instance cache missing network info. {{(pid=61806) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 555.183967] env[61806]: DEBUG oslo_concurrency.lockutils [req-b670eeef-4a14-48c0-bd92-75fa8a1927ae req-5cfc66b1-e883-4135-b2fc-ddc16a8254b6 service nova] Releasing lock "refresh_cache-7fcee323-8ea6-4f69-9f1e-f2105af0fe96" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 555.280395] env[61806]: DEBUG nova.compute.resource_tracker [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Instance 52b19347-a08f-44a7-8aad-22a3af5e291d has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61806) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 555.330838] env[61806]: DEBUG oslo_concurrency.lockutils [None req-3e50272a-5153-4e11-beee-d774a917f2df tempest-ServerPasswordTestJSON-1741413915 tempest-ServerPasswordTestJSON-1741413915-project-member] Acquiring lock "0d757385-ff1c-49c9-825f-f9d542ae074f" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 555.332298] env[61806]: DEBUG oslo_concurrency.lockutils [None req-3e50272a-5153-4e11-beee-d774a917f2df tempest-ServerPasswordTestJSON-1741413915 tempest-ServerPasswordTestJSON-1741413915-project-member] Lock "0d757385-ff1c-49c9-825f-f9d542ae074f" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 555.433261] env[61806]: DEBUG nova.network.neutron [None req-3d02814b-d8ff-4079-980b-6c580616ce84 tempest-AttachInterfacesUnderV243Test-1683422373 tempest-AttachInterfacesUnderV243Test-1683422373-project-member] [instance: df2576da-dc49-4ced-9462-fa71710e2e2a] Updating instance_info_cache with network_info: [] {{(pid=61806) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 555.545256] env[61806]: DEBUG nova.network.neutron [req-0dee6e58-ecae-4166-be6f-633d7bc307e3 req-c810f074-cc7f-4f62-a40c-3c11e6000901 service nova] [instance: 26a0baf9-68bd-4380-87b3-53f50524bbbf] Instance cache missing network info. {{(pid=61806) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 555.576992] env[61806]: DEBUG oslo_concurrency.lockutils [None req-0858196d-347c-4201-9cb8-c7f147f5d99d tempest-MigrationsAdminTest-470098049 tempest-MigrationsAdminTest-470098049-project-member] Lock "52b19347-a08f-44a7-8aad-22a3af5e291d" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 31.846s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 555.600073] env[61806]: DEBUG nova.network.neutron [-] [instance: 26a0baf9-68bd-4380-87b3-53f50524bbbf] Updating instance_info_cache with network_info: [] {{(pid=61806) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 555.746914] env[61806]: DEBUG nova.network.neutron [req-0dee6e58-ecae-4166-be6f-633d7bc307e3 req-c810f074-cc7f-4f62-a40c-3c11e6000901 service nova] [instance: 26a0baf9-68bd-4380-87b3-53f50524bbbf] Updating instance_info_cache with network_info: [] {{(pid=61806) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 555.784425] env[61806]: DEBUG nova.compute.resource_tracker [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Instance df2576da-dc49-4ced-9462-fa71710e2e2a has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61806) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 555.784425] env[61806]: DEBUG nova.compute.resource_tracker [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Instance a1aca638-e9c4-4e3b-9855-f92e7bd67204 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61806) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 555.784425] env[61806]: DEBUG nova.compute.resource_tracker [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Instance 30e44576-dcd3-4036-9070-9032b775c90a actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61806) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 555.784425] env[61806]: DEBUG nova.compute.resource_tracker [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Instance 9ddec22f-0fce-4d28-b97e-92b179f829f6 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61806) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 555.784621] env[61806]: DEBUG nova.compute.resource_tracker [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Instance 7fcee323-8ea6-4f69-9f1e-f2105af0fe96 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 256, 'VCPU': 1}}. {{(pid=61806) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 555.784621] env[61806]: DEBUG nova.compute.resource_tracker [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Instance 26a0baf9-68bd-4380-87b3-53f50524bbbf actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61806) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 555.941841] env[61806]: INFO nova.compute.manager [None req-3d02814b-d8ff-4079-980b-6c580616ce84 tempest-AttachInterfacesUnderV243Test-1683422373 tempest-AttachInterfacesUnderV243Test-1683422373-project-member] [instance: df2576da-dc49-4ced-9462-fa71710e2e2a] Took 1.04 seconds to deallocate network for instance. [ 555.993128] env[61806]: DEBUG nova.compute.manager [req-ccdec538-096b-4081-99c7-d731f0db097d req-2c0b48da-cfa2-41b6-99dd-a1d8a8bc2550 service nova] [instance: 7fcee323-8ea6-4f69-9f1e-f2105af0fe96] Received event network-vif-deleted-eed1949d-29dd-4127-a2a4-d7dc2369df55 {{(pid=61806) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 556.079827] env[61806]: DEBUG nova.compute.manager [None req-5c11e673-bd02-4846-9be6-b5679ed1117d tempest-VolumesAssistedSnapshotsTest-2037321669 tempest-VolumesAssistedSnapshotsTest-2037321669-project-member] [instance: 76722306-a9ad-4ef9-b579-d958c36cf6bc] Starting instance... {{(pid=61806) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 556.107060] env[61806]: INFO nova.compute.manager [-] [instance: 26a0baf9-68bd-4380-87b3-53f50524bbbf] Took 1.05 seconds to deallocate network for instance. [ 556.108757] env[61806]: DEBUG nova.compute.claims [None req-55a61261-0862-4d73-b971-6d11cbe88c5e tempest-ServersAdminTestJSON-14199920 tempest-ServersAdminTestJSON-14199920-project-member] [instance: 26a0baf9-68bd-4380-87b3-53f50524bbbf] Aborting claim: {{(pid=61806) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 556.108757] env[61806]: DEBUG oslo_concurrency.lockutils [None req-55a61261-0862-4d73-b971-6d11cbe88c5e tempest-ServersAdminTestJSON-14199920 tempest-ServersAdminTestJSON-14199920-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 556.249528] env[61806]: DEBUG oslo_concurrency.lockutils [req-0dee6e58-ecae-4166-be6f-633d7bc307e3 req-c810f074-cc7f-4f62-a40c-3c11e6000901 service nova] Releasing lock "refresh_cache-26a0baf9-68bd-4380-87b3-53f50524bbbf" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 556.288559] env[61806]: DEBUG nova.compute.resource_tracker [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Instance 123f15be-3b56-4999-8010-40d2e86a6813 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61806) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 556.576611] env[61806]: DEBUG oslo_concurrency.lockutils [None req-f8b2f38e-90f5-4b03-89bc-ca1926adb40c tempest-FloatingIPsAssociationTestJSON-35718468 tempest-FloatingIPsAssociationTestJSON-35718468-project-member] Acquiring lock "88cb0961-ad34-4ffc-8fb2-6034f08ee740" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 556.576872] env[61806]: DEBUG oslo_concurrency.lockutils [None req-f8b2f38e-90f5-4b03-89bc-ca1926adb40c tempest-FloatingIPsAssociationTestJSON-35718468 tempest-FloatingIPsAssociationTestJSON-35718468-project-member] Lock "88cb0961-ad34-4ffc-8fb2-6034f08ee740" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 556.614896] env[61806]: DEBUG oslo_concurrency.lockutils [None req-5c11e673-bd02-4846-9be6-b5679ed1117d tempest-VolumesAssistedSnapshotsTest-2037321669 tempest-VolumesAssistedSnapshotsTest-2037321669-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 556.784492] env[61806]: DEBUG nova.compute.manager [req-69bd2c0d-d779-4b9a-8ace-561c71d17c5f req-36ea377d-c542-4c54-b3e1-be0ada0121ec service nova] [instance: 26a0baf9-68bd-4380-87b3-53f50524bbbf] Received event network-vif-deleted-31cd9389-2251-4222-885e-268c99301a23 {{(pid=61806) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 556.793434] env[61806]: DEBUG nova.compute.resource_tracker [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Instance 8905f20f-28a6-49f4-88bf-e52177c6a4b6 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61806) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 556.996403] env[61806]: INFO nova.scheduler.client.report [None req-3d02814b-d8ff-4079-980b-6c580616ce84 tempest-AttachInterfacesUnderV243Test-1683422373 tempest-AttachInterfacesUnderV243Test-1683422373-project-member] Deleted allocations for instance df2576da-dc49-4ced-9462-fa71710e2e2a [ 557.297832] env[61806]: DEBUG nova.compute.resource_tracker [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Instance 89ce0fd8-6a22-4b1c-bdff-d1584abd538a has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61806) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 557.512508] env[61806]: DEBUG oslo_concurrency.lockutils [None req-3d02814b-d8ff-4079-980b-6c580616ce84 tempest-AttachInterfacesUnderV243Test-1683422373 tempest-AttachInterfacesUnderV243Test-1683422373-project-member] Lock "df2576da-dc49-4ced-9462-fa71710e2e2a" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 32.797s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 557.804027] env[61806]: DEBUG nova.compute.resource_tracker [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Instance 76722306-a9ad-4ef9-b579-d958c36cf6bc has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61806) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 558.016675] env[61806]: DEBUG nova.compute.manager [None req-b50d0802-a025-4c5c-b8e9-95dd624e7ff5 tempest-ServersWithSpecificFlavorTestJSON-815343160 tempest-ServersWithSpecificFlavorTestJSON-815343160-project-member] [instance: b68b509e-96dc-489d-9cf1-93ba91095af1] Starting instance... {{(pid=61806) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 558.240765] env[61806]: DEBUG oslo_concurrency.lockutils [None req-e8db3b70-c93a-419a-89f8-b994313357e1 tempest-ServerShowV254Test-608938657 tempest-ServerShowV254Test-608938657-project-member] Acquiring lock "8016bafc-b287-4df1-bd4b-a00210177a86" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 558.240907] env[61806]: DEBUG oslo_concurrency.lockutils [None req-e8db3b70-c93a-419a-89f8-b994313357e1 tempest-ServerShowV254Test-608938657 tempest-ServerShowV254Test-608938657-project-member] Lock "8016bafc-b287-4df1-bd4b-a00210177a86" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 558.305910] env[61806]: DEBUG nova.compute.resource_tracker [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Instance b68b509e-96dc-489d-9cf1-93ba91095af1 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61806) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 558.544053] env[61806]: DEBUG oslo_concurrency.lockutils [None req-b50d0802-a025-4c5c-b8e9-95dd624e7ff5 tempest-ServersWithSpecificFlavorTestJSON-815343160 tempest-ServersWithSpecificFlavorTestJSON-815343160-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 558.811835] env[61806]: DEBUG nova.compute.resource_tracker [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Instance 7e730579-8d0c-4226-86bc-292db16c49ee has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61806) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 559.316907] env[61806]: DEBUG nova.compute.resource_tracker [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Instance 53d3e2f2-f63d-45c7-a5d1-1fa9a9f0ac0f has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61806) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 559.823820] env[61806]: DEBUG nova.compute.resource_tracker [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Instance ddc801a0-0782-4c6d-bd72-95d0afd01339 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61806) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 560.326657] env[61806]: DEBUG nova.compute.resource_tracker [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Instance 79fc9ea2-c427-47ce-acdb-1002ee5d78c9 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61806) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 560.609021] env[61806]: DEBUG oslo_concurrency.lockutils [None req-88cd9ecf-9bda-478c-b5e6-5a21edd9271f tempest-MigrationsAdminTest-470098049 tempest-MigrationsAdminTest-470098049-project-member] Acquiring lock "e66c921c-320e-4347-96c9-27aa80ba7f41" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 560.609428] env[61806]: DEBUG oslo_concurrency.lockutils [None req-88cd9ecf-9bda-478c-b5e6-5a21edd9271f tempest-MigrationsAdminTest-470098049 tempest-MigrationsAdminTest-470098049-project-member] Lock "e66c921c-320e-4347-96c9-27aa80ba7f41" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 560.831875] env[61806]: DEBUG nova.compute.resource_tracker [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Instance c2d1c9f6-f458-4843-982c-afdb2a90edae has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61806) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 561.337510] env[61806]: DEBUG nova.compute.resource_tracker [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Instance 65f3b2f3-6d25-44f0-95cb-f9a7e75f8e82 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61806) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 561.417423] env[61806]: DEBUG oslo_concurrency.lockutils [None req-44eba412-ca3f-4386-8c36-f42e9734628d tempest-ServersTestBootFromVolume-1815739798 tempest-ServersTestBootFromVolume-1815739798-project-member] Acquiring lock "56d26d08-3c6d-45e8-b238-6dbca6024561" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 561.417679] env[61806]: DEBUG oslo_concurrency.lockutils [None req-44eba412-ca3f-4386-8c36-f42e9734628d tempest-ServersTestBootFromVolume-1815739798 tempest-ServersTestBootFromVolume-1815739798-project-member] Lock "56d26d08-3c6d-45e8-b238-6dbca6024561" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 561.842388] env[61806]: DEBUG nova.compute.resource_tracker [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Instance 38a79453-939d-4649-9419-a2799636180d has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61806) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 562.164524] env[61806]: DEBUG oslo_concurrency.lockutils [None req-4803fdfc-9054-49cf-8525-e438672eeedd tempest-ServersAdmin275Test-1945525728 tempest-ServersAdmin275Test-1945525728-project-member] Acquiring lock "9db80c3c-e2c3-4034-a79a-d7be924c4ff6" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 562.164845] env[61806]: DEBUG oslo_concurrency.lockutils [None req-4803fdfc-9054-49cf-8525-e438672eeedd tempest-ServersAdmin275Test-1945525728 tempest-ServersAdmin275Test-1945525728-project-member] Lock "9db80c3c-e2c3-4034-a79a-d7be924c4ff6" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 562.192892] env[61806]: DEBUG oslo_concurrency.lockutils [None req-a93ef3a4-6e83-4dfe-9292-bac21f14d678 tempest-ServerExternalEventsTest-1847900285 tempest-ServerExternalEventsTest-1847900285-project-member] Acquiring lock "d5434dd2-206f-4a1e-b97b-bd6f691b6ddf" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 562.192892] env[61806]: DEBUG oslo_concurrency.lockutils [None req-a93ef3a4-6e83-4dfe-9292-bac21f14d678 tempest-ServerExternalEventsTest-1847900285 tempest-ServerExternalEventsTest-1847900285-project-member] Lock "d5434dd2-206f-4a1e-b97b-bd6f691b6ddf" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 562.347522] env[61806]: DEBUG nova.compute.resource_tracker [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Instance b8249dc5-5561-454a-aa01-477f9f1fc0e1 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61806) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 562.849273] env[61806]: DEBUG nova.compute.resource_tracker [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Instance 0d757385-ff1c-49c9-825f-f9d542ae074f has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61806) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 562.851298] env[61806]: DEBUG nova.compute.resource_tracker [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Total usable vcpus: 48, total allocated vcpus: 5 {{(pid=61806) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 562.851298] env[61806]: DEBUG nova.compute.resource_tracker [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=1536MB phys_disk=200GB used_disk=5GB total_vcpus=48 used_vcpus=5 pci_stats=[] {{(pid=61806) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 563.301044] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-310f18d9-8a87-4629-9551-5c1e2e3ad7ba {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 563.310052] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5e01224b-50ac-4e20-b3a0-15e9c0562b1a {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 563.343941] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cc7bd8d2-0a6f-4206-bc05-52c2887bf5c5 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 563.353402] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7e6121fb-f55b-4704-8a14-b084431df24b {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 563.370191] env[61806]: DEBUG nova.compute.provider_tree [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Inventory has not changed in ProviderTree for provider: a2858be1-fd22-4e08-979e-87ad25293407 {{(pid=61806) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 563.624775] env[61806]: DEBUG oslo_concurrency.lockutils [None req-b4447f02-a3c7-4b6e-89ba-6bc34ec03205 tempest-InstanceActionsTestJSON-949444324 tempest-InstanceActionsTestJSON-949444324-project-member] Acquiring lock "2681766c-0468-4170-8c51-03389e2bd1ab" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 563.625019] env[61806]: DEBUG oslo_concurrency.lockutils [None req-b4447f02-a3c7-4b6e-89ba-6bc34ec03205 tempest-InstanceActionsTestJSON-949444324 tempest-InstanceActionsTestJSON-949444324-project-member] Lock "2681766c-0468-4170-8c51-03389e2bd1ab" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 563.873551] env[61806]: DEBUG nova.scheduler.client.report [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Inventory has not changed for provider a2858be1-fd22-4e08-979e-87ad25293407 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 139, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61806) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 564.382664] env[61806]: DEBUG nova.compute.resource_tracker [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=61806) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 564.382936] env[61806]: DEBUG oslo_concurrency.lockutils [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 10.652s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 564.385208] env[61806]: DEBUG oslo_concurrency.lockutils [None req-be9179b6-8d17-4f04-8823-d48174fcc5c1 tempest-InstanceActionsV221TestJSON-379107765 tempest-InstanceActionsV221TestJSON-379107765-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 24.134s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 564.385208] env[61806]: INFO nova.compute.claims [None req-be9179b6-8d17-4f04-8823-d48174fcc5c1 tempest-InstanceActionsV221TestJSON-379107765 tempest-InstanceActionsV221TestJSON-379107765-project-member] [instance: 123f15be-3b56-4999-8010-40d2e86a6813] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 565.604416] env[61806]: DEBUG oslo_concurrency.lockutils [None req-7bb45eee-0cbe-4995-9898-530114f6f104 tempest-ImagesOneServerNegativeTestJSON-606465252 tempest-ImagesOneServerNegativeTestJSON-606465252-project-member] Acquiring lock "8d0cdc81-56ed-4a7a-8221-30bb906e16b2" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 565.605273] env[61806]: DEBUG oslo_concurrency.lockutils [None req-7bb45eee-0cbe-4995-9898-530114f6f104 tempest-ImagesOneServerNegativeTestJSON-606465252 tempest-ImagesOneServerNegativeTestJSON-606465252-project-member] Lock "8d0cdc81-56ed-4a7a-8221-30bb906e16b2" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 565.897030] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9c7dc21c-ed9a-401a-aeb4-ce9ef6aa26fe {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 565.908779] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-39bf3764-56c2-42a1-b2de-24d13e07f776 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 565.948032] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c3ecc490-851e-4c7b-9e1c-57f67e501dcf {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 565.959753] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-81674f05-65cc-4f69-8f9f-483981f1a050 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 565.975023] env[61806]: DEBUG nova.compute.provider_tree [None req-be9179b6-8d17-4f04-8823-d48174fcc5c1 tempest-InstanceActionsV221TestJSON-379107765 tempest-InstanceActionsV221TestJSON-379107765-project-member] Inventory has not changed in ProviderTree for provider: a2858be1-fd22-4e08-979e-87ad25293407 {{(pid=61806) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 566.478438] env[61806]: DEBUG nova.scheduler.client.report [None req-be9179b6-8d17-4f04-8823-d48174fcc5c1 tempest-InstanceActionsV221TestJSON-379107765 tempest-InstanceActionsV221TestJSON-379107765-project-member] Inventory has not changed for provider a2858be1-fd22-4e08-979e-87ad25293407 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 139, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61806) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 566.628984] env[61806]: DEBUG oslo_concurrency.lockutils [None req-865353ee-538a-40fc-bfba-502aba0b6a20 tempest-VolumesAdminNegativeTest-1387177911 tempest-VolumesAdminNegativeTest-1387177911-project-member] Acquiring lock "b7d55e87-95db-4c45-9417-bbece4fbb98d" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 566.629274] env[61806]: DEBUG oslo_concurrency.lockutils [None req-865353ee-538a-40fc-bfba-502aba0b6a20 tempest-VolumesAdminNegativeTest-1387177911 tempest-VolumesAdminNegativeTest-1387177911-project-member] Lock "b7d55e87-95db-4c45-9417-bbece4fbb98d" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 566.990399] env[61806]: DEBUG oslo_concurrency.lockutils [None req-be9179b6-8d17-4f04-8823-d48174fcc5c1 tempest-InstanceActionsV221TestJSON-379107765 tempest-InstanceActionsV221TestJSON-379107765-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.606s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 566.990399] env[61806]: DEBUG nova.compute.manager [None req-be9179b6-8d17-4f04-8823-d48174fcc5c1 tempest-InstanceActionsV221TestJSON-379107765 tempest-InstanceActionsV221TestJSON-379107765-project-member] [instance: 123f15be-3b56-4999-8010-40d2e86a6813] Start building networks asynchronously for instance. {{(pid=61806) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 566.993969] env[61806]: DEBUG oslo_concurrency.lockutils [None req-d8e4a121-0976-4eaa-b4c6-51deda86d081 tempest-InstanceActionsNegativeTestJSON-503825571 tempest-InstanceActionsNegativeTestJSON-503825571-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 25.443s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 567.498221] env[61806]: DEBUG nova.compute.utils [None req-be9179b6-8d17-4f04-8823-d48174fcc5c1 tempest-InstanceActionsV221TestJSON-379107765 tempest-InstanceActionsV221TestJSON-379107765-project-member] Using /dev/sd instead of None {{(pid=61806) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 567.506392] env[61806]: DEBUG nova.compute.manager [None req-be9179b6-8d17-4f04-8823-d48174fcc5c1 tempest-InstanceActionsV221TestJSON-379107765 tempest-InstanceActionsV221TestJSON-379107765-project-member] [instance: 123f15be-3b56-4999-8010-40d2e86a6813] Allocating IP information in the background. {{(pid=61806) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 567.506566] env[61806]: DEBUG nova.network.neutron [None req-be9179b6-8d17-4f04-8823-d48174fcc5c1 tempest-InstanceActionsV221TestJSON-379107765 tempest-InstanceActionsV221TestJSON-379107765-project-member] [instance: 123f15be-3b56-4999-8010-40d2e86a6813] allocate_for_instance() {{(pid=61806) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 567.595270] env[61806]: DEBUG nova.policy [None req-be9179b6-8d17-4f04-8823-d48174fcc5c1 tempest-InstanceActionsV221TestJSON-379107765 tempest-InstanceActionsV221TestJSON-379107765-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '201df270706c47dab68637483ce4a3e4', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'd676693c21bd4d888177fa2c28fe1bcc', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61806) authorize /opt/stack/nova/nova/policy.py:201}} [ 567.999917] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7ca291a7-4fc2-435b-aafe-4ea201df20b2 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 568.007873] env[61806]: DEBUG nova.compute.manager [None req-be9179b6-8d17-4f04-8823-d48174fcc5c1 tempest-InstanceActionsV221TestJSON-379107765 tempest-InstanceActionsV221TestJSON-379107765-project-member] [instance: 123f15be-3b56-4999-8010-40d2e86a6813] Start building block device mappings for instance. {{(pid=61806) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 568.011681] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-07659e81-8244-4c68-ba23-c8ac7318c1d2 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 568.043270] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b2480045-3c95-46de-87c1-d0dad4bef0af {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 568.046647] env[61806]: DEBUG nova.network.neutron [None req-be9179b6-8d17-4f04-8823-d48174fcc5c1 tempest-InstanceActionsV221TestJSON-379107765 tempest-InstanceActionsV221TestJSON-379107765-project-member] [instance: 123f15be-3b56-4999-8010-40d2e86a6813] Successfully created port: 8fbea74a-6823-4774-b3af-5bda4ccd8820 {{(pid=61806) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 568.055930] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-86ff21da-e031-4401-90c9-6edc173332a8 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 568.071734] env[61806]: DEBUG nova.compute.provider_tree [None req-d8e4a121-0976-4eaa-b4c6-51deda86d081 tempest-InstanceActionsNegativeTestJSON-503825571 tempest-InstanceActionsNegativeTestJSON-503825571-project-member] Inventory has not changed in ProviderTree for provider: a2858be1-fd22-4e08-979e-87ad25293407 {{(pid=61806) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 568.577021] env[61806]: DEBUG nova.scheduler.client.report [None req-d8e4a121-0976-4eaa-b4c6-51deda86d081 tempest-InstanceActionsNegativeTestJSON-503825571 tempest-InstanceActionsNegativeTestJSON-503825571-project-member] Inventory has not changed for provider a2858be1-fd22-4e08-979e-87ad25293407 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 139, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61806) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 569.025453] env[61806]: DEBUG nova.compute.manager [None req-be9179b6-8d17-4f04-8823-d48174fcc5c1 tempest-InstanceActionsV221TestJSON-379107765 tempest-InstanceActionsV221TestJSON-379107765-project-member] [instance: 123f15be-3b56-4999-8010-40d2e86a6813] Start spawning the instance on the hypervisor. {{(pid=61806) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 569.065117] env[61806]: DEBUG nova.virt.hardware [None req-be9179b6-8d17-4f04-8823-d48174fcc5c1 tempest-InstanceActionsV221TestJSON-379107765 tempest-InstanceActionsV221TestJSON-379107765-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-15T18:05:10Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-15T18:04:52Z,direct_url=,disk_format='vmdk',id=ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='84ba83fce2c349a6988173c5d6fc3b07',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-15T18:04:53Z,virtual_size=,visibility=), allow threads: False {{(pid=61806) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 569.065376] env[61806]: DEBUG nova.virt.hardware [None req-be9179b6-8d17-4f04-8823-d48174fcc5c1 tempest-InstanceActionsV221TestJSON-379107765 tempest-InstanceActionsV221TestJSON-379107765-project-member] Flavor limits 0:0:0 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 569.065980] env[61806]: DEBUG nova.virt.hardware [None req-be9179b6-8d17-4f04-8823-d48174fcc5c1 tempest-InstanceActionsV221TestJSON-379107765 tempest-InstanceActionsV221TestJSON-379107765-project-member] Image limits 0:0:0 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 569.066236] env[61806]: DEBUG nova.virt.hardware [None req-be9179b6-8d17-4f04-8823-d48174fcc5c1 tempest-InstanceActionsV221TestJSON-379107765 tempest-InstanceActionsV221TestJSON-379107765-project-member] Flavor pref 0:0:0 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 569.066386] env[61806]: DEBUG nova.virt.hardware [None req-be9179b6-8d17-4f04-8823-d48174fcc5c1 tempest-InstanceActionsV221TestJSON-379107765 tempest-InstanceActionsV221TestJSON-379107765-project-member] Image pref 0:0:0 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 569.066533] env[61806]: DEBUG nova.virt.hardware [None req-be9179b6-8d17-4f04-8823-d48174fcc5c1 tempest-InstanceActionsV221TestJSON-379107765 tempest-InstanceActionsV221TestJSON-379107765-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 569.067224] env[61806]: DEBUG nova.virt.hardware [None req-be9179b6-8d17-4f04-8823-d48174fcc5c1 tempest-InstanceActionsV221TestJSON-379107765 tempest-InstanceActionsV221TestJSON-379107765-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61806) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 569.067451] env[61806]: DEBUG nova.virt.hardware [None req-be9179b6-8d17-4f04-8823-d48174fcc5c1 tempest-InstanceActionsV221TestJSON-379107765 tempest-InstanceActionsV221TestJSON-379107765-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61806) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 569.067641] env[61806]: DEBUG nova.virt.hardware [None req-be9179b6-8d17-4f04-8823-d48174fcc5c1 tempest-InstanceActionsV221TestJSON-379107765 tempest-InstanceActionsV221TestJSON-379107765-project-member] Got 1 possible topologies {{(pid=61806) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 569.068310] env[61806]: DEBUG nova.virt.hardware [None req-be9179b6-8d17-4f04-8823-d48174fcc5c1 tempest-InstanceActionsV221TestJSON-379107765 tempest-InstanceActionsV221TestJSON-379107765-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61806) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 569.068310] env[61806]: DEBUG nova.virt.hardware [None req-be9179b6-8d17-4f04-8823-d48174fcc5c1 tempest-InstanceActionsV221TestJSON-379107765 tempest-InstanceActionsV221TestJSON-379107765-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61806) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 569.069527] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9b21dc86-357b-4e7f-bfee-0e46a3065763 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 569.081414] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-08af681f-b78e-48a8-bb87-9dfca350dc79 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 569.089809] env[61806]: DEBUG oslo_concurrency.lockutils [None req-d8e4a121-0976-4eaa-b4c6-51deda86d081 tempest-InstanceActionsNegativeTestJSON-503825571 tempest-InstanceActionsNegativeTestJSON-503825571-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.094s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 569.089870] env[61806]: ERROR nova.compute.manager [None req-d8e4a121-0976-4eaa-b4c6-51deda86d081 tempest-InstanceActionsNegativeTestJSON-503825571 tempest-InstanceActionsNegativeTestJSON-503825571-project-member] [instance: a1aca638-e9c4-4e3b-9855-f92e7bd67204] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 6cc2dbc9-ee9c-42fa-b51e-2fd8a2c5b6d9, please check neutron logs for more information. [ 569.089870] env[61806]: ERROR nova.compute.manager [instance: a1aca638-e9c4-4e3b-9855-f92e7bd67204] Traceback (most recent call last): [ 569.089870] env[61806]: ERROR nova.compute.manager [instance: a1aca638-e9c4-4e3b-9855-f92e7bd67204] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 569.089870] env[61806]: ERROR nova.compute.manager [instance: a1aca638-e9c4-4e3b-9855-f92e7bd67204] self.driver.spawn(context, instance, image_meta, [ 569.089870] env[61806]: ERROR nova.compute.manager [instance: a1aca638-e9c4-4e3b-9855-f92e7bd67204] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 569.089870] env[61806]: ERROR nova.compute.manager [instance: a1aca638-e9c4-4e3b-9855-f92e7bd67204] self._vmops.spawn(context, instance, image_meta, injected_files, [ 569.089870] env[61806]: ERROR nova.compute.manager [instance: a1aca638-e9c4-4e3b-9855-f92e7bd67204] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 569.089870] env[61806]: ERROR nova.compute.manager [instance: a1aca638-e9c4-4e3b-9855-f92e7bd67204] vm_ref = self.build_virtual_machine(instance, [ 569.089870] env[61806]: ERROR nova.compute.manager [instance: a1aca638-e9c4-4e3b-9855-f92e7bd67204] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 569.089870] env[61806]: ERROR nova.compute.manager [instance: a1aca638-e9c4-4e3b-9855-f92e7bd67204] vif_infos = vmwarevif.get_vif_info(self._session, [ 569.089870] env[61806]: ERROR nova.compute.manager [instance: a1aca638-e9c4-4e3b-9855-f92e7bd67204] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 569.090216] env[61806]: ERROR nova.compute.manager [instance: a1aca638-e9c4-4e3b-9855-f92e7bd67204] for vif in network_info: [ 569.090216] env[61806]: ERROR nova.compute.manager [instance: a1aca638-e9c4-4e3b-9855-f92e7bd67204] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 569.090216] env[61806]: ERROR nova.compute.manager [instance: a1aca638-e9c4-4e3b-9855-f92e7bd67204] return self._sync_wrapper(fn, *args, **kwargs) [ 569.090216] env[61806]: ERROR nova.compute.manager [instance: a1aca638-e9c4-4e3b-9855-f92e7bd67204] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 569.090216] env[61806]: ERROR nova.compute.manager [instance: a1aca638-e9c4-4e3b-9855-f92e7bd67204] self.wait() [ 569.090216] env[61806]: ERROR nova.compute.manager [instance: a1aca638-e9c4-4e3b-9855-f92e7bd67204] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 569.090216] env[61806]: ERROR nova.compute.manager [instance: a1aca638-e9c4-4e3b-9855-f92e7bd67204] self[:] = self._gt.wait() [ 569.090216] env[61806]: ERROR nova.compute.manager [instance: a1aca638-e9c4-4e3b-9855-f92e7bd67204] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 569.090216] env[61806]: ERROR nova.compute.manager [instance: a1aca638-e9c4-4e3b-9855-f92e7bd67204] return self._exit_event.wait() [ 569.090216] env[61806]: ERROR nova.compute.manager [instance: a1aca638-e9c4-4e3b-9855-f92e7bd67204] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 569.090216] env[61806]: ERROR nova.compute.manager [instance: a1aca638-e9c4-4e3b-9855-f92e7bd67204] result = hub.switch() [ 569.090216] env[61806]: ERROR nova.compute.manager [instance: a1aca638-e9c4-4e3b-9855-f92e7bd67204] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 569.090216] env[61806]: ERROR nova.compute.manager [instance: a1aca638-e9c4-4e3b-9855-f92e7bd67204] return self.greenlet.switch() [ 569.090625] env[61806]: ERROR nova.compute.manager [instance: a1aca638-e9c4-4e3b-9855-f92e7bd67204] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 569.090625] env[61806]: ERROR nova.compute.manager [instance: a1aca638-e9c4-4e3b-9855-f92e7bd67204] result = function(*args, **kwargs) [ 569.090625] env[61806]: ERROR nova.compute.manager [instance: a1aca638-e9c4-4e3b-9855-f92e7bd67204] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 569.090625] env[61806]: ERROR nova.compute.manager [instance: a1aca638-e9c4-4e3b-9855-f92e7bd67204] return func(*args, **kwargs) [ 569.090625] env[61806]: ERROR nova.compute.manager [instance: a1aca638-e9c4-4e3b-9855-f92e7bd67204] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 569.090625] env[61806]: ERROR nova.compute.manager [instance: a1aca638-e9c4-4e3b-9855-f92e7bd67204] raise e [ 569.090625] env[61806]: ERROR nova.compute.manager [instance: a1aca638-e9c4-4e3b-9855-f92e7bd67204] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 569.090625] env[61806]: ERROR nova.compute.manager [instance: a1aca638-e9c4-4e3b-9855-f92e7bd67204] nwinfo = self.network_api.allocate_for_instance( [ 569.090625] env[61806]: ERROR nova.compute.manager [instance: a1aca638-e9c4-4e3b-9855-f92e7bd67204] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 569.090625] env[61806]: ERROR nova.compute.manager [instance: a1aca638-e9c4-4e3b-9855-f92e7bd67204] created_port_ids = self._update_ports_for_instance( [ 569.090625] env[61806]: ERROR nova.compute.manager [instance: a1aca638-e9c4-4e3b-9855-f92e7bd67204] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 569.090625] env[61806]: ERROR nova.compute.manager [instance: a1aca638-e9c4-4e3b-9855-f92e7bd67204] with excutils.save_and_reraise_exception(): [ 569.090625] env[61806]: ERROR nova.compute.manager [instance: a1aca638-e9c4-4e3b-9855-f92e7bd67204] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 569.091068] env[61806]: ERROR nova.compute.manager [instance: a1aca638-e9c4-4e3b-9855-f92e7bd67204] self.force_reraise() [ 569.091068] env[61806]: ERROR nova.compute.manager [instance: a1aca638-e9c4-4e3b-9855-f92e7bd67204] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 569.091068] env[61806]: ERROR nova.compute.manager [instance: a1aca638-e9c4-4e3b-9855-f92e7bd67204] raise self.value [ 569.091068] env[61806]: ERROR nova.compute.manager [instance: a1aca638-e9c4-4e3b-9855-f92e7bd67204] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 569.091068] env[61806]: ERROR nova.compute.manager [instance: a1aca638-e9c4-4e3b-9855-f92e7bd67204] updated_port = self._update_port( [ 569.091068] env[61806]: ERROR nova.compute.manager [instance: a1aca638-e9c4-4e3b-9855-f92e7bd67204] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 569.091068] env[61806]: ERROR nova.compute.manager [instance: a1aca638-e9c4-4e3b-9855-f92e7bd67204] _ensure_no_port_binding_failure(port) [ 569.091068] env[61806]: ERROR nova.compute.manager [instance: a1aca638-e9c4-4e3b-9855-f92e7bd67204] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 569.091068] env[61806]: ERROR nova.compute.manager [instance: a1aca638-e9c4-4e3b-9855-f92e7bd67204] raise exception.PortBindingFailed(port_id=port['id']) [ 569.091068] env[61806]: ERROR nova.compute.manager [instance: a1aca638-e9c4-4e3b-9855-f92e7bd67204] nova.exception.PortBindingFailed: Binding failed for port 6cc2dbc9-ee9c-42fa-b51e-2fd8a2c5b6d9, please check neutron logs for more information. [ 569.091068] env[61806]: ERROR nova.compute.manager [instance: a1aca638-e9c4-4e3b-9855-f92e7bd67204] [ 569.091440] env[61806]: DEBUG nova.compute.utils [None req-d8e4a121-0976-4eaa-b4c6-51deda86d081 tempest-InstanceActionsNegativeTestJSON-503825571 tempest-InstanceActionsNegativeTestJSON-503825571-project-member] [instance: a1aca638-e9c4-4e3b-9855-f92e7bd67204] Binding failed for port 6cc2dbc9-ee9c-42fa-b51e-2fd8a2c5b6d9, please check neutron logs for more information. {{(pid=61806) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 569.093622] env[61806]: DEBUG nova.compute.manager [req-3b678f53-eeb4-46b1-9fc1-8b6e9352dd63 req-9ad69bf5-731c-4af0-ae79-f7399f8f1658 service nova] [instance: 123f15be-3b56-4999-8010-40d2e86a6813] Received event network-changed-8fbea74a-6823-4774-b3af-5bda4ccd8820 {{(pid=61806) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 569.093622] env[61806]: DEBUG nova.compute.manager [req-3b678f53-eeb4-46b1-9fc1-8b6e9352dd63 req-9ad69bf5-731c-4af0-ae79-f7399f8f1658 service nova] [instance: 123f15be-3b56-4999-8010-40d2e86a6813] Refreshing instance network info cache due to event network-changed-8fbea74a-6823-4774-b3af-5bda4ccd8820. {{(pid=61806) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 569.093622] env[61806]: DEBUG oslo_concurrency.lockutils [req-3b678f53-eeb4-46b1-9fc1-8b6e9352dd63 req-9ad69bf5-731c-4af0-ae79-f7399f8f1658 service nova] Acquiring lock "refresh_cache-123f15be-3b56-4999-8010-40d2e86a6813" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 569.093747] env[61806]: DEBUG oslo_concurrency.lockutils [req-3b678f53-eeb4-46b1-9fc1-8b6e9352dd63 req-9ad69bf5-731c-4af0-ae79-f7399f8f1658 service nova] Acquired lock "refresh_cache-123f15be-3b56-4999-8010-40d2e86a6813" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 569.093861] env[61806]: DEBUG nova.network.neutron [req-3b678f53-eeb4-46b1-9fc1-8b6e9352dd63 req-9ad69bf5-731c-4af0-ae79-f7399f8f1658 service nova] [instance: 123f15be-3b56-4999-8010-40d2e86a6813] Refreshing network info cache for port 8fbea74a-6823-4774-b3af-5bda4ccd8820 {{(pid=61806) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 569.095228] env[61806]: DEBUG nova.compute.manager [None req-d8e4a121-0976-4eaa-b4c6-51deda86d081 tempest-InstanceActionsNegativeTestJSON-503825571 tempest-InstanceActionsNegativeTestJSON-503825571-project-member] [instance: a1aca638-e9c4-4e3b-9855-f92e7bd67204] Build of instance a1aca638-e9c4-4e3b-9855-f92e7bd67204 was re-scheduled: Binding failed for port 6cc2dbc9-ee9c-42fa-b51e-2fd8a2c5b6d9, please check neutron logs for more information. {{(pid=61806) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 569.099052] env[61806]: DEBUG nova.compute.manager [None req-d8e4a121-0976-4eaa-b4c6-51deda86d081 tempest-InstanceActionsNegativeTestJSON-503825571 tempest-InstanceActionsNegativeTestJSON-503825571-project-member] [instance: a1aca638-e9c4-4e3b-9855-f92e7bd67204] Unplugging VIFs for instance {{(pid=61806) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 569.099341] env[61806]: DEBUG oslo_concurrency.lockutils [None req-d8e4a121-0976-4eaa-b4c6-51deda86d081 tempest-InstanceActionsNegativeTestJSON-503825571 tempest-InstanceActionsNegativeTestJSON-503825571-project-member] Acquiring lock "refresh_cache-a1aca638-e9c4-4e3b-9855-f92e7bd67204" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 569.099538] env[61806]: DEBUG oslo_concurrency.lockutils [None req-d8e4a121-0976-4eaa-b4c6-51deda86d081 tempest-InstanceActionsNegativeTestJSON-503825571 tempest-InstanceActionsNegativeTestJSON-503825571-project-member] Acquired lock "refresh_cache-a1aca638-e9c4-4e3b-9855-f92e7bd67204" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 569.099707] env[61806]: DEBUG nova.network.neutron [None req-d8e4a121-0976-4eaa-b4c6-51deda86d081 tempest-InstanceActionsNegativeTestJSON-503825571 tempest-InstanceActionsNegativeTestJSON-503825571-project-member] [instance: a1aca638-e9c4-4e3b-9855-f92e7bd67204] Building network info cache for instance {{(pid=61806) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 569.100707] env[61806]: DEBUG oslo_concurrency.lockutils [None req-960d1abd-1b06-4b5a-8ddd-4b463d493a45 tempest-ListServerFiltersTestJSON-241228521 tempest-ListServerFiltersTestJSON-241228521-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 25.529s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 569.135765] env[61806]: ERROR nova.compute.manager [None req-be9179b6-8d17-4f04-8823-d48174fcc5c1 tempest-InstanceActionsV221TestJSON-379107765 tempest-InstanceActionsV221TestJSON-379107765-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 8fbea74a-6823-4774-b3af-5bda4ccd8820, please check neutron logs for more information. [ 569.135765] env[61806]: ERROR nova.compute.manager Traceback (most recent call last): [ 569.135765] env[61806]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 569.135765] env[61806]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 569.135765] env[61806]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 569.135765] env[61806]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 569.135765] env[61806]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 569.135765] env[61806]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 569.135765] env[61806]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 569.135765] env[61806]: ERROR nova.compute.manager self.force_reraise() [ 569.135765] env[61806]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 569.135765] env[61806]: ERROR nova.compute.manager raise self.value [ 569.135765] env[61806]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 569.135765] env[61806]: ERROR nova.compute.manager updated_port = self._update_port( [ 569.135765] env[61806]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 569.135765] env[61806]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 569.136359] env[61806]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 569.136359] env[61806]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 569.136359] env[61806]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 8fbea74a-6823-4774-b3af-5bda4ccd8820, please check neutron logs for more information. [ 569.136359] env[61806]: ERROR nova.compute.manager [ 569.136359] env[61806]: Traceback (most recent call last): [ 569.136359] env[61806]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 569.136359] env[61806]: listener.cb(fileno) [ 569.136359] env[61806]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 569.136359] env[61806]: result = function(*args, **kwargs) [ 569.136359] env[61806]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 569.136359] env[61806]: return func(*args, **kwargs) [ 569.136359] env[61806]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 569.136359] env[61806]: raise e [ 569.136359] env[61806]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 569.136359] env[61806]: nwinfo = self.network_api.allocate_for_instance( [ 569.136359] env[61806]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 569.136359] env[61806]: created_port_ids = self._update_ports_for_instance( [ 569.136359] env[61806]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 569.136359] env[61806]: with excutils.save_and_reraise_exception(): [ 569.136359] env[61806]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 569.136359] env[61806]: self.force_reraise() [ 569.136359] env[61806]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 569.136359] env[61806]: raise self.value [ 569.136359] env[61806]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 569.136359] env[61806]: updated_port = self._update_port( [ 569.136359] env[61806]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 569.136359] env[61806]: _ensure_no_port_binding_failure(port) [ 569.136359] env[61806]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 569.136359] env[61806]: raise exception.PortBindingFailed(port_id=port['id']) [ 569.137221] env[61806]: nova.exception.PortBindingFailed: Binding failed for port 8fbea74a-6823-4774-b3af-5bda4ccd8820, please check neutron logs for more information. [ 569.137221] env[61806]: Removing descriptor: 16 [ 569.137221] env[61806]: ERROR nova.compute.manager [None req-be9179b6-8d17-4f04-8823-d48174fcc5c1 tempest-InstanceActionsV221TestJSON-379107765 tempest-InstanceActionsV221TestJSON-379107765-project-member] [instance: 123f15be-3b56-4999-8010-40d2e86a6813] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 8fbea74a-6823-4774-b3af-5bda4ccd8820, please check neutron logs for more information. [ 569.137221] env[61806]: ERROR nova.compute.manager [instance: 123f15be-3b56-4999-8010-40d2e86a6813] Traceback (most recent call last): [ 569.137221] env[61806]: ERROR nova.compute.manager [instance: 123f15be-3b56-4999-8010-40d2e86a6813] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 569.137221] env[61806]: ERROR nova.compute.manager [instance: 123f15be-3b56-4999-8010-40d2e86a6813] yield resources [ 569.137221] env[61806]: ERROR nova.compute.manager [instance: 123f15be-3b56-4999-8010-40d2e86a6813] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 569.137221] env[61806]: ERROR nova.compute.manager [instance: 123f15be-3b56-4999-8010-40d2e86a6813] self.driver.spawn(context, instance, image_meta, [ 569.137221] env[61806]: ERROR nova.compute.manager [instance: 123f15be-3b56-4999-8010-40d2e86a6813] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 569.137221] env[61806]: ERROR nova.compute.manager [instance: 123f15be-3b56-4999-8010-40d2e86a6813] self._vmops.spawn(context, instance, image_meta, injected_files, [ 569.137221] env[61806]: ERROR nova.compute.manager [instance: 123f15be-3b56-4999-8010-40d2e86a6813] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 569.137221] env[61806]: ERROR nova.compute.manager [instance: 123f15be-3b56-4999-8010-40d2e86a6813] vm_ref = self.build_virtual_machine(instance, [ 569.138102] env[61806]: ERROR nova.compute.manager [instance: 123f15be-3b56-4999-8010-40d2e86a6813] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 569.138102] env[61806]: ERROR nova.compute.manager [instance: 123f15be-3b56-4999-8010-40d2e86a6813] vif_infos = vmwarevif.get_vif_info(self._session, [ 569.138102] env[61806]: ERROR nova.compute.manager [instance: 123f15be-3b56-4999-8010-40d2e86a6813] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 569.138102] env[61806]: ERROR nova.compute.manager [instance: 123f15be-3b56-4999-8010-40d2e86a6813] for vif in network_info: [ 569.138102] env[61806]: ERROR nova.compute.manager [instance: 123f15be-3b56-4999-8010-40d2e86a6813] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 569.138102] env[61806]: ERROR nova.compute.manager [instance: 123f15be-3b56-4999-8010-40d2e86a6813] return self._sync_wrapper(fn, *args, **kwargs) [ 569.138102] env[61806]: ERROR nova.compute.manager [instance: 123f15be-3b56-4999-8010-40d2e86a6813] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 569.138102] env[61806]: ERROR nova.compute.manager [instance: 123f15be-3b56-4999-8010-40d2e86a6813] self.wait() [ 569.138102] env[61806]: ERROR nova.compute.manager [instance: 123f15be-3b56-4999-8010-40d2e86a6813] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 569.138102] env[61806]: ERROR nova.compute.manager [instance: 123f15be-3b56-4999-8010-40d2e86a6813] self[:] = self._gt.wait() [ 569.138102] env[61806]: ERROR nova.compute.manager [instance: 123f15be-3b56-4999-8010-40d2e86a6813] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 569.138102] env[61806]: ERROR nova.compute.manager [instance: 123f15be-3b56-4999-8010-40d2e86a6813] return self._exit_event.wait() [ 569.138102] env[61806]: ERROR nova.compute.manager [instance: 123f15be-3b56-4999-8010-40d2e86a6813] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 569.138538] env[61806]: ERROR nova.compute.manager [instance: 123f15be-3b56-4999-8010-40d2e86a6813] result = hub.switch() [ 569.138538] env[61806]: ERROR nova.compute.manager [instance: 123f15be-3b56-4999-8010-40d2e86a6813] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 569.138538] env[61806]: ERROR nova.compute.manager [instance: 123f15be-3b56-4999-8010-40d2e86a6813] return self.greenlet.switch() [ 569.138538] env[61806]: ERROR nova.compute.manager [instance: 123f15be-3b56-4999-8010-40d2e86a6813] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 569.138538] env[61806]: ERROR nova.compute.manager [instance: 123f15be-3b56-4999-8010-40d2e86a6813] result = function(*args, **kwargs) [ 569.138538] env[61806]: ERROR nova.compute.manager [instance: 123f15be-3b56-4999-8010-40d2e86a6813] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 569.138538] env[61806]: ERROR nova.compute.manager [instance: 123f15be-3b56-4999-8010-40d2e86a6813] return func(*args, **kwargs) [ 569.138538] env[61806]: ERROR nova.compute.manager [instance: 123f15be-3b56-4999-8010-40d2e86a6813] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 569.138538] env[61806]: ERROR nova.compute.manager [instance: 123f15be-3b56-4999-8010-40d2e86a6813] raise e [ 569.138538] env[61806]: ERROR nova.compute.manager [instance: 123f15be-3b56-4999-8010-40d2e86a6813] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 569.138538] env[61806]: ERROR nova.compute.manager [instance: 123f15be-3b56-4999-8010-40d2e86a6813] nwinfo = self.network_api.allocate_for_instance( [ 569.138538] env[61806]: ERROR nova.compute.manager [instance: 123f15be-3b56-4999-8010-40d2e86a6813] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 569.138538] env[61806]: ERROR nova.compute.manager [instance: 123f15be-3b56-4999-8010-40d2e86a6813] created_port_ids = self._update_ports_for_instance( [ 569.139221] env[61806]: ERROR nova.compute.manager [instance: 123f15be-3b56-4999-8010-40d2e86a6813] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 569.139221] env[61806]: ERROR nova.compute.manager [instance: 123f15be-3b56-4999-8010-40d2e86a6813] with excutils.save_and_reraise_exception(): [ 569.139221] env[61806]: ERROR nova.compute.manager [instance: 123f15be-3b56-4999-8010-40d2e86a6813] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 569.139221] env[61806]: ERROR nova.compute.manager [instance: 123f15be-3b56-4999-8010-40d2e86a6813] self.force_reraise() [ 569.139221] env[61806]: ERROR nova.compute.manager [instance: 123f15be-3b56-4999-8010-40d2e86a6813] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 569.139221] env[61806]: ERROR nova.compute.manager [instance: 123f15be-3b56-4999-8010-40d2e86a6813] raise self.value [ 569.139221] env[61806]: ERROR nova.compute.manager [instance: 123f15be-3b56-4999-8010-40d2e86a6813] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 569.139221] env[61806]: ERROR nova.compute.manager [instance: 123f15be-3b56-4999-8010-40d2e86a6813] updated_port = self._update_port( [ 569.139221] env[61806]: ERROR nova.compute.manager [instance: 123f15be-3b56-4999-8010-40d2e86a6813] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 569.139221] env[61806]: ERROR nova.compute.manager [instance: 123f15be-3b56-4999-8010-40d2e86a6813] _ensure_no_port_binding_failure(port) [ 569.139221] env[61806]: ERROR nova.compute.manager [instance: 123f15be-3b56-4999-8010-40d2e86a6813] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 569.139221] env[61806]: ERROR nova.compute.manager [instance: 123f15be-3b56-4999-8010-40d2e86a6813] raise exception.PortBindingFailed(port_id=port['id']) [ 569.139608] env[61806]: ERROR nova.compute.manager [instance: 123f15be-3b56-4999-8010-40d2e86a6813] nova.exception.PortBindingFailed: Binding failed for port 8fbea74a-6823-4774-b3af-5bda4ccd8820, please check neutron logs for more information. [ 569.139608] env[61806]: ERROR nova.compute.manager [instance: 123f15be-3b56-4999-8010-40d2e86a6813] [ 569.139608] env[61806]: INFO nova.compute.manager [None req-be9179b6-8d17-4f04-8823-d48174fcc5c1 tempest-InstanceActionsV221TestJSON-379107765 tempest-InstanceActionsV221TestJSON-379107765-project-member] [instance: 123f15be-3b56-4999-8010-40d2e86a6813] Terminating instance [ 569.142848] env[61806]: DEBUG oslo_concurrency.lockutils [None req-be9179b6-8d17-4f04-8823-d48174fcc5c1 tempest-InstanceActionsV221TestJSON-379107765 tempest-InstanceActionsV221TestJSON-379107765-project-member] Acquiring lock "refresh_cache-123f15be-3b56-4999-8010-40d2e86a6813" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 569.491014] env[61806]: DEBUG oslo_concurrency.lockutils [None req-81a55da1-45c4-45f8-af56-f8f7d424c8aa tempest-ServersTestJSON-430716070 tempest-ServersTestJSON-430716070-project-member] Acquiring lock "eca2e3da-b38c-482e-b9b6-4ad75d83caa1" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 569.491014] env[61806]: DEBUG oslo_concurrency.lockutils [None req-81a55da1-45c4-45f8-af56-f8f7d424c8aa tempest-ServersTestJSON-430716070 tempest-ServersTestJSON-430716070-project-member] Lock "eca2e3da-b38c-482e-b9b6-4ad75d83caa1" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 569.644177] env[61806]: DEBUG nova.network.neutron [None req-d8e4a121-0976-4eaa-b4c6-51deda86d081 tempest-InstanceActionsNegativeTestJSON-503825571 tempest-InstanceActionsNegativeTestJSON-503825571-project-member] [instance: a1aca638-e9c4-4e3b-9855-f92e7bd67204] Instance cache missing network info. {{(pid=61806) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 569.653087] env[61806]: DEBUG nova.network.neutron [req-3b678f53-eeb4-46b1-9fc1-8b6e9352dd63 req-9ad69bf5-731c-4af0-ae79-f7399f8f1658 service nova] [instance: 123f15be-3b56-4999-8010-40d2e86a6813] Instance cache missing network info. {{(pid=61806) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 569.820335] env[61806]: DEBUG nova.network.neutron [req-3b678f53-eeb4-46b1-9fc1-8b6e9352dd63 req-9ad69bf5-731c-4af0-ae79-f7399f8f1658 service nova] [instance: 123f15be-3b56-4999-8010-40d2e86a6813] Updating instance_info_cache with network_info: [] {{(pid=61806) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 569.884221] env[61806]: DEBUG nova.network.neutron [None req-d8e4a121-0976-4eaa-b4c6-51deda86d081 tempest-InstanceActionsNegativeTestJSON-503825571 tempest-InstanceActionsNegativeTestJSON-503825571-project-member] [instance: a1aca638-e9c4-4e3b-9855-f92e7bd67204] Updating instance_info_cache with network_info: [] {{(pid=61806) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 570.162951] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7622d3b5-12ca-4e5e-99b2-2cd057b4131a {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 570.175167] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-500cee77-5f9b-41ab-a47a-0aa04a37e6eb {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 570.213283] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a86e753f-8d87-4940-b7c1-e9a4717ac20a {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 570.227486] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d6ffe45c-397d-4c6e-8491-dcc682a58c66 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 570.244642] env[61806]: DEBUG nova.compute.provider_tree [None req-960d1abd-1b06-4b5a-8ddd-4b463d493a45 tempest-ListServerFiltersTestJSON-241228521 tempest-ListServerFiltersTestJSON-241228521-project-member] Inventory has not changed in ProviderTree for provider: a2858be1-fd22-4e08-979e-87ad25293407 {{(pid=61806) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 570.323705] env[61806]: DEBUG oslo_concurrency.lockutils [req-3b678f53-eeb4-46b1-9fc1-8b6e9352dd63 req-9ad69bf5-731c-4af0-ae79-f7399f8f1658 service nova] Releasing lock "refresh_cache-123f15be-3b56-4999-8010-40d2e86a6813" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 570.324214] env[61806]: DEBUG oslo_concurrency.lockutils [None req-be9179b6-8d17-4f04-8823-d48174fcc5c1 tempest-InstanceActionsV221TestJSON-379107765 tempest-InstanceActionsV221TestJSON-379107765-project-member] Acquired lock "refresh_cache-123f15be-3b56-4999-8010-40d2e86a6813" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 570.324483] env[61806]: DEBUG nova.network.neutron [None req-be9179b6-8d17-4f04-8823-d48174fcc5c1 tempest-InstanceActionsV221TestJSON-379107765 tempest-InstanceActionsV221TestJSON-379107765-project-member] [instance: 123f15be-3b56-4999-8010-40d2e86a6813] Building network info cache for instance {{(pid=61806) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 570.388741] env[61806]: DEBUG oslo_concurrency.lockutils [None req-d8e4a121-0976-4eaa-b4c6-51deda86d081 tempest-InstanceActionsNegativeTestJSON-503825571 tempest-InstanceActionsNegativeTestJSON-503825571-project-member] Releasing lock "refresh_cache-a1aca638-e9c4-4e3b-9855-f92e7bd67204" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 570.388989] env[61806]: DEBUG nova.compute.manager [None req-d8e4a121-0976-4eaa-b4c6-51deda86d081 tempest-InstanceActionsNegativeTestJSON-503825571 tempest-InstanceActionsNegativeTestJSON-503825571-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61806) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 570.389193] env[61806]: DEBUG nova.compute.manager [None req-d8e4a121-0976-4eaa-b4c6-51deda86d081 tempest-InstanceActionsNegativeTestJSON-503825571 tempest-InstanceActionsNegativeTestJSON-503825571-project-member] [instance: a1aca638-e9c4-4e3b-9855-f92e7bd67204] Deallocating network for instance {{(pid=61806) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 570.389384] env[61806]: DEBUG nova.network.neutron [None req-d8e4a121-0976-4eaa-b4c6-51deda86d081 tempest-InstanceActionsNegativeTestJSON-503825571 tempest-InstanceActionsNegativeTestJSON-503825571-project-member] [instance: a1aca638-e9c4-4e3b-9855-f92e7bd67204] deallocate_for_instance() {{(pid=61806) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 570.416904] env[61806]: DEBUG nova.network.neutron [None req-d8e4a121-0976-4eaa-b4c6-51deda86d081 tempest-InstanceActionsNegativeTestJSON-503825571 tempest-InstanceActionsNegativeTestJSON-503825571-project-member] [instance: a1aca638-e9c4-4e3b-9855-f92e7bd67204] Instance cache missing network info. {{(pid=61806) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 570.747811] env[61806]: DEBUG nova.scheduler.client.report [None req-960d1abd-1b06-4b5a-8ddd-4b463d493a45 tempest-ListServerFiltersTestJSON-241228521 tempest-ListServerFiltersTestJSON-241228521-project-member] Inventory has not changed for provider a2858be1-fd22-4e08-979e-87ad25293407 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 139, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61806) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 570.865896] env[61806]: DEBUG nova.network.neutron [None req-be9179b6-8d17-4f04-8823-d48174fcc5c1 tempest-InstanceActionsV221TestJSON-379107765 tempest-InstanceActionsV221TestJSON-379107765-project-member] [instance: 123f15be-3b56-4999-8010-40d2e86a6813] Instance cache missing network info. {{(pid=61806) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 570.919608] env[61806]: DEBUG nova.network.neutron [None req-d8e4a121-0976-4eaa-b4c6-51deda86d081 tempest-InstanceActionsNegativeTestJSON-503825571 tempest-InstanceActionsNegativeTestJSON-503825571-project-member] [instance: a1aca638-e9c4-4e3b-9855-f92e7bd67204] Updating instance_info_cache with network_info: [] {{(pid=61806) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 570.973590] env[61806]: DEBUG nova.network.neutron [None req-be9179b6-8d17-4f04-8823-d48174fcc5c1 tempest-InstanceActionsV221TestJSON-379107765 tempest-InstanceActionsV221TestJSON-379107765-project-member] [instance: 123f15be-3b56-4999-8010-40d2e86a6813] Updating instance_info_cache with network_info: [] {{(pid=61806) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 571.111454] env[61806]: DEBUG nova.compute.manager [req-a0cb2c84-3c71-4b46-b23d-532608f63e3e req-03ab5d06-9191-4795-8461-0986322a7b9e service nova] [instance: 123f15be-3b56-4999-8010-40d2e86a6813] Received event network-vif-deleted-8fbea74a-6823-4774-b3af-5bda4ccd8820 {{(pid=61806) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 571.257619] env[61806]: DEBUG oslo_concurrency.lockutils [None req-960d1abd-1b06-4b5a-8ddd-4b463d493a45 tempest-ListServerFiltersTestJSON-241228521 tempest-ListServerFiltersTestJSON-241228521-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.155s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 571.257619] env[61806]: ERROR nova.compute.manager [None req-960d1abd-1b06-4b5a-8ddd-4b463d493a45 tempest-ListServerFiltersTestJSON-241228521 tempest-ListServerFiltersTestJSON-241228521-project-member] [instance: 30e44576-dcd3-4036-9070-9032b775c90a] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 5990bc22-0a46-437d-b518-9a9263b60655, please check neutron logs for more information. [ 571.257619] env[61806]: ERROR nova.compute.manager [instance: 30e44576-dcd3-4036-9070-9032b775c90a] Traceback (most recent call last): [ 571.257619] env[61806]: ERROR nova.compute.manager [instance: 30e44576-dcd3-4036-9070-9032b775c90a] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 571.257619] env[61806]: ERROR nova.compute.manager [instance: 30e44576-dcd3-4036-9070-9032b775c90a] self.driver.spawn(context, instance, image_meta, [ 571.257619] env[61806]: ERROR nova.compute.manager [instance: 30e44576-dcd3-4036-9070-9032b775c90a] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 571.257619] env[61806]: ERROR nova.compute.manager [instance: 30e44576-dcd3-4036-9070-9032b775c90a] self._vmops.spawn(context, instance, image_meta, injected_files, [ 571.257619] env[61806]: ERROR nova.compute.manager [instance: 30e44576-dcd3-4036-9070-9032b775c90a] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 571.257619] env[61806]: ERROR nova.compute.manager [instance: 30e44576-dcd3-4036-9070-9032b775c90a] vm_ref = self.build_virtual_machine(instance, [ 571.258505] env[61806]: ERROR nova.compute.manager [instance: 30e44576-dcd3-4036-9070-9032b775c90a] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 571.258505] env[61806]: ERROR nova.compute.manager [instance: 30e44576-dcd3-4036-9070-9032b775c90a] vif_infos = vmwarevif.get_vif_info(self._session, [ 571.258505] env[61806]: ERROR nova.compute.manager [instance: 30e44576-dcd3-4036-9070-9032b775c90a] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 571.258505] env[61806]: ERROR nova.compute.manager [instance: 30e44576-dcd3-4036-9070-9032b775c90a] for vif in network_info: [ 571.258505] env[61806]: ERROR nova.compute.manager [instance: 30e44576-dcd3-4036-9070-9032b775c90a] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 571.258505] env[61806]: ERROR nova.compute.manager [instance: 30e44576-dcd3-4036-9070-9032b775c90a] return self._sync_wrapper(fn, *args, **kwargs) [ 571.258505] env[61806]: ERROR nova.compute.manager [instance: 30e44576-dcd3-4036-9070-9032b775c90a] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 571.258505] env[61806]: ERROR nova.compute.manager [instance: 30e44576-dcd3-4036-9070-9032b775c90a] self.wait() [ 571.258505] env[61806]: ERROR nova.compute.manager [instance: 30e44576-dcd3-4036-9070-9032b775c90a] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 571.258505] env[61806]: ERROR nova.compute.manager [instance: 30e44576-dcd3-4036-9070-9032b775c90a] self[:] = self._gt.wait() [ 571.258505] env[61806]: ERROR nova.compute.manager [instance: 30e44576-dcd3-4036-9070-9032b775c90a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 571.258505] env[61806]: ERROR nova.compute.manager [instance: 30e44576-dcd3-4036-9070-9032b775c90a] return self._exit_event.wait() [ 571.258505] env[61806]: ERROR nova.compute.manager [instance: 30e44576-dcd3-4036-9070-9032b775c90a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 571.258969] env[61806]: ERROR nova.compute.manager [instance: 30e44576-dcd3-4036-9070-9032b775c90a] result = hub.switch() [ 571.258969] env[61806]: ERROR nova.compute.manager [instance: 30e44576-dcd3-4036-9070-9032b775c90a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 571.258969] env[61806]: ERROR nova.compute.manager [instance: 30e44576-dcd3-4036-9070-9032b775c90a] return self.greenlet.switch() [ 571.258969] env[61806]: ERROR nova.compute.manager [instance: 30e44576-dcd3-4036-9070-9032b775c90a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 571.258969] env[61806]: ERROR nova.compute.manager [instance: 30e44576-dcd3-4036-9070-9032b775c90a] result = function(*args, **kwargs) [ 571.258969] env[61806]: ERROR nova.compute.manager [instance: 30e44576-dcd3-4036-9070-9032b775c90a] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 571.258969] env[61806]: ERROR nova.compute.manager [instance: 30e44576-dcd3-4036-9070-9032b775c90a] return func(*args, **kwargs) [ 571.258969] env[61806]: ERROR nova.compute.manager [instance: 30e44576-dcd3-4036-9070-9032b775c90a] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 571.258969] env[61806]: ERROR nova.compute.manager [instance: 30e44576-dcd3-4036-9070-9032b775c90a] raise e [ 571.258969] env[61806]: ERROR nova.compute.manager [instance: 30e44576-dcd3-4036-9070-9032b775c90a] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 571.258969] env[61806]: ERROR nova.compute.manager [instance: 30e44576-dcd3-4036-9070-9032b775c90a] nwinfo = self.network_api.allocate_for_instance( [ 571.258969] env[61806]: ERROR nova.compute.manager [instance: 30e44576-dcd3-4036-9070-9032b775c90a] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 571.258969] env[61806]: ERROR nova.compute.manager [instance: 30e44576-dcd3-4036-9070-9032b775c90a] created_port_ids = self._update_ports_for_instance( [ 571.259411] env[61806]: ERROR nova.compute.manager [instance: 30e44576-dcd3-4036-9070-9032b775c90a] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 571.259411] env[61806]: ERROR nova.compute.manager [instance: 30e44576-dcd3-4036-9070-9032b775c90a] with excutils.save_and_reraise_exception(): [ 571.259411] env[61806]: ERROR nova.compute.manager [instance: 30e44576-dcd3-4036-9070-9032b775c90a] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 571.259411] env[61806]: ERROR nova.compute.manager [instance: 30e44576-dcd3-4036-9070-9032b775c90a] self.force_reraise() [ 571.259411] env[61806]: ERROR nova.compute.manager [instance: 30e44576-dcd3-4036-9070-9032b775c90a] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 571.259411] env[61806]: ERROR nova.compute.manager [instance: 30e44576-dcd3-4036-9070-9032b775c90a] raise self.value [ 571.259411] env[61806]: ERROR nova.compute.manager [instance: 30e44576-dcd3-4036-9070-9032b775c90a] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 571.259411] env[61806]: ERROR nova.compute.manager [instance: 30e44576-dcd3-4036-9070-9032b775c90a] updated_port = self._update_port( [ 571.259411] env[61806]: ERROR nova.compute.manager [instance: 30e44576-dcd3-4036-9070-9032b775c90a] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 571.259411] env[61806]: ERROR nova.compute.manager [instance: 30e44576-dcd3-4036-9070-9032b775c90a] _ensure_no_port_binding_failure(port) [ 571.259411] env[61806]: ERROR nova.compute.manager [instance: 30e44576-dcd3-4036-9070-9032b775c90a] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 571.259411] env[61806]: ERROR nova.compute.manager [instance: 30e44576-dcd3-4036-9070-9032b775c90a] raise exception.PortBindingFailed(port_id=port['id']) [ 571.259916] env[61806]: ERROR nova.compute.manager [instance: 30e44576-dcd3-4036-9070-9032b775c90a] nova.exception.PortBindingFailed: Binding failed for port 5990bc22-0a46-437d-b518-9a9263b60655, please check neutron logs for more information. [ 571.259916] env[61806]: ERROR nova.compute.manager [instance: 30e44576-dcd3-4036-9070-9032b775c90a] [ 571.259916] env[61806]: DEBUG nova.compute.utils [None req-960d1abd-1b06-4b5a-8ddd-4b463d493a45 tempest-ListServerFiltersTestJSON-241228521 tempest-ListServerFiltersTestJSON-241228521-project-member] [instance: 30e44576-dcd3-4036-9070-9032b775c90a] Binding failed for port 5990bc22-0a46-437d-b518-9a9263b60655, please check neutron logs for more information. {{(pid=61806) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 571.259916] env[61806]: DEBUG oslo_concurrency.lockutils [None req-7d4ae408-54ff-4248-838b-a36cbf45b0c5 tempest-ListServerFiltersTestJSON-241228521 tempest-ListServerFiltersTestJSON-241228521-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 22.324s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 571.266507] env[61806]: DEBUG nova.compute.manager [None req-960d1abd-1b06-4b5a-8ddd-4b463d493a45 tempest-ListServerFiltersTestJSON-241228521 tempest-ListServerFiltersTestJSON-241228521-project-member] [instance: 30e44576-dcd3-4036-9070-9032b775c90a] Build of instance 30e44576-dcd3-4036-9070-9032b775c90a was re-scheduled: Binding failed for port 5990bc22-0a46-437d-b518-9a9263b60655, please check neutron logs for more information. {{(pid=61806) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 571.266978] env[61806]: DEBUG nova.compute.manager [None req-960d1abd-1b06-4b5a-8ddd-4b463d493a45 tempest-ListServerFiltersTestJSON-241228521 tempest-ListServerFiltersTestJSON-241228521-project-member] [instance: 30e44576-dcd3-4036-9070-9032b775c90a] Unplugging VIFs for instance {{(pid=61806) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 571.267596] env[61806]: DEBUG oslo_concurrency.lockutils [None req-960d1abd-1b06-4b5a-8ddd-4b463d493a45 tempest-ListServerFiltersTestJSON-241228521 tempest-ListServerFiltersTestJSON-241228521-project-member] Acquiring lock "refresh_cache-30e44576-dcd3-4036-9070-9032b775c90a" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 571.267596] env[61806]: DEBUG oslo_concurrency.lockutils [None req-960d1abd-1b06-4b5a-8ddd-4b463d493a45 tempest-ListServerFiltersTestJSON-241228521 tempest-ListServerFiltersTestJSON-241228521-project-member] Acquired lock "refresh_cache-30e44576-dcd3-4036-9070-9032b775c90a" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 571.267596] env[61806]: DEBUG nova.network.neutron [None req-960d1abd-1b06-4b5a-8ddd-4b463d493a45 tempest-ListServerFiltersTestJSON-241228521 tempest-ListServerFiltersTestJSON-241228521-project-member] [instance: 30e44576-dcd3-4036-9070-9032b775c90a] Building network info cache for instance {{(pid=61806) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 571.409957] env[61806]: DEBUG oslo_concurrency.lockutils [None req-2bf81f39-4b8d-4851-a793-760af008ffe2 tempest-ImagesOneServerTestJSON-1003762139 tempest-ImagesOneServerTestJSON-1003762139-project-member] Acquiring lock "294f0a37-7589-405d-a324-13099c80f458" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 571.410307] env[61806]: DEBUG oslo_concurrency.lockutils [None req-2bf81f39-4b8d-4851-a793-760af008ffe2 tempest-ImagesOneServerTestJSON-1003762139 tempest-ImagesOneServerTestJSON-1003762139-project-member] Lock "294f0a37-7589-405d-a324-13099c80f458" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.002s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 571.423779] env[61806]: INFO nova.compute.manager [None req-d8e4a121-0976-4eaa-b4c6-51deda86d081 tempest-InstanceActionsNegativeTestJSON-503825571 tempest-InstanceActionsNegativeTestJSON-503825571-project-member] [instance: a1aca638-e9c4-4e3b-9855-f92e7bd67204] Took 1.03 seconds to deallocate network for instance. [ 571.477774] env[61806]: DEBUG oslo_concurrency.lockutils [None req-be9179b6-8d17-4f04-8823-d48174fcc5c1 tempest-InstanceActionsV221TestJSON-379107765 tempest-InstanceActionsV221TestJSON-379107765-project-member] Releasing lock "refresh_cache-123f15be-3b56-4999-8010-40d2e86a6813" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 571.478106] env[61806]: DEBUG nova.compute.manager [None req-be9179b6-8d17-4f04-8823-d48174fcc5c1 tempest-InstanceActionsV221TestJSON-379107765 tempest-InstanceActionsV221TestJSON-379107765-project-member] [instance: 123f15be-3b56-4999-8010-40d2e86a6813] Start destroying the instance on the hypervisor. {{(pid=61806) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 571.478215] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-be9179b6-8d17-4f04-8823-d48174fcc5c1 tempest-InstanceActionsV221TestJSON-379107765 tempest-InstanceActionsV221TestJSON-379107765-project-member] [instance: 123f15be-3b56-4999-8010-40d2e86a6813] Destroying instance {{(pid=61806) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 571.478861] env[61806]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-1d64611e-8b98-4b3c-879c-697664487323 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 571.491104] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d2a6d8ad-d7ff-41a5-830f-bef2ce7332d1 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 571.524842] env[61806]: WARNING nova.virt.vmwareapi.vmops [None req-be9179b6-8d17-4f04-8823-d48174fcc5c1 tempest-InstanceActionsV221TestJSON-379107765 tempest-InstanceActionsV221TestJSON-379107765-project-member] [instance: 123f15be-3b56-4999-8010-40d2e86a6813] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 123f15be-3b56-4999-8010-40d2e86a6813 could not be found. [ 571.525067] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-be9179b6-8d17-4f04-8823-d48174fcc5c1 tempest-InstanceActionsV221TestJSON-379107765 tempest-InstanceActionsV221TestJSON-379107765-project-member] [instance: 123f15be-3b56-4999-8010-40d2e86a6813] Instance destroyed {{(pid=61806) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 571.525263] env[61806]: INFO nova.compute.manager [None req-be9179b6-8d17-4f04-8823-d48174fcc5c1 tempest-InstanceActionsV221TestJSON-379107765 tempest-InstanceActionsV221TestJSON-379107765-project-member] [instance: 123f15be-3b56-4999-8010-40d2e86a6813] Took 0.05 seconds to destroy the instance on the hypervisor. [ 571.525512] env[61806]: DEBUG oslo.service.loopingcall [None req-be9179b6-8d17-4f04-8823-d48174fcc5c1 tempest-InstanceActionsV221TestJSON-379107765 tempest-InstanceActionsV221TestJSON-379107765-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61806) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 571.525818] env[61806]: DEBUG nova.compute.manager [-] [instance: 123f15be-3b56-4999-8010-40d2e86a6813] Deallocating network for instance {{(pid=61806) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 571.528042] env[61806]: DEBUG nova.network.neutron [-] [instance: 123f15be-3b56-4999-8010-40d2e86a6813] deallocate_for_instance() {{(pid=61806) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 571.553211] env[61806]: DEBUG nova.network.neutron [-] [instance: 123f15be-3b56-4999-8010-40d2e86a6813] Instance cache missing network info. {{(pid=61806) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 571.806087] env[61806]: DEBUG nova.network.neutron [None req-960d1abd-1b06-4b5a-8ddd-4b463d493a45 tempest-ListServerFiltersTestJSON-241228521 tempest-ListServerFiltersTestJSON-241228521-project-member] [instance: 30e44576-dcd3-4036-9070-9032b775c90a] Instance cache missing network info. {{(pid=61806) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 572.055986] env[61806]: DEBUG nova.network.neutron [-] [instance: 123f15be-3b56-4999-8010-40d2e86a6813] Updating instance_info_cache with network_info: [] {{(pid=61806) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 572.118571] env[61806]: DEBUG nova.network.neutron [None req-960d1abd-1b06-4b5a-8ddd-4b463d493a45 tempest-ListServerFiltersTestJSON-241228521 tempest-ListServerFiltersTestJSON-241228521-project-member] [instance: 30e44576-dcd3-4036-9070-9032b775c90a] Updating instance_info_cache with network_info: [] {{(pid=61806) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 572.221211] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cb8bc5de-e021-4494-8a32-facb8c347abf {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 572.228726] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c95612f4-60f6-4297-b2b1-ef33314c638d {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 572.268474] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1af16162-2f87-4669-8bcd-87c832b77f23 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 572.277063] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f30552c7-a44d-4217-8169-f4d8dcc1da10 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 572.294910] env[61806]: DEBUG nova.compute.provider_tree [None req-7d4ae408-54ff-4248-838b-a36cbf45b0c5 tempest-ListServerFiltersTestJSON-241228521 tempest-ListServerFiltersTestJSON-241228521-project-member] Inventory has not changed in ProviderTree for provider: a2858be1-fd22-4e08-979e-87ad25293407 {{(pid=61806) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 572.305262] env[61806]: DEBUG oslo_concurrency.lockutils [None req-0f0a69c6-2310-4272-abc1-520940b4bcf0 tempest-ServersTestFqdnHostnames-356624376 tempest-ServersTestFqdnHostnames-356624376-project-member] Acquiring lock "46c34ed0-9bac-4803-8d5c-3a3346a64117" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 572.305262] env[61806]: DEBUG oslo_concurrency.lockutils [None req-0f0a69c6-2310-4272-abc1-520940b4bcf0 tempest-ServersTestFqdnHostnames-356624376 tempest-ServersTestFqdnHostnames-356624376-project-member] Lock "46c34ed0-9bac-4803-8d5c-3a3346a64117" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 572.459907] env[61806]: INFO nova.scheduler.client.report [None req-d8e4a121-0976-4eaa-b4c6-51deda86d081 tempest-InstanceActionsNegativeTestJSON-503825571 tempest-InstanceActionsNegativeTestJSON-503825571-project-member] Deleted allocations for instance a1aca638-e9c4-4e3b-9855-f92e7bd67204 [ 572.558830] env[61806]: INFO nova.compute.manager [-] [instance: 123f15be-3b56-4999-8010-40d2e86a6813] Took 1.03 seconds to deallocate network for instance. [ 572.562407] env[61806]: DEBUG nova.compute.claims [None req-be9179b6-8d17-4f04-8823-d48174fcc5c1 tempest-InstanceActionsV221TestJSON-379107765 tempest-InstanceActionsV221TestJSON-379107765-project-member] [instance: 123f15be-3b56-4999-8010-40d2e86a6813] Aborting claim: {{(pid=61806) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 572.562578] env[61806]: DEBUG oslo_concurrency.lockutils [None req-be9179b6-8d17-4f04-8823-d48174fcc5c1 tempest-InstanceActionsV221TestJSON-379107765 tempest-InstanceActionsV221TestJSON-379107765-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 572.621347] env[61806]: DEBUG oslo_concurrency.lockutils [None req-960d1abd-1b06-4b5a-8ddd-4b463d493a45 tempest-ListServerFiltersTestJSON-241228521 tempest-ListServerFiltersTestJSON-241228521-project-member] Releasing lock "refresh_cache-30e44576-dcd3-4036-9070-9032b775c90a" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 572.621682] env[61806]: DEBUG nova.compute.manager [None req-960d1abd-1b06-4b5a-8ddd-4b463d493a45 tempest-ListServerFiltersTestJSON-241228521 tempest-ListServerFiltersTestJSON-241228521-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61806) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 572.621909] env[61806]: DEBUG nova.compute.manager [None req-960d1abd-1b06-4b5a-8ddd-4b463d493a45 tempest-ListServerFiltersTestJSON-241228521 tempest-ListServerFiltersTestJSON-241228521-project-member] [instance: 30e44576-dcd3-4036-9070-9032b775c90a] Deallocating network for instance {{(pid=61806) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 572.622105] env[61806]: DEBUG nova.network.neutron [None req-960d1abd-1b06-4b5a-8ddd-4b463d493a45 tempest-ListServerFiltersTestJSON-241228521 tempest-ListServerFiltersTestJSON-241228521-project-member] [instance: 30e44576-dcd3-4036-9070-9032b775c90a] deallocate_for_instance() {{(pid=61806) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 572.639485] env[61806]: DEBUG nova.network.neutron [None req-960d1abd-1b06-4b5a-8ddd-4b463d493a45 tempest-ListServerFiltersTestJSON-241228521 tempest-ListServerFiltersTestJSON-241228521-project-member] [instance: 30e44576-dcd3-4036-9070-9032b775c90a] Instance cache missing network info. {{(pid=61806) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 572.799849] env[61806]: DEBUG nova.scheduler.client.report [None req-7d4ae408-54ff-4248-838b-a36cbf45b0c5 tempest-ListServerFiltersTestJSON-241228521 tempest-ListServerFiltersTestJSON-241228521-project-member] Inventory has not changed for provider a2858be1-fd22-4e08-979e-87ad25293407 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 139, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61806) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 572.968567] env[61806]: DEBUG oslo_concurrency.lockutils [None req-d8e4a121-0976-4eaa-b4c6-51deda86d081 tempest-InstanceActionsNegativeTestJSON-503825571 tempest-InstanceActionsNegativeTestJSON-503825571-project-member] Lock "a1aca638-e9c4-4e3b-9855-f92e7bd67204" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 47.943s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 573.142680] env[61806]: DEBUG nova.network.neutron [None req-960d1abd-1b06-4b5a-8ddd-4b463d493a45 tempest-ListServerFiltersTestJSON-241228521 tempest-ListServerFiltersTestJSON-241228521-project-member] [instance: 30e44576-dcd3-4036-9070-9032b775c90a] Updating instance_info_cache with network_info: [] {{(pid=61806) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 573.314046] env[61806]: DEBUG oslo_concurrency.lockutils [None req-7d4ae408-54ff-4248-838b-a36cbf45b0c5 tempest-ListServerFiltersTestJSON-241228521 tempest-ListServerFiltersTestJSON-241228521-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.055s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 573.314671] env[61806]: ERROR nova.compute.manager [None req-7d4ae408-54ff-4248-838b-a36cbf45b0c5 tempest-ListServerFiltersTestJSON-241228521 tempest-ListServerFiltersTestJSON-241228521-project-member] [instance: 9ddec22f-0fce-4d28-b97e-92b179f829f6] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port c8f9a917-21ab-46bb-bbd3-1717da5a257b, please check neutron logs for more information. [ 573.314671] env[61806]: ERROR nova.compute.manager [instance: 9ddec22f-0fce-4d28-b97e-92b179f829f6] Traceback (most recent call last): [ 573.314671] env[61806]: ERROR nova.compute.manager [instance: 9ddec22f-0fce-4d28-b97e-92b179f829f6] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 573.314671] env[61806]: ERROR nova.compute.manager [instance: 9ddec22f-0fce-4d28-b97e-92b179f829f6] self.driver.spawn(context, instance, image_meta, [ 573.314671] env[61806]: ERROR nova.compute.manager [instance: 9ddec22f-0fce-4d28-b97e-92b179f829f6] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 573.314671] env[61806]: ERROR nova.compute.manager [instance: 9ddec22f-0fce-4d28-b97e-92b179f829f6] self._vmops.spawn(context, instance, image_meta, injected_files, [ 573.314671] env[61806]: ERROR nova.compute.manager [instance: 9ddec22f-0fce-4d28-b97e-92b179f829f6] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 573.314671] env[61806]: ERROR nova.compute.manager [instance: 9ddec22f-0fce-4d28-b97e-92b179f829f6] vm_ref = self.build_virtual_machine(instance, [ 573.314671] env[61806]: ERROR nova.compute.manager [instance: 9ddec22f-0fce-4d28-b97e-92b179f829f6] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 573.314671] env[61806]: ERROR nova.compute.manager [instance: 9ddec22f-0fce-4d28-b97e-92b179f829f6] vif_infos = vmwarevif.get_vif_info(self._session, [ 573.314671] env[61806]: ERROR nova.compute.manager [instance: 9ddec22f-0fce-4d28-b97e-92b179f829f6] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 573.316901] env[61806]: ERROR nova.compute.manager [instance: 9ddec22f-0fce-4d28-b97e-92b179f829f6] for vif in network_info: [ 573.316901] env[61806]: ERROR nova.compute.manager [instance: 9ddec22f-0fce-4d28-b97e-92b179f829f6] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 573.316901] env[61806]: ERROR nova.compute.manager [instance: 9ddec22f-0fce-4d28-b97e-92b179f829f6] return self._sync_wrapper(fn, *args, **kwargs) [ 573.316901] env[61806]: ERROR nova.compute.manager [instance: 9ddec22f-0fce-4d28-b97e-92b179f829f6] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 573.316901] env[61806]: ERROR nova.compute.manager [instance: 9ddec22f-0fce-4d28-b97e-92b179f829f6] self.wait() [ 573.316901] env[61806]: ERROR nova.compute.manager [instance: 9ddec22f-0fce-4d28-b97e-92b179f829f6] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 573.316901] env[61806]: ERROR nova.compute.manager [instance: 9ddec22f-0fce-4d28-b97e-92b179f829f6] self[:] = self._gt.wait() [ 573.316901] env[61806]: ERROR nova.compute.manager [instance: 9ddec22f-0fce-4d28-b97e-92b179f829f6] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 573.316901] env[61806]: ERROR nova.compute.manager [instance: 9ddec22f-0fce-4d28-b97e-92b179f829f6] return self._exit_event.wait() [ 573.316901] env[61806]: ERROR nova.compute.manager [instance: 9ddec22f-0fce-4d28-b97e-92b179f829f6] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 573.316901] env[61806]: ERROR nova.compute.manager [instance: 9ddec22f-0fce-4d28-b97e-92b179f829f6] result = hub.switch() [ 573.316901] env[61806]: ERROR nova.compute.manager [instance: 9ddec22f-0fce-4d28-b97e-92b179f829f6] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 573.316901] env[61806]: ERROR nova.compute.manager [instance: 9ddec22f-0fce-4d28-b97e-92b179f829f6] return self.greenlet.switch() [ 573.318084] env[61806]: ERROR nova.compute.manager [instance: 9ddec22f-0fce-4d28-b97e-92b179f829f6] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 573.318084] env[61806]: ERROR nova.compute.manager [instance: 9ddec22f-0fce-4d28-b97e-92b179f829f6] result = function(*args, **kwargs) [ 573.318084] env[61806]: ERROR nova.compute.manager [instance: 9ddec22f-0fce-4d28-b97e-92b179f829f6] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 573.318084] env[61806]: ERROR nova.compute.manager [instance: 9ddec22f-0fce-4d28-b97e-92b179f829f6] return func(*args, **kwargs) [ 573.318084] env[61806]: ERROR nova.compute.manager [instance: 9ddec22f-0fce-4d28-b97e-92b179f829f6] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 573.318084] env[61806]: ERROR nova.compute.manager [instance: 9ddec22f-0fce-4d28-b97e-92b179f829f6] raise e [ 573.318084] env[61806]: ERROR nova.compute.manager [instance: 9ddec22f-0fce-4d28-b97e-92b179f829f6] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 573.318084] env[61806]: ERROR nova.compute.manager [instance: 9ddec22f-0fce-4d28-b97e-92b179f829f6] nwinfo = self.network_api.allocate_for_instance( [ 573.318084] env[61806]: ERROR nova.compute.manager [instance: 9ddec22f-0fce-4d28-b97e-92b179f829f6] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 573.318084] env[61806]: ERROR nova.compute.manager [instance: 9ddec22f-0fce-4d28-b97e-92b179f829f6] created_port_ids = self._update_ports_for_instance( [ 573.318084] env[61806]: ERROR nova.compute.manager [instance: 9ddec22f-0fce-4d28-b97e-92b179f829f6] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 573.318084] env[61806]: ERROR nova.compute.manager [instance: 9ddec22f-0fce-4d28-b97e-92b179f829f6] with excutils.save_and_reraise_exception(): [ 573.318084] env[61806]: ERROR nova.compute.manager [instance: 9ddec22f-0fce-4d28-b97e-92b179f829f6] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 573.318662] env[61806]: ERROR nova.compute.manager [instance: 9ddec22f-0fce-4d28-b97e-92b179f829f6] self.force_reraise() [ 573.318662] env[61806]: ERROR nova.compute.manager [instance: 9ddec22f-0fce-4d28-b97e-92b179f829f6] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 573.318662] env[61806]: ERROR nova.compute.manager [instance: 9ddec22f-0fce-4d28-b97e-92b179f829f6] raise self.value [ 573.318662] env[61806]: ERROR nova.compute.manager [instance: 9ddec22f-0fce-4d28-b97e-92b179f829f6] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 573.318662] env[61806]: ERROR nova.compute.manager [instance: 9ddec22f-0fce-4d28-b97e-92b179f829f6] updated_port = self._update_port( [ 573.318662] env[61806]: ERROR nova.compute.manager [instance: 9ddec22f-0fce-4d28-b97e-92b179f829f6] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 573.318662] env[61806]: ERROR nova.compute.manager [instance: 9ddec22f-0fce-4d28-b97e-92b179f829f6] _ensure_no_port_binding_failure(port) [ 573.318662] env[61806]: ERROR nova.compute.manager [instance: 9ddec22f-0fce-4d28-b97e-92b179f829f6] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 573.318662] env[61806]: ERROR nova.compute.manager [instance: 9ddec22f-0fce-4d28-b97e-92b179f829f6] raise exception.PortBindingFailed(port_id=port['id']) [ 573.318662] env[61806]: ERROR nova.compute.manager [instance: 9ddec22f-0fce-4d28-b97e-92b179f829f6] nova.exception.PortBindingFailed: Binding failed for port c8f9a917-21ab-46bb-bbd3-1717da5a257b, please check neutron logs for more information. [ 573.318662] env[61806]: ERROR nova.compute.manager [instance: 9ddec22f-0fce-4d28-b97e-92b179f829f6] [ 573.319850] env[61806]: DEBUG nova.compute.utils [None req-7d4ae408-54ff-4248-838b-a36cbf45b0c5 tempest-ListServerFiltersTestJSON-241228521 tempest-ListServerFiltersTestJSON-241228521-project-member] [instance: 9ddec22f-0fce-4d28-b97e-92b179f829f6] Binding failed for port c8f9a917-21ab-46bb-bbd3-1717da5a257b, please check neutron logs for more information. {{(pid=61806) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 573.319850] env[61806]: DEBUG nova.compute.manager [None req-7d4ae408-54ff-4248-838b-a36cbf45b0c5 tempest-ListServerFiltersTestJSON-241228521 tempest-ListServerFiltersTestJSON-241228521-project-member] [instance: 9ddec22f-0fce-4d28-b97e-92b179f829f6] Build of instance 9ddec22f-0fce-4d28-b97e-92b179f829f6 was re-scheduled: Binding failed for port c8f9a917-21ab-46bb-bbd3-1717da5a257b, please check neutron logs for more information. {{(pid=61806) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 573.319850] env[61806]: DEBUG nova.compute.manager [None req-7d4ae408-54ff-4248-838b-a36cbf45b0c5 tempest-ListServerFiltersTestJSON-241228521 tempest-ListServerFiltersTestJSON-241228521-project-member] [instance: 9ddec22f-0fce-4d28-b97e-92b179f829f6] Unplugging VIFs for instance {{(pid=61806) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 573.319850] env[61806]: DEBUG oslo_concurrency.lockutils [None req-7d4ae408-54ff-4248-838b-a36cbf45b0c5 tempest-ListServerFiltersTestJSON-241228521 tempest-ListServerFiltersTestJSON-241228521-project-member] Acquiring lock "refresh_cache-9ddec22f-0fce-4d28-b97e-92b179f829f6" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 573.320061] env[61806]: DEBUG oslo_concurrency.lockutils [None req-7d4ae408-54ff-4248-838b-a36cbf45b0c5 tempest-ListServerFiltersTestJSON-241228521 tempest-ListServerFiltersTestJSON-241228521-project-member] Acquired lock "refresh_cache-9ddec22f-0fce-4d28-b97e-92b179f829f6" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 573.320061] env[61806]: DEBUG nova.network.neutron [None req-7d4ae408-54ff-4248-838b-a36cbf45b0c5 tempest-ListServerFiltersTestJSON-241228521 tempest-ListServerFiltersTestJSON-241228521-project-member] [instance: 9ddec22f-0fce-4d28-b97e-92b179f829f6] Building network info cache for instance {{(pid=61806) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 573.320061] env[61806]: DEBUG oslo_concurrency.lockutils [None req-38c4eaec-22d8-4bf0-bd45-321ef43329a0 tempest-ServersAdminTestJSON-14199920 tempest-ServersAdminTestJSON-14199920-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 23.999s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 573.321541] env[61806]: INFO nova.compute.claims [None req-38c4eaec-22d8-4bf0-bd45-321ef43329a0 tempest-ServersAdminTestJSON-14199920 tempest-ServersAdminTestJSON-14199920-project-member] [instance: 8905f20f-28a6-49f4-88bf-e52177c6a4b6] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 573.471190] env[61806]: DEBUG nova.compute.manager [None req-9101cb9e-b191-4241-8bc1-9b126210ca50 tempest-ListImageFiltersTestJSON-680116525 tempest-ListImageFiltersTestJSON-680116525-project-member] [instance: 7e730579-8d0c-4226-86bc-292db16c49ee] Starting instance... {{(pid=61806) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 573.646156] env[61806]: INFO nova.compute.manager [None req-960d1abd-1b06-4b5a-8ddd-4b463d493a45 tempest-ListServerFiltersTestJSON-241228521 tempest-ListServerFiltersTestJSON-241228521-project-member] [instance: 30e44576-dcd3-4036-9070-9032b775c90a] Took 1.02 seconds to deallocate network for instance. [ 573.847639] env[61806]: DEBUG nova.network.neutron [None req-7d4ae408-54ff-4248-838b-a36cbf45b0c5 tempest-ListServerFiltersTestJSON-241228521 tempest-ListServerFiltersTestJSON-241228521-project-member] [instance: 9ddec22f-0fce-4d28-b97e-92b179f829f6] Instance cache missing network info. {{(pid=61806) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 573.993303] env[61806]: DEBUG oslo_concurrency.lockutils [None req-9101cb9e-b191-4241-8bc1-9b126210ca50 tempest-ListImageFiltersTestJSON-680116525 tempest-ListImageFiltersTestJSON-680116525-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 574.002913] env[61806]: DEBUG nova.network.neutron [None req-7d4ae408-54ff-4248-838b-a36cbf45b0c5 tempest-ListServerFiltersTestJSON-241228521 tempest-ListServerFiltersTestJSON-241228521-project-member] [instance: 9ddec22f-0fce-4d28-b97e-92b179f829f6] Updating instance_info_cache with network_info: [] {{(pid=61806) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 574.506856] env[61806]: DEBUG oslo_concurrency.lockutils [None req-7d4ae408-54ff-4248-838b-a36cbf45b0c5 tempest-ListServerFiltersTestJSON-241228521 tempest-ListServerFiltersTestJSON-241228521-project-member] Releasing lock "refresh_cache-9ddec22f-0fce-4d28-b97e-92b179f829f6" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 574.507156] env[61806]: DEBUG nova.compute.manager [None req-7d4ae408-54ff-4248-838b-a36cbf45b0c5 tempest-ListServerFiltersTestJSON-241228521 tempest-ListServerFiltersTestJSON-241228521-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61806) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 574.507346] env[61806]: DEBUG nova.compute.manager [None req-7d4ae408-54ff-4248-838b-a36cbf45b0c5 tempest-ListServerFiltersTestJSON-241228521 tempest-ListServerFiltersTestJSON-241228521-project-member] [instance: 9ddec22f-0fce-4d28-b97e-92b179f829f6] Deallocating network for instance {{(pid=61806) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 574.507526] env[61806]: DEBUG nova.network.neutron [None req-7d4ae408-54ff-4248-838b-a36cbf45b0c5 tempest-ListServerFiltersTestJSON-241228521 tempest-ListServerFiltersTestJSON-241228521-project-member] [instance: 9ddec22f-0fce-4d28-b97e-92b179f829f6] deallocate_for_instance() {{(pid=61806) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 574.527009] env[61806]: DEBUG nova.network.neutron [None req-7d4ae408-54ff-4248-838b-a36cbf45b0c5 tempest-ListServerFiltersTestJSON-241228521 tempest-ListServerFiltersTestJSON-241228521-project-member] [instance: 9ddec22f-0fce-4d28-b97e-92b179f829f6] Instance cache missing network info. {{(pid=61806) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 574.680892] env[61806]: INFO nova.scheduler.client.report [None req-960d1abd-1b06-4b5a-8ddd-4b463d493a45 tempest-ListServerFiltersTestJSON-241228521 tempest-ListServerFiltersTestJSON-241228521-project-member] Deleted allocations for instance 30e44576-dcd3-4036-9070-9032b775c90a [ 574.876244] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-42298f78-9e7b-4ace-98f1-08c200a50e7b {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 574.884469] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b535a94f-2157-474d-8e49-e34bf1464b49 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 574.921332] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-201ba24e-5ce9-456c-af55-7742a4b73006 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 574.930550] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7d46395b-5fe5-4692-bc5b-0fa405bac25f {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 574.948222] env[61806]: DEBUG nova.compute.provider_tree [None req-38c4eaec-22d8-4bf0-bd45-321ef43329a0 tempest-ServersAdminTestJSON-14199920 tempest-ServersAdminTestJSON-14199920-project-member] Inventory has not changed in ProviderTree for provider: a2858be1-fd22-4e08-979e-87ad25293407 {{(pid=61806) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 575.032601] env[61806]: DEBUG nova.network.neutron [None req-7d4ae408-54ff-4248-838b-a36cbf45b0c5 tempest-ListServerFiltersTestJSON-241228521 tempest-ListServerFiltersTestJSON-241228521-project-member] [instance: 9ddec22f-0fce-4d28-b97e-92b179f829f6] Updating instance_info_cache with network_info: [] {{(pid=61806) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 575.192582] env[61806]: DEBUG oslo_concurrency.lockutils [None req-960d1abd-1b06-4b5a-8ddd-4b463d493a45 tempest-ListServerFiltersTestJSON-241228521 tempest-ListServerFiltersTestJSON-241228521-project-member] Lock "30e44576-dcd3-4036-9070-9032b775c90a" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 48.711s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 575.451894] env[61806]: DEBUG nova.scheduler.client.report [None req-38c4eaec-22d8-4bf0-bd45-321ef43329a0 tempest-ServersAdminTestJSON-14199920 tempest-ServersAdminTestJSON-14199920-project-member] Inventory has not changed for provider a2858be1-fd22-4e08-979e-87ad25293407 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 139, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61806) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 575.535345] env[61806]: INFO nova.compute.manager [None req-7d4ae408-54ff-4248-838b-a36cbf45b0c5 tempest-ListServerFiltersTestJSON-241228521 tempest-ListServerFiltersTestJSON-241228521-project-member] [instance: 9ddec22f-0fce-4d28-b97e-92b179f829f6] Took 1.03 seconds to deallocate network for instance. [ 575.697189] env[61806]: DEBUG nova.compute.manager [None req-962d5603-e71b-4a26-895d-6dce1320b70e tempest-ListImageFiltersTestJSON-680116525 tempest-ListImageFiltersTestJSON-680116525-project-member] [instance: 53d3e2f2-f63d-45c7-a5d1-1fa9a9f0ac0f] Starting instance... {{(pid=61806) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 575.770913] env[61806]: DEBUG oslo_concurrency.lockutils [None req-05ef4399-8310-428c-ba55-af11abf4e445 tempest-SecurityGroupsTestJSON-1213485581 tempest-SecurityGroupsTestJSON-1213485581-project-member] Acquiring lock "294adb58-f23e-4510-a25a-de6b909e3189" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 575.771169] env[61806]: DEBUG oslo_concurrency.lockutils [None req-05ef4399-8310-428c-ba55-af11abf4e445 tempest-SecurityGroupsTestJSON-1213485581 tempest-SecurityGroupsTestJSON-1213485581-project-member] Lock "294adb58-f23e-4510-a25a-de6b909e3189" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 575.956276] env[61806]: DEBUG oslo_concurrency.lockutils [None req-38c4eaec-22d8-4bf0-bd45-321ef43329a0 tempest-ServersAdminTestJSON-14199920 tempest-ServersAdminTestJSON-14199920-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.637s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 575.956725] env[61806]: DEBUG nova.compute.manager [None req-38c4eaec-22d8-4bf0-bd45-321ef43329a0 tempest-ServersAdminTestJSON-14199920 tempest-ServersAdminTestJSON-14199920-project-member] [instance: 8905f20f-28a6-49f4-88bf-e52177c6a4b6] Start building networks asynchronously for instance. {{(pid=61806) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 575.959336] env[61806]: DEBUG oslo_concurrency.lockutils [None req-37e71da7-cb5e-431d-ba3c-95daaf8af7f3 tempest-ServersAdminNegativeTestJSON-2128564044 tempest-ServersAdminNegativeTestJSON-2128564044-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 26.276s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 575.960854] env[61806]: INFO nova.compute.claims [None req-37e71da7-cb5e-431d-ba3c-95daaf8af7f3 tempest-ServersAdminNegativeTestJSON-2128564044 tempest-ServersAdminNegativeTestJSON-2128564044-project-member] [instance: 89ce0fd8-6a22-4b1c-bdff-d1584abd538a] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 576.223269] env[61806]: DEBUG oslo_concurrency.lockutils [None req-962d5603-e71b-4a26-895d-6dce1320b70e tempest-ListImageFiltersTestJSON-680116525 tempest-ListImageFiltersTestJSON-680116525-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 576.471855] env[61806]: DEBUG nova.compute.utils [None req-38c4eaec-22d8-4bf0-bd45-321ef43329a0 tempest-ServersAdminTestJSON-14199920 tempest-ServersAdminTestJSON-14199920-project-member] Using /dev/sd instead of None {{(pid=61806) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 576.474198] env[61806]: DEBUG nova.compute.manager [None req-38c4eaec-22d8-4bf0-bd45-321ef43329a0 tempest-ServersAdminTestJSON-14199920 tempest-ServersAdminTestJSON-14199920-project-member] [instance: 8905f20f-28a6-49f4-88bf-e52177c6a4b6] Allocating IP information in the background. {{(pid=61806) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 576.474198] env[61806]: DEBUG nova.network.neutron [None req-38c4eaec-22d8-4bf0-bd45-321ef43329a0 tempest-ServersAdminTestJSON-14199920 tempest-ServersAdminTestJSON-14199920-project-member] [instance: 8905f20f-28a6-49f4-88bf-e52177c6a4b6] allocate_for_instance() {{(pid=61806) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 576.552863] env[61806]: DEBUG nova.policy [None req-38c4eaec-22d8-4bf0-bd45-321ef43329a0 tempest-ServersAdminTestJSON-14199920 tempest-ServersAdminTestJSON-14199920-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '4147cb03f0904970bf0784dccb850d0b', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '738e028f263e4026ba95a5d491905d22', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61806) authorize /opt/stack/nova/nova/policy.py:201}} [ 576.580978] env[61806]: INFO nova.scheduler.client.report [None req-7d4ae408-54ff-4248-838b-a36cbf45b0c5 tempest-ListServerFiltersTestJSON-241228521 tempest-ListServerFiltersTestJSON-241228521-project-member] Deleted allocations for instance 9ddec22f-0fce-4d28-b97e-92b179f829f6 [ 576.871604] env[61806]: DEBUG nova.network.neutron [None req-38c4eaec-22d8-4bf0-bd45-321ef43329a0 tempest-ServersAdminTestJSON-14199920 tempest-ServersAdminTestJSON-14199920-project-member] [instance: 8905f20f-28a6-49f4-88bf-e52177c6a4b6] Successfully created port: 6c13d932-a28d-4aa7-99c6-9cda907b02db {{(pid=61806) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 576.986127] env[61806]: DEBUG nova.compute.manager [None req-38c4eaec-22d8-4bf0-bd45-321ef43329a0 tempest-ServersAdminTestJSON-14199920 tempest-ServersAdminTestJSON-14199920-project-member] [instance: 8905f20f-28a6-49f4-88bf-e52177c6a4b6] Start building block device mappings for instance. {{(pid=61806) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 577.093800] env[61806]: DEBUG oslo_concurrency.lockutils [None req-7d4ae408-54ff-4248-838b-a36cbf45b0c5 tempest-ListServerFiltersTestJSON-241228521 tempest-ListServerFiltersTestJSON-241228521-project-member] Lock "9ddec22f-0fce-4d28-b97e-92b179f829f6" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 48.283s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 577.424054] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-22fbd51a-561f-47de-bd1b-9593abe723ce {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 577.432582] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7c787f04-1720-4c45-8284-f9d73db58990 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 577.475183] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-761162a7-3d5f-48be-836f-883c9bbe0fbe {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 577.484853] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ee282204-e113-4e12-b13a-b74d19fc9646 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 577.510057] env[61806]: DEBUG nova.compute.provider_tree [None req-37e71da7-cb5e-431d-ba3c-95daaf8af7f3 tempest-ServersAdminNegativeTestJSON-2128564044 tempest-ServersAdminNegativeTestJSON-2128564044-project-member] Inventory has not changed in ProviderTree for provider: a2858be1-fd22-4e08-979e-87ad25293407 {{(pid=61806) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 577.599364] env[61806]: DEBUG nova.compute.manager [None req-11b5d1a0-518c-42d9-940e-4bf652a6d358 tempest-DeleteServersAdminTestJSON-1553667494 tempest-DeleteServersAdminTestJSON-1553667494-project-member] [instance: ddc801a0-0782-4c6d-bd72-95d0afd01339] Starting instance... {{(pid=61806) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 577.861385] env[61806]: DEBUG nova.compute.manager [req-22b0a96b-244c-4141-8023-eabe420855c6 req-7ddb6a57-87c2-4bf2-9296-dce50e4f93c8 service nova] [instance: 8905f20f-28a6-49f4-88bf-e52177c6a4b6] Received event network-changed-6c13d932-a28d-4aa7-99c6-9cda907b02db {{(pid=61806) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 577.861474] env[61806]: DEBUG nova.compute.manager [req-22b0a96b-244c-4141-8023-eabe420855c6 req-7ddb6a57-87c2-4bf2-9296-dce50e4f93c8 service nova] [instance: 8905f20f-28a6-49f4-88bf-e52177c6a4b6] Refreshing instance network info cache due to event network-changed-6c13d932-a28d-4aa7-99c6-9cda907b02db. {{(pid=61806) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 577.861620] env[61806]: DEBUG oslo_concurrency.lockutils [req-22b0a96b-244c-4141-8023-eabe420855c6 req-7ddb6a57-87c2-4bf2-9296-dce50e4f93c8 service nova] Acquiring lock "refresh_cache-8905f20f-28a6-49f4-88bf-e52177c6a4b6" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 577.861766] env[61806]: DEBUG oslo_concurrency.lockutils [req-22b0a96b-244c-4141-8023-eabe420855c6 req-7ddb6a57-87c2-4bf2-9296-dce50e4f93c8 service nova] Acquired lock "refresh_cache-8905f20f-28a6-49f4-88bf-e52177c6a4b6" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 577.861999] env[61806]: DEBUG nova.network.neutron [req-22b0a96b-244c-4141-8023-eabe420855c6 req-7ddb6a57-87c2-4bf2-9296-dce50e4f93c8 service nova] [instance: 8905f20f-28a6-49f4-88bf-e52177c6a4b6] Refreshing network info cache for port 6c13d932-a28d-4aa7-99c6-9cda907b02db {{(pid=61806) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 577.988254] env[61806]: ERROR nova.compute.manager [None req-38c4eaec-22d8-4bf0-bd45-321ef43329a0 tempest-ServersAdminTestJSON-14199920 tempest-ServersAdminTestJSON-14199920-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 6c13d932-a28d-4aa7-99c6-9cda907b02db, please check neutron logs for more information. [ 577.988254] env[61806]: ERROR nova.compute.manager Traceback (most recent call last): [ 577.988254] env[61806]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 577.988254] env[61806]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 577.988254] env[61806]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 577.988254] env[61806]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 577.988254] env[61806]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 577.988254] env[61806]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 577.988254] env[61806]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 577.988254] env[61806]: ERROR nova.compute.manager self.force_reraise() [ 577.988254] env[61806]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 577.988254] env[61806]: ERROR nova.compute.manager raise self.value [ 577.988254] env[61806]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 577.988254] env[61806]: ERROR nova.compute.manager updated_port = self._update_port( [ 577.988254] env[61806]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 577.988254] env[61806]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 577.989030] env[61806]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 577.989030] env[61806]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 577.989030] env[61806]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 6c13d932-a28d-4aa7-99c6-9cda907b02db, please check neutron logs for more information. [ 577.989030] env[61806]: ERROR nova.compute.manager [ 577.989030] env[61806]: Traceback (most recent call last): [ 577.989030] env[61806]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 577.989030] env[61806]: listener.cb(fileno) [ 577.989030] env[61806]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 577.989030] env[61806]: result = function(*args, **kwargs) [ 577.989030] env[61806]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 577.989030] env[61806]: return func(*args, **kwargs) [ 577.989030] env[61806]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 577.989030] env[61806]: raise e [ 577.989030] env[61806]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 577.989030] env[61806]: nwinfo = self.network_api.allocate_for_instance( [ 577.989030] env[61806]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 577.989030] env[61806]: created_port_ids = self._update_ports_for_instance( [ 577.989030] env[61806]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 577.989030] env[61806]: with excutils.save_and_reraise_exception(): [ 577.989030] env[61806]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 577.989030] env[61806]: self.force_reraise() [ 577.989030] env[61806]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 577.989030] env[61806]: raise self.value [ 577.989030] env[61806]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 577.989030] env[61806]: updated_port = self._update_port( [ 577.989030] env[61806]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 577.989030] env[61806]: _ensure_no_port_binding_failure(port) [ 577.989030] env[61806]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 577.989030] env[61806]: raise exception.PortBindingFailed(port_id=port['id']) [ 577.990103] env[61806]: nova.exception.PortBindingFailed: Binding failed for port 6c13d932-a28d-4aa7-99c6-9cda907b02db, please check neutron logs for more information. [ 577.990103] env[61806]: Removing descriptor: 17 [ 577.995736] env[61806]: DEBUG nova.compute.manager [None req-38c4eaec-22d8-4bf0-bd45-321ef43329a0 tempest-ServersAdminTestJSON-14199920 tempest-ServersAdminTestJSON-14199920-project-member] [instance: 8905f20f-28a6-49f4-88bf-e52177c6a4b6] Start spawning the instance on the hypervisor. {{(pid=61806) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 578.015624] env[61806]: DEBUG nova.scheduler.client.report [None req-37e71da7-cb5e-431d-ba3c-95daaf8af7f3 tempest-ServersAdminNegativeTestJSON-2128564044 tempest-ServersAdminNegativeTestJSON-2128564044-project-member] Inventory has not changed for provider a2858be1-fd22-4e08-979e-87ad25293407 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 139, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61806) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 578.028202] env[61806]: DEBUG nova.virt.hardware [None req-38c4eaec-22d8-4bf0-bd45-321ef43329a0 tempest-ServersAdminTestJSON-14199920 tempest-ServersAdminTestJSON-14199920-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-15T18:05:10Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-15T18:04:52Z,direct_url=,disk_format='vmdk',id=ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='84ba83fce2c349a6988173c5d6fc3b07',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-15T18:04:53Z,virtual_size=,visibility=), allow threads: False {{(pid=61806) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 578.028449] env[61806]: DEBUG nova.virt.hardware [None req-38c4eaec-22d8-4bf0-bd45-321ef43329a0 tempest-ServersAdminTestJSON-14199920 tempest-ServersAdminTestJSON-14199920-project-member] Flavor limits 0:0:0 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 578.028609] env[61806]: DEBUG nova.virt.hardware [None req-38c4eaec-22d8-4bf0-bd45-321ef43329a0 tempest-ServersAdminTestJSON-14199920 tempest-ServersAdminTestJSON-14199920-project-member] Image limits 0:0:0 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 578.028898] env[61806]: DEBUG nova.virt.hardware [None req-38c4eaec-22d8-4bf0-bd45-321ef43329a0 tempest-ServersAdminTestJSON-14199920 tempest-ServersAdminTestJSON-14199920-project-member] Flavor pref 0:0:0 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 578.028988] env[61806]: DEBUG nova.virt.hardware [None req-38c4eaec-22d8-4bf0-bd45-321ef43329a0 tempest-ServersAdminTestJSON-14199920 tempest-ServersAdminTestJSON-14199920-project-member] Image pref 0:0:0 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 578.029235] env[61806]: DEBUG nova.virt.hardware [None req-38c4eaec-22d8-4bf0-bd45-321ef43329a0 tempest-ServersAdminTestJSON-14199920 tempest-ServersAdminTestJSON-14199920-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 578.029467] env[61806]: DEBUG nova.virt.hardware [None req-38c4eaec-22d8-4bf0-bd45-321ef43329a0 tempest-ServersAdminTestJSON-14199920 tempest-ServersAdminTestJSON-14199920-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61806) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 578.029649] env[61806]: DEBUG nova.virt.hardware [None req-38c4eaec-22d8-4bf0-bd45-321ef43329a0 tempest-ServersAdminTestJSON-14199920 tempest-ServersAdminTestJSON-14199920-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61806) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 578.029827] env[61806]: DEBUG nova.virt.hardware [None req-38c4eaec-22d8-4bf0-bd45-321ef43329a0 tempest-ServersAdminTestJSON-14199920 tempest-ServersAdminTestJSON-14199920-project-member] Got 1 possible topologies {{(pid=61806) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 578.029992] env[61806]: DEBUG nova.virt.hardware [None req-38c4eaec-22d8-4bf0-bd45-321ef43329a0 tempest-ServersAdminTestJSON-14199920 tempest-ServersAdminTestJSON-14199920-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61806) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 578.030183] env[61806]: DEBUG nova.virt.hardware [None req-38c4eaec-22d8-4bf0-bd45-321ef43329a0 tempest-ServersAdminTestJSON-14199920 tempest-ServersAdminTestJSON-14199920-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61806) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 578.031419] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b835eb52-0dfd-4500-b5e3-fe5c59de031d {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 578.043225] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1c7da0de-4588-4724-8024-53b9c9ddd25b {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 578.059026] env[61806]: ERROR nova.compute.manager [None req-38c4eaec-22d8-4bf0-bd45-321ef43329a0 tempest-ServersAdminTestJSON-14199920 tempest-ServersAdminTestJSON-14199920-project-member] [instance: 8905f20f-28a6-49f4-88bf-e52177c6a4b6] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 6c13d932-a28d-4aa7-99c6-9cda907b02db, please check neutron logs for more information. [ 578.059026] env[61806]: ERROR nova.compute.manager [instance: 8905f20f-28a6-49f4-88bf-e52177c6a4b6] Traceback (most recent call last): [ 578.059026] env[61806]: ERROR nova.compute.manager [instance: 8905f20f-28a6-49f4-88bf-e52177c6a4b6] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 578.059026] env[61806]: ERROR nova.compute.manager [instance: 8905f20f-28a6-49f4-88bf-e52177c6a4b6] yield resources [ 578.059026] env[61806]: ERROR nova.compute.manager [instance: 8905f20f-28a6-49f4-88bf-e52177c6a4b6] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 578.059026] env[61806]: ERROR nova.compute.manager [instance: 8905f20f-28a6-49f4-88bf-e52177c6a4b6] self.driver.spawn(context, instance, image_meta, [ 578.059026] env[61806]: ERROR nova.compute.manager [instance: 8905f20f-28a6-49f4-88bf-e52177c6a4b6] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 578.059026] env[61806]: ERROR nova.compute.manager [instance: 8905f20f-28a6-49f4-88bf-e52177c6a4b6] self._vmops.spawn(context, instance, image_meta, injected_files, [ 578.059026] env[61806]: ERROR nova.compute.manager [instance: 8905f20f-28a6-49f4-88bf-e52177c6a4b6] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 578.059026] env[61806]: ERROR nova.compute.manager [instance: 8905f20f-28a6-49f4-88bf-e52177c6a4b6] vm_ref = self.build_virtual_machine(instance, [ 578.059026] env[61806]: ERROR nova.compute.manager [instance: 8905f20f-28a6-49f4-88bf-e52177c6a4b6] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 578.059471] env[61806]: ERROR nova.compute.manager [instance: 8905f20f-28a6-49f4-88bf-e52177c6a4b6] vif_infos = vmwarevif.get_vif_info(self._session, [ 578.059471] env[61806]: ERROR nova.compute.manager [instance: 8905f20f-28a6-49f4-88bf-e52177c6a4b6] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 578.059471] env[61806]: ERROR nova.compute.manager [instance: 8905f20f-28a6-49f4-88bf-e52177c6a4b6] for vif in network_info: [ 578.059471] env[61806]: ERROR nova.compute.manager [instance: 8905f20f-28a6-49f4-88bf-e52177c6a4b6] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 578.059471] env[61806]: ERROR nova.compute.manager [instance: 8905f20f-28a6-49f4-88bf-e52177c6a4b6] return self._sync_wrapper(fn, *args, **kwargs) [ 578.059471] env[61806]: ERROR nova.compute.manager [instance: 8905f20f-28a6-49f4-88bf-e52177c6a4b6] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 578.059471] env[61806]: ERROR nova.compute.manager [instance: 8905f20f-28a6-49f4-88bf-e52177c6a4b6] self.wait() [ 578.059471] env[61806]: ERROR nova.compute.manager [instance: 8905f20f-28a6-49f4-88bf-e52177c6a4b6] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 578.059471] env[61806]: ERROR nova.compute.manager [instance: 8905f20f-28a6-49f4-88bf-e52177c6a4b6] self[:] = self._gt.wait() [ 578.059471] env[61806]: ERROR nova.compute.manager [instance: 8905f20f-28a6-49f4-88bf-e52177c6a4b6] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 578.059471] env[61806]: ERROR nova.compute.manager [instance: 8905f20f-28a6-49f4-88bf-e52177c6a4b6] return self._exit_event.wait() [ 578.059471] env[61806]: ERROR nova.compute.manager [instance: 8905f20f-28a6-49f4-88bf-e52177c6a4b6] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 578.059471] env[61806]: ERROR nova.compute.manager [instance: 8905f20f-28a6-49f4-88bf-e52177c6a4b6] current.throw(*self._exc) [ 578.060379] env[61806]: ERROR nova.compute.manager [instance: 8905f20f-28a6-49f4-88bf-e52177c6a4b6] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 578.060379] env[61806]: ERROR nova.compute.manager [instance: 8905f20f-28a6-49f4-88bf-e52177c6a4b6] result = function(*args, **kwargs) [ 578.060379] env[61806]: ERROR nova.compute.manager [instance: 8905f20f-28a6-49f4-88bf-e52177c6a4b6] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 578.060379] env[61806]: ERROR nova.compute.manager [instance: 8905f20f-28a6-49f4-88bf-e52177c6a4b6] return func(*args, **kwargs) [ 578.060379] env[61806]: ERROR nova.compute.manager [instance: 8905f20f-28a6-49f4-88bf-e52177c6a4b6] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 578.060379] env[61806]: ERROR nova.compute.manager [instance: 8905f20f-28a6-49f4-88bf-e52177c6a4b6] raise e [ 578.060379] env[61806]: ERROR nova.compute.manager [instance: 8905f20f-28a6-49f4-88bf-e52177c6a4b6] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 578.060379] env[61806]: ERROR nova.compute.manager [instance: 8905f20f-28a6-49f4-88bf-e52177c6a4b6] nwinfo = self.network_api.allocate_for_instance( [ 578.060379] env[61806]: ERROR nova.compute.manager [instance: 8905f20f-28a6-49f4-88bf-e52177c6a4b6] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 578.060379] env[61806]: ERROR nova.compute.manager [instance: 8905f20f-28a6-49f4-88bf-e52177c6a4b6] created_port_ids = self._update_ports_for_instance( [ 578.060379] env[61806]: ERROR nova.compute.manager [instance: 8905f20f-28a6-49f4-88bf-e52177c6a4b6] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 578.060379] env[61806]: ERROR nova.compute.manager [instance: 8905f20f-28a6-49f4-88bf-e52177c6a4b6] with excutils.save_and_reraise_exception(): [ 578.060379] env[61806]: ERROR nova.compute.manager [instance: 8905f20f-28a6-49f4-88bf-e52177c6a4b6] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 578.060840] env[61806]: ERROR nova.compute.manager [instance: 8905f20f-28a6-49f4-88bf-e52177c6a4b6] self.force_reraise() [ 578.060840] env[61806]: ERROR nova.compute.manager [instance: 8905f20f-28a6-49f4-88bf-e52177c6a4b6] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 578.060840] env[61806]: ERROR nova.compute.manager [instance: 8905f20f-28a6-49f4-88bf-e52177c6a4b6] raise self.value [ 578.060840] env[61806]: ERROR nova.compute.manager [instance: 8905f20f-28a6-49f4-88bf-e52177c6a4b6] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 578.060840] env[61806]: ERROR nova.compute.manager [instance: 8905f20f-28a6-49f4-88bf-e52177c6a4b6] updated_port = self._update_port( [ 578.060840] env[61806]: ERROR nova.compute.manager [instance: 8905f20f-28a6-49f4-88bf-e52177c6a4b6] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 578.060840] env[61806]: ERROR nova.compute.manager [instance: 8905f20f-28a6-49f4-88bf-e52177c6a4b6] _ensure_no_port_binding_failure(port) [ 578.060840] env[61806]: ERROR nova.compute.manager [instance: 8905f20f-28a6-49f4-88bf-e52177c6a4b6] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 578.060840] env[61806]: ERROR nova.compute.manager [instance: 8905f20f-28a6-49f4-88bf-e52177c6a4b6] raise exception.PortBindingFailed(port_id=port['id']) [ 578.060840] env[61806]: ERROR nova.compute.manager [instance: 8905f20f-28a6-49f4-88bf-e52177c6a4b6] nova.exception.PortBindingFailed: Binding failed for port 6c13d932-a28d-4aa7-99c6-9cda907b02db, please check neutron logs for more information. [ 578.060840] env[61806]: ERROR nova.compute.manager [instance: 8905f20f-28a6-49f4-88bf-e52177c6a4b6] [ 578.060840] env[61806]: INFO nova.compute.manager [None req-38c4eaec-22d8-4bf0-bd45-321ef43329a0 tempest-ServersAdminTestJSON-14199920 tempest-ServersAdminTestJSON-14199920-project-member] [instance: 8905f20f-28a6-49f4-88bf-e52177c6a4b6] Terminating instance [ 578.062408] env[61806]: DEBUG oslo_concurrency.lockutils [None req-38c4eaec-22d8-4bf0-bd45-321ef43329a0 tempest-ServersAdminTestJSON-14199920 tempest-ServersAdminTestJSON-14199920-project-member] Acquiring lock "refresh_cache-8905f20f-28a6-49f4-88bf-e52177c6a4b6" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 578.129013] env[61806]: DEBUG oslo_concurrency.lockutils [None req-11b5d1a0-518c-42d9-940e-4bf652a6d358 tempest-DeleteServersAdminTestJSON-1553667494 tempest-DeleteServersAdminTestJSON-1553667494-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 578.398140] env[61806]: DEBUG nova.network.neutron [req-22b0a96b-244c-4141-8023-eabe420855c6 req-7ddb6a57-87c2-4bf2-9296-dce50e4f93c8 service nova] [instance: 8905f20f-28a6-49f4-88bf-e52177c6a4b6] Instance cache missing network info. {{(pid=61806) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 578.502308] env[61806]: DEBUG nova.network.neutron [req-22b0a96b-244c-4141-8023-eabe420855c6 req-7ddb6a57-87c2-4bf2-9296-dce50e4f93c8 service nova] [instance: 8905f20f-28a6-49f4-88bf-e52177c6a4b6] Updating instance_info_cache with network_info: [] {{(pid=61806) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 578.526888] env[61806]: DEBUG oslo_concurrency.lockutils [None req-37e71da7-cb5e-431d-ba3c-95daaf8af7f3 tempest-ServersAdminNegativeTestJSON-2128564044 tempest-ServersAdminNegativeTestJSON-2128564044-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.567s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 578.527077] env[61806]: DEBUG nova.compute.manager [None req-37e71da7-cb5e-431d-ba3c-95daaf8af7f3 tempest-ServersAdminNegativeTestJSON-2128564044 tempest-ServersAdminNegativeTestJSON-2128564044-project-member] [instance: 89ce0fd8-6a22-4b1c-bdff-d1584abd538a] Start building networks asynchronously for instance. {{(pid=61806) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 578.529701] env[61806]: DEBUG oslo_concurrency.lockutils [None req-aac795d3-c37f-4811-9be9-aa67ad520ab0 tempest-ListServerFiltersTestJSON-241228521 tempest-ListServerFiltersTestJSON-241228521-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 23.451s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 579.005684] env[61806]: DEBUG oslo_concurrency.lockutils [req-22b0a96b-244c-4141-8023-eabe420855c6 req-7ddb6a57-87c2-4bf2-9296-dce50e4f93c8 service nova] Releasing lock "refresh_cache-8905f20f-28a6-49f4-88bf-e52177c6a4b6" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 579.006140] env[61806]: DEBUG oslo_concurrency.lockutils [None req-38c4eaec-22d8-4bf0-bd45-321ef43329a0 tempest-ServersAdminTestJSON-14199920 tempest-ServersAdminTestJSON-14199920-project-member] Acquired lock "refresh_cache-8905f20f-28a6-49f4-88bf-e52177c6a4b6" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 579.006326] env[61806]: DEBUG nova.network.neutron [None req-38c4eaec-22d8-4bf0-bd45-321ef43329a0 tempest-ServersAdminTestJSON-14199920 tempest-ServersAdminTestJSON-14199920-project-member] [instance: 8905f20f-28a6-49f4-88bf-e52177c6a4b6] Building network info cache for instance {{(pid=61806) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 579.035058] env[61806]: DEBUG nova.compute.utils [None req-37e71da7-cb5e-431d-ba3c-95daaf8af7f3 tempest-ServersAdminNegativeTestJSON-2128564044 tempest-ServersAdminNegativeTestJSON-2128564044-project-member] Using /dev/sd instead of None {{(pid=61806) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 579.039431] env[61806]: DEBUG nova.compute.manager [None req-37e71da7-cb5e-431d-ba3c-95daaf8af7f3 tempest-ServersAdminNegativeTestJSON-2128564044 tempest-ServersAdminNegativeTestJSON-2128564044-project-member] [instance: 89ce0fd8-6a22-4b1c-bdff-d1584abd538a] Allocating IP information in the background. {{(pid=61806) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 579.039683] env[61806]: DEBUG nova.network.neutron [None req-37e71da7-cb5e-431d-ba3c-95daaf8af7f3 tempest-ServersAdminNegativeTestJSON-2128564044 tempest-ServersAdminNegativeTestJSON-2128564044-project-member] [instance: 89ce0fd8-6a22-4b1c-bdff-d1584abd538a] allocate_for_instance() {{(pid=61806) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 579.129735] env[61806]: DEBUG nova.policy [None req-37e71da7-cb5e-431d-ba3c-95daaf8af7f3 tempest-ServersAdminNegativeTestJSON-2128564044 tempest-ServersAdminNegativeTestJSON-2128564044-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'd07036c78f9b453eb11fb11ec2fe11c6', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'f4cc1450314b43749b933fb92369b922', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61806) authorize /opt/stack/nova/nova/policy.py:201}} [ 579.493745] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-abcaecf0-bdee-4e6e-a936-f8ec7a1c6779 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 579.502321] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c0dc011f-2ff3-478b-acdc-5c74a2771db0 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 579.541775] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-73db6742-5ca1-4de8-96f6-6a06585bd28a {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 579.545404] env[61806]: DEBUG nova.compute.manager [None req-37e71da7-cb5e-431d-ba3c-95daaf8af7f3 tempest-ServersAdminNegativeTestJSON-2128564044 tempest-ServersAdminNegativeTestJSON-2128564044-project-member] [instance: 89ce0fd8-6a22-4b1c-bdff-d1584abd538a] Start building block device mappings for instance. {{(pid=61806) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 579.555502] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e23a071b-9087-44fb-88e9-b0bf30ddd8bf {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 579.560821] env[61806]: DEBUG nova.network.neutron [None req-38c4eaec-22d8-4bf0-bd45-321ef43329a0 tempest-ServersAdminTestJSON-14199920 tempest-ServersAdminTestJSON-14199920-project-member] [instance: 8905f20f-28a6-49f4-88bf-e52177c6a4b6] Instance cache missing network info. {{(pid=61806) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 579.576098] env[61806]: DEBUG nova.compute.provider_tree [None req-aac795d3-c37f-4811-9be9-aa67ad520ab0 tempest-ListServerFiltersTestJSON-241228521 tempest-ListServerFiltersTestJSON-241228521-project-member] Inventory has not changed in ProviderTree for provider: a2858be1-fd22-4e08-979e-87ad25293407 {{(pid=61806) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 579.613447] env[61806]: DEBUG nova.network.neutron [None req-37e71da7-cb5e-431d-ba3c-95daaf8af7f3 tempest-ServersAdminNegativeTestJSON-2128564044 tempest-ServersAdminNegativeTestJSON-2128564044-project-member] [instance: 89ce0fd8-6a22-4b1c-bdff-d1584abd538a] Successfully created port: 63122136-e392-4570-8f9c-7785b771ac82 {{(pid=61806) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 579.640284] env[61806]: DEBUG oslo_concurrency.lockutils [None req-fcbc4cf1-fc11-42b1-9d49-ef556b52782e tempest-ServerRescueNegativeTestJSON-1483258717 tempest-ServerRescueNegativeTestJSON-1483258717-project-member] Acquiring lock "7ec2497a-7329-4ed4-99ab-dec3d1e8117c" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 579.640684] env[61806]: DEBUG oslo_concurrency.lockutils [None req-fcbc4cf1-fc11-42b1-9d49-ef556b52782e tempest-ServerRescueNegativeTestJSON-1483258717 tempest-ServerRescueNegativeTestJSON-1483258717-project-member] Lock "7ec2497a-7329-4ed4-99ab-dec3d1e8117c" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 579.692399] env[61806]: DEBUG nova.network.neutron [None req-38c4eaec-22d8-4bf0-bd45-321ef43329a0 tempest-ServersAdminTestJSON-14199920 tempest-ServersAdminTestJSON-14199920-project-member] [instance: 8905f20f-28a6-49f4-88bf-e52177c6a4b6] Updating instance_info_cache with network_info: [] {{(pid=61806) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 580.050277] env[61806]: DEBUG nova.compute.manager [req-1aa4f5ca-6249-4d6a-92d1-7df5ba48a3eb req-d8819cf3-2353-420b-9514-ccb17115e7e5 service nova] [instance: 8905f20f-28a6-49f4-88bf-e52177c6a4b6] Received event network-vif-deleted-6c13d932-a28d-4aa7-99c6-9cda907b02db {{(pid=61806) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 580.081993] env[61806]: DEBUG nova.scheduler.client.report [None req-aac795d3-c37f-4811-9be9-aa67ad520ab0 tempest-ListServerFiltersTestJSON-241228521 tempest-ListServerFiltersTestJSON-241228521-project-member] Inventory has not changed for provider a2858be1-fd22-4e08-979e-87ad25293407 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 139, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61806) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 580.196451] env[61806]: DEBUG oslo_concurrency.lockutils [None req-38c4eaec-22d8-4bf0-bd45-321ef43329a0 tempest-ServersAdminTestJSON-14199920 tempest-ServersAdminTestJSON-14199920-project-member] Releasing lock "refresh_cache-8905f20f-28a6-49f4-88bf-e52177c6a4b6" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 580.197531] env[61806]: DEBUG nova.compute.manager [None req-38c4eaec-22d8-4bf0-bd45-321ef43329a0 tempest-ServersAdminTestJSON-14199920 tempest-ServersAdminTestJSON-14199920-project-member] [instance: 8905f20f-28a6-49f4-88bf-e52177c6a4b6] Start destroying the instance on the hypervisor. {{(pid=61806) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 580.197531] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-38c4eaec-22d8-4bf0-bd45-321ef43329a0 tempest-ServersAdminTestJSON-14199920 tempest-ServersAdminTestJSON-14199920-project-member] [instance: 8905f20f-28a6-49f4-88bf-e52177c6a4b6] Destroying instance {{(pid=61806) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 580.197531] env[61806]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-ad0f671d-cf86-4708-bfec-298d250c5420 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 580.210240] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d12a1b2d-09cf-439f-b1de-5b393fd4fd98 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 580.242327] env[61806]: WARNING nova.virt.vmwareapi.vmops [None req-38c4eaec-22d8-4bf0-bd45-321ef43329a0 tempest-ServersAdminTestJSON-14199920 tempest-ServersAdminTestJSON-14199920-project-member] [instance: 8905f20f-28a6-49f4-88bf-e52177c6a4b6] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 8905f20f-28a6-49f4-88bf-e52177c6a4b6 could not be found. [ 580.242386] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-38c4eaec-22d8-4bf0-bd45-321ef43329a0 tempest-ServersAdminTestJSON-14199920 tempest-ServersAdminTestJSON-14199920-project-member] [instance: 8905f20f-28a6-49f4-88bf-e52177c6a4b6] Instance destroyed {{(pid=61806) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 580.242836] env[61806]: INFO nova.compute.manager [None req-38c4eaec-22d8-4bf0-bd45-321ef43329a0 tempest-ServersAdminTestJSON-14199920 tempest-ServersAdminTestJSON-14199920-project-member] [instance: 8905f20f-28a6-49f4-88bf-e52177c6a4b6] Took 0.05 seconds to destroy the instance on the hypervisor. [ 580.242960] env[61806]: DEBUG oslo.service.loopingcall [None req-38c4eaec-22d8-4bf0-bd45-321ef43329a0 tempest-ServersAdminTestJSON-14199920 tempest-ServersAdminTestJSON-14199920-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61806) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 580.243281] env[61806]: DEBUG nova.compute.manager [-] [instance: 8905f20f-28a6-49f4-88bf-e52177c6a4b6] Deallocating network for instance {{(pid=61806) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 580.243399] env[61806]: DEBUG nova.network.neutron [-] [instance: 8905f20f-28a6-49f4-88bf-e52177c6a4b6] deallocate_for_instance() {{(pid=61806) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 580.266913] env[61806]: DEBUG nova.network.neutron [-] [instance: 8905f20f-28a6-49f4-88bf-e52177c6a4b6] Instance cache missing network info. {{(pid=61806) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 580.492148] env[61806]: DEBUG oslo_concurrency.lockutils [None req-6b55cbbf-00f7-4d4d-87f3-a59cc1c553b0 tempest-ServerRescueNegativeTestJSON-1483258717 tempest-ServerRescueNegativeTestJSON-1483258717-project-member] Acquiring lock "7ea919b6-12f7-4f4e-a123-413044bbbffc" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 580.492148] env[61806]: DEBUG oslo_concurrency.lockutils [None req-6b55cbbf-00f7-4d4d-87f3-a59cc1c553b0 tempest-ServerRescueNegativeTestJSON-1483258717 tempest-ServerRescueNegativeTestJSON-1483258717-project-member] Lock "7ea919b6-12f7-4f4e-a123-413044bbbffc" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 580.555613] env[61806]: DEBUG nova.compute.manager [None req-37e71da7-cb5e-431d-ba3c-95daaf8af7f3 tempest-ServersAdminNegativeTestJSON-2128564044 tempest-ServersAdminNegativeTestJSON-2128564044-project-member] [instance: 89ce0fd8-6a22-4b1c-bdff-d1584abd538a] Start spawning the instance on the hypervisor. {{(pid=61806) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 580.587183] env[61806]: DEBUG nova.virt.hardware [None req-37e71da7-cb5e-431d-ba3c-95daaf8af7f3 tempest-ServersAdminNegativeTestJSON-2128564044 tempest-ServersAdminNegativeTestJSON-2128564044-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-15T18:05:10Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-15T18:04:52Z,direct_url=,disk_format='vmdk',id=ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='84ba83fce2c349a6988173c5d6fc3b07',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-15T18:04:53Z,virtual_size=,visibility=), allow threads: False {{(pid=61806) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 580.587437] env[61806]: DEBUG nova.virt.hardware [None req-37e71da7-cb5e-431d-ba3c-95daaf8af7f3 tempest-ServersAdminNegativeTestJSON-2128564044 tempest-ServersAdminNegativeTestJSON-2128564044-project-member] Flavor limits 0:0:0 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 580.587614] env[61806]: DEBUG nova.virt.hardware [None req-37e71da7-cb5e-431d-ba3c-95daaf8af7f3 tempest-ServersAdminNegativeTestJSON-2128564044 tempest-ServersAdminNegativeTestJSON-2128564044-project-member] Image limits 0:0:0 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 580.587815] env[61806]: DEBUG nova.virt.hardware [None req-37e71da7-cb5e-431d-ba3c-95daaf8af7f3 tempest-ServersAdminNegativeTestJSON-2128564044 tempest-ServersAdminNegativeTestJSON-2128564044-project-member] Flavor pref 0:0:0 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 580.587963] env[61806]: DEBUG nova.virt.hardware [None req-37e71da7-cb5e-431d-ba3c-95daaf8af7f3 tempest-ServersAdminNegativeTestJSON-2128564044 tempest-ServersAdminNegativeTestJSON-2128564044-project-member] Image pref 0:0:0 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 580.588131] env[61806]: DEBUG nova.virt.hardware [None req-37e71da7-cb5e-431d-ba3c-95daaf8af7f3 tempest-ServersAdminNegativeTestJSON-2128564044 tempest-ServersAdminNegativeTestJSON-2128564044-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 580.588334] env[61806]: DEBUG nova.virt.hardware [None req-37e71da7-cb5e-431d-ba3c-95daaf8af7f3 tempest-ServersAdminNegativeTestJSON-2128564044 tempest-ServersAdminNegativeTestJSON-2128564044-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61806) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 580.588488] env[61806]: DEBUG nova.virt.hardware [None req-37e71da7-cb5e-431d-ba3c-95daaf8af7f3 tempest-ServersAdminNegativeTestJSON-2128564044 tempest-ServersAdminNegativeTestJSON-2128564044-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61806) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 580.588652] env[61806]: DEBUG nova.virt.hardware [None req-37e71da7-cb5e-431d-ba3c-95daaf8af7f3 tempest-ServersAdminNegativeTestJSON-2128564044 tempest-ServersAdminNegativeTestJSON-2128564044-project-member] Got 1 possible topologies {{(pid=61806) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 580.588811] env[61806]: DEBUG nova.virt.hardware [None req-37e71da7-cb5e-431d-ba3c-95daaf8af7f3 tempest-ServersAdminNegativeTestJSON-2128564044 tempest-ServersAdminNegativeTestJSON-2128564044-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61806) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 580.588979] env[61806]: DEBUG nova.virt.hardware [None req-37e71da7-cb5e-431d-ba3c-95daaf8af7f3 tempest-ServersAdminNegativeTestJSON-2128564044 tempest-ServersAdminNegativeTestJSON-2128564044-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61806) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 580.589706] env[61806]: DEBUG oslo_concurrency.lockutils [None req-aac795d3-c37f-4811-9be9-aa67ad520ab0 tempest-ListServerFiltersTestJSON-241228521 tempest-ListServerFiltersTestJSON-241228521-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.060s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 580.590298] env[61806]: ERROR nova.compute.manager [None req-aac795d3-c37f-4811-9be9-aa67ad520ab0 tempest-ListServerFiltersTestJSON-241228521 tempest-ListServerFiltersTestJSON-241228521-project-member] [instance: 7fcee323-8ea6-4f69-9f1e-f2105af0fe96] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port eed1949d-29dd-4127-a2a4-d7dc2369df55, please check neutron logs for more information. [ 580.590298] env[61806]: ERROR nova.compute.manager [instance: 7fcee323-8ea6-4f69-9f1e-f2105af0fe96] Traceback (most recent call last): [ 580.590298] env[61806]: ERROR nova.compute.manager [instance: 7fcee323-8ea6-4f69-9f1e-f2105af0fe96] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 580.590298] env[61806]: ERROR nova.compute.manager [instance: 7fcee323-8ea6-4f69-9f1e-f2105af0fe96] self.driver.spawn(context, instance, image_meta, [ 580.590298] env[61806]: ERROR nova.compute.manager [instance: 7fcee323-8ea6-4f69-9f1e-f2105af0fe96] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 580.590298] env[61806]: ERROR nova.compute.manager [instance: 7fcee323-8ea6-4f69-9f1e-f2105af0fe96] self._vmops.spawn(context, instance, image_meta, injected_files, [ 580.590298] env[61806]: ERROR nova.compute.manager [instance: 7fcee323-8ea6-4f69-9f1e-f2105af0fe96] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 580.590298] env[61806]: ERROR nova.compute.manager [instance: 7fcee323-8ea6-4f69-9f1e-f2105af0fe96] vm_ref = self.build_virtual_machine(instance, [ 580.590298] env[61806]: ERROR nova.compute.manager [instance: 7fcee323-8ea6-4f69-9f1e-f2105af0fe96] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 580.590298] env[61806]: ERROR nova.compute.manager [instance: 7fcee323-8ea6-4f69-9f1e-f2105af0fe96] vif_infos = vmwarevif.get_vif_info(self._session, [ 580.590298] env[61806]: ERROR nova.compute.manager [instance: 7fcee323-8ea6-4f69-9f1e-f2105af0fe96] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 580.590716] env[61806]: ERROR nova.compute.manager [instance: 7fcee323-8ea6-4f69-9f1e-f2105af0fe96] for vif in network_info: [ 580.590716] env[61806]: ERROR nova.compute.manager [instance: 7fcee323-8ea6-4f69-9f1e-f2105af0fe96] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 580.590716] env[61806]: ERROR nova.compute.manager [instance: 7fcee323-8ea6-4f69-9f1e-f2105af0fe96] return self._sync_wrapper(fn, *args, **kwargs) [ 580.590716] env[61806]: ERROR nova.compute.manager [instance: 7fcee323-8ea6-4f69-9f1e-f2105af0fe96] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 580.590716] env[61806]: ERROR nova.compute.manager [instance: 7fcee323-8ea6-4f69-9f1e-f2105af0fe96] self.wait() [ 580.590716] env[61806]: ERROR nova.compute.manager [instance: 7fcee323-8ea6-4f69-9f1e-f2105af0fe96] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 580.590716] env[61806]: ERROR nova.compute.manager [instance: 7fcee323-8ea6-4f69-9f1e-f2105af0fe96] self[:] = self._gt.wait() [ 580.590716] env[61806]: ERROR nova.compute.manager [instance: 7fcee323-8ea6-4f69-9f1e-f2105af0fe96] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 580.590716] env[61806]: ERROR nova.compute.manager [instance: 7fcee323-8ea6-4f69-9f1e-f2105af0fe96] return self._exit_event.wait() [ 580.590716] env[61806]: ERROR nova.compute.manager [instance: 7fcee323-8ea6-4f69-9f1e-f2105af0fe96] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 580.590716] env[61806]: ERROR nova.compute.manager [instance: 7fcee323-8ea6-4f69-9f1e-f2105af0fe96] result = hub.switch() [ 580.590716] env[61806]: ERROR nova.compute.manager [instance: 7fcee323-8ea6-4f69-9f1e-f2105af0fe96] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 580.590716] env[61806]: ERROR nova.compute.manager [instance: 7fcee323-8ea6-4f69-9f1e-f2105af0fe96] return self.greenlet.switch() [ 580.591173] env[61806]: ERROR nova.compute.manager [instance: 7fcee323-8ea6-4f69-9f1e-f2105af0fe96] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 580.591173] env[61806]: ERROR nova.compute.manager [instance: 7fcee323-8ea6-4f69-9f1e-f2105af0fe96] result = function(*args, **kwargs) [ 580.591173] env[61806]: ERROR nova.compute.manager [instance: 7fcee323-8ea6-4f69-9f1e-f2105af0fe96] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 580.591173] env[61806]: ERROR nova.compute.manager [instance: 7fcee323-8ea6-4f69-9f1e-f2105af0fe96] return func(*args, **kwargs) [ 580.591173] env[61806]: ERROR nova.compute.manager [instance: 7fcee323-8ea6-4f69-9f1e-f2105af0fe96] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 580.591173] env[61806]: ERROR nova.compute.manager [instance: 7fcee323-8ea6-4f69-9f1e-f2105af0fe96] raise e [ 580.591173] env[61806]: ERROR nova.compute.manager [instance: 7fcee323-8ea6-4f69-9f1e-f2105af0fe96] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 580.591173] env[61806]: ERROR nova.compute.manager [instance: 7fcee323-8ea6-4f69-9f1e-f2105af0fe96] nwinfo = self.network_api.allocate_for_instance( [ 580.591173] env[61806]: ERROR nova.compute.manager [instance: 7fcee323-8ea6-4f69-9f1e-f2105af0fe96] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 580.591173] env[61806]: ERROR nova.compute.manager [instance: 7fcee323-8ea6-4f69-9f1e-f2105af0fe96] created_port_ids = self._update_ports_for_instance( [ 580.591173] env[61806]: ERROR nova.compute.manager [instance: 7fcee323-8ea6-4f69-9f1e-f2105af0fe96] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 580.591173] env[61806]: ERROR nova.compute.manager [instance: 7fcee323-8ea6-4f69-9f1e-f2105af0fe96] with excutils.save_and_reraise_exception(): [ 580.591173] env[61806]: ERROR nova.compute.manager [instance: 7fcee323-8ea6-4f69-9f1e-f2105af0fe96] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 580.591596] env[61806]: ERROR nova.compute.manager [instance: 7fcee323-8ea6-4f69-9f1e-f2105af0fe96] self.force_reraise() [ 580.591596] env[61806]: ERROR nova.compute.manager [instance: 7fcee323-8ea6-4f69-9f1e-f2105af0fe96] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 580.591596] env[61806]: ERROR nova.compute.manager [instance: 7fcee323-8ea6-4f69-9f1e-f2105af0fe96] raise self.value [ 580.591596] env[61806]: ERROR nova.compute.manager [instance: 7fcee323-8ea6-4f69-9f1e-f2105af0fe96] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 580.591596] env[61806]: ERROR nova.compute.manager [instance: 7fcee323-8ea6-4f69-9f1e-f2105af0fe96] updated_port = self._update_port( [ 580.591596] env[61806]: ERROR nova.compute.manager [instance: 7fcee323-8ea6-4f69-9f1e-f2105af0fe96] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 580.591596] env[61806]: ERROR nova.compute.manager [instance: 7fcee323-8ea6-4f69-9f1e-f2105af0fe96] _ensure_no_port_binding_failure(port) [ 580.591596] env[61806]: ERROR nova.compute.manager [instance: 7fcee323-8ea6-4f69-9f1e-f2105af0fe96] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 580.591596] env[61806]: ERROR nova.compute.manager [instance: 7fcee323-8ea6-4f69-9f1e-f2105af0fe96] raise exception.PortBindingFailed(port_id=port['id']) [ 580.591596] env[61806]: ERROR nova.compute.manager [instance: 7fcee323-8ea6-4f69-9f1e-f2105af0fe96] nova.exception.PortBindingFailed: Binding failed for port eed1949d-29dd-4127-a2a4-d7dc2369df55, please check neutron logs for more information. [ 580.591596] env[61806]: ERROR nova.compute.manager [instance: 7fcee323-8ea6-4f69-9f1e-f2105af0fe96] [ 580.591991] env[61806]: DEBUG nova.compute.utils [None req-aac795d3-c37f-4811-9be9-aa67ad520ab0 tempest-ListServerFiltersTestJSON-241228521 tempest-ListServerFiltersTestJSON-241228521-project-member] [instance: 7fcee323-8ea6-4f69-9f1e-f2105af0fe96] Binding failed for port eed1949d-29dd-4127-a2a4-d7dc2369df55, please check neutron logs for more information. {{(pid=61806) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 580.592643] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-515b6f7c-079d-4b77-a58f-24e4d6aa7eca {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 580.595978] env[61806]: DEBUG nova.compute.manager [None req-aac795d3-c37f-4811-9be9-aa67ad520ab0 tempest-ListServerFiltersTestJSON-241228521 tempest-ListServerFiltersTestJSON-241228521-project-member] [instance: 7fcee323-8ea6-4f69-9f1e-f2105af0fe96] Build of instance 7fcee323-8ea6-4f69-9f1e-f2105af0fe96 was re-scheduled: Binding failed for port eed1949d-29dd-4127-a2a4-d7dc2369df55, please check neutron logs for more information. {{(pid=61806) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 580.596428] env[61806]: DEBUG nova.compute.manager [None req-aac795d3-c37f-4811-9be9-aa67ad520ab0 tempest-ListServerFiltersTestJSON-241228521 tempest-ListServerFiltersTestJSON-241228521-project-member] [instance: 7fcee323-8ea6-4f69-9f1e-f2105af0fe96] Unplugging VIFs for instance {{(pid=61806) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 580.596650] env[61806]: DEBUG oslo_concurrency.lockutils [None req-aac795d3-c37f-4811-9be9-aa67ad520ab0 tempest-ListServerFiltersTestJSON-241228521 tempest-ListServerFiltersTestJSON-241228521-project-member] Acquiring lock "refresh_cache-7fcee323-8ea6-4f69-9f1e-f2105af0fe96" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 580.596817] env[61806]: DEBUG oslo_concurrency.lockutils [None req-aac795d3-c37f-4811-9be9-aa67ad520ab0 tempest-ListServerFiltersTestJSON-241228521 tempest-ListServerFiltersTestJSON-241228521-project-member] Acquired lock "refresh_cache-7fcee323-8ea6-4f69-9f1e-f2105af0fe96" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 580.600020] env[61806]: DEBUG nova.network.neutron [None req-aac795d3-c37f-4811-9be9-aa67ad520ab0 tempest-ListServerFiltersTestJSON-241228521 tempest-ListServerFiltersTestJSON-241228521-project-member] [instance: 7fcee323-8ea6-4f69-9f1e-f2105af0fe96] Building network info cache for instance {{(pid=61806) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 580.600020] env[61806]: DEBUG oslo_concurrency.lockutils [None req-55a61261-0862-4d73-b971-6d11cbe88c5e tempest-ServersAdminTestJSON-14199920 tempest-ServersAdminTestJSON-14199920-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 24.489s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 580.611732] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7ada3310-2b25-4643-a4fb-fbdf007f736f {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 580.771374] env[61806]: DEBUG nova.network.neutron [-] [instance: 8905f20f-28a6-49f4-88bf-e52177c6a4b6] Updating instance_info_cache with network_info: [] {{(pid=61806) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 580.805621] env[61806]: ERROR nova.compute.manager [None req-37e71da7-cb5e-431d-ba3c-95daaf8af7f3 tempest-ServersAdminNegativeTestJSON-2128564044 tempest-ServersAdminNegativeTestJSON-2128564044-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 63122136-e392-4570-8f9c-7785b771ac82, please check neutron logs for more information. [ 580.805621] env[61806]: ERROR nova.compute.manager Traceback (most recent call last): [ 580.805621] env[61806]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 580.805621] env[61806]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 580.805621] env[61806]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 580.805621] env[61806]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 580.805621] env[61806]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 580.805621] env[61806]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 580.805621] env[61806]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 580.805621] env[61806]: ERROR nova.compute.manager self.force_reraise() [ 580.805621] env[61806]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 580.805621] env[61806]: ERROR nova.compute.manager raise self.value [ 580.805621] env[61806]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 580.805621] env[61806]: ERROR nova.compute.manager updated_port = self._update_port( [ 580.805621] env[61806]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 580.805621] env[61806]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 580.806442] env[61806]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 580.806442] env[61806]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 580.806442] env[61806]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 63122136-e392-4570-8f9c-7785b771ac82, please check neutron logs for more information. [ 580.806442] env[61806]: ERROR nova.compute.manager [ 580.806442] env[61806]: Traceback (most recent call last): [ 580.806442] env[61806]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 580.806442] env[61806]: listener.cb(fileno) [ 580.806442] env[61806]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 580.806442] env[61806]: result = function(*args, **kwargs) [ 580.806442] env[61806]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 580.806442] env[61806]: return func(*args, **kwargs) [ 580.806442] env[61806]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 580.806442] env[61806]: raise e [ 580.806442] env[61806]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 580.806442] env[61806]: nwinfo = self.network_api.allocate_for_instance( [ 580.806442] env[61806]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 580.806442] env[61806]: created_port_ids = self._update_ports_for_instance( [ 580.806442] env[61806]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 580.806442] env[61806]: with excutils.save_and_reraise_exception(): [ 580.806442] env[61806]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 580.806442] env[61806]: self.force_reraise() [ 580.806442] env[61806]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 580.806442] env[61806]: raise self.value [ 580.806442] env[61806]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 580.806442] env[61806]: updated_port = self._update_port( [ 580.806442] env[61806]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 580.806442] env[61806]: _ensure_no_port_binding_failure(port) [ 580.806442] env[61806]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 580.806442] env[61806]: raise exception.PortBindingFailed(port_id=port['id']) [ 580.807332] env[61806]: nova.exception.PortBindingFailed: Binding failed for port 63122136-e392-4570-8f9c-7785b771ac82, please check neutron logs for more information. [ 580.807332] env[61806]: Removing descriptor: 17 [ 580.807332] env[61806]: ERROR nova.compute.manager [None req-37e71da7-cb5e-431d-ba3c-95daaf8af7f3 tempest-ServersAdminNegativeTestJSON-2128564044 tempest-ServersAdminNegativeTestJSON-2128564044-project-member] [instance: 89ce0fd8-6a22-4b1c-bdff-d1584abd538a] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 63122136-e392-4570-8f9c-7785b771ac82, please check neutron logs for more information. [ 580.807332] env[61806]: ERROR nova.compute.manager [instance: 89ce0fd8-6a22-4b1c-bdff-d1584abd538a] Traceback (most recent call last): [ 580.807332] env[61806]: ERROR nova.compute.manager [instance: 89ce0fd8-6a22-4b1c-bdff-d1584abd538a] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 580.807332] env[61806]: ERROR nova.compute.manager [instance: 89ce0fd8-6a22-4b1c-bdff-d1584abd538a] yield resources [ 580.807332] env[61806]: ERROR nova.compute.manager [instance: 89ce0fd8-6a22-4b1c-bdff-d1584abd538a] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 580.807332] env[61806]: ERROR nova.compute.manager [instance: 89ce0fd8-6a22-4b1c-bdff-d1584abd538a] self.driver.spawn(context, instance, image_meta, [ 580.807332] env[61806]: ERROR nova.compute.manager [instance: 89ce0fd8-6a22-4b1c-bdff-d1584abd538a] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 580.807332] env[61806]: ERROR nova.compute.manager [instance: 89ce0fd8-6a22-4b1c-bdff-d1584abd538a] self._vmops.spawn(context, instance, image_meta, injected_files, [ 580.807332] env[61806]: ERROR nova.compute.manager [instance: 89ce0fd8-6a22-4b1c-bdff-d1584abd538a] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 580.807332] env[61806]: ERROR nova.compute.manager [instance: 89ce0fd8-6a22-4b1c-bdff-d1584abd538a] vm_ref = self.build_virtual_machine(instance, [ 580.807760] env[61806]: ERROR nova.compute.manager [instance: 89ce0fd8-6a22-4b1c-bdff-d1584abd538a] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 580.807760] env[61806]: ERROR nova.compute.manager [instance: 89ce0fd8-6a22-4b1c-bdff-d1584abd538a] vif_infos = vmwarevif.get_vif_info(self._session, [ 580.807760] env[61806]: ERROR nova.compute.manager [instance: 89ce0fd8-6a22-4b1c-bdff-d1584abd538a] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 580.807760] env[61806]: ERROR nova.compute.manager [instance: 89ce0fd8-6a22-4b1c-bdff-d1584abd538a] for vif in network_info: [ 580.807760] env[61806]: ERROR nova.compute.manager [instance: 89ce0fd8-6a22-4b1c-bdff-d1584abd538a] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 580.807760] env[61806]: ERROR nova.compute.manager [instance: 89ce0fd8-6a22-4b1c-bdff-d1584abd538a] return self._sync_wrapper(fn, *args, **kwargs) [ 580.807760] env[61806]: ERROR nova.compute.manager [instance: 89ce0fd8-6a22-4b1c-bdff-d1584abd538a] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 580.807760] env[61806]: ERROR nova.compute.manager [instance: 89ce0fd8-6a22-4b1c-bdff-d1584abd538a] self.wait() [ 580.807760] env[61806]: ERROR nova.compute.manager [instance: 89ce0fd8-6a22-4b1c-bdff-d1584abd538a] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 580.807760] env[61806]: ERROR nova.compute.manager [instance: 89ce0fd8-6a22-4b1c-bdff-d1584abd538a] self[:] = self._gt.wait() [ 580.807760] env[61806]: ERROR nova.compute.manager [instance: 89ce0fd8-6a22-4b1c-bdff-d1584abd538a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 580.807760] env[61806]: ERROR nova.compute.manager [instance: 89ce0fd8-6a22-4b1c-bdff-d1584abd538a] return self._exit_event.wait() [ 580.807760] env[61806]: ERROR nova.compute.manager [instance: 89ce0fd8-6a22-4b1c-bdff-d1584abd538a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 580.808244] env[61806]: ERROR nova.compute.manager [instance: 89ce0fd8-6a22-4b1c-bdff-d1584abd538a] result = hub.switch() [ 580.808244] env[61806]: ERROR nova.compute.manager [instance: 89ce0fd8-6a22-4b1c-bdff-d1584abd538a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 580.808244] env[61806]: ERROR nova.compute.manager [instance: 89ce0fd8-6a22-4b1c-bdff-d1584abd538a] return self.greenlet.switch() [ 580.808244] env[61806]: ERROR nova.compute.manager [instance: 89ce0fd8-6a22-4b1c-bdff-d1584abd538a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 580.808244] env[61806]: ERROR nova.compute.manager [instance: 89ce0fd8-6a22-4b1c-bdff-d1584abd538a] result = function(*args, **kwargs) [ 580.808244] env[61806]: ERROR nova.compute.manager [instance: 89ce0fd8-6a22-4b1c-bdff-d1584abd538a] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 580.808244] env[61806]: ERROR nova.compute.manager [instance: 89ce0fd8-6a22-4b1c-bdff-d1584abd538a] return func(*args, **kwargs) [ 580.808244] env[61806]: ERROR nova.compute.manager [instance: 89ce0fd8-6a22-4b1c-bdff-d1584abd538a] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 580.808244] env[61806]: ERROR nova.compute.manager [instance: 89ce0fd8-6a22-4b1c-bdff-d1584abd538a] raise e [ 580.808244] env[61806]: ERROR nova.compute.manager [instance: 89ce0fd8-6a22-4b1c-bdff-d1584abd538a] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 580.808244] env[61806]: ERROR nova.compute.manager [instance: 89ce0fd8-6a22-4b1c-bdff-d1584abd538a] nwinfo = self.network_api.allocate_for_instance( [ 580.808244] env[61806]: ERROR nova.compute.manager [instance: 89ce0fd8-6a22-4b1c-bdff-d1584abd538a] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 580.808244] env[61806]: ERROR nova.compute.manager [instance: 89ce0fd8-6a22-4b1c-bdff-d1584abd538a] created_port_ids = self._update_ports_for_instance( [ 580.808692] env[61806]: ERROR nova.compute.manager [instance: 89ce0fd8-6a22-4b1c-bdff-d1584abd538a] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 580.808692] env[61806]: ERROR nova.compute.manager [instance: 89ce0fd8-6a22-4b1c-bdff-d1584abd538a] with excutils.save_and_reraise_exception(): [ 580.808692] env[61806]: ERROR nova.compute.manager [instance: 89ce0fd8-6a22-4b1c-bdff-d1584abd538a] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 580.808692] env[61806]: ERROR nova.compute.manager [instance: 89ce0fd8-6a22-4b1c-bdff-d1584abd538a] self.force_reraise() [ 580.808692] env[61806]: ERROR nova.compute.manager [instance: 89ce0fd8-6a22-4b1c-bdff-d1584abd538a] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 580.808692] env[61806]: ERROR nova.compute.manager [instance: 89ce0fd8-6a22-4b1c-bdff-d1584abd538a] raise self.value [ 580.808692] env[61806]: ERROR nova.compute.manager [instance: 89ce0fd8-6a22-4b1c-bdff-d1584abd538a] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 580.808692] env[61806]: ERROR nova.compute.manager [instance: 89ce0fd8-6a22-4b1c-bdff-d1584abd538a] updated_port = self._update_port( [ 580.808692] env[61806]: ERROR nova.compute.manager [instance: 89ce0fd8-6a22-4b1c-bdff-d1584abd538a] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 580.808692] env[61806]: ERROR nova.compute.manager [instance: 89ce0fd8-6a22-4b1c-bdff-d1584abd538a] _ensure_no_port_binding_failure(port) [ 580.808692] env[61806]: ERROR nova.compute.manager [instance: 89ce0fd8-6a22-4b1c-bdff-d1584abd538a] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 580.808692] env[61806]: ERROR nova.compute.manager [instance: 89ce0fd8-6a22-4b1c-bdff-d1584abd538a] raise exception.PortBindingFailed(port_id=port['id']) [ 580.809139] env[61806]: ERROR nova.compute.manager [instance: 89ce0fd8-6a22-4b1c-bdff-d1584abd538a] nova.exception.PortBindingFailed: Binding failed for port 63122136-e392-4570-8f9c-7785b771ac82, please check neutron logs for more information. [ 580.809139] env[61806]: ERROR nova.compute.manager [instance: 89ce0fd8-6a22-4b1c-bdff-d1584abd538a] [ 580.809139] env[61806]: INFO nova.compute.manager [None req-37e71da7-cb5e-431d-ba3c-95daaf8af7f3 tempest-ServersAdminNegativeTestJSON-2128564044 tempest-ServersAdminNegativeTestJSON-2128564044-project-member] [instance: 89ce0fd8-6a22-4b1c-bdff-d1584abd538a] Terminating instance [ 580.809139] env[61806]: DEBUG oslo_concurrency.lockutils [None req-37e71da7-cb5e-431d-ba3c-95daaf8af7f3 tempest-ServersAdminNegativeTestJSON-2128564044 tempest-ServersAdminNegativeTestJSON-2128564044-project-member] Acquiring lock "refresh_cache-89ce0fd8-6a22-4b1c-bdff-d1584abd538a" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 580.809139] env[61806]: DEBUG oslo_concurrency.lockutils [None req-37e71da7-cb5e-431d-ba3c-95daaf8af7f3 tempest-ServersAdminNegativeTestJSON-2128564044 tempest-ServersAdminNegativeTestJSON-2128564044-project-member] Acquired lock "refresh_cache-89ce0fd8-6a22-4b1c-bdff-d1584abd538a" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 580.809139] env[61806]: DEBUG nova.network.neutron [None req-37e71da7-cb5e-431d-ba3c-95daaf8af7f3 tempest-ServersAdminNegativeTestJSON-2128564044 tempest-ServersAdminNegativeTestJSON-2128564044-project-member] [instance: 89ce0fd8-6a22-4b1c-bdff-d1584abd538a] Building network info cache for instance {{(pid=61806) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 581.131302] env[61806]: DEBUG nova.network.neutron [None req-aac795d3-c37f-4811-9be9-aa67ad520ab0 tempest-ListServerFiltersTestJSON-241228521 tempest-ListServerFiltersTestJSON-241228521-project-member] [instance: 7fcee323-8ea6-4f69-9f1e-f2105af0fe96] Instance cache missing network info. {{(pid=61806) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 581.223201] env[61806]: DEBUG nova.network.neutron [None req-aac795d3-c37f-4811-9be9-aa67ad520ab0 tempest-ListServerFiltersTestJSON-241228521 tempest-ListServerFiltersTestJSON-241228521-project-member] [instance: 7fcee323-8ea6-4f69-9f1e-f2105af0fe96] Updating instance_info_cache with network_info: [] {{(pid=61806) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 581.275406] env[61806]: INFO nova.compute.manager [-] [instance: 8905f20f-28a6-49f4-88bf-e52177c6a4b6] Took 1.03 seconds to deallocate network for instance. [ 581.277870] env[61806]: DEBUG nova.compute.claims [None req-38c4eaec-22d8-4bf0-bd45-321ef43329a0 tempest-ServersAdminTestJSON-14199920 tempest-ServersAdminTestJSON-14199920-project-member] [instance: 8905f20f-28a6-49f4-88bf-e52177c6a4b6] Aborting claim: {{(pid=61806) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 581.278066] env[61806]: DEBUG oslo_concurrency.lockutils [None req-38c4eaec-22d8-4bf0-bd45-321ef43329a0 tempest-ServersAdminTestJSON-14199920 tempest-ServersAdminTestJSON-14199920-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 581.326793] env[61806]: DEBUG nova.network.neutron [None req-37e71da7-cb5e-431d-ba3c-95daaf8af7f3 tempest-ServersAdminNegativeTestJSON-2128564044 tempest-ServersAdminNegativeTestJSON-2128564044-project-member] [instance: 89ce0fd8-6a22-4b1c-bdff-d1584abd538a] Instance cache missing network info. {{(pid=61806) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 581.379504] env[61806]: DEBUG nova.network.neutron [None req-37e71da7-cb5e-431d-ba3c-95daaf8af7f3 tempest-ServersAdminNegativeTestJSON-2128564044 tempest-ServersAdminNegativeTestJSON-2128564044-project-member] [instance: 89ce0fd8-6a22-4b1c-bdff-d1584abd538a] Updating instance_info_cache with network_info: [] {{(pid=61806) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 581.474227] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-01cb0ea2-aac5-4ca7-a65d-7758b3d0176a {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 581.483179] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6ef2af91-9415-4124-8c15-3d566cec1029 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 581.512563] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f08630e7-3559-483c-bcda-e7a3104f286c {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 581.520055] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-461175f2-0c46-4662-a3d4-e87ebeaa2c9e {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 581.534825] env[61806]: DEBUG nova.compute.provider_tree [None req-55a61261-0862-4d73-b971-6d11cbe88c5e tempest-ServersAdminTestJSON-14199920 tempest-ServersAdminTestJSON-14199920-project-member] Inventory has not changed in ProviderTree for provider: a2858be1-fd22-4e08-979e-87ad25293407 {{(pid=61806) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 581.728608] env[61806]: DEBUG oslo_concurrency.lockutils [None req-aac795d3-c37f-4811-9be9-aa67ad520ab0 tempest-ListServerFiltersTestJSON-241228521 tempest-ListServerFiltersTestJSON-241228521-project-member] Releasing lock "refresh_cache-7fcee323-8ea6-4f69-9f1e-f2105af0fe96" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 581.728868] env[61806]: DEBUG nova.compute.manager [None req-aac795d3-c37f-4811-9be9-aa67ad520ab0 tempest-ListServerFiltersTestJSON-241228521 tempest-ListServerFiltersTestJSON-241228521-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61806) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 581.729068] env[61806]: DEBUG nova.compute.manager [None req-aac795d3-c37f-4811-9be9-aa67ad520ab0 tempest-ListServerFiltersTestJSON-241228521 tempest-ListServerFiltersTestJSON-241228521-project-member] [instance: 7fcee323-8ea6-4f69-9f1e-f2105af0fe96] Deallocating network for instance {{(pid=61806) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 581.729277] env[61806]: DEBUG nova.network.neutron [None req-aac795d3-c37f-4811-9be9-aa67ad520ab0 tempest-ListServerFiltersTestJSON-241228521 tempest-ListServerFiltersTestJSON-241228521-project-member] [instance: 7fcee323-8ea6-4f69-9f1e-f2105af0fe96] deallocate_for_instance() {{(pid=61806) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 581.745323] env[61806]: DEBUG nova.network.neutron [None req-aac795d3-c37f-4811-9be9-aa67ad520ab0 tempest-ListServerFiltersTestJSON-241228521 tempest-ListServerFiltersTestJSON-241228521-project-member] [instance: 7fcee323-8ea6-4f69-9f1e-f2105af0fe96] Instance cache missing network info. {{(pid=61806) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 581.883172] env[61806]: DEBUG oslo_concurrency.lockutils [None req-37e71da7-cb5e-431d-ba3c-95daaf8af7f3 tempest-ServersAdminNegativeTestJSON-2128564044 tempest-ServersAdminNegativeTestJSON-2128564044-project-member] Releasing lock "refresh_cache-89ce0fd8-6a22-4b1c-bdff-d1584abd538a" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 581.883172] env[61806]: DEBUG nova.compute.manager [None req-37e71da7-cb5e-431d-ba3c-95daaf8af7f3 tempest-ServersAdminNegativeTestJSON-2128564044 tempest-ServersAdminNegativeTestJSON-2128564044-project-member] [instance: 89ce0fd8-6a22-4b1c-bdff-d1584abd538a] Start destroying the instance on the hypervisor. {{(pid=61806) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 581.883172] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-37e71da7-cb5e-431d-ba3c-95daaf8af7f3 tempest-ServersAdminNegativeTestJSON-2128564044 tempest-ServersAdminNegativeTestJSON-2128564044-project-member] [instance: 89ce0fd8-6a22-4b1c-bdff-d1584abd538a] Destroying instance {{(pid=61806) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 581.883172] env[61806]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-f6818f2e-d3ab-4076-b2ac-cc231027586e {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 581.892085] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1dcd6858-1d81-4a38-8cb6-9485d4abf92a {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 581.916216] env[61806]: WARNING nova.virt.vmwareapi.vmops [None req-37e71da7-cb5e-431d-ba3c-95daaf8af7f3 tempest-ServersAdminNegativeTestJSON-2128564044 tempest-ServersAdminNegativeTestJSON-2128564044-project-member] [instance: 89ce0fd8-6a22-4b1c-bdff-d1584abd538a] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 89ce0fd8-6a22-4b1c-bdff-d1584abd538a could not be found. [ 581.916582] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-37e71da7-cb5e-431d-ba3c-95daaf8af7f3 tempest-ServersAdminNegativeTestJSON-2128564044 tempest-ServersAdminNegativeTestJSON-2128564044-project-member] [instance: 89ce0fd8-6a22-4b1c-bdff-d1584abd538a] Instance destroyed {{(pid=61806) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 581.916781] env[61806]: INFO nova.compute.manager [None req-37e71da7-cb5e-431d-ba3c-95daaf8af7f3 tempest-ServersAdminNegativeTestJSON-2128564044 tempest-ServersAdminNegativeTestJSON-2128564044-project-member] [instance: 89ce0fd8-6a22-4b1c-bdff-d1584abd538a] Took 0.03 seconds to destroy the instance on the hypervisor. [ 581.917038] env[61806]: DEBUG oslo.service.loopingcall [None req-37e71da7-cb5e-431d-ba3c-95daaf8af7f3 tempest-ServersAdminNegativeTestJSON-2128564044 tempest-ServersAdminNegativeTestJSON-2128564044-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61806) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 581.917237] env[61806]: DEBUG nova.compute.manager [-] [instance: 89ce0fd8-6a22-4b1c-bdff-d1584abd538a] Deallocating network for instance {{(pid=61806) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 581.917328] env[61806]: DEBUG nova.network.neutron [-] [instance: 89ce0fd8-6a22-4b1c-bdff-d1584abd538a] deallocate_for_instance() {{(pid=61806) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 581.937087] env[61806]: DEBUG nova.network.neutron [-] [instance: 89ce0fd8-6a22-4b1c-bdff-d1584abd538a] Instance cache missing network info. {{(pid=61806) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 582.036832] env[61806]: DEBUG nova.scheduler.client.report [None req-55a61261-0862-4d73-b971-6d11cbe88c5e tempest-ServersAdminTestJSON-14199920 tempest-ServersAdminTestJSON-14199920-project-member] Inventory has not changed for provider a2858be1-fd22-4e08-979e-87ad25293407 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 139, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61806) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 582.090090] env[61806]: DEBUG nova.compute.manager [req-ed54a5fa-db5b-418c-bd9a-b726b86c7538 req-a19ea44c-8946-4f2b-ad6f-3587bd5dce88 service nova] [instance: 89ce0fd8-6a22-4b1c-bdff-d1584abd538a] Received event network-changed-63122136-e392-4570-8f9c-7785b771ac82 {{(pid=61806) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 582.090337] env[61806]: DEBUG nova.compute.manager [req-ed54a5fa-db5b-418c-bd9a-b726b86c7538 req-a19ea44c-8946-4f2b-ad6f-3587bd5dce88 service nova] [instance: 89ce0fd8-6a22-4b1c-bdff-d1584abd538a] Refreshing instance network info cache due to event network-changed-63122136-e392-4570-8f9c-7785b771ac82. {{(pid=61806) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 582.090638] env[61806]: DEBUG oslo_concurrency.lockutils [req-ed54a5fa-db5b-418c-bd9a-b726b86c7538 req-a19ea44c-8946-4f2b-ad6f-3587bd5dce88 service nova] Acquiring lock "refresh_cache-89ce0fd8-6a22-4b1c-bdff-d1584abd538a" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 582.090880] env[61806]: DEBUG oslo_concurrency.lockutils [req-ed54a5fa-db5b-418c-bd9a-b726b86c7538 req-a19ea44c-8946-4f2b-ad6f-3587bd5dce88 service nova] Acquired lock "refresh_cache-89ce0fd8-6a22-4b1c-bdff-d1584abd538a" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 582.091210] env[61806]: DEBUG nova.network.neutron [req-ed54a5fa-db5b-418c-bd9a-b726b86c7538 req-a19ea44c-8946-4f2b-ad6f-3587bd5dce88 service nova] [instance: 89ce0fd8-6a22-4b1c-bdff-d1584abd538a] Refreshing network info cache for port 63122136-e392-4570-8f9c-7785b771ac82 {{(pid=61806) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 582.248843] env[61806]: DEBUG nova.network.neutron [None req-aac795d3-c37f-4811-9be9-aa67ad520ab0 tempest-ListServerFiltersTestJSON-241228521 tempest-ListServerFiltersTestJSON-241228521-project-member] [instance: 7fcee323-8ea6-4f69-9f1e-f2105af0fe96] Updating instance_info_cache with network_info: [] {{(pid=61806) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 582.441466] env[61806]: DEBUG nova.network.neutron [-] [instance: 89ce0fd8-6a22-4b1c-bdff-d1584abd538a] Updating instance_info_cache with network_info: [] {{(pid=61806) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 582.543953] env[61806]: DEBUG oslo_concurrency.lockutils [None req-55a61261-0862-4d73-b971-6d11cbe88c5e tempest-ServersAdminTestJSON-14199920 tempest-ServersAdminTestJSON-14199920-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.945s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 582.544140] env[61806]: ERROR nova.compute.manager [None req-55a61261-0862-4d73-b971-6d11cbe88c5e tempest-ServersAdminTestJSON-14199920 tempest-ServersAdminTestJSON-14199920-project-member] [instance: 26a0baf9-68bd-4380-87b3-53f50524bbbf] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 31cd9389-2251-4222-885e-268c99301a23, please check neutron logs for more information. [ 582.544140] env[61806]: ERROR nova.compute.manager [instance: 26a0baf9-68bd-4380-87b3-53f50524bbbf] Traceback (most recent call last): [ 582.544140] env[61806]: ERROR nova.compute.manager [instance: 26a0baf9-68bd-4380-87b3-53f50524bbbf] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 582.544140] env[61806]: ERROR nova.compute.manager [instance: 26a0baf9-68bd-4380-87b3-53f50524bbbf] self.driver.spawn(context, instance, image_meta, [ 582.544140] env[61806]: ERROR nova.compute.manager [instance: 26a0baf9-68bd-4380-87b3-53f50524bbbf] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 582.544140] env[61806]: ERROR nova.compute.manager [instance: 26a0baf9-68bd-4380-87b3-53f50524bbbf] self._vmops.spawn(context, instance, image_meta, injected_files, [ 582.544140] env[61806]: ERROR nova.compute.manager [instance: 26a0baf9-68bd-4380-87b3-53f50524bbbf] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 582.544140] env[61806]: ERROR nova.compute.manager [instance: 26a0baf9-68bd-4380-87b3-53f50524bbbf] vm_ref = self.build_virtual_machine(instance, [ 582.544140] env[61806]: ERROR nova.compute.manager [instance: 26a0baf9-68bd-4380-87b3-53f50524bbbf] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 582.544140] env[61806]: ERROR nova.compute.manager [instance: 26a0baf9-68bd-4380-87b3-53f50524bbbf] vif_infos = vmwarevif.get_vif_info(self._session, [ 582.544140] env[61806]: ERROR nova.compute.manager [instance: 26a0baf9-68bd-4380-87b3-53f50524bbbf] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 582.544506] env[61806]: ERROR nova.compute.manager [instance: 26a0baf9-68bd-4380-87b3-53f50524bbbf] for vif in network_info: [ 582.544506] env[61806]: ERROR nova.compute.manager [instance: 26a0baf9-68bd-4380-87b3-53f50524bbbf] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 582.544506] env[61806]: ERROR nova.compute.manager [instance: 26a0baf9-68bd-4380-87b3-53f50524bbbf] return self._sync_wrapper(fn, *args, **kwargs) [ 582.544506] env[61806]: ERROR nova.compute.manager [instance: 26a0baf9-68bd-4380-87b3-53f50524bbbf] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 582.544506] env[61806]: ERROR nova.compute.manager [instance: 26a0baf9-68bd-4380-87b3-53f50524bbbf] self.wait() [ 582.544506] env[61806]: ERROR nova.compute.manager [instance: 26a0baf9-68bd-4380-87b3-53f50524bbbf] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 582.544506] env[61806]: ERROR nova.compute.manager [instance: 26a0baf9-68bd-4380-87b3-53f50524bbbf] self[:] = self._gt.wait() [ 582.544506] env[61806]: ERROR nova.compute.manager [instance: 26a0baf9-68bd-4380-87b3-53f50524bbbf] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 582.544506] env[61806]: ERROR nova.compute.manager [instance: 26a0baf9-68bd-4380-87b3-53f50524bbbf] return self._exit_event.wait() [ 582.544506] env[61806]: ERROR nova.compute.manager [instance: 26a0baf9-68bd-4380-87b3-53f50524bbbf] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 582.544506] env[61806]: ERROR nova.compute.manager [instance: 26a0baf9-68bd-4380-87b3-53f50524bbbf] result = hub.switch() [ 582.544506] env[61806]: ERROR nova.compute.manager [instance: 26a0baf9-68bd-4380-87b3-53f50524bbbf] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 582.544506] env[61806]: ERROR nova.compute.manager [instance: 26a0baf9-68bd-4380-87b3-53f50524bbbf] return self.greenlet.switch() [ 582.544969] env[61806]: ERROR nova.compute.manager [instance: 26a0baf9-68bd-4380-87b3-53f50524bbbf] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 582.544969] env[61806]: ERROR nova.compute.manager [instance: 26a0baf9-68bd-4380-87b3-53f50524bbbf] result = function(*args, **kwargs) [ 582.544969] env[61806]: ERROR nova.compute.manager [instance: 26a0baf9-68bd-4380-87b3-53f50524bbbf] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 582.544969] env[61806]: ERROR nova.compute.manager [instance: 26a0baf9-68bd-4380-87b3-53f50524bbbf] return func(*args, **kwargs) [ 582.544969] env[61806]: ERROR nova.compute.manager [instance: 26a0baf9-68bd-4380-87b3-53f50524bbbf] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 582.544969] env[61806]: ERROR nova.compute.manager [instance: 26a0baf9-68bd-4380-87b3-53f50524bbbf] raise e [ 582.544969] env[61806]: ERROR nova.compute.manager [instance: 26a0baf9-68bd-4380-87b3-53f50524bbbf] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 582.544969] env[61806]: ERROR nova.compute.manager [instance: 26a0baf9-68bd-4380-87b3-53f50524bbbf] nwinfo = self.network_api.allocate_for_instance( [ 582.544969] env[61806]: ERROR nova.compute.manager [instance: 26a0baf9-68bd-4380-87b3-53f50524bbbf] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 582.544969] env[61806]: ERROR nova.compute.manager [instance: 26a0baf9-68bd-4380-87b3-53f50524bbbf] created_port_ids = self._update_ports_for_instance( [ 582.544969] env[61806]: ERROR nova.compute.manager [instance: 26a0baf9-68bd-4380-87b3-53f50524bbbf] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 582.544969] env[61806]: ERROR nova.compute.manager [instance: 26a0baf9-68bd-4380-87b3-53f50524bbbf] with excutils.save_and_reraise_exception(): [ 582.544969] env[61806]: ERROR nova.compute.manager [instance: 26a0baf9-68bd-4380-87b3-53f50524bbbf] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 582.545471] env[61806]: ERROR nova.compute.manager [instance: 26a0baf9-68bd-4380-87b3-53f50524bbbf] self.force_reraise() [ 582.545471] env[61806]: ERROR nova.compute.manager [instance: 26a0baf9-68bd-4380-87b3-53f50524bbbf] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 582.545471] env[61806]: ERROR nova.compute.manager [instance: 26a0baf9-68bd-4380-87b3-53f50524bbbf] raise self.value [ 582.545471] env[61806]: ERROR nova.compute.manager [instance: 26a0baf9-68bd-4380-87b3-53f50524bbbf] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 582.545471] env[61806]: ERROR nova.compute.manager [instance: 26a0baf9-68bd-4380-87b3-53f50524bbbf] updated_port = self._update_port( [ 582.545471] env[61806]: ERROR nova.compute.manager [instance: 26a0baf9-68bd-4380-87b3-53f50524bbbf] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 582.545471] env[61806]: ERROR nova.compute.manager [instance: 26a0baf9-68bd-4380-87b3-53f50524bbbf] _ensure_no_port_binding_failure(port) [ 582.545471] env[61806]: ERROR nova.compute.manager [instance: 26a0baf9-68bd-4380-87b3-53f50524bbbf] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 582.545471] env[61806]: ERROR nova.compute.manager [instance: 26a0baf9-68bd-4380-87b3-53f50524bbbf] raise exception.PortBindingFailed(port_id=port['id']) [ 582.545471] env[61806]: ERROR nova.compute.manager [instance: 26a0baf9-68bd-4380-87b3-53f50524bbbf] nova.exception.PortBindingFailed: Binding failed for port 31cd9389-2251-4222-885e-268c99301a23, please check neutron logs for more information. [ 582.545471] env[61806]: ERROR nova.compute.manager [instance: 26a0baf9-68bd-4380-87b3-53f50524bbbf] [ 582.545790] env[61806]: DEBUG nova.compute.utils [None req-55a61261-0862-4d73-b971-6d11cbe88c5e tempest-ServersAdminTestJSON-14199920 tempest-ServersAdminTestJSON-14199920-project-member] [instance: 26a0baf9-68bd-4380-87b3-53f50524bbbf] Binding failed for port 31cd9389-2251-4222-885e-268c99301a23, please check neutron logs for more information. {{(pid=61806) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 582.545952] env[61806]: DEBUG oslo_concurrency.lockutils [None req-5c11e673-bd02-4846-9be6-b5679ed1117d tempest-VolumesAssistedSnapshotsTest-2037321669 tempest-VolumesAssistedSnapshotsTest-2037321669-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 25.932s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 582.547406] env[61806]: INFO nova.compute.claims [None req-5c11e673-bd02-4846-9be6-b5679ed1117d tempest-VolumesAssistedSnapshotsTest-2037321669 tempest-VolumesAssistedSnapshotsTest-2037321669-project-member] [instance: 76722306-a9ad-4ef9-b579-d958c36cf6bc] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 582.550176] env[61806]: DEBUG nova.compute.manager [None req-55a61261-0862-4d73-b971-6d11cbe88c5e tempest-ServersAdminTestJSON-14199920 tempest-ServersAdminTestJSON-14199920-project-member] [instance: 26a0baf9-68bd-4380-87b3-53f50524bbbf] Build of instance 26a0baf9-68bd-4380-87b3-53f50524bbbf was re-scheduled: Binding failed for port 31cd9389-2251-4222-885e-268c99301a23, please check neutron logs for more information. {{(pid=61806) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 582.551111] env[61806]: DEBUG nova.compute.manager [None req-55a61261-0862-4d73-b971-6d11cbe88c5e tempest-ServersAdminTestJSON-14199920 tempest-ServersAdminTestJSON-14199920-project-member] [instance: 26a0baf9-68bd-4380-87b3-53f50524bbbf] Unplugging VIFs for instance {{(pid=61806) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 582.551111] env[61806]: DEBUG oslo_concurrency.lockutils [None req-55a61261-0862-4d73-b971-6d11cbe88c5e tempest-ServersAdminTestJSON-14199920 tempest-ServersAdminTestJSON-14199920-project-member] Acquiring lock "refresh_cache-26a0baf9-68bd-4380-87b3-53f50524bbbf" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 582.551111] env[61806]: DEBUG oslo_concurrency.lockutils [None req-55a61261-0862-4d73-b971-6d11cbe88c5e tempest-ServersAdminTestJSON-14199920 tempest-ServersAdminTestJSON-14199920-project-member] Acquired lock "refresh_cache-26a0baf9-68bd-4380-87b3-53f50524bbbf" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 582.551291] env[61806]: DEBUG nova.network.neutron [None req-55a61261-0862-4d73-b971-6d11cbe88c5e tempest-ServersAdminTestJSON-14199920 tempest-ServersAdminTestJSON-14199920-project-member] [instance: 26a0baf9-68bd-4380-87b3-53f50524bbbf] Building network info cache for instance {{(pid=61806) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 582.618193] env[61806]: DEBUG nova.network.neutron [req-ed54a5fa-db5b-418c-bd9a-b726b86c7538 req-a19ea44c-8946-4f2b-ad6f-3587bd5dce88 service nova] [instance: 89ce0fd8-6a22-4b1c-bdff-d1584abd538a] Instance cache missing network info. {{(pid=61806) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 582.675116] env[61806]: DEBUG nova.network.neutron [req-ed54a5fa-db5b-418c-bd9a-b726b86c7538 req-a19ea44c-8946-4f2b-ad6f-3587bd5dce88 service nova] [instance: 89ce0fd8-6a22-4b1c-bdff-d1584abd538a] Updating instance_info_cache with network_info: [] {{(pid=61806) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 582.751529] env[61806]: INFO nova.compute.manager [None req-aac795d3-c37f-4811-9be9-aa67ad520ab0 tempest-ListServerFiltersTestJSON-241228521 tempest-ListServerFiltersTestJSON-241228521-project-member] [instance: 7fcee323-8ea6-4f69-9f1e-f2105af0fe96] Took 1.02 seconds to deallocate network for instance. [ 582.944560] env[61806]: INFO nova.compute.manager [-] [instance: 89ce0fd8-6a22-4b1c-bdff-d1584abd538a] Took 1.03 seconds to deallocate network for instance. [ 582.946857] env[61806]: DEBUG nova.compute.claims [None req-37e71da7-cb5e-431d-ba3c-95daaf8af7f3 tempest-ServersAdminNegativeTestJSON-2128564044 tempest-ServersAdminNegativeTestJSON-2128564044-project-member] [instance: 89ce0fd8-6a22-4b1c-bdff-d1584abd538a] Aborting claim: {{(pid=61806) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 582.947048] env[61806]: DEBUG oslo_concurrency.lockutils [None req-37e71da7-cb5e-431d-ba3c-95daaf8af7f3 tempest-ServersAdminNegativeTestJSON-2128564044 tempest-ServersAdminNegativeTestJSON-2128564044-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 583.072717] env[61806]: DEBUG nova.network.neutron [None req-55a61261-0862-4d73-b971-6d11cbe88c5e tempest-ServersAdminTestJSON-14199920 tempest-ServersAdminTestJSON-14199920-project-member] [instance: 26a0baf9-68bd-4380-87b3-53f50524bbbf] Instance cache missing network info. {{(pid=61806) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 583.152590] env[61806]: DEBUG nova.network.neutron [None req-55a61261-0862-4d73-b971-6d11cbe88c5e tempest-ServersAdminTestJSON-14199920 tempest-ServersAdminTestJSON-14199920-project-member] [instance: 26a0baf9-68bd-4380-87b3-53f50524bbbf] Updating instance_info_cache with network_info: [] {{(pid=61806) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 583.177182] env[61806]: DEBUG oslo_concurrency.lockutils [req-ed54a5fa-db5b-418c-bd9a-b726b86c7538 req-a19ea44c-8946-4f2b-ad6f-3587bd5dce88 service nova] Releasing lock "refresh_cache-89ce0fd8-6a22-4b1c-bdff-d1584abd538a" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 583.177446] env[61806]: DEBUG nova.compute.manager [req-ed54a5fa-db5b-418c-bd9a-b726b86c7538 req-a19ea44c-8946-4f2b-ad6f-3587bd5dce88 service nova] [instance: 89ce0fd8-6a22-4b1c-bdff-d1584abd538a] Received event network-vif-deleted-63122136-e392-4570-8f9c-7785b771ac82 {{(pid=61806) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 583.654888] env[61806]: DEBUG oslo_concurrency.lockutils [None req-55a61261-0862-4d73-b971-6d11cbe88c5e tempest-ServersAdminTestJSON-14199920 tempest-ServersAdminTestJSON-14199920-project-member] Releasing lock "refresh_cache-26a0baf9-68bd-4380-87b3-53f50524bbbf" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 583.655165] env[61806]: DEBUG nova.compute.manager [None req-55a61261-0862-4d73-b971-6d11cbe88c5e tempest-ServersAdminTestJSON-14199920 tempest-ServersAdminTestJSON-14199920-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61806) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 583.655551] env[61806]: DEBUG nova.compute.manager [None req-55a61261-0862-4d73-b971-6d11cbe88c5e tempest-ServersAdminTestJSON-14199920 tempest-ServersAdminTestJSON-14199920-project-member] [instance: 26a0baf9-68bd-4380-87b3-53f50524bbbf] Deallocating network for instance {{(pid=61806) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 583.655551] env[61806]: DEBUG nova.network.neutron [None req-55a61261-0862-4d73-b971-6d11cbe88c5e tempest-ServersAdminTestJSON-14199920 tempest-ServersAdminTestJSON-14199920-project-member] [instance: 26a0baf9-68bd-4380-87b3-53f50524bbbf] deallocate_for_instance() {{(pid=61806) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 583.669946] env[61806]: DEBUG nova.network.neutron [None req-55a61261-0862-4d73-b971-6d11cbe88c5e tempest-ServersAdminTestJSON-14199920 tempest-ServersAdminTestJSON-14199920-project-member] [instance: 26a0baf9-68bd-4380-87b3-53f50524bbbf] Instance cache missing network info. {{(pid=61806) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 583.779879] env[61806]: INFO nova.scheduler.client.report [None req-aac795d3-c37f-4811-9be9-aa67ad520ab0 tempest-ListServerFiltersTestJSON-241228521 tempest-ListServerFiltersTestJSON-241228521-project-member] Deleted allocations for instance 7fcee323-8ea6-4f69-9f1e-f2105af0fe96 [ 583.929290] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-63731376-6235-4b8f-924c-edb68fcae2e1 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 583.937199] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bc27ac76-0e43-4323-939e-54332b48e1d3 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 583.966660] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c81dc36a-5414-45a9-9cef-dff035d9bf43 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 583.974702] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-77dd399a-23cc-4d2c-93ba-8f7945caf2b5 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 583.987712] env[61806]: DEBUG nova.compute.provider_tree [None req-5c11e673-bd02-4846-9be6-b5679ed1117d tempest-VolumesAssistedSnapshotsTest-2037321669 tempest-VolumesAssistedSnapshotsTest-2037321669-project-member] Inventory has not changed in ProviderTree for provider: a2858be1-fd22-4e08-979e-87ad25293407 {{(pid=61806) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 584.173267] env[61806]: DEBUG nova.network.neutron [None req-55a61261-0862-4d73-b971-6d11cbe88c5e tempest-ServersAdminTestJSON-14199920 tempest-ServersAdminTestJSON-14199920-project-member] [instance: 26a0baf9-68bd-4380-87b3-53f50524bbbf] Updating instance_info_cache with network_info: [] {{(pid=61806) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 584.290403] env[61806]: DEBUG oslo_concurrency.lockutils [None req-aac795d3-c37f-4811-9be9-aa67ad520ab0 tempest-ListServerFiltersTestJSON-241228521 tempest-ListServerFiltersTestJSON-241228521-project-member] Lock "7fcee323-8ea6-4f69-9f1e-f2105af0fe96" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 52.943s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 584.490899] env[61806]: DEBUG nova.scheduler.client.report [None req-5c11e673-bd02-4846-9be6-b5679ed1117d tempest-VolumesAssistedSnapshotsTest-2037321669 tempest-VolumesAssistedSnapshotsTest-2037321669-project-member] Inventory has not changed for provider a2858be1-fd22-4e08-979e-87ad25293407 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 139, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61806) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 584.677068] env[61806]: INFO nova.compute.manager [None req-55a61261-0862-4d73-b971-6d11cbe88c5e tempest-ServersAdminTestJSON-14199920 tempest-ServersAdminTestJSON-14199920-project-member] [instance: 26a0baf9-68bd-4380-87b3-53f50524bbbf] Took 1.02 seconds to deallocate network for instance. [ 584.795325] env[61806]: DEBUG nova.compute.manager [None req-ebff8349-d5ea-4a95-aee1-e4c9e8b6d3a4 tempest-ServerDiagnosticsTest-195484958 tempest-ServerDiagnosticsTest-195484958-project-member] [instance: 79fc9ea2-c427-47ce-acdb-1002ee5d78c9] Starting instance... {{(pid=61806) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 584.997064] env[61806]: DEBUG oslo_concurrency.lockutils [None req-5c11e673-bd02-4846-9be6-b5679ed1117d tempest-VolumesAssistedSnapshotsTest-2037321669 tempest-VolumesAssistedSnapshotsTest-2037321669-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.451s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 584.997352] env[61806]: DEBUG nova.compute.manager [None req-5c11e673-bd02-4846-9be6-b5679ed1117d tempest-VolumesAssistedSnapshotsTest-2037321669 tempest-VolumesAssistedSnapshotsTest-2037321669-project-member] [instance: 76722306-a9ad-4ef9-b579-d958c36cf6bc] Start building networks asynchronously for instance. {{(pid=61806) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 585.001253] env[61806]: DEBUG oslo_concurrency.lockutils [None req-b50d0802-a025-4c5c-b8e9-95dd624e7ff5 tempest-ServersWithSpecificFlavorTestJSON-815343160 tempest-ServersWithSpecificFlavorTestJSON-815343160-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 26.456s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 585.001528] env[61806]: INFO nova.compute.claims [None req-b50d0802-a025-4c5c-b8e9-95dd624e7ff5 tempest-ServersWithSpecificFlavorTestJSON-815343160 tempest-ServersWithSpecificFlavorTestJSON-815343160-project-member] [instance: b68b509e-96dc-489d-9cf1-93ba91095af1] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 585.319415] env[61806]: DEBUG oslo_concurrency.lockutils [None req-ebff8349-d5ea-4a95-aee1-e4c9e8b6d3a4 tempest-ServerDiagnosticsTest-195484958 tempest-ServerDiagnosticsTest-195484958-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 585.505642] env[61806]: DEBUG nova.compute.utils [None req-5c11e673-bd02-4846-9be6-b5679ed1117d tempest-VolumesAssistedSnapshotsTest-2037321669 tempest-VolumesAssistedSnapshotsTest-2037321669-project-member] Using /dev/sd instead of None {{(pid=61806) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 585.509152] env[61806]: DEBUG nova.compute.manager [None req-5c11e673-bd02-4846-9be6-b5679ed1117d tempest-VolumesAssistedSnapshotsTest-2037321669 tempest-VolumesAssistedSnapshotsTest-2037321669-project-member] [instance: 76722306-a9ad-4ef9-b579-d958c36cf6bc] Allocating IP information in the background. {{(pid=61806) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 585.509333] env[61806]: DEBUG nova.network.neutron [None req-5c11e673-bd02-4846-9be6-b5679ed1117d tempest-VolumesAssistedSnapshotsTest-2037321669 tempest-VolumesAssistedSnapshotsTest-2037321669-project-member] [instance: 76722306-a9ad-4ef9-b579-d958c36cf6bc] allocate_for_instance() {{(pid=61806) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 585.570793] env[61806]: DEBUG nova.policy [None req-5c11e673-bd02-4846-9be6-b5679ed1117d tempest-VolumesAssistedSnapshotsTest-2037321669 tempest-VolumesAssistedSnapshotsTest-2037321669-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'dfbdb5e9935c40669cad49fe5eed94a3', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '9fbd6bdf468f4668a4eb6ddd6f00ad06', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61806) authorize /opt/stack/nova/nova/policy.py:201}} [ 585.713588] env[61806]: INFO nova.scheduler.client.report [None req-55a61261-0862-4d73-b971-6d11cbe88c5e tempest-ServersAdminTestJSON-14199920 tempest-ServersAdminTestJSON-14199920-project-member] Deleted allocations for instance 26a0baf9-68bd-4380-87b3-53f50524bbbf [ 585.927586] env[61806]: DEBUG nova.network.neutron [None req-5c11e673-bd02-4846-9be6-b5679ed1117d tempest-VolumesAssistedSnapshotsTest-2037321669 tempest-VolumesAssistedSnapshotsTest-2037321669-project-member] [instance: 76722306-a9ad-4ef9-b579-d958c36cf6bc] Successfully created port: 0094ad04-66d2-46ea-8ce7-09664ac7c614 {{(pid=61806) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 586.010264] env[61806]: DEBUG nova.compute.manager [None req-5c11e673-bd02-4846-9be6-b5679ed1117d tempest-VolumesAssistedSnapshotsTest-2037321669 tempest-VolumesAssistedSnapshotsTest-2037321669-project-member] [instance: 76722306-a9ad-4ef9-b579-d958c36cf6bc] Start building block device mappings for instance. {{(pid=61806) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 586.225026] env[61806]: DEBUG oslo_concurrency.lockutils [None req-55a61261-0862-4d73-b971-6d11cbe88c5e tempest-ServersAdminTestJSON-14199920 tempest-ServersAdminTestJSON-14199920-project-member] Lock "26a0baf9-68bd-4380-87b3-53f50524bbbf" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 53.188s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 586.431159] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7b1fbd2d-4686-444f-8f14-0d366329b5b5 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 586.439015] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-828dbccc-d4ce-49e1-8c48-415045fc4678 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 586.471468] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9152038d-b90c-4e58-9603-b7cf6a352594 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 586.481671] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fee08800-a22f-457c-818e-09c7d8bfd6b5 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 586.498664] env[61806]: DEBUG nova.compute.provider_tree [None req-b50d0802-a025-4c5c-b8e9-95dd624e7ff5 tempest-ServersWithSpecificFlavorTestJSON-815343160 tempest-ServersWithSpecificFlavorTestJSON-815343160-project-member] Inventory has not changed in ProviderTree for provider: a2858be1-fd22-4e08-979e-87ad25293407 {{(pid=61806) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 586.730690] env[61806]: DEBUG nova.compute.manager [None req-24088d08-d3d7-4f62-ad63-91029bd9b22d tempest-ServersTestMultiNic-1480543082 tempest-ServersTestMultiNic-1480543082-project-member] [instance: c2d1c9f6-f458-4843-982c-afdb2a90edae] Starting instance... {{(pid=61806) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 587.001552] env[61806]: DEBUG nova.scheduler.client.report [None req-b50d0802-a025-4c5c-b8e9-95dd624e7ff5 tempest-ServersWithSpecificFlavorTestJSON-815343160 tempest-ServersWithSpecificFlavorTestJSON-815343160-project-member] Inventory has not changed for provider a2858be1-fd22-4e08-979e-87ad25293407 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 139, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61806) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 587.032960] env[61806]: DEBUG nova.compute.manager [None req-5c11e673-bd02-4846-9be6-b5679ed1117d tempest-VolumesAssistedSnapshotsTest-2037321669 tempest-VolumesAssistedSnapshotsTest-2037321669-project-member] [instance: 76722306-a9ad-4ef9-b579-d958c36cf6bc] Start spawning the instance on the hypervisor. {{(pid=61806) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 587.063319] env[61806]: DEBUG nova.virt.hardware [None req-5c11e673-bd02-4846-9be6-b5679ed1117d tempest-VolumesAssistedSnapshotsTest-2037321669 tempest-VolumesAssistedSnapshotsTest-2037321669-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-15T18:05:10Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-15T18:04:52Z,direct_url=,disk_format='vmdk',id=ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='84ba83fce2c349a6988173c5d6fc3b07',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-15T18:04:53Z,virtual_size=,visibility=), allow threads: False {{(pid=61806) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 587.063560] env[61806]: DEBUG nova.virt.hardware [None req-5c11e673-bd02-4846-9be6-b5679ed1117d tempest-VolumesAssistedSnapshotsTest-2037321669 tempest-VolumesAssistedSnapshotsTest-2037321669-project-member] Flavor limits 0:0:0 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 587.063724] env[61806]: DEBUG nova.virt.hardware [None req-5c11e673-bd02-4846-9be6-b5679ed1117d tempest-VolumesAssistedSnapshotsTest-2037321669 tempest-VolumesAssistedSnapshotsTest-2037321669-project-member] Image limits 0:0:0 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 587.063903] env[61806]: DEBUG nova.virt.hardware [None req-5c11e673-bd02-4846-9be6-b5679ed1117d tempest-VolumesAssistedSnapshotsTest-2037321669 tempest-VolumesAssistedSnapshotsTest-2037321669-project-member] Flavor pref 0:0:0 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 587.064145] env[61806]: DEBUG nova.virt.hardware [None req-5c11e673-bd02-4846-9be6-b5679ed1117d tempest-VolumesAssistedSnapshotsTest-2037321669 tempest-VolumesAssistedSnapshotsTest-2037321669-project-member] Image pref 0:0:0 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 587.064470] env[61806]: DEBUG nova.virt.hardware [None req-5c11e673-bd02-4846-9be6-b5679ed1117d tempest-VolumesAssistedSnapshotsTest-2037321669 tempest-VolumesAssistedSnapshotsTest-2037321669-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 587.064620] env[61806]: DEBUG nova.virt.hardware [None req-5c11e673-bd02-4846-9be6-b5679ed1117d tempest-VolumesAssistedSnapshotsTest-2037321669 tempest-VolumesAssistedSnapshotsTest-2037321669-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61806) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 587.064810] env[61806]: DEBUG nova.virt.hardware [None req-5c11e673-bd02-4846-9be6-b5679ed1117d tempest-VolumesAssistedSnapshotsTest-2037321669 tempest-VolumesAssistedSnapshotsTest-2037321669-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61806) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 587.065019] env[61806]: DEBUG nova.virt.hardware [None req-5c11e673-bd02-4846-9be6-b5679ed1117d tempest-VolumesAssistedSnapshotsTest-2037321669 tempest-VolumesAssistedSnapshotsTest-2037321669-project-member] Got 1 possible topologies {{(pid=61806) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 587.065223] env[61806]: DEBUG nova.virt.hardware [None req-5c11e673-bd02-4846-9be6-b5679ed1117d tempest-VolumesAssistedSnapshotsTest-2037321669 tempest-VolumesAssistedSnapshotsTest-2037321669-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61806) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 587.065444] env[61806]: DEBUG nova.virt.hardware [None req-5c11e673-bd02-4846-9be6-b5679ed1117d tempest-VolumesAssistedSnapshotsTest-2037321669 tempest-VolumesAssistedSnapshotsTest-2037321669-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61806) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 587.066766] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fea4c419-eef4-4fc2-87b2-5aa828327c20 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 587.071843] env[61806]: DEBUG nova.compute.manager [req-bf5d6437-3fa3-4ef4-881a-2d082bc30cd0 req-0f005592-c1bb-4280-ace6-06bab323e7f1 service nova] [instance: 76722306-a9ad-4ef9-b579-d958c36cf6bc] Received event network-changed-0094ad04-66d2-46ea-8ce7-09664ac7c614 {{(pid=61806) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 587.072050] env[61806]: DEBUG nova.compute.manager [req-bf5d6437-3fa3-4ef4-881a-2d082bc30cd0 req-0f005592-c1bb-4280-ace6-06bab323e7f1 service nova] [instance: 76722306-a9ad-4ef9-b579-d958c36cf6bc] Refreshing instance network info cache due to event network-changed-0094ad04-66d2-46ea-8ce7-09664ac7c614. {{(pid=61806) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 587.072455] env[61806]: DEBUG oslo_concurrency.lockutils [req-bf5d6437-3fa3-4ef4-881a-2d082bc30cd0 req-0f005592-c1bb-4280-ace6-06bab323e7f1 service nova] Acquiring lock "refresh_cache-76722306-a9ad-4ef9-b579-d958c36cf6bc" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 587.072455] env[61806]: DEBUG oslo_concurrency.lockutils [req-bf5d6437-3fa3-4ef4-881a-2d082bc30cd0 req-0f005592-c1bb-4280-ace6-06bab323e7f1 service nova] Acquired lock "refresh_cache-76722306-a9ad-4ef9-b579-d958c36cf6bc" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 587.072669] env[61806]: DEBUG nova.network.neutron [req-bf5d6437-3fa3-4ef4-881a-2d082bc30cd0 req-0f005592-c1bb-4280-ace6-06bab323e7f1 service nova] [instance: 76722306-a9ad-4ef9-b579-d958c36cf6bc] Refreshing network info cache for port 0094ad04-66d2-46ea-8ce7-09664ac7c614 {{(pid=61806) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 587.079364] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-403b6642-6180-4d64-8ab2-e3379da2b3e8 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 587.219261] env[61806]: ERROR nova.compute.manager [None req-5c11e673-bd02-4846-9be6-b5679ed1117d tempest-VolumesAssistedSnapshotsTest-2037321669 tempest-VolumesAssistedSnapshotsTest-2037321669-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 0094ad04-66d2-46ea-8ce7-09664ac7c614, please check neutron logs for more information. [ 587.219261] env[61806]: ERROR nova.compute.manager Traceback (most recent call last): [ 587.219261] env[61806]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 587.219261] env[61806]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 587.219261] env[61806]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 587.219261] env[61806]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 587.219261] env[61806]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 587.219261] env[61806]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 587.219261] env[61806]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 587.219261] env[61806]: ERROR nova.compute.manager self.force_reraise() [ 587.219261] env[61806]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 587.219261] env[61806]: ERROR nova.compute.manager raise self.value [ 587.219261] env[61806]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 587.219261] env[61806]: ERROR nova.compute.manager updated_port = self._update_port( [ 587.219261] env[61806]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 587.219261] env[61806]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 587.219998] env[61806]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 587.219998] env[61806]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 587.219998] env[61806]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 0094ad04-66d2-46ea-8ce7-09664ac7c614, please check neutron logs for more information. [ 587.219998] env[61806]: ERROR nova.compute.manager [ 587.219998] env[61806]: Traceback (most recent call last): [ 587.219998] env[61806]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 587.219998] env[61806]: listener.cb(fileno) [ 587.219998] env[61806]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 587.219998] env[61806]: result = function(*args, **kwargs) [ 587.219998] env[61806]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 587.219998] env[61806]: return func(*args, **kwargs) [ 587.219998] env[61806]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 587.219998] env[61806]: raise e [ 587.219998] env[61806]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 587.219998] env[61806]: nwinfo = self.network_api.allocate_for_instance( [ 587.219998] env[61806]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 587.219998] env[61806]: created_port_ids = self._update_ports_for_instance( [ 587.219998] env[61806]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 587.219998] env[61806]: with excutils.save_and_reraise_exception(): [ 587.219998] env[61806]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 587.219998] env[61806]: self.force_reraise() [ 587.219998] env[61806]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 587.219998] env[61806]: raise self.value [ 587.219998] env[61806]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 587.219998] env[61806]: updated_port = self._update_port( [ 587.219998] env[61806]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 587.219998] env[61806]: _ensure_no_port_binding_failure(port) [ 587.219998] env[61806]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 587.219998] env[61806]: raise exception.PortBindingFailed(port_id=port['id']) [ 587.221095] env[61806]: nova.exception.PortBindingFailed: Binding failed for port 0094ad04-66d2-46ea-8ce7-09664ac7c614, please check neutron logs for more information. [ 587.221095] env[61806]: Removing descriptor: 17 [ 587.221095] env[61806]: ERROR nova.compute.manager [None req-5c11e673-bd02-4846-9be6-b5679ed1117d tempest-VolumesAssistedSnapshotsTest-2037321669 tempest-VolumesAssistedSnapshotsTest-2037321669-project-member] [instance: 76722306-a9ad-4ef9-b579-d958c36cf6bc] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 0094ad04-66d2-46ea-8ce7-09664ac7c614, please check neutron logs for more information. [ 587.221095] env[61806]: ERROR nova.compute.manager [instance: 76722306-a9ad-4ef9-b579-d958c36cf6bc] Traceback (most recent call last): [ 587.221095] env[61806]: ERROR nova.compute.manager [instance: 76722306-a9ad-4ef9-b579-d958c36cf6bc] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 587.221095] env[61806]: ERROR nova.compute.manager [instance: 76722306-a9ad-4ef9-b579-d958c36cf6bc] yield resources [ 587.221095] env[61806]: ERROR nova.compute.manager [instance: 76722306-a9ad-4ef9-b579-d958c36cf6bc] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 587.221095] env[61806]: ERROR nova.compute.manager [instance: 76722306-a9ad-4ef9-b579-d958c36cf6bc] self.driver.spawn(context, instance, image_meta, [ 587.221095] env[61806]: ERROR nova.compute.manager [instance: 76722306-a9ad-4ef9-b579-d958c36cf6bc] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 587.221095] env[61806]: ERROR nova.compute.manager [instance: 76722306-a9ad-4ef9-b579-d958c36cf6bc] self._vmops.spawn(context, instance, image_meta, injected_files, [ 587.221095] env[61806]: ERROR nova.compute.manager [instance: 76722306-a9ad-4ef9-b579-d958c36cf6bc] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 587.221095] env[61806]: ERROR nova.compute.manager [instance: 76722306-a9ad-4ef9-b579-d958c36cf6bc] vm_ref = self.build_virtual_machine(instance, [ 587.221464] env[61806]: ERROR nova.compute.manager [instance: 76722306-a9ad-4ef9-b579-d958c36cf6bc] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 587.221464] env[61806]: ERROR nova.compute.manager [instance: 76722306-a9ad-4ef9-b579-d958c36cf6bc] vif_infos = vmwarevif.get_vif_info(self._session, [ 587.221464] env[61806]: ERROR nova.compute.manager [instance: 76722306-a9ad-4ef9-b579-d958c36cf6bc] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 587.221464] env[61806]: ERROR nova.compute.manager [instance: 76722306-a9ad-4ef9-b579-d958c36cf6bc] for vif in network_info: [ 587.221464] env[61806]: ERROR nova.compute.manager [instance: 76722306-a9ad-4ef9-b579-d958c36cf6bc] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 587.221464] env[61806]: ERROR nova.compute.manager [instance: 76722306-a9ad-4ef9-b579-d958c36cf6bc] return self._sync_wrapper(fn, *args, **kwargs) [ 587.221464] env[61806]: ERROR nova.compute.manager [instance: 76722306-a9ad-4ef9-b579-d958c36cf6bc] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 587.221464] env[61806]: ERROR nova.compute.manager [instance: 76722306-a9ad-4ef9-b579-d958c36cf6bc] self.wait() [ 587.221464] env[61806]: ERROR nova.compute.manager [instance: 76722306-a9ad-4ef9-b579-d958c36cf6bc] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 587.221464] env[61806]: ERROR nova.compute.manager [instance: 76722306-a9ad-4ef9-b579-d958c36cf6bc] self[:] = self._gt.wait() [ 587.221464] env[61806]: ERROR nova.compute.manager [instance: 76722306-a9ad-4ef9-b579-d958c36cf6bc] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 587.221464] env[61806]: ERROR nova.compute.manager [instance: 76722306-a9ad-4ef9-b579-d958c36cf6bc] return self._exit_event.wait() [ 587.221464] env[61806]: ERROR nova.compute.manager [instance: 76722306-a9ad-4ef9-b579-d958c36cf6bc] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 587.221861] env[61806]: ERROR nova.compute.manager [instance: 76722306-a9ad-4ef9-b579-d958c36cf6bc] result = hub.switch() [ 587.221861] env[61806]: ERROR nova.compute.manager [instance: 76722306-a9ad-4ef9-b579-d958c36cf6bc] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 587.221861] env[61806]: ERROR nova.compute.manager [instance: 76722306-a9ad-4ef9-b579-d958c36cf6bc] return self.greenlet.switch() [ 587.221861] env[61806]: ERROR nova.compute.manager [instance: 76722306-a9ad-4ef9-b579-d958c36cf6bc] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 587.221861] env[61806]: ERROR nova.compute.manager [instance: 76722306-a9ad-4ef9-b579-d958c36cf6bc] result = function(*args, **kwargs) [ 587.221861] env[61806]: ERROR nova.compute.manager [instance: 76722306-a9ad-4ef9-b579-d958c36cf6bc] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 587.221861] env[61806]: ERROR nova.compute.manager [instance: 76722306-a9ad-4ef9-b579-d958c36cf6bc] return func(*args, **kwargs) [ 587.221861] env[61806]: ERROR nova.compute.manager [instance: 76722306-a9ad-4ef9-b579-d958c36cf6bc] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 587.221861] env[61806]: ERROR nova.compute.manager [instance: 76722306-a9ad-4ef9-b579-d958c36cf6bc] raise e [ 587.221861] env[61806]: ERROR nova.compute.manager [instance: 76722306-a9ad-4ef9-b579-d958c36cf6bc] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 587.221861] env[61806]: ERROR nova.compute.manager [instance: 76722306-a9ad-4ef9-b579-d958c36cf6bc] nwinfo = self.network_api.allocate_for_instance( [ 587.221861] env[61806]: ERROR nova.compute.manager [instance: 76722306-a9ad-4ef9-b579-d958c36cf6bc] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 587.221861] env[61806]: ERROR nova.compute.manager [instance: 76722306-a9ad-4ef9-b579-d958c36cf6bc] created_port_ids = self._update_ports_for_instance( [ 587.222266] env[61806]: ERROR nova.compute.manager [instance: 76722306-a9ad-4ef9-b579-d958c36cf6bc] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 587.222266] env[61806]: ERROR nova.compute.manager [instance: 76722306-a9ad-4ef9-b579-d958c36cf6bc] with excutils.save_and_reraise_exception(): [ 587.222266] env[61806]: ERROR nova.compute.manager [instance: 76722306-a9ad-4ef9-b579-d958c36cf6bc] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 587.222266] env[61806]: ERROR nova.compute.manager [instance: 76722306-a9ad-4ef9-b579-d958c36cf6bc] self.force_reraise() [ 587.222266] env[61806]: ERROR nova.compute.manager [instance: 76722306-a9ad-4ef9-b579-d958c36cf6bc] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 587.222266] env[61806]: ERROR nova.compute.manager [instance: 76722306-a9ad-4ef9-b579-d958c36cf6bc] raise self.value [ 587.222266] env[61806]: ERROR nova.compute.manager [instance: 76722306-a9ad-4ef9-b579-d958c36cf6bc] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 587.222266] env[61806]: ERROR nova.compute.manager [instance: 76722306-a9ad-4ef9-b579-d958c36cf6bc] updated_port = self._update_port( [ 587.222266] env[61806]: ERROR nova.compute.manager [instance: 76722306-a9ad-4ef9-b579-d958c36cf6bc] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 587.222266] env[61806]: ERROR nova.compute.manager [instance: 76722306-a9ad-4ef9-b579-d958c36cf6bc] _ensure_no_port_binding_failure(port) [ 587.222266] env[61806]: ERROR nova.compute.manager [instance: 76722306-a9ad-4ef9-b579-d958c36cf6bc] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 587.222266] env[61806]: ERROR nova.compute.manager [instance: 76722306-a9ad-4ef9-b579-d958c36cf6bc] raise exception.PortBindingFailed(port_id=port['id']) [ 587.222625] env[61806]: ERROR nova.compute.manager [instance: 76722306-a9ad-4ef9-b579-d958c36cf6bc] nova.exception.PortBindingFailed: Binding failed for port 0094ad04-66d2-46ea-8ce7-09664ac7c614, please check neutron logs for more information. [ 587.222625] env[61806]: ERROR nova.compute.manager [instance: 76722306-a9ad-4ef9-b579-d958c36cf6bc] [ 587.222625] env[61806]: INFO nova.compute.manager [None req-5c11e673-bd02-4846-9be6-b5679ed1117d tempest-VolumesAssistedSnapshotsTest-2037321669 tempest-VolumesAssistedSnapshotsTest-2037321669-project-member] [instance: 76722306-a9ad-4ef9-b579-d958c36cf6bc] Terminating instance [ 587.222721] env[61806]: DEBUG oslo_concurrency.lockutils [None req-5c11e673-bd02-4846-9be6-b5679ed1117d tempest-VolumesAssistedSnapshotsTest-2037321669 tempest-VolumesAssistedSnapshotsTest-2037321669-project-member] Acquiring lock "refresh_cache-76722306-a9ad-4ef9-b579-d958c36cf6bc" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 587.268282] env[61806]: DEBUG oslo_concurrency.lockutils [None req-24088d08-d3d7-4f62-ad63-91029bd9b22d tempest-ServersTestMultiNic-1480543082 tempest-ServersTestMultiNic-1480543082-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 587.512947] env[61806]: DEBUG oslo_concurrency.lockutils [None req-b50d0802-a025-4c5c-b8e9-95dd624e7ff5 tempest-ServersWithSpecificFlavorTestJSON-815343160 tempest-ServersWithSpecificFlavorTestJSON-815343160-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.513s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 587.513492] env[61806]: DEBUG nova.compute.manager [None req-b50d0802-a025-4c5c-b8e9-95dd624e7ff5 tempest-ServersWithSpecificFlavorTestJSON-815343160 tempest-ServersWithSpecificFlavorTestJSON-815343160-project-member] [instance: b68b509e-96dc-489d-9cf1-93ba91095af1] Start building networks asynchronously for instance. {{(pid=61806) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 587.516165] env[61806]: DEBUG oslo_concurrency.lockutils [None req-be9179b6-8d17-4f04-8823-d48174fcc5c1 tempest-InstanceActionsV221TestJSON-379107765 tempest-InstanceActionsV221TestJSON-379107765-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 14.953s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 587.602108] env[61806]: DEBUG nova.network.neutron [req-bf5d6437-3fa3-4ef4-881a-2d082bc30cd0 req-0f005592-c1bb-4280-ace6-06bab323e7f1 service nova] [instance: 76722306-a9ad-4ef9-b579-d958c36cf6bc] Instance cache missing network info. {{(pid=61806) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 587.705395] env[61806]: DEBUG nova.network.neutron [req-bf5d6437-3fa3-4ef4-881a-2d082bc30cd0 req-0f005592-c1bb-4280-ace6-06bab323e7f1 service nova] [instance: 76722306-a9ad-4ef9-b579-d958c36cf6bc] Updating instance_info_cache with network_info: [] {{(pid=61806) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 588.024798] env[61806]: DEBUG nova.compute.utils [None req-b50d0802-a025-4c5c-b8e9-95dd624e7ff5 tempest-ServersWithSpecificFlavorTestJSON-815343160 tempest-ServersWithSpecificFlavorTestJSON-815343160-project-member] Using /dev/sd instead of None {{(pid=61806) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 588.026514] env[61806]: DEBUG nova.compute.manager [None req-b50d0802-a025-4c5c-b8e9-95dd624e7ff5 tempest-ServersWithSpecificFlavorTestJSON-815343160 tempest-ServersWithSpecificFlavorTestJSON-815343160-project-member] [instance: b68b509e-96dc-489d-9cf1-93ba91095af1] Allocating IP information in the background. {{(pid=61806) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 588.026693] env[61806]: DEBUG nova.network.neutron [None req-b50d0802-a025-4c5c-b8e9-95dd624e7ff5 tempest-ServersWithSpecificFlavorTestJSON-815343160 tempest-ServersWithSpecificFlavorTestJSON-815343160-project-member] [instance: b68b509e-96dc-489d-9cf1-93ba91095af1] allocate_for_instance() {{(pid=61806) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 588.077193] env[61806]: DEBUG nova.policy [None req-b50d0802-a025-4c5c-b8e9-95dd624e7ff5 tempest-ServersWithSpecificFlavorTestJSON-815343160 tempest-ServersWithSpecificFlavorTestJSON-815343160-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'cbb4b75195f54103a75549a14413133a', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'd0f900c22e334c51b36c7ec2fd050f99', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61806) authorize /opt/stack/nova/nova/policy.py:201}} [ 588.210211] env[61806]: DEBUG oslo_concurrency.lockutils [req-bf5d6437-3fa3-4ef4-881a-2d082bc30cd0 req-0f005592-c1bb-4280-ace6-06bab323e7f1 service nova] Releasing lock "refresh_cache-76722306-a9ad-4ef9-b579-d958c36cf6bc" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 588.210624] env[61806]: DEBUG oslo_concurrency.lockutils [None req-5c11e673-bd02-4846-9be6-b5679ed1117d tempest-VolumesAssistedSnapshotsTest-2037321669 tempest-VolumesAssistedSnapshotsTest-2037321669-project-member] Acquired lock "refresh_cache-76722306-a9ad-4ef9-b579-d958c36cf6bc" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 588.210849] env[61806]: DEBUG nova.network.neutron [None req-5c11e673-bd02-4846-9be6-b5679ed1117d tempest-VolumesAssistedSnapshotsTest-2037321669 tempest-VolumesAssistedSnapshotsTest-2037321669-project-member] [instance: 76722306-a9ad-4ef9-b579-d958c36cf6bc] Building network info cache for instance {{(pid=61806) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 588.398956] env[61806]: DEBUG nova.network.neutron [None req-b50d0802-a025-4c5c-b8e9-95dd624e7ff5 tempest-ServersWithSpecificFlavorTestJSON-815343160 tempest-ServersWithSpecificFlavorTestJSON-815343160-project-member] [instance: b68b509e-96dc-489d-9cf1-93ba91095af1] Successfully created port: ec4cd098-60bf-4a9b-9f8a-2d85646b89ab {{(pid=61806) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 588.441382] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3e304aaa-abcb-4055-af22-b33202ede4fe {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 588.452215] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-096d7416-5bee-49d9-8866-5b8d7b5275fc {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 588.485807] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8136c494-6b61-4d65-8628-bd6852c636f1 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 588.493655] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-98701181-1719-48e5-91c9-0ebd5a205055 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 588.507741] env[61806]: DEBUG nova.compute.provider_tree [None req-be9179b6-8d17-4f04-8823-d48174fcc5c1 tempest-InstanceActionsV221TestJSON-379107765 tempest-InstanceActionsV221TestJSON-379107765-project-member] Inventory has not changed in ProviderTree for provider: a2858be1-fd22-4e08-979e-87ad25293407 {{(pid=61806) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 588.532588] env[61806]: DEBUG nova.compute.manager [None req-b50d0802-a025-4c5c-b8e9-95dd624e7ff5 tempest-ServersWithSpecificFlavorTestJSON-815343160 tempest-ServersWithSpecificFlavorTestJSON-815343160-project-member] [instance: b68b509e-96dc-489d-9cf1-93ba91095af1] Start building block device mappings for instance. {{(pid=61806) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 588.730040] env[61806]: DEBUG nova.network.neutron [None req-5c11e673-bd02-4846-9be6-b5679ed1117d tempest-VolumesAssistedSnapshotsTest-2037321669 tempest-VolumesAssistedSnapshotsTest-2037321669-project-member] [instance: 76722306-a9ad-4ef9-b579-d958c36cf6bc] Instance cache missing network info. {{(pid=61806) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 588.849492] env[61806]: DEBUG nova.network.neutron [None req-5c11e673-bd02-4846-9be6-b5679ed1117d tempest-VolumesAssistedSnapshotsTest-2037321669 tempest-VolumesAssistedSnapshotsTest-2037321669-project-member] [instance: 76722306-a9ad-4ef9-b579-d958c36cf6bc] Updating instance_info_cache with network_info: [] {{(pid=61806) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 589.010871] env[61806]: DEBUG nova.scheduler.client.report [None req-be9179b6-8d17-4f04-8823-d48174fcc5c1 tempest-InstanceActionsV221TestJSON-379107765 tempest-InstanceActionsV221TestJSON-379107765-project-member] Inventory has not changed for provider a2858be1-fd22-4e08-979e-87ad25293407 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 139, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61806) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 589.272821] env[61806]: DEBUG nova.compute.manager [req-027ff6fb-470f-4cf7-b85a-0d37aa0ee041 req-319b2a2b-a646-4da2-ae10-c893c68e3787 service nova] [instance: 76722306-a9ad-4ef9-b579-d958c36cf6bc] Received event network-vif-deleted-0094ad04-66d2-46ea-8ce7-09664ac7c614 {{(pid=61806) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 589.353241] env[61806]: DEBUG oslo_concurrency.lockutils [None req-5c11e673-bd02-4846-9be6-b5679ed1117d tempest-VolumesAssistedSnapshotsTest-2037321669 tempest-VolumesAssistedSnapshotsTest-2037321669-project-member] Releasing lock "refresh_cache-76722306-a9ad-4ef9-b579-d958c36cf6bc" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 589.353687] env[61806]: DEBUG nova.compute.manager [None req-5c11e673-bd02-4846-9be6-b5679ed1117d tempest-VolumesAssistedSnapshotsTest-2037321669 tempest-VolumesAssistedSnapshotsTest-2037321669-project-member] [instance: 76722306-a9ad-4ef9-b579-d958c36cf6bc] Start destroying the instance on the hypervisor. {{(pid=61806) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 589.353910] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-5c11e673-bd02-4846-9be6-b5679ed1117d tempest-VolumesAssistedSnapshotsTest-2037321669 tempest-VolumesAssistedSnapshotsTest-2037321669-project-member] [instance: 76722306-a9ad-4ef9-b579-d958c36cf6bc] Destroying instance {{(pid=61806) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 589.354330] env[61806]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-60560865-6779-413e-b081-932a58bb5c27 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 589.365051] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f4890544-89d6-456e-800d-efeb06fda397 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 589.389245] env[61806]: WARNING nova.virt.vmwareapi.vmops [None req-5c11e673-bd02-4846-9be6-b5679ed1117d tempest-VolumesAssistedSnapshotsTest-2037321669 tempest-VolumesAssistedSnapshotsTest-2037321669-project-member] [instance: 76722306-a9ad-4ef9-b579-d958c36cf6bc] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 76722306-a9ad-4ef9-b579-d958c36cf6bc could not be found. [ 589.391692] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-5c11e673-bd02-4846-9be6-b5679ed1117d tempest-VolumesAssistedSnapshotsTest-2037321669 tempest-VolumesAssistedSnapshotsTest-2037321669-project-member] [instance: 76722306-a9ad-4ef9-b579-d958c36cf6bc] Instance destroyed {{(pid=61806) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 589.391692] env[61806]: INFO nova.compute.manager [None req-5c11e673-bd02-4846-9be6-b5679ed1117d tempest-VolumesAssistedSnapshotsTest-2037321669 tempest-VolumesAssistedSnapshotsTest-2037321669-project-member] [instance: 76722306-a9ad-4ef9-b579-d958c36cf6bc] Took 0.04 seconds to destroy the instance on the hypervisor. [ 589.391692] env[61806]: DEBUG oslo.service.loopingcall [None req-5c11e673-bd02-4846-9be6-b5679ed1117d tempest-VolumesAssistedSnapshotsTest-2037321669 tempest-VolumesAssistedSnapshotsTest-2037321669-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61806) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 589.391692] env[61806]: DEBUG nova.compute.manager [-] [instance: 76722306-a9ad-4ef9-b579-d958c36cf6bc] Deallocating network for instance {{(pid=61806) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 589.391692] env[61806]: DEBUG nova.network.neutron [-] [instance: 76722306-a9ad-4ef9-b579-d958c36cf6bc] deallocate_for_instance() {{(pid=61806) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 589.409950] env[61806]: DEBUG nova.network.neutron [-] [instance: 76722306-a9ad-4ef9-b579-d958c36cf6bc] Instance cache missing network info. {{(pid=61806) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 589.516195] env[61806]: DEBUG oslo_concurrency.lockutils [None req-be9179b6-8d17-4f04-8823-d48174fcc5c1 tempest-InstanceActionsV221TestJSON-379107765 tempest-InstanceActionsV221TestJSON-379107765-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.999s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 589.516195] env[61806]: ERROR nova.compute.manager [None req-be9179b6-8d17-4f04-8823-d48174fcc5c1 tempest-InstanceActionsV221TestJSON-379107765 tempest-InstanceActionsV221TestJSON-379107765-project-member] [instance: 123f15be-3b56-4999-8010-40d2e86a6813] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 8fbea74a-6823-4774-b3af-5bda4ccd8820, please check neutron logs for more information. [ 589.516195] env[61806]: ERROR nova.compute.manager [instance: 123f15be-3b56-4999-8010-40d2e86a6813] Traceback (most recent call last): [ 589.516195] env[61806]: ERROR nova.compute.manager [instance: 123f15be-3b56-4999-8010-40d2e86a6813] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 589.516195] env[61806]: ERROR nova.compute.manager [instance: 123f15be-3b56-4999-8010-40d2e86a6813] self.driver.spawn(context, instance, image_meta, [ 589.516195] env[61806]: ERROR nova.compute.manager [instance: 123f15be-3b56-4999-8010-40d2e86a6813] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 589.516195] env[61806]: ERROR nova.compute.manager [instance: 123f15be-3b56-4999-8010-40d2e86a6813] self._vmops.spawn(context, instance, image_meta, injected_files, [ 589.516195] env[61806]: ERROR nova.compute.manager [instance: 123f15be-3b56-4999-8010-40d2e86a6813] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 589.516195] env[61806]: ERROR nova.compute.manager [instance: 123f15be-3b56-4999-8010-40d2e86a6813] vm_ref = self.build_virtual_machine(instance, [ 589.516744] env[61806]: ERROR nova.compute.manager [instance: 123f15be-3b56-4999-8010-40d2e86a6813] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 589.516744] env[61806]: ERROR nova.compute.manager [instance: 123f15be-3b56-4999-8010-40d2e86a6813] vif_infos = vmwarevif.get_vif_info(self._session, [ 589.516744] env[61806]: ERROR nova.compute.manager [instance: 123f15be-3b56-4999-8010-40d2e86a6813] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 589.516744] env[61806]: ERROR nova.compute.manager [instance: 123f15be-3b56-4999-8010-40d2e86a6813] for vif in network_info: [ 589.516744] env[61806]: ERROR nova.compute.manager [instance: 123f15be-3b56-4999-8010-40d2e86a6813] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 589.516744] env[61806]: ERROR nova.compute.manager [instance: 123f15be-3b56-4999-8010-40d2e86a6813] return self._sync_wrapper(fn, *args, **kwargs) [ 589.516744] env[61806]: ERROR nova.compute.manager [instance: 123f15be-3b56-4999-8010-40d2e86a6813] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 589.516744] env[61806]: ERROR nova.compute.manager [instance: 123f15be-3b56-4999-8010-40d2e86a6813] self.wait() [ 589.516744] env[61806]: ERROR nova.compute.manager [instance: 123f15be-3b56-4999-8010-40d2e86a6813] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 589.516744] env[61806]: ERROR nova.compute.manager [instance: 123f15be-3b56-4999-8010-40d2e86a6813] self[:] = self._gt.wait() [ 589.516744] env[61806]: ERROR nova.compute.manager [instance: 123f15be-3b56-4999-8010-40d2e86a6813] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 589.516744] env[61806]: ERROR nova.compute.manager [instance: 123f15be-3b56-4999-8010-40d2e86a6813] return self._exit_event.wait() [ 589.516744] env[61806]: ERROR nova.compute.manager [instance: 123f15be-3b56-4999-8010-40d2e86a6813] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 589.517211] env[61806]: ERROR nova.compute.manager [instance: 123f15be-3b56-4999-8010-40d2e86a6813] result = hub.switch() [ 589.517211] env[61806]: ERROR nova.compute.manager [instance: 123f15be-3b56-4999-8010-40d2e86a6813] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 589.517211] env[61806]: ERROR nova.compute.manager [instance: 123f15be-3b56-4999-8010-40d2e86a6813] return self.greenlet.switch() [ 589.517211] env[61806]: ERROR nova.compute.manager [instance: 123f15be-3b56-4999-8010-40d2e86a6813] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 589.517211] env[61806]: ERROR nova.compute.manager [instance: 123f15be-3b56-4999-8010-40d2e86a6813] result = function(*args, **kwargs) [ 589.517211] env[61806]: ERROR nova.compute.manager [instance: 123f15be-3b56-4999-8010-40d2e86a6813] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 589.517211] env[61806]: ERROR nova.compute.manager [instance: 123f15be-3b56-4999-8010-40d2e86a6813] return func(*args, **kwargs) [ 589.517211] env[61806]: ERROR nova.compute.manager [instance: 123f15be-3b56-4999-8010-40d2e86a6813] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 589.517211] env[61806]: ERROR nova.compute.manager [instance: 123f15be-3b56-4999-8010-40d2e86a6813] raise e [ 589.517211] env[61806]: ERROR nova.compute.manager [instance: 123f15be-3b56-4999-8010-40d2e86a6813] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 589.517211] env[61806]: ERROR nova.compute.manager [instance: 123f15be-3b56-4999-8010-40d2e86a6813] nwinfo = self.network_api.allocate_for_instance( [ 589.517211] env[61806]: ERROR nova.compute.manager [instance: 123f15be-3b56-4999-8010-40d2e86a6813] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 589.517211] env[61806]: ERROR nova.compute.manager [instance: 123f15be-3b56-4999-8010-40d2e86a6813] created_port_ids = self._update_ports_for_instance( [ 589.517720] env[61806]: ERROR nova.compute.manager [instance: 123f15be-3b56-4999-8010-40d2e86a6813] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 589.517720] env[61806]: ERROR nova.compute.manager [instance: 123f15be-3b56-4999-8010-40d2e86a6813] with excutils.save_and_reraise_exception(): [ 589.517720] env[61806]: ERROR nova.compute.manager [instance: 123f15be-3b56-4999-8010-40d2e86a6813] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 589.517720] env[61806]: ERROR nova.compute.manager [instance: 123f15be-3b56-4999-8010-40d2e86a6813] self.force_reraise() [ 589.517720] env[61806]: ERROR nova.compute.manager [instance: 123f15be-3b56-4999-8010-40d2e86a6813] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 589.517720] env[61806]: ERROR nova.compute.manager [instance: 123f15be-3b56-4999-8010-40d2e86a6813] raise self.value [ 589.517720] env[61806]: ERROR nova.compute.manager [instance: 123f15be-3b56-4999-8010-40d2e86a6813] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 589.517720] env[61806]: ERROR nova.compute.manager [instance: 123f15be-3b56-4999-8010-40d2e86a6813] updated_port = self._update_port( [ 589.517720] env[61806]: ERROR nova.compute.manager [instance: 123f15be-3b56-4999-8010-40d2e86a6813] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 589.517720] env[61806]: ERROR nova.compute.manager [instance: 123f15be-3b56-4999-8010-40d2e86a6813] _ensure_no_port_binding_failure(port) [ 589.517720] env[61806]: ERROR nova.compute.manager [instance: 123f15be-3b56-4999-8010-40d2e86a6813] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 589.517720] env[61806]: ERROR nova.compute.manager [instance: 123f15be-3b56-4999-8010-40d2e86a6813] raise exception.PortBindingFailed(port_id=port['id']) [ 589.518147] env[61806]: ERROR nova.compute.manager [instance: 123f15be-3b56-4999-8010-40d2e86a6813] nova.exception.PortBindingFailed: Binding failed for port 8fbea74a-6823-4774-b3af-5bda4ccd8820, please check neutron logs for more information. [ 589.518147] env[61806]: ERROR nova.compute.manager [instance: 123f15be-3b56-4999-8010-40d2e86a6813] [ 589.518147] env[61806]: DEBUG nova.compute.utils [None req-be9179b6-8d17-4f04-8823-d48174fcc5c1 tempest-InstanceActionsV221TestJSON-379107765 tempest-InstanceActionsV221TestJSON-379107765-project-member] [instance: 123f15be-3b56-4999-8010-40d2e86a6813] Binding failed for port 8fbea74a-6823-4774-b3af-5bda4ccd8820, please check neutron logs for more information. {{(pid=61806) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 589.518147] env[61806]: DEBUG oslo_concurrency.lockutils [None req-9101cb9e-b191-4241-8bc1-9b126210ca50 tempest-ListImageFiltersTestJSON-680116525 tempest-ListImageFiltersTestJSON-680116525-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 15.525s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 589.520959] env[61806]: INFO nova.compute.claims [None req-9101cb9e-b191-4241-8bc1-9b126210ca50 tempest-ListImageFiltersTestJSON-680116525 tempest-ListImageFiltersTestJSON-680116525-project-member] [instance: 7e730579-8d0c-4226-86bc-292db16c49ee] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 589.526974] env[61806]: DEBUG nova.compute.manager [None req-be9179b6-8d17-4f04-8823-d48174fcc5c1 tempest-InstanceActionsV221TestJSON-379107765 tempest-InstanceActionsV221TestJSON-379107765-project-member] [instance: 123f15be-3b56-4999-8010-40d2e86a6813] Build of instance 123f15be-3b56-4999-8010-40d2e86a6813 was re-scheduled: Binding failed for port 8fbea74a-6823-4774-b3af-5bda4ccd8820, please check neutron logs for more information. {{(pid=61806) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 589.526974] env[61806]: DEBUG nova.compute.manager [None req-be9179b6-8d17-4f04-8823-d48174fcc5c1 tempest-InstanceActionsV221TestJSON-379107765 tempest-InstanceActionsV221TestJSON-379107765-project-member] [instance: 123f15be-3b56-4999-8010-40d2e86a6813] Unplugging VIFs for instance {{(pid=61806) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 589.526974] env[61806]: DEBUG oslo_concurrency.lockutils [None req-be9179b6-8d17-4f04-8823-d48174fcc5c1 tempest-InstanceActionsV221TestJSON-379107765 tempest-InstanceActionsV221TestJSON-379107765-project-member] Acquiring lock "refresh_cache-123f15be-3b56-4999-8010-40d2e86a6813" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 589.526974] env[61806]: DEBUG oslo_concurrency.lockutils [None req-be9179b6-8d17-4f04-8823-d48174fcc5c1 tempest-InstanceActionsV221TestJSON-379107765 tempest-InstanceActionsV221TestJSON-379107765-project-member] Acquired lock "refresh_cache-123f15be-3b56-4999-8010-40d2e86a6813" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 589.527201] env[61806]: DEBUG nova.network.neutron [None req-be9179b6-8d17-4f04-8823-d48174fcc5c1 tempest-InstanceActionsV221TestJSON-379107765 tempest-InstanceActionsV221TestJSON-379107765-project-member] [instance: 123f15be-3b56-4999-8010-40d2e86a6813] Building network info cache for instance {{(pid=61806) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 589.546404] env[61806]: DEBUG nova.compute.manager [None req-b50d0802-a025-4c5c-b8e9-95dd624e7ff5 tempest-ServersWithSpecificFlavorTestJSON-815343160 tempest-ServersWithSpecificFlavorTestJSON-815343160-project-member] [instance: b68b509e-96dc-489d-9cf1-93ba91095af1] Start spawning the instance on the hypervisor. {{(pid=61806) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 589.586155] env[61806]: DEBUG nova.virt.hardware [None req-b50d0802-a025-4c5c-b8e9-95dd624e7ff5 tempest-ServersWithSpecificFlavorTestJSON-815343160 tempest-ServersWithSpecificFlavorTestJSON-815343160-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-15T18:07:10Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='360186094',id=23,is_public=True,memory_mb=192,name='tempest-flavor_with_ephemeral_0-2027440506',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-15T18:04:52Z,direct_url=,disk_format='vmdk',id=ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='84ba83fce2c349a6988173c5d6fc3b07',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-15T18:04:53Z,virtual_size=,visibility=), allow threads: False {{(pid=61806) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 589.586444] env[61806]: DEBUG nova.virt.hardware [None req-b50d0802-a025-4c5c-b8e9-95dd624e7ff5 tempest-ServersWithSpecificFlavorTestJSON-815343160 tempest-ServersWithSpecificFlavorTestJSON-815343160-project-member] Flavor limits 0:0:0 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 589.586753] env[61806]: DEBUG nova.virt.hardware [None req-b50d0802-a025-4c5c-b8e9-95dd624e7ff5 tempest-ServersWithSpecificFlavorTestJSON-815343160 tempest-ServersWithSpecificFlavorTestJSON-815343160-project-member] Image limits 0:0:0 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 589.586826] env[61806]: DEBUG nova.virt.hardware [None req-b50d0802-a025-4c5c-b8e9-95dd624e7ff5 tempest-ServersWithSpecificFlavorTestJSON-815343160 tempest-ServersWithSpecificFlavorTestJSON-815343160-project-member] Flavor pref 0:0:0 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 589.586930] env[61806]: DEBUG nova.virt.hardware [None req-b50d0802-a025-4c5c-b8e9-95dd624e7ff5 tempest-ServersWithSpecificFlavorTestJSON-815343160 tempest-ServersWithSpecificFlavorTestJSON-815343160-project-member] Image pref 0:0:0 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 589.587090] env[61806]: DEBUG nova.virt.hardware [None req-b50d0802-a025-4c5c-b8e9-95dd624e7ff5 tempest-ServersWithSpecificFlavorTestJSON-815343160 tempest-ServersWithSpecificFlavorTestJSON-815343160-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 589.587304] env[61806]: DEBUG nova.virt.hardware [None req-b50d0802-a025-4c5c-b8e9-95dd624e7ff5 tempest-ServersWithSpecificFlavorTestJSON-815343160 tempest-ServersWithSpecificFlavorTestJSON-815343160-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61806) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 589.587463] env[61806]: DEBUG nova.virt.hardware [None req-b50d0802-a025-4c5c-b8e9-95dd624e7ff5 tempest-ServersWithSpecificFlavorTestJSON-815343160 tempest-ServersWithSpecificFlavorTestJSON-815343160-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61806) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 589.587687] env[61806]: DEBUG nova.virt.hardware [None req-b50d0802-a025-4c5c-b8e9-95dd624e7ff5 tempest-ServersWithSpecificFlavorTestJSON-815343160 tempest-ServersWithSpecificFlavorTestJSON-815343160-project-member] Got 1 possible topologies {{(pid=61806) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 589.588020] env[61806]: DEBUG nova.virt.hardware [None req-b50d0802-a025-4c5c-b8e9-95dd624e7ff5 tempest-ServersWithSpecificFlavorTestJSON-815343160 tempest-ServersWithSpecificFlavorTestJSON-815343160-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61806) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 589.588139] env[61806]: DEBUG nova.virt.hardware [None req-b50d0802-a025-4c5c-b8e9-95dd624e7ff5 tempest-ServersWithSpecificFlavorTestJSON-815343160 tempest-ServersWithSpecificFlavorTestJSON-815343160-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61806) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 589.589931] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-882b81a0-3495-417a-9227-71a259bac685 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 589.601440] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c914a8fc-a7e4-4733-b168-64cd85a6bda8 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 589.671984] env[61806]: DEBUG oslo_concurrency.lockutils [None req-3c3981d2-fa25-4f4b-87d4-a5411e67199e tempest-MultipleCreateTestJSON-1589015462 tempest-MultipleCreateTestJSON-1589015462-project-member] Acquiring lock "7211d568-cc68-4e50-80ad-b78878d1deab" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 589.671984] env[61806]: DEBUG oslo_concurrency.lockutils [None req-3c3981d2-fa25-4f4b-87d4-a5411e67199e tempest-MultipleCreateTestJSON-1589015462 tempest-MultipleCreateTestJSON-1589015462-project-member] Lock "7211d568-cc68-4e50-80ad-b78878d1deab" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 589.676158] env[61806]: ERROR nova.compute.manager [None req-b50d0802-a025-4c5c-b8e9-95dd624e7ff5 tempest-ServersWithSpecificFlavorTestJSON-815343160 tempest-ServersWithSpecificFlavorTestJSON-815343160-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port ec4cd098-60bf-4a9b-9f8a-2d85646b89ab, please check neutron logs for more information. [ 589.676158] env[61806]: ERROR nova.compute.manager Traceback (most recent call last): [ 589.676158] env[61806]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 589.676158] env[61806]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 589.676158] env[61806]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 589.676158] env[61806]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 589.676158] env[61806]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 589.676158] env[61806]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 589.676158] env[61806]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 589.676158] env[61806]: ERROR nova.compute.manager self.force_reraise() [ 589.676158] env[61806]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 589.676158] env[61806]: ERROR nova.compute.manager raise self.value [ 589.676158] env[61806]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 589.676158] env[61806]: ERROR nova.compute.manager updated_port = self._update_port( [ 589.676158] env[61806]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 589.676158] env[61806]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 589.677123] env[61806]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 589.677123] env[61806]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 589.677123] env[61806]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port ec4cd098-60bf-4a9b-9f8a-2d85646b89ab, please check neutron logs for more information. [ 589.677123] env[61806]: ERROR nova.compute.manager [ 589.677123] env[61806]: Traceback (most recent call last): [ 589.677123] env[61806]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 589.677123] env[61806]: listener.cb(fileno) [ 589.677123] env[61806]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 589.677123] env[61806]: result = function(*args, **kwargs) [ 589.677123] env[61806]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 589.677123] env[61806]: return func(*args, **kwargs) [ 589.677123] env[61806]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 589.677123] env[61806]: raise e [ 589.677123] env[61806]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 589.677123] env[61806]: nwinfo = self.network_api.allocate_for_instance( [ 589.677123] env[61806]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 589.677123] env[61806]: created_port_ids = self._update_ports_for_instance( [ 589.677123] env[61806]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 589.677123] env[61806]: with excutils.save_and_reraise_exception(): [ 589.677123] env[61806]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 589.677123] env[61806]: self.force_reraise() [ 589.677123] env[61806]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 589.677123] env[61806]: raise self.value [ 589.677123] env[61806]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 589.677123] env[61806]: updated_port = self._update_port( [ 589.677123] env[61806]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 589.677123] env[61806]: _ensure_no_port_binding_failure(port) [ 589.677123] env[61806]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 589.677123] env[61806]: raise exception.PortBindingFailed(port_id=port['id']) [ 589.678334] env[61806]: nova.exception.PortBindingFailed: Binding failed for port ec4cd098-60bf-4a9b-9f8a-2d85646b89ab, please check neutron logs for more information. [ 589.678334] env[61806]: Removing descriptor: 17 [ 589.678334] env[61806]: ERROR nova.compute.manager [None req-b50d0802-a025-4c5c-b8e9-95dd624e7ff5 tempest-ServersWithSpecificFlavorTestJSON-815343160 tempest-ServersWithSpecificFlavorTestJSON-815343160-project-member] [instance: b68b509e-96dc-489d-9cf1-93ba91095af1] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port ec4cd098-60bf-4a9b-9f8a-2d85646b89ab, please check neutron logs for more information. [ 589.678334] env[61806]: ERROR nova.compute.manager [instance: b68b509e-96dc-489d-9cf1-93ba91095af1] Traceback (most recent call last): [ 589.678334] env[61806]: ERROR nova.compute.manager [instance: b68b509e-96dc-489d-9cf1-93ba91095af1] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 589.678334] env[61806]: ERROR nova.compute.manager [instance: b68b509e-96dc-489d-9cf1-93ba91095af1] yield resources [ 589.678334] env[61806]: ERROR nova.compute.manager [instance: b68b509e-96dc-489d-9cf1-93ba91095af1] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 589.678334] env[61806]: ERROR nova.compute.manager [instance: b68b509e-96dc-489d-9cf1-93ba91095af1] self.driver.spawn(context, instance, image_meta, [ 589.678334] env[61806]: ERROR nova.compute.manager [instance: b68b509e-96dc-489d-9cf1-93ba91095af1] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 589.678334] env[61806]: ERROR nova.compute.manager [instance: b68b509e-96dc-489d-9cf1-93ba91095af1] self._vmops.spawn(context, instance, image_meta, injected_files, [ 589.678334] env[61806]: ERROR nova.compute.manager [instance: b68b509e-96dc-489d-9cf1-93ba91095af1] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 589.678334] env[61806]: ERROR nova.compute.manager [instance: b68b509e-96dc-489d-9cf1-93ba91095af1] vm_ref = self.build_virtual_machine(instance, [ 589.678728] env[61806]: ERROR nova.compute.manager [instance: b68b509e-96dc-489d-9cf1-93ba91095af1] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 589.678728] env[61806]: ERROR nova.compute.manager [instance: b68b509e-96dc-489d-9cf1-93ba91095af1] vif_infos = vmwarevif.get_vif_info(self._session, [ 589.678728] env[61806]: ERROR nova.compute.manager [instance: b68b509e-96dc-489d-9cf1-93ba91095af1] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 589.678728] env[61806]: ERROR nova.compute.manager [instance: b68b509e-96dc-489d-9cf1-93ba91095af1] for vif in network_info: [ 589.678728] env[61806]: ERROR nova.compute.manager [instance: b68b509e-96dc-489d-9cf1-93ba91095af1] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 589.678728] env[61806]: ERROR nova.compute.manager [instance: b68b509e-96dc-489d-9cf1-93ba91095af1] return self._sync_wrapper(fn, *args, **kwargs) [ 589.678728] env[61806]: ERROR nova.compute.manager [instance: b68b509e-96dc-489d-9cf1-93ba91095af1] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 589.678728] env[61806]: ERROR nova.compute.manager [instance: b68b509e-96dc-489d-9cf1-93ba91095af1] self.wait() [ 589.678728] env[61806]: ERROR nova.compute.manager [instance: b68b509e-96dc-489d-9cf1-93ba91095af1] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 589.678728] env[61806]: ERROR nova.compute.manager [instance: b68b509e-96dc-489d-9cf1-93ba91095af1] self[:] = self._gt.wait() [ 589.678728] env[61806]: ERROR nova.compute.manager [instance: b68b509e-96dc-489d-9cf1-93ba91095af1] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 589.678728] env[61806]: ERROR nova.compute.manager [instance: b68b509e-96dc-489d-9cf1-93ba91095af1] return self._exit_event.wait() [ 589.678728] env[61806]: ERROR nova.compute.manager [instance: b68b509e-96dc-489d-9cf1-93ba91095af1] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 589.679142] env[61806]: ERROR nova.compute.manager [instance: b68b509e-96dc-489d-9cf1-93ba91095af1] result = hub.switch() [ 589.679142] env[61806]: ERROR nova.compute.manager [instance: b68b509e-96dc-489d-9cf1-93ba91095af1] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 589.679142] env[61806]: ERROR nova.compute.manager [instance: b68b509e-96dc-489d-9cf1-93ba91095af1] return self.greenlet.switch() [ 589.679142] env[61806]: ERROR nova.compute.manager [instance: b68b509e-96dc-489d-9cf1-93ba91095af1] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 589.679142] env[61806]: ERROR nova.compute.manager [instance: b68b509e-96dc-489d-9cf1-93ba91095af1] result = function(*args, **kwargs) [ 589.679142] env[61806]: ERROR nova.compute.manager [instance: b68b509e-96dc-489d-9cf1-93ba91095af1] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 589.679142] env[61806]: ERROR nova.compute.manager [instance: b68b509e-96dc-489d-9cf1-93ba91095af1] return func(*args, **kwargs) [ 589.679142] env[61806]: ERROR nova.compute.manager [instance: b68b509e-96dc-489d-9cf1-93ba91095af1] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 589.679142] env[61806]: ERROR nova.compute.manager [instance: b68b509e-96dc-489d-9cf1-93ba91095af1] raise e [ 589.679142] env[61806]: ERROR nova.compute.manager [instance: b68b509e-96dc-489d-9cf1-93ba91095af1] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 589.679142] env[61806]: ERROR nova.compute.manager [instance: b68b509e-96dc-489d-9cf1-93ba91095af1] nwinfo = self.network_api.allocate_for_instance( [ 589.679142] env[61806]: ERROR nova.compute.manager [instance: b68b509e-96dc-489d-9cf1-93ba91095af1] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 589.679142] env[61806]: ERROR nova.compute.manager [instance: b68b509e-96dc-489d-9cf1-93ba91095af1] created_port_ids = self._update_ports_for_instance( [ 589.679536] env[61806]: ERROR nova.compute.manager [instance: b68b509e-96dc-489d-9cf1-93ba91095af1] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 589.679536] env[61806]: ERROR nova.compute.manager [instance: b68b509e-96dc-489d-9cf1-93ba91095af1] with excutils.save_and_reraise_exception(): [ 589.679536] env[61806]: ERROR nova.compute.manager [instance: b68b509e-96dc-489d-9cf1-93ba91095af1] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 589.679536] env[61806]: ERROR nova.compute.manager [instance: b68b509e-96dc-489d-9cf1-93ba91095af1] self.force_reraise() [ 589.679536] env[61806]: ERROR nova.compute.manager [instance: b68b509e-96dc-489d-9cf1-93ba91095af1] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 589.679536] env[61806]: ERROR nova.compute.manager [instance: b68b509e-96dc-489d-9cf1-93ba91095af1] raise self.value [ 589.679536] env[61806]: ERROR nova.compute.manager [instance: b68b509e-96dc-489d-9cf1-93ba91095af1] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 589.679536] env[61806]: ERROR nova.compute.manager [instance: b68b509e-96dc-489d-9cf1-93ba91095af1] updated_port = self._update_port( [ 589.679536] env[61806]: ERROR nova.compute.manager [instance: b68b509e-96dc-489d-9cf1-93ba91095af1] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 589.679536] env[61806]: ERROR nova.compute.manager [instance: b68b509e-96dc-489d-9cf1-93ba91095af1] _ensure_no_port_binding_failure(port) [ 589.679536] env[61806]: ERROR nova.compute.manager [instance: b68b509e-96dc-489d-9cf1-93ba91095af1] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 589.679536] env[61806]: ERROR nova.compute.manager [instance: b68b509e-96dc-489d-9cf1-93ba91095af1] raise exception.PortBindingFailed(port_id=port['id']) [ 589.679917] env[61806]: ERROR nova.compute.manager [instance: b68b509e-96dc-489d-9cf1-93ba91095af1] nova.exception.PortBindingFailed: Binding failed for port ec4cd098-60bf-4a9b-9f8a-2d85646b89ab, please check neutron logs for more information. [ 589.679917] env[61806]: ERROR nova.compute.manager [instance: b68b509e-96dc-489d-9cf1-93ba91095af1] [ 589.679917] env[61806]: INFO nova.compute.manager [None req-b50d0802-a025-4c5c-b8e9-95dd624e7ff5 tempest-ServersWithSpecificFlavorTestJSON-815343160 tempest-ServersWithSpecificFlavorTestJSON-815343160-project-member] [instance: b68b509e-96dc-489d-9cf1-93ba91095af1] Terminating instance [ 589.679917] env[61806]: DEBUG oslo_concurrency.lockutils [None req-b50d0802-a025-4c5c-b8e9-95dd624e7ff5 tempest-ServersWithSpecificFlavorTestJSON-815343160 tempest-ServersWithSpecificFlavorTestJSON-815343160-project-member] Acquiring lock "refresh_cache-b68b509e-96dc-489d-9cf1-93ba91095af1" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 589.679917] env[61806]: DEBUG oslo_concurrency.lockutils [None req-b50d0802-a025-4c5c-b8e9-95dd624e7ff5 tempest-ServersWithSpecificFlavorTestJSON-815343160 tempest-ServersWithSpecificFlavorTestJSON-815343160-project-member] Acquired lock "refresh_cache-b68b509e-96dc-489d-9cf1-93ba91095af1" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 589.679917] env[61806]: DEBUG nova.network.neutron [None req-b50d0802-a025-4c5c-b8e9-95dd624e7ff5 tempest-ServersWithSpecificFlavorTestJSON-815343160 tempest-ServersWithSpecificFlavorTestJSON-815343160-project-member] [instance: b68b509e-96dc-489d-9cf1-93ba91095af1] Building network info cache for instance {{(pid=61806) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 589.704427] env[61806]: DEBUG oslo_concurrency.lockutils [None req-3c3981d2-fa25-4f4b-87d4-a5411e67199e tempest-MultipleCreateTestJSON-1589015462 tempest-MultipleCreateTestJSON-1589015462-project-member] Acquiring lock "5b22c2fc-525c-481b-b84e-e0a7f68f633e" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 589.704427] env[61806]: DEBUG oslo_concurrency.lockutils [None req-3c3981d2-fa25-4f4b-87d4-a5411e67199e tempest-MultipleCreateTestJSON-1589015462 tempest-MultipleCreateTestJSON-1589015462-project-member] Lock "5b22c2fc-525c-481b-b84e-e0a7f68f633e" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 589.914110] env[61806]: DEBUG nova.network.neutron [-] [instance: 76722306-a9ad-4ef9-b579-d958c36cf6bc] Updating instance_info_cache with network_info: [] {{(pid=61806) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 590.050838] env[61806]: DEBUG nova.network.neutron [None req-be9179b6-8d17-4f04-8823-d48174fcc5c1 tempest-InstanceActionsV221TestJSON-379107765 tempest-InstanceActionsV221TestJSON-379107765-project-member] [instance: 123f15be-3b56-4999-8010-40d2e86a6813] Instance cache missing network info. {{(pid=61806) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 590.140594] env[61806]: DEBUG nova.network.neutron [None req-be9179b6-8d17-4f04-8823-d48174fcc5c1 tempest-InstanceActionsV221TestJSON-379107765 tempest-InstanceActionsV221TestJSON-379107765-project-member] [instance: 123f15be-3b56-4999-8010-40d2e86a6813] Updating instance_info_cache with network_info: [] {{(pid=61806) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 590.196763] env[61806]: DEBUG nova.network.neutron [None req-b50d0802-a025-4c5c-b8e9-95dd624e7ff5 tempest-ServersWithSpecificFlavorTestJSON-815343160 tempest-ServersWithSpecificFlavorTestJSON-815343160-project-member] [instance: b68b509e-96dc-489d-9cf1-93ba91095af1] Instance cache missing network info. {{(pid=61806) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 590.286441] env[61806]: DEBUG nova.network.neutron [None req-b50d0802-a025-4c5c-b8e9-95dd624e7ff5 tempest-ServersWithSpecificFlavorTestJSON-815343160 tempest-ServersWithSpecificFlavorTestJSON-815343160-project-member] [instance: b68b509e-96dc-489d-9cf1-93ba91095af1] Updating instance_info_cache with network_info: [] {{(pid=61806) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 590.416610] env[61806]: INFO nova.compute.manager [-] [instance: 76722306-a9ad-4ef9-b579-d958c36cf6bc] Took 1.03 seconds to deallocate network for instance. [ 590.419293] env[61806]: DEBUG nova.compute.claims [None req-5c11e673-bd02-4846-9be6-b5679ed1117d tempest-VolumesAssistedSnapshotsTest-2037321669 tempest-VolumesAssistedSnapshotsTest-2037321669-project-member] [instance: 76722306-a9ad-4ef9-b579-d958c36cf6bc] Aborting claim: {{(pid=61806) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 590.419503] env[61806]: DEBUG oslo_concurrency.lockutils [None req-5c11e673-bd02-4846-9be6-b5679ed1117d tempest-VolumesAssistedSnapshotsTest-2037321669 tempest-VolumesAssistedSnapshotsTest-2037321669-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 590.644246] env[61806]: DEBUG oslo_concurrency.lockutils [None req-be9179b6-8d17-4f04-8823-d48174fcc5c1 tempest-InstanceActionsV221TestJSON-379107765 tempest-InstanceActionsV221TestJSON-379107765-project-member] Releasing lock "refresh_cache-123f15be-3b56-4999-8010-40d2e86a6813" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 590.644444] env[61806]: DEBUG nova.compute.manager [None req-be9179b6-8d17-4f04-8823-d48174fcc5c1 tempest-InstanceActionsV221TestJSON-379107765 tempest-InstanceActionsV221TestJSON-379107765-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61806) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 590.644625] env[61806]: DEBUG nova.compute.manager [None req-be9179b6-8d17-4f04-8823-d48174fcc5c1 tempest-InstanceActionsV221TestJSON-379107765 tempest-InstanceActionsV221TestJSON-379107765-project-member] [instance: 123f15be-3b56-4999-8010-40d2e86a6813] Deallocating network for instance {{(pid=61806) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 590.644795] env[61806]: DEBUG nova.network.neutron [None req-be9179b6-8d17-4f04-8823-d48174fcc5c1 tempest-InstanceActionsV221TestJSON-379107765 tempest-InstanceActionsV221TestJSON-379107765-project-member] [instance: 123f15be-3b56-4999-8010-40d2e86a6813] deallocate_for_instance() {{(pid=61806) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 590.659181] env[61806]: DEBUG nova.network.neutron [None req-be9179b6-8d17-4f04-8823-d48174fcc5c1 tempest-InstanceActionsV221TestJSON-379107765 tempest-InstanceActionsV221TestJSON-379107765-project-member] [instance: 123f15be-3b56-4999-8010-40d2e86a6813] Instance cache missing network info. {{(pid=61806) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 590.789700] env[61806]: DEBUG oslo_concurrency.lockutils [None req-b50d0802-a025-4c5c-b8e9-95dd624e7ff5 tempest-ServersWithSpecificFlavorTestJSON-815343160 tempest-ServersWithSpecificFlavorTestJSON-815343160-project-member] Releasing lock "refresh_cache-b68b509e-96dc-489d-9cf1-93ba91095af1" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 590.789700] env[61806]: DEBUG nova.compute.manager [None req-b50d0802-a025-4c5c-b8e9-95dd624e7ff5 tempest-ServersWithSpecificFlavorTestJSON-815343160 tempest-ServersWithSpecificFlavorTestJSON-815343160-project-member] [instance: b68b509e-96dc-489d-9cf1-93ba91095af1] Start destroying the instance on the hypervisor. {{(pid=61806) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 590.789837] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-b50d0802-a025-4c5c-b8e9-95dd624e7ff5 tempest-ServersWithSpecificFlavorTestJSON-815343160 tempest-ServersWithSpecificFlavorTestJSON-815343160-project-member] [instance: b68b509e-96dc-489d-9cf1-93ba91095af1] Destroying instance {{(pid=61806) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 590.790621] env[61806]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-17f27930-ee21-4de1-aee0-b2b64b00c6e2 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 590.801411] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eeacb6dc-68d0-4678-8bf2-848f57c3b9d4 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 590.824504] env[61806]: WARNING nova.virt.vmwareapi.vmops [None req-b50d0802-a025-4c5c-b8e9-95dd624e7ff5 tempest-ServersWithSpecificFlavorTestJSON-815343160 tempest-ServersWithSpecificFlavorTestJSON-815343160-project-member] [instance: b68b509e-96dc-489d-9cf1-93ba91095af1] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance b68b509e-96dc-489d-9cf1-93ba91095af1 could not be found. [ 590.824732] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-b50d0802-a025-4c5c-b8e9-95dd624e7ff5 tempest-ServersWithSpecificFlavorTestJSON-815343160 tempest-ServersWithSpecificFlavorTestJSON-815343160-project-member] [instance: b68b509e-96dc-489d-9cf1-93ba91095af1] Instance destroyed {{(pid=61806) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 590.824917] env[61806]: INFO nova.compute.manager [None req-b50d0802-a025-4c5c-b8e9-95dd624e7ff5 tempest-ServersWithSpecificFlavorTestJSON-815343160 tempest-ServersWithSpecificFlavorTestJSON-815343160-project-member] [instance: b68b509e-96dc-489d-9cf1-93ba91095af1] Took 0.04 seconds to destroy the instance on the hypervisor. [ 590.825178] env[61806]: DEBUG oslo.service.loopingcall [None req-b50d0802-a025-4c5c-b8e9-95dd624e7ff5 tempest-ServersWithSpecificFlavorTestJSON-815343160 tempest-ServersWithSpecificFlavorTestJSON-815343160-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61806) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 590.827416] env[61806]: DEBUG nova.compute.manager [-] [instance: b68b509e-96dc-489d-9cf1-93ba91095af1] Deallocating network for instance {{(pid=61806) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 590.827534] env[61806]: DEBUG nova.network.neutron [-] [instance: b68b509e-96dc-489d-9cf1-93ba91095af1] deallocate_for_instance() {{(pid=61806) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 590.847303] env[61806]: DEBUG nova.network.neutron [-] [instance: b68b509e-96dc-489d-9cf1-93ba91095af1] Instance cache missing network info. {{(pid=61806) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 590.918479] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bf5ec248-0788-4088-9d1d-e14ab271a63d {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 590.926121] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b8b9e129-511d-419d-b6be-d198b3c7592b {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 590.959050] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0e4b2c29-865c-4d58-973c-f081616e054a {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 590.966432] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e1731443-299d-45e5-a906-ec9e407bd7bc {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 590.982689] env[61806]: DEBUG nova.compute.provider_tree [None req-9101cb9e-b191-4241-8bc1-9b126210ca50 tempest-ListImageFiltersTestJSON-680116525 tempest-ListImageFiltersTestJSON-680116525-project-member] Inventory has not changed in ProviderTree for provider: a2858be1-fd22-4e08-979e-87ad25293407 {{(pid=61806) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 591.162125] env[61806]: DEBUG nova.network.neutron [None req-be9179b6-8d17-4f04-8823-d48174fcc5c1 tempest-InstanceActionsV221TestJSON-379107765 tempest-InstanceActionsV221TestJSON-379107765-project-member] [instance: 123f15be-3b56-4999-8010-40d2e86a6813] Updating instance_info_cache with network_info: [] {{(pid=61806) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 591.304580] env[61806]: DEBUG nova.compute.manager [req-8e737cd1-da71-4a09-95db-9c565a9dc344 req-e01fc057-a4df-4596-b399-e1780b060ea8 service nova] [instance: b68b509e-96dc-489d-9cf1-93ba91095af1] Received event network-changed-ec4cd098-60bf-4a9b-9f8a-2d85646b89ab {{(pid=61806) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 591.304821] env[61806]: DEBUG nova.compute.manager [req-8e737cd1-da71-4a09-95db-9c565a9dc344 req-e01fc057-a4df-4596-b399-e1780b060ea8 service nova] [instance: b68b509e-96dc-489d-9cf1-93ba91095af1] Refreshing instance network info cache due to event network-changed-ec4cd098-60bf-4a9b-9f8a-2d85646b89ab. {{(pid=61806) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 591.305087] env[61806]: DEBUG oslo_concurrency.lockutils [req-8e737cd1-da71-4a09-95db-9c565a9dc344 req-e01fc057-a4df-4596-b399-e1780b060ea8 service nova] Acquiring lock "refresh_cache-b68b509e-96dc-489d-9cf1-93ba91095af1" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 591.305136] env[61806]: DEBUG oslo_concurrency.lockutils [req-8e737cd1-da71-4a09-95db-9c565a9dc344 req-e01fc057-a4df-4596-b399-e1780b060ea8 service nova] Acquired lock "refresh_cache-b68b509e-96dc-489d-9cf1-93ba91095af1" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 591.305296] env[61806]: DEBUG nova.network.neutron [req-8e737cd1-da71-4a09-95db-9c565a9dc344 req-e01fc057-a4df-4596-b399-e1780b060ea8 service nova] [instance: b68b509e-96dc-489d-9cf1-93ba91095af1] Refreshing network info cache for port ec4cd098-60bf-4a9b-9f8a-2d85646b89ab {{(pid=61806) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 591.350370] env[61806]: DEBUG nova.network.neutron [-] [instance: b68b509e-96dc-489d-9cf1-93ba91095af1] Updating instance_info_cache with network_info: [] {{(pid=61806) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 591.487504] env[61806]: DEBUG nova.scheduler.client.report [None req-9101cb9e-b191-4241-8bc1-9b126210ca50 tempest-ListImageFiltersTestJSON-680116525 tempest-ListImageFiltersTestJSON-680116525-project-member] Inventory has not changed for provider a2858be1-fd22-4e08-979e-87ad25293407 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 139, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61806) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 591.665437] env[61806]: INFO nova.compute.manager [None req-be9179b6-8d17-4f04-8823-d48174fcc5c1 tempest-InstanceActionsV221TestJSON-379107765 tempest-InstanceActionsV221TestJSON-379107765-project-member] [instance: 123f15be-3b56-4999-8010-40d2e86a6813] Took 1.02 seconds to deallocate network for instance. [ 591.821801] env[61806]: DEBUG nova.network.neutron [req-8e737cd1-da71-4a09-95db-9c565a9dc344 req-e01fc057-a4df-4596-b399-e1780b060ea8 service nova] [instance: b68b509e-96dc-489d-9cf1-93ba91095af1] Instance cache missing network info. {{(pid=61806) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 591.853584] env[61806]: INFO nova.compute.manager [-] [instance: b68b509e-96dc-489d-9cf1-93ba91095af1] Took 1.03 seconds to deallocate network for instance. [ 591.855341] env[61806]: DEBUG nova.compute.claims [None req-b50d0802-a025-4c5c-b8e9-95dd624e7ff5 tempest-ServersWithSpecificFlavorTestJSON-815343160 tempest-ServersWithSpecificFlavorTestJSON-815343160-project-member] [instance: b68b509e-96dc-489d-9cf1-93ba91095af1] Aborting claim: {{(pid=61806) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 591.855522] env[61806]: DEBUG oslo_concurrency.lockutils [None req-b50d0802-a025-4c5c-b8e9-95dd624e7ff5 tempest-ServersWithSpecificFlavorTestJSON-815343160 tempest-ServersWithSpecificFlavorTestJSON-815343160-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 591.908015] env[61806]: DEBUG nova.network.neutron [req-8e737cd1-da71-4a09-95db-9c565a9dc344 req-e01fc057-a4df-4596-b399-e1780b060ea8 service nova] [instance: b68b509e-96dc-489d-9cf1-93ba91095af1] Updating instance_info_cache with network_info: [] {{(pid=61806) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 591.992423] env[61806]: DEBUG oslo_concurrency.lockutils [None req-9101cb9e-b191-4241-8bc1-9b126210ca50 tempest-ListImageFiltersTestJSON-680116525 tempest-ListImageFiltersTestJSON-680116525-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.474s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 591.992937] env[61806]: DEBUG nova.compute.manager [None req-9101cb9e-b191-4241-8bc1-9b126210ca50 tempest-ListImageFiltersTestJSON-680116525 tempest-ListImageFiltersTestJSON-680116525-project-member] [instance: 7e730579-8d0c-4226-86bc-292db16c49ee] Start building networks asynchronously for instance. {{(pid=61806) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 591.995348] env[61806]: DEBUG oslo_concurrency.lockutils [None req-962d5603-e71b-4a26-895d-6dce1320b70e tempest-ListImageFiltersTestJSON-680116525 tempest-ListImageFiltersTestJSON-680116525-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 15.772s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 591.997865] env[61806]: INFO nova.compute.claims [None req-962d5603-e71b-4a26-895d-6dce1320b70e tempest-ListImageFiltersTestJSON-680116525 tempest-ListImageFiltersTestJSON-680116525-project-member] [instance: 53d3e2f2-f63d-45c7-a5d1-1fa9a9f0ac0f] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 592.410445] env[61806]: DEBUG oslo_concurrency.lockutils [req-8e737cd1-da71-4a09-95db-9c565a9dc344 req-e01fc057-a4df-4596-b399-e1780b060ea8 service nova] Releasing lock "refresh_cache-b68b509e-96dc-489d-9cf1-93ba91095af1" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 592.410715] env[61806]: DEBUG nova.compute.manager [req-8e737cd1-da71-4a09-95db-9c565a9dc344 req-e01fc057-a4df-4596-b399-e1780b060ea8 service nova] [instance: b68b509e-96dc-489d-9cf1-93ba91095af1] Received event network-vif-deleted-ec4cd098-60bf-4a9b-9f8a-2d85646b89ab {{(pid=61806) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 592.502191] env[61806]: DEBUG nova.compute.utils [None req-9101cb9e-b191-4241-8bc1-9b126210ca50 tempest-ListImageFiltersTestJSON-680116525 tempest-ListImageFiltersTestJSON-680116525-project-member] Using /dev/sd instead of None {{(pid=61806) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 592.506618] env[61806]: DEBUG nova.compute.manager [None req-9101cb9e-b191-4241-8bc1-9b126210ca50 tempest-ListImageFiltersTestJSON-680116525 tempest-ListImageFiltersTestJSON-680116525-project-member] [instance: 7e730579-8d0c-4226-86bc-292db16c49ee] Allocating IP information in the background. {{(pid=61806) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 592.507687] env[61806]: DEBUG nova.network.neutron [None req-9101cb9e-b191-4241-8bc1-9b126210ca50 tempest-ListImageFiltersTestJSON-680116525 tempest-ListImageFiltersTestJSON-680116525-project-member] [instance: 7e730579-8d0c-4226-86bc-292db16c49ee] allocate_for_instance() {{(pid=61806) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 592.563942] env[61806]: DEBUG nova.policy [None req-9101cb9e-b191-4241-8bc1-9b126210ca50 tempest-ListImageFiltersTestJSON-680116525 tempest-ListImageFiltersTestJSON-680116525-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '157d177055ee46d49a8dba2d46dcfe69', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '84e544566d8d4acab78fed67715d2c54', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61806) authorize /opt/stack/nova/nova/policy.py:201}} [ 592.697563] env[61806]: INFO nova.scheduler.client.report [None req-be9179b6-8d17-4f04-8823-d48174fcc5c1 tempest-InstanceActionsV221TestJSON-379107765 tempest-InstanceActionsV221TestJSON-379107765-project-member] Deleted allocations for instance 123f15be-3b56-4999-8010-40d2e86a6813 [ 592.918129] env[61806]: DEBUG nova.network.neutron [None req-9101cb9e-b191-4241-8bc1-9b126210ca50 tempest-ListImageFiltersTestJSON-680116525 tempest-ListImageFiltersTestJSON-680116525-project-member] [instance: 7e730579-8d0c-4226-86bc-292db16c49ee] Successfully created port: 24a71d6d-5a30-4980-900b-c78cb8631d22 {{(pid=61806) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 593.007665] env[61806]: DEBUG nova.compute.manager [None req-9101cb9e-b191-4241-8bc1-9b126210ca50 tempest-ListImageFiltersTestJSON-680116525 tempest-ListImageFiltersTestJSON-680116525-project-member] [instance: 7e730579-8d0c-4226-86bc-292db16c49ee] Start building block device mappings for instance. {{(pid=61806) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 593.211046] env[61806]: DEBUG oslo_concurrency.lockutils [None req-be9179b6-8d17-4f04-8823-d48174fcc5c1 tempest-InstanceActionsV221TestJSON-379107765 tempest-InstanceActionsV221TestJSON-379107765-project-member] Lock "123f15be-3b56-4999-8010-40d2e86a6813" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 60.169s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 593.468971] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bf01de0c-9f65-496c-b228-1d95e513dc7d {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 593.477093] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7b804b0a-039c-4656-86df-68621eefc9a3 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 593.515886] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6bb1348a-485b-4f87-a8b8-abbb7cb5a605 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 593.527263] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-85a50126-c0a1-4b47-8e48-f9c58c8af9aa {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 593.543232] env[61806]: DEBUG nova.compute.provider_tree [None req-962d5603-e71b-4a26-895d-6dce1320b70e tempest-ListImageFiltersTestJSON-680116525 tempest-ListImageFiltersTestJSON-680116525-project-member] Inventory has not changed in ProviderTree for provider: a2858be1-fd22-4e08-979e-87ad25293407 {{(pid=61806) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 593.714344] env[61806]: DEBUG nova.compute.manager [None req-e34d2bcf-266f-4faa-8f31-6bd93f6ab176 tempest-ImagesTestJSON-618852556 tempest-ImagesTestJSON-618852556-project-member] [instance: 65f3b2f3-6d25-44f0-95cb-f9a7e75f8e82] Starting instance... {{(pid=61806) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 593.786166] env[61806]: DEBUG nova.compute.manager [req-723307aa-e43b-4b28-a30e-5acc5126f909 req-a5944a4b-37eb-44f3-9cb1-7cdaf5f32298 service nova] [instance: 7e730579-8d0c-4226-86bc-292db16c49ee] Received event network-changed-24a71d6d-5a30-4980-900b-c78cb8631d22 {{(pid=61806) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 593.786166] env[61806]: DEBUG nova.compute.manager [req-723307aa-e43b-4b28-a30e-5acc5126f909 req-a5944a4b-37eb-44f3-9cb1-7cdaf5f32298 service nova] [instance: 7e730579-8d0c-4226-86bc-292db16c49ee] Refreshing instance network info cache due to event network-changed-24a71d6d-5a30-4980-900b-c78cb8631d22. {{(pid=61806) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 593.786166] env[61806]: DEBUG oslo_concurrency.lockutils [req-723307aa-e43b-4b28-a30e-5acc5126f909 req-a5944a4b-37eb-44f3-9cb1-7cdaf5f32298 service nova] Acquiring lock "refresh_cache-7e730579-8d0c-4226-86bc-292db16c49ee" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 593.786166] env[61806]: DEBUG oslo_concurrency.lockutils [req-723307aa-e43b-4b28-a30e-5acc5126f909 req-a5944a4b-37eb-44f3-9cb1-7cdaf5f32298 service nova] Acquired lock "refresh_cache-7e730579-8d0c-4226-86bc-292db16c49ee" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 593.786166] env[61806]: DEBUG nova.network.neutron [req-723307aa-e43b-4b28-a30e-5acc5126f909 req-a5944a4b-37eb-44f3-9cb1-7cdaf5f32298 service nova] [instance: 7e730579-8d0c-4226-86bc-292db16c49ee] Refreshing network info cache for port 24a71d6d-5a30-4980-900b-c78cb8631d22 {{(pid=61806) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 593.970914] env[61806]: ERROR nova.compute.manager [None req-9101cb9e-b191-4241-8bc1-9b126210ca50 tempest-ListImageFiltersTestJSON-680116525 tempest-ListImageFiltersTestJSON-680116525-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 24a71d6d-5a30-4980-900b-c78cb8631d22, please check neutron logs for more information. [ 593.970914] env[61806]: ERROR nova.compute.manager Traceback (most recent call last): [ 593.970914] env[61806]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 593.970914] env[61806]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 593.970914] env[61806]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 593.970914] env[61806]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 593.970914] env[61806]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 593.970914] env[61806]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 593.970914] env[61806]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 593.970914] env[61806]: ERROR nova.compute.manager self.force_reraise() [ 593.970914] env[61806]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 593.970914] env[61806]: ERROR nova.compute.manager raise self.value [ 593.970914] env[61806]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 593.970914] env[61806]: ERROR nova.compute.manager updated_port = self._update_port( [ 593.970914] env[61806]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 593.970914] env[61806]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 593.971732] env[61806]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 593.971732] env[61806]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 593.971732] env[61806]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 24a71d6d-5a30-4980-900b-c78cb8631d22, please check neutron logs for more information. [ 593.971732] env[61806]: ERROR nova.compute.manager [ 593.971732] env[61806]: Traceback (most recent call last): [ 593.971732] env[61806]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 593.971732] env[61806]: listener.cb(fileno) [ 593.971732] env[61806]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 593.971732] env[61806]: result = function(*args, **kwargs) [ 593.971732] env[61806]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 593.971732] env[61806]: return func(*args, **kwargs) [ 593.971732] env[61806]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 593.971732] env[61806]: raise e [ 593.971732] env[61806]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 593.971732] env[61806]: nwinfo = self.network_api.allocate_for_instance( [ 593.971732] env[61806]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 593.971732] env[61806]: created_port_ids = self._update_ports_for_instance( [ 593.971732] env[61806]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 593.971732] env[61806]: with excutils.save_and_reraise_exception(): [ 593.971732] env[61806]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 593.971732] env[61806]: self.force_reraise() [ 593.971732] env[61806]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 593.971732] env[61806]: raise self.value [ 593.971732] env[61806]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 593.971732] env[61806]: updated_port = self._update_port( [ 593.971732] env[61806]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 593.971732] env[61806]: _ensure_no_port_binding_failure(port) [ 593.971732] env[61806]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 593.971732] env[61806]: raise exception.PortBindingFailed(port_id=port['id']) [ 593.973165] env[61806]: nova.exception.PortBindingFailed: Binding failed for port 24a71d6d-5a30-4980-900b-c78cb8631d22, please check neutron logs for more information. [ 593.973165] env[61806]: Removing descriptor: 17 [ 594.025700] env[61806]: DEBUG nova.compute.manager [None req-9101cb9e-b191-4241-8bc1-9b126210ca50 tempest-ListImageFiltersTestJSON-680116525 tempest-ListImageFiltersTestJSON-680116525-project-member] [instance: 7e730579-8d0c-4226-86bc-292db16c49ee] Start spawning the instance on the hypervisor. {{(pid=61806) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 594.047260] env[61806]: DEBUG nova.scheduler.client.report [None req-962d5603-e71b-4a26-895d-6dce1320b70e tempest-ListImageFiltersTestJSON-680116525 tempest-ListImageFiltersTestJSON-680116525-project-member] Inventory has not changed for provider a2858be1-fd22-4e08-979e-87ad25293407 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 139, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61806) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 594.055984] env[61806]: DEBUG nova.virt.hardware [None req-9101cb9e-b191-4241-8bc1-9b126210ca50 tempest-ListImageFiltersTestJSON-680116525 tempest-ListImageFiltersTestJSON-680116525-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-15T18:05:10Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-15T18:04:52Z,direct_url=,disk_format='vmdk',id=ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='84ba83fce2c349a6988173c5d6fc3b07',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-15T18:04:53Z,virtual_size=,visibility=), allow threads: False {{(pid=61806) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 594.056587] env[61806]: DEBUG nova.virt.hardware [None req-9101cb9e-b191-4241-8bc1-9b126210ca50 tempest-ListImageFiltersTestJSON-680116525 tempest-ListImageFiltersTestJSON-680116525-project-member] Flavor limits 0:0:0 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 594.057065] env[61806]: DEBUG nova.virt.hardware [None req-9101cb9e-b191-4241-8bc1-9b126210ca50 tempest-ListImageFiltersTestJSON-680116525 tempest-ListImageFiltersTestJSON-680116525-project-member] Image limits 0:0:0 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 594.057669] env[61806]: DEBUG nova.virt.hardware [None req-9101cb9e-b191-4241-8bc1-9b126210ca50 tempest-ListImageFiltersTestJSON-680116525 tempest-ListImageFiltersTestJSON-680116525-project-member] Flavor pref 0:0:0 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 594.058094] env[61806]: DEBUG nova.virt.hardware [None req-9101cb9e-b191-4241-8bc1-9b126210ca50 tempest-ListImageFiltersTestJSON-680116525 tempest-ListImageFiltersTestJSON-680116525-project-member] Image pref 0:0:0 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 594.058541] env[61806]: DEBUG nova.virt.hardware [None req-9101cb9e-b191-4241-8bc1-9b126210ca50 tempest-ListImageFiltersTestJSON-680116525 tempest-ListImageFiltersTestJSON-680116525-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 594.058957] env[61806]: DEBUG nova.virt.hardware [None req-9101cb9e-b191-4241-8bc1-9b126210ca50 tempest-ListImageFiltersTestJSON-680116525 tempest-ListImageFiltersTestJSON-680116525-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61806) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 594.059298] env[61806]: DEBUG nova.virt.hardware [None req-9101cb9e-b191-4241-8bc1-9b126210ca50 tempest-ListImageFiltersTestJSON-680116525 tempest-ListImageFiltersTestJSON-680116525-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61806) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 594.059631] env[61806]: DEBUG nova.virt.hardware [None req-9101cb9e-b191-4241-8bc1-9b126210ca50 tempest-ListImageFiltersTestJSON-680116525 tempest-ListImageFiltersTestJSON-680116525-project-member] Got 1 possible topologies {{(pid=61806) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 594.060236] env[61806]: DEBUG nova.virt.hardware [None req-9101cb9e-b191-4241-8bc1-9b126210ca50 tempest-ListImageFiltersTestJSON-680116525 tempest-ListImageFiltersTestJSON-680116525-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61806) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 594.060598] env[61806]: DEBUG nova.virt.hardware [None req-9101cb9e-b191-4241-8bc1-9b126210ca50 tempest-ListImageFiltersTestJSON-680116525 tempest-ListImageFiltersTestJSON-680116525-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61806) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 594.062273] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9747f750-c82e-4a8e-8c46-94f0e1031097 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 594.071386] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f255ecf3-57e0-45ca-984c-e90277a8abf4 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 594.086746] env[61806]: ERROR nova.compute.manager [None req-9101cb9e-b191-4241-8bc1-9b126210ca50 tempest-ListImageFiltersTestJSON-680116525 tempest-ListImageFiltersTestJSON-680116525-project-member] [instance: 7e730579-8d0c-4226-86bc-292db16c49ee] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 24a71d6d-5a30-4980-900b-c78cb8631d22, please check neutron logs for more information. [ 594.086746] env[61806]: ERROR nova.compute.manager [instance: 7e730579-8d0c-4226-86bc-292db16c49ee] Traceback (most recent call last): [ 594.086746] env[61806]: ERROR nova.compute.manager [instance: 7e730579-8d0c-4226-86bc-292db16c49ee] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 594.086746] env[61806]: ERROR nova.compute.manager [instance: 7e730579-8d0c-4226-86bc-292db16c49ee] yield resources [ 594.086746] env[61806]: ERROR nova.compute.manager [instance: 7e730579-8d0c-4226-86bc-292db16c49ee] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 594.086746] env[61806]: ERROR nova.compute.manager [instance: 7e730579-8d0c-4226-86bc-292db16c49ee] self.driver.spawn(context, instance, image_meta, [ 594.086746] env[61806]: ERROR nova.compute.manager [instance: 7e730579-8d0c-4226-86bc-292db16c49ee] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 594.086746] env[61806]: ERROR nova.compute.manager [instance: 7e730579-8d0c-4226-86bc-292db16c49ee] self._vmops.spawn(context, instance, image_meta, injected_files, [ 594.086746] env[61806]: ERROR nova.compute.manager [instance: 7e730579-8d0c-4226-86bc-292db16c49ee] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 594.086746] env[61806]: ERROR nova.compute.manager [instance: 7e730579-8d0c-4226-86bc-292db16c49ee] vm_ref = self.build_virtual_machine(instance, [ 594.086746] env[61806]: ERROR nova.compute.manager [instance: 7e730579-8d0c-4226-86bc-292db16c49ee] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 594.087217] env[61806]: ERROR nova.compute.manager [instance: 7e730579-8d0c-4226-86bc-292db16c49ee] vif_infos = vmwarevif.get_vif_info(self._session, [ 594.087217] env[61806]: ERROR nova.compute.manager [instance: 7e730579-8d0c-4226-86bc-292db16c49ee] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 594.087217] env[61806]: ERROR nova.compute.manager [instance: 7e730579-8d0c-4226-86bc-292db16c49ee] for vif in network_info: [ 594.087217] env[61806]: ERROR nova.compute.manager [instance: 7e730579-8d0c-4226-86bc-292db16c49ee] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 594.087217] env[61806]: ERROR nova.compute.manager [instance: 7e730579-8d0c-4226-86bc-292db16c49ee] return self._sync_wrapper(fn, *args, **kwargs) [ 594.087217] env[61806]: ERROR nova.compute.manager [instance: 7e730579-8d0c-4226-86bc-292db16c49ee] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 594.087217] env[61806]: ERROR nova.compute.manager [instance: 7e730579-8d0c-4226-86bc-292db16c49ee] self.wait() [ 594.087217] env[61806]: ERROR nova.compute.manager [instance: 7e730579-8d0c-4226-86bc-292db16c49ee] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 594.087217] env[61806]: ERROR nova.compute.manager [instance: 7e730579-8d0c-4226-86bc-292db16c49ee] self[:] = self._gt.wait() [ 594.087217] env[61806]: ERROR nova.compute.manager [instance: 7e730579-8d0c-4226-86bc-292db16c49ee] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 594.087217] env[61806]: ERROR nova.compute.manager [instance: 7e730579-8d0c-4226-86bc-292db16c49ee] return self._exit_event.wait() [ 594.087217] env[61806]: ERROR nova.compute.manager [instance: 7e730579-8d0c-4226-86bc-292db16c49ee] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 594.087217] env[61806]: ERROR nova.compute.manager [instance: 7e730579-8d0c-4226-86bc-292db16c49ee] current.throw(*self._exc) [ 594.087650] env[61806]: ERROR nova.compute.manager [instance: 7e730579-8d0c-4226-86bc-292db16c49ee] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 594.087650] env[61806]: ERROR nova.compute.manager [instance: 7e730579-8d0c-4226-86bc-292db16c49ee] result = function(*args, **kwargs) [ 594.087650] env[61806]: ERROR nova.compute.manager [instance: 7e730579-8d0c-4226-86bc-292db16c49ee] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 594.087650] env[61806]: ERROR nova.compute.manager [instance: 7e730579-8d0c-4226-86bc-292db16c49ee] return func(*args, **kwargs) [ 594.087650] env[61806]: ERROR nova.compute.manager [instance: 7e730579-8d0c-4226-86bc-292db16c49ee] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 594.087650] env[61806]: ERROR nova.compute.manager [instance: 7e730579-8d0c-4226-86bc-292db16c49ee] raise e [ 594.087650] env[61806]: ERROR nova.compute.manager [instance: 7e730579-8d0c-4226-86bc-292db16c49ee] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 594.087650] env[61806]: ERROR nova.compute.manager [instance: 7e730579-8d0c-4226-86bc-292db16c49ee] nwinfo = self.network_api.allocate_for_instance( [ 594.087650] env[61806]: ERROR nova.compute.manager [instance: 7e730579-8d0c-4226-86bc-292db16c49ee] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 594.087650] env[61806]: ERROR nova.compute.manager [instance: 7e730579-8d0c-4226-86bc-292db16c49ee] created_port_ids = self._update_ports_for_instance( [ 594.087650] env[61806]: ERROR nova.compute.manager [instance: 7e730579-8d0c-4226-86bc-292db16c49ee] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 594.087650] env[61806]: ERROR nova.compute.manager [instance: 7e730579-8d0c-4226-86bc-292db16c49ee] with excutils.save_and_reraise_exception(): [ 594.087650] env[61806]: ERROR nova.compute.manager [instance: 7e730579-8d0c-4226-86bc-292db16c49ee] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 594.088090] env[61806]: ERROR nova.compute.manager [instance: 7e730579-8d0c-4226-86bc-292db16c49ee] self.force_reraise() [ 594.088090] env[61806]: ERROR nova.compute.manager [instance: 7e730579-8d0c-4226-86bc-292db16c49ee] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 594.088090] env[61806]: ERROR nova.compute.manager [instance: 7e730579-8d0c-4226-86bc-292db16c49ee] raise self.value [ 594.088090] env[61806]: ERROR nova.compute.manager [instance: 7e730579-8d0c-4226-86bc-292db16c49ee] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 594.088090] env[61806]: ERROR nova.compute.manager [instance: 7e730579-8d0c-4226-86bc-292db16c49ee] updated_port = self._update_port( [ 594.088090] env[61806]: ERROR nova.compute.manager [instance: 7e730579-8d0c-4226-86bc-292db16c49ee] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 594.088090] env[61806]: ERROR nova.compute.manager [instance: 7e730579-8d0c-4226-86bc-292db16c49ee] _ensure_no_port_binding_failure(port) [ 594.088090] env[61806]: ERROR nova.compute.manager [instance: 7e730579-8d0c-4226-86bc-292db16c49ee] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 594.088090] env[61806]: ERROR nova.compute.manager [instance: 7e730579-8d0c-4226-86bc-292db16c49ee] raise exception.PortBindingFailed(port_id=port['id']) [ 594.088090] env[61806]: ERROR nova.compute.manager [instance: 7e730579-8d0c-4226-86bc-292db16c49ee] nova.exception.PortBindingFailed: Binding failed for port 24a71d6d-5a30-4980-900b-c78cb8631d22, please check neutron logs for more information. [ 594.088090] env[61806]: ERROR nova.compute.manager [instance: 7e730579-8d0c-4226-86bc-292db16c49ee] [ 594.088090] env[61806]: INFO nova.compute.manager [None req-9101cb9e-b191-4241-8bc1-9b126210ca50 tempest-ListImageFiltersTestJSON-680116525 tempest-ListImageFiltersTestJSON-680116525-project-member] [instance: 7e730579-8d0c-4226-86bc-292db16c49ee] Terminating instance [ 594.091743] env[61806]: DEBUG oslo_concurrency.lockutils [None req-9101cb9e-b191-4241-8bc1-9b126210ca50 tempest-ListImageFiltersTestJSON-680116525 tempest-ListImageFiltersTestJSON-680116525-project-member] Acquiring lock "refresh_cache-7e730579-8d0c-4226-86bc-292db16c49ee" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 594.232113] env[61806]: DEBUG oslo_concurrency.lockutils [None req-e34d2bcf-266f-4faa-8f31-6bd93f6ab176 tempest-ImagesTestJSON-618852556 tempest-ImagesTestJSON-618852556-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 594.307505] env[61806]: DEBUG nova.network.neutron [req-723307aa-e43b-4b28-a30e-5acc5126f909 req-a5944a4b-37eb-44f3-9cb1-7cdaf5f32298 service nova] [instance: 7e730579-8d0c-4226-86bc-292db16c49ee] Instance cache missing network info. {{(pid=61806) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 594.367717] env[61806]: DEBUG nova.network.neutron [req-723307aa-e43b-4b28-a30e-5acc5126f909 req-a5944a4b-37eb-44f3-9cb1-7cdaf5f32298 service nova] [instance: 7e730579-8d0c-4226-86bc-292db16c49ee] Updating instance_info_cache with network_info: [] {{(pid=61806) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 594.552069] env[61806]: DEBUG oslo_concurrency.lockutils [None req-962d5603-e71b-4a26-895d-6dce1320b70e tempest-ListImageFiltersTestJSON-680116525 tempest-ListImageFiltersTestJSON-680116525-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.556s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 594.552069] env[61806]: DEBUG nova.compute.manager [None req-962d5603-e71b-4a26-895d-6dce1320b70e tempest-ListImageFiltersTestJSON-680116525 tempest-ListImageFiltersTestJSON-680116525-project-member] [instance: 53d3e2f2-f63d-45c7-a5d1-1fa9a9f0ac0f] Start building networks asynchronously for instance. {{(pid=61806) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 594.556663] env[61806]: DEBUG oslo_concurrency.lockutils [None req-11b5d1a0-518c-42d9-940e-4bf652a6d358 tempest-DeleteServersAdminTestJSON-1553667494 tempest-DeleteServersAdminTestJSON-1553667494-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 16.428s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 594.558295] env[61806]: INFO nova.compute.claims [None req-11b5d1a0-518c-42d9-940e-4bf652a6d358 tempest-DeleteServersAdminTestJSON-1553667494 tempest-DeleteServersAdminTestJSON-1553667494-project-member] [instance: ddc801a0-0782-4c6d-bd72-95d0afd01339] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 594.874676] env[61806]: DEBUG oslo_concurrency.lockutils [req-723307aa-e43b-4b28-a30e-5acc5126f909 req-a5944a4b-37eb-44f3-9cb1-7cdaf5f32298 service nova] Releasing lock "refresh_cache-7e730579-8d0c-4226-86bc-292db16c49ee" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 594.874676] env[61806]: DEBUG oslo_concurrency.lockutils [None req-9101cb9e-b191-4241-8bc1-9b126210ca50 tempest-ListImageFiltersTestJSON-680116525 tempest-ListImageFiltersTestJSON-680116525-project-member] Acquired lock "refresh_cache-7e730579-8d0c-4226-86bc-292db16c49ee" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 594.874676] env[61806]: DEBUG nova.network.neutron [None req-9101cb9e-b191-4241-8bc1-9b126210ca50 tempest-ListImageFiltersTestJSON-680116525 tempest-ListImageFiltersTestJSON-680116525-project-member] [instance: 7e730579-8d0c-4226-86bc-292db16c49ee] Building network info cache for instance {{(pid=61806) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 595.060326] env[61806]: DEBUG nova.compute.utils [None req-962d5603-e71b-4a26-895d-6dce1320b70e tempest-ListImageFiltersTestJSON-680116525 tempest-ListImageFiltersTestJSON-680116525-project-member] Using /dev/sd instead of None {{(pid=61806) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 595.060326] env[61806]: DEBUG nova.compute.manager [None req-962d5603-e71b-4a26-895d-6dce1320b70e tempest-ListImageFiltersTestJSON-680116525 tempest-ListImageFiltersTestJSON-680116525-project-member] [instance: 53d3e2f2-f63d-45c7-a5d1-1fa9a9f0ac0f] Allocating IP information in the background. {{(pid=61806) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 595.060326] env[61806]: DEBUG nova.network.neutron [None req-962d5603-e71b-4a26-895d-6dce1320b70e tempest-ListImageFiltersTestJSON-680116525 tempest-ListImageFiltersTestJSON-680116525-project-member] [instance: 53d3e2f2-f63d-45c7-a5d1-1fa9a9f0ac0f] allocate_for_instance() {{(pid=61806) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 595.132527] env[61806]: DEBUG nova.policy [None req-962d5603-e71b-4a26-895d-6dce1320b70e tempest-ListImageFiltersTestJSON-680116525 tempest-ListImageFiltersTestJSON-680116525-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '157d177055ee46d49a8dba2d46dcfe69', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '84e544566d8d4acab78fed67715d2c54', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61806) authorize /opt/stack/nova/nova/policy.py:201}} [ 595.392381] env[61806]: DEBUG nova.network.neutron [None req-9101cb9e-b191-4241-8bc1-9b126210ca50 tempest-ListImageFiltersTestJSON-680116525 tempest-ListImageFiltersTestJSON-680116525-project-member] [instance: 7e730579-8d0c-4226-86bc-292db16c49ee] Instance cache missing network info. {{(pid=61806) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 595.449907] env[61806]: DEBUG nova.network.neutron [None req-9101cb9e-b191-4241-8bc1-9b126210ca50 tempest-ListImageFiltersTestJSON-680116525 tempest-ListImageFiltersTestJSON-680116525-project-member] [instance: 7e730579-8d0c-4226-86bc-292db16c49ee] Updating instance_info_cache with network_info: [] {{(pid=61806) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 595.544498] env[61806]: DEBUG nova.network.neutron [None req-962d5603-e71b-4a26-895d-6dce1320b70e tempest-ListImageFiltersTestJSON-680116525 tempest-ListImageFiltersTestJSON-680116525-project-member] [instance: 53d3e2f2-f63d-45c7-a5d1-1fa9a9f0ac0f] Successfully created port: f650d966-513e-4686-9faa-9dd139c8c888 {{(pid=61806) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 595.563157] env[61806]: DEBUG nova.compute.manager [None req-962d5603-e71b-4a26-895d-6dce1320b70e tempest-ListImageFiltersTestJSON-680116525 tempest-ListImageFiltersTestJSON-680116525-project-member] [instance: 53d3e2f2-f63d-45c7-a5d1-1fa9a9f0ac0f] Start building block device mappings for instance. {{(pid=61806) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 595.855104] env[61806]: DEBUG nova.compute.manager [req-9315456f-0e24-4191-8022-8f40f7f7afc0 req-f58209e8-6c14-486c-af2c-97ff41ffae2c service nova] [instance: 7e730579-8d0c-4226-86bc-292db16c49ee] Received event network-vif-deleted-24a71d6d-5a30-4980-900b-c78cb8631d22 {{(pid=61806) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 595.954829] env[61806]: DEBUG oslo_concurrency.lockutils [None req-9101cb9e-b191-4241-8bc1-9b126210ca50 tempest-ListImageFiltersTestJSON-680116525 tempest-ListImageFiltersTestJSON-680116525-project-member] Releasing lock "refresh_cache-7e730579-8d0c-4226-86bc-292db16c49ee" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 595.955259] env[61806]: DEBUG nova.compute.manager [None req-9101cb9e-b191-4241-8bc1-9b126210ca50 tempest-ListImageFiltersTestJSON-680116525 tempest-ListImageFiltersTestJSON-680116525-project-member] [instance: 7e730579-8d0c-4226-86bc-292db16c49ee] Start destroying the instance on the hypervisor. {{(pid=61806) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 595.955451] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-9101cb9e-b191-4241-8bc1-9b126210ca50 tempest-ListImageFiltersTestJSON-680116525 tempest-ListImageFiltersTestJSON-680116525-project-member] [instance: 7e730579-8d0c-4226-86bc-292db16c49ee] Destroying instance {{(pid=61806) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 595.955969] env[61806]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-438c5e02-e150-4900-83a5-9c4921c544f8 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 595.968742] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dfae787b-8fd9-43d9-919d-bd405f377971 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 595.993249] env[61806]: WARNING nova.virt.vmwareapi.vmops [None req-9101cb9e-b191-4241-8bc1-9b126210ca50 tempest-ListImageFiltersTestJSON-680116525 tempest-ListImageFiltersTestJSON-680116525-project-member] [instance: 7e730579-8d0c-4226-86bc-292db16c49ee] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 7e730579-8d0c-4226-86bc-292db16c49ee could not be found. [ 595.993529] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-9101cb9e-b191-4241-8bc1-9b126210ca50 tempest-ListImageFiltersTestJSON-680116525 tempest-ListImageFiltersTestJSON-680116525-project-member] [instance: 7e730579-8d0c-4226-86bc-292db16c49ee] Instance destroyed {{(pid=61806) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 595.993771] env[61806]: INFO nova.compute.manager [None req-9101cb9e-b191-4241-8bc1-9b126210ca50 tempest-ListImageFiltersTestJSON-680116525 tempest-ListImageFiltersTestJSON-680116525-project-member] [instance: 7e730579-8d0c-4226-86bc-292db16c49ee] Took 0.04 seconds to destroy the instance on the hypervisor. [ 595.994080] env[61806]: DEBUG oslo.service.loopingcall [None req-9101cb9e-b191-4241-8bc1-9b126210ca50 tempest-ListImageFiltersTestJSON-680116525 tempest-ListImageFiltersTestJSON-680116525-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61806) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 595.996013] env[61806]: DEBUG nova.compute.manager [-] [instance: 7e730579-8d0c-4226-86bc-292db16c49ee] Deallocating network for instance {{(pid=61806) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 595.996013] env[61806]: DEBUG nova.network.neutron [-] [instance: 7e730579-8d0c-4226-86bc-292db16c49ee] deallocate_for_instance() {{(pid=61806) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 596.013086] env[61806]: DEBUG nova.network.neutron [-] [instance: 7e730579-8d0c-4226-86bc-292db16c49ee] Instance cache missing network info. {{(pid=61806) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 596.064208] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-423a9670-3455-4b0d-a3f0-337976b750c9 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 596.078082] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8f3491d9-239e-4786-8243-78b9a8f6176b {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 596.113585] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1b1e44d1-f711-49ad-ba9e-822eed418484 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 596.119756] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4e4dfe28-a9b2-4710-a3a3-8bb48bd2664d {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 596.134767] env[61806]: DEBUG nova.compute.provider_tree [None req-11b5d1a0-518c-42d9-940e-4bf652a6d358 tempest-DeleteServersAdminTestJSON-1553667494 tempest-DeleteServersAdminTestJSON-1553667494-project-member] Inventory has not changed in ProviderTree for provider: a2858be1-fd22-4e08-979e-87ad25293407 {{(pid=61806) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 596.517617] env[61806]: DEBUG nova.network.neutron [-] [instance: 7e730579-8d0c-4226-86bc-292db16c49ee] Updating instance_info_cache with network_info: [] {{(pid=61806) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 596.574109] env[61806]: DEBUG nova.compute.manager [None req-962d5603-e71b-4a26-895d-6dce1320b70e tempest-ListImageFiltersTestJSON-680116525 tempest-ListImageFiltersTestJSON-680116525-project-member] [instance: 53d3e2f2-f63d-45c7-a5d1-1fa9a9f0ac0f] Start spawning the instance on the hypervisor. {{(pid=61806) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 596.599852] env[61806]: DEBUG nova.virt.hardware [None req-962d5603-e71b-4a26-895d-6dce1320b70e tempest-ListImageFiltersTestJSON-680116525 tempest-ListImageFiltersTestJSON-680116525-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-15T18:05:10Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-15T18:04:52Z,direct_url=,disk_format='vmdk',id=ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='84ba83fce2c349a6988173c5d6fc3b07',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-15T18:04:53Z,virtual_size=,visibility=), allow threads: False {{(pid=61806) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 596.600125] env[61806]: DEBUG nova.virt.hardware [None req-962d5603-e71b-4a26-895d-6dce1320b70e tempest-ListImageFiltersTestJSON-680116525 tempest-ListImageFiltersTestJSON-680116525-project-member] Flavor limits 0:0:0 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 596.600289] env[61806]: DEBUG nova.virt.hardware [None req-962d5603-e71b-4a26-895d-6dce1320b70e tempest-ListImageFiltersTestJSON-680116525 tempest-ListImageFiltersTestJSON-680116525-project-member] Image limits 0:0:0 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 596.600468] env[61806]: DEBUG nova.virt.hardware [None req-962d5603-e71b-4a26-895d-6dce1320b70e tempest-ListImageFiltersTestJSON-680116525 tempest-ListImageFiltersTestJSON-680116525-project-member] Flavor pref 0:0:0 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 596.600612] env[61806]: DEBUG nova.virt.hardware [None req-962d5603-e71b-4a26-895d-6dce1320b70e tempest-ListImageFiltersTestJSON-680116525 tempest-ListImageFiltersTestJSON-680116525-project-member] Image pref 0:0:0 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 596.600765] env[61806]: DEBUG nova.virt.hardware [None req-962d5603-e71b-4a26-895d-6dce1320b70e tempest-ListImageFiltersTestJSON-680116525 tempest-ListImageFiltersTestJSON-680116525-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 596.601079] env[61806]: DEBUG nova.virt.hardware [None req-962d5603-e71b-4a26-895d-6dce1320b70e tempest-ListImageFiltersTestJSON-680116525 tempest-ListImageFiltersTestJSON-680116525-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61806) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 596.601264] env[61806]: DEBUG nova.virt.hardware [None req-962d5603-e71b-4a26-895d-6dce1320b70e tempest-ListImageFiltersTestJSON-680116525 tempest-ListImageFiltersTestJSON-680116525-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61806) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 596.601435] env[61806]: DEBUG nova.virt.hardware [None req-962d5603-e71b-4a26-895d-6dce1320b70e tempest-ListImageFiltersTestJSON-680116525 tempest-ListImageFiltersTestJSON-680116525-project-member] Got 1 possible topologies {{(pid=61806) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 596.601598] env[61806]: DEBUG nova.virt.hardware [None req-962d5603-e71b-4a26-895d-6dce1320b70e tempest-ListImageFiltersTestJSON-680116525 tempest-ListImageFiltersTestJSON-680116525-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61806) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 596.601771] env[61806]: DEBUG nova.virt.hardware [None req-962d5603-e71b-4a26-895d-6dce1320b70e tempest-ListImageFiltersTestJSON-680116525 tempest-ListImageFiltersTestJSON-680116525-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61806) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 596.602651] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6ee4af23-f27c-4e4d-ba92-ff0fb1b0dc34 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 596.610394] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-537dc129-e749-46f0-acc3-f8441bfdb293 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 596.637819] env[61806]: DEBUG nova.scheduler.client.report [None req-11b5d1a0-518c-42d9-940e-4bf652a6d358 tempest-DeleteServersAdminTestJSON-1553667494 tempest-DeleteServersAdminTestJSON-1553667494-project-member] Inventory has not changed for provider a2858be1-fd22-4e08-979e-87ad25293407 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 139, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61806) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 596.873773] env[61806]: ERROR nova.compute.manager [None req-962d5603-e71b-4a26-895d-6dce1320b70e tempest-ListImageFiltersTestJSON-680116525 tempest-ListImageFiltersTestJSON-680116525-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port f650d966-513e-4686-9faa-9dd139c8c888, please check neutron logs for more information. [ 596.873773] env[61806]: ERROR nova.compute.manager Traceback (most recent call last): [ 596.873773] env[61806]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 596.873773] env[61806]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 596.873773] env[61806]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 596.873773] env[61806]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 596.873773] env[61806]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 596.873773] env[61806]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 596.873773] env[61806]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 596.873773] env[61806]: ERROR nova.compute.manager self.force_reraise() [ 596.873773] env[61806]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 596.873773] env[61806]: ERROR nova.compute.manager raise self.value [ 596.873773] env[61806]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 596.873773] env[61806]: ERROR nova.compute.manager updated_port = self._update_port( [ 596.873773] env[61806]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 596.873773] env[61806]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 596.874353] env[61806]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 596.874353] env[61806]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 596.874353] env[61806]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port f650d966-513e-4686-9faa-9dd139c8c888, please check neutron logs for more information. [ 596.874353] env[61806]: ERROR nova.compute.manager [ 596.874353] env[61806]: Traceback (most recent call last): [ 596.874353] env[61806]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 596.874353] env[61806]: listener.cb(fileno) [ 596.874353] env[61806]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 596.874353] env[61806]: result = function(*args, **kwargs) [ 596.874353] env[61806]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 596.874353] env[61806]: return func(*args, **kwargs) [ 596.874353] env[61806]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 596.874353] env[61806]: raise e [ 596.874353] env[61806]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 596.874353] env[61806]: nwinfo = self.network_api.allocate_for_instance( [ 596.874353] env[61806]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 596.874353] env[61806]: created_port_ids = self._update_ports_for_instance( [ 596.874353] env[61806]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 596.874353] env[61806]: with excutils.save_and_reraise_exception(): [ 596.874353] env[61806]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 596.874353] env[61806]: self.force_reraise() [ 596.874353] env[61806]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 596.874353] env[61806]: raise self.value [ 596.874353] env[61806]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 596.874353] env[61806]: updated_port = self._update_port( [ 596.874353] env[61806]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 596.874353] env[61806]: _ensure_no_port_binding_failure(port) [ 596.874353] env[61806]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 596.874353] env[61806]: raise exception.PortBindingFailed(port_id=port['id']) [ 596.875343] env[61806]: nova.exception.PortBindingFailed: Binding failed for port f650d966-513e-4686-9faa-9dd139c8c888, please check neutron logs for more information. [ 596.875343] env[61806]: Removing descriptor: 17 [ 596.875343] env[61806]: ERROR nova.compute.manager [None req-962d5603-e71b-4a26-895d-6dce1320b70e tempest-ListImageFiltersTestJSON-680116525 tempest-ListImageFiltersTestJSON-680116525-project-member] [instance: 53d3e2f2-f63d-45c7-a5d1-1fa9a9f0ac0f] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port f650d966-513e-4686-9faa-9dd139c8c888, please check neutron logs for more information. [ 596.875343] env[61806]: ERROR nova.compute.manager [instance: 53d3e2f2-f63d-45c7-a5d1-1fa9a9f0ac0f] Traceback (most recent call last): [ 596.875343] env[61806]: ERROR nova.compute.manager [instance: 53d3e2f2-f63d-45c7-a5d1-1fa9a9f0ac0f] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 596.875343] env[61806]: ERROR nova.compute.manager [instance: 53d3e2f2-f63d-45c7-a5d1-1fa9a9f0ac0f] yield resources [ 596.875343] env[61806]: ERROR nova.compute.manager [instance: 53d3e2f2-f63d-45c7-a5d1-1fa9a9f0ac0f] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 596.875343] env[61806]: ERROR nova.compute.manager [instance: 53d3e2f2-f63d-45c7-a5d1-1fa9a9f0ac0f] self.driver.spawn(context, instance, image_meta, [ 596.875343] env[61806]: ERROR nova.compute.manager [instance: 53d3e2f2-f63d-45c7-a5d1-1fa9a9f0ac0f] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 596.875343] env[61806]: ERROR nova.compute.manager [instance: 53d3e2f2-f63d-45c7-a5d1-1fa9a9f0ac0f] self._vmops.spawn(context, instance, image_meta, injected_files, [ 596.875343] env[61806]: ERROR nova.compute.manager [instance: 53d3e2f2-f63d-45c7-a5d1-1fa9a9f0ac0f] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 596.875343] env[61806]: ERROR nova.compute.manager [instance: 53d3e2f2-f63d-45c7-a5d1-1fa9a9f0ac0f] vm_ref = self.build_virtual_machine(instance, [ 596.875761] env[61806]: ERROR nova.compute.manager [instance: 53d3e2f2-f63d-45c7-a5d1-1fa9a9f0ac0f] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 596.875761] env[61806]: ERROR nova.compute.manager [instance: 53d3e2f2-f63d-45c7-a5d1-1fa9a9f0ac0f] vif_infos = vmwarevif.get_vif_info(self._session, [ 596.875761] env[61806]: ERROR nova.compute.manager [instance: 53d3e2f2-f63d-45c7-a5d1-1fa9a9f0ac0f] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 596.875761] env[61806]: ERROR nova.compute.manager [instance: 53d3e2f2-f63d-45c7-a5d1-1fa9a9f0ac0f] for vif in network_info: [ 596.875761] env[61806]: ERROR nova.compute.manager [instance: 53d3e2f2-f63d-45c7-a5d1-1fa9a9f0ac0f] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 596.875761] env[61806]: ERROR nova.compute.manager [instance: 53d3e2f2-f63d-45c7-a5d1-1fa9a9f0ac0f] return self._sync_wrapper(fn, *args, **kwargs) [ 596.875761] env[61806]: ERROR nova.compute.manager [instance: 53d3e2f2-f63d-45c7-a5d1-1fa9a9f0ac0f] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 596.875761] env[61806]: ERROR nova.compute.manager [instance: 53d3e2f2-f63d-45c7-a5d1-1fa9a9f0ac0f] self.wait() [ 596.875761] env[61806]: ERROR nova.compute.manager [instance: 53d3e2f2-f63d-45c7-a5d1-1fa9a9f0ac0f] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 596.875761] env[61806]: ERROR nova.compute.manager [instance: 53d3e2f2-f63d-45c7-a5d1-1fa9a9f0ac0f] self[:] = self._gt.wait() [ 596.875761] env[61806]: ERROR nova.compute.manager [instance: 53d3e2f2-f63d-45c7-a5d1-1fa9a9f0ac0f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 596.875761] env[61806]: ERROR nova.compute.manager [instance: 53d3e2f2-f63d-45c7-a5d1-1fa9a9f0ac0f] return self._exit_event.wait() [ 596.875761] env[61806]: ERROR nova.compute.manager [instance: 53d3e2f2-f63d-45c7-a5d1-1fa9a9f0ac0f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 596.876259] env[61806]: ERROR nova.compute.manager [instance: 53d3e2f2-f63d-45c7-a5d1-1fa9a9f0ac0f] result = hub.switch() [ 596.876259] env[61806]: ERROR nova.compute.manager [instance: 53d3e2f2-f63d-45c7-a5d1-1fa9a9f0ac0f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 596.876259] env[61806]: ERROR nova.compute.manager [instance: 53d3e2f2-f63d-45c7-a5d1-1fa9a9f0ac0f] return self.greenlet.switch() [ 596.876259] env[61806]: ERROR nova.compute.manager [instance: 53d3e2f2-f63d-45c7-a5d1-1fa9a9f0ac0f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 596.876259] env[61806]: ERROR nova.compute.manager [instance: 53d3e2f2-f63d-45c7-a5d1-1fa9a9f0ac0f] result = function(*args, **kwargs) [ 596.876259] env[61806]: ERROR nova.compute.manager [instance: 53d3e2f2-f63d-45c7-a5d1-1fa9a9f0ac0f] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 596.876259] env[61806]: ERROR nova.compute.manager [instance: 53d3e2f2-f63d-45c7-a5d1-1fa9a9f0ac0f] return func(*args, **kwargs) [ 596.876259] env[61806]: ERROR nova.compute.manager [instance: 53d3e2f2-f63d-45c7-a5d1-1fa9a9f0ac0f] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 596.876259] env[61806]: ERROR nova.compute.manager [instance: 53d3e2f2-f63d-45c7-a5d1-1fa9a9f0ac0f] raise e [ 596.876259] env[61806]: ERROR nova.compute.manager [instance: 53d3e2f2-f63d-45c7-a5d1-1fa9a9f0ac0f] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 596.876259] env[61806]: ERROR nova.compute.manager [instance: 53d3e2f2-f63d-45c7-a5d1-1fa9a9f0ac0f] nwinfo = self.network_api.allocate_for_instance( [ 596.876259] env[61806]: ERROR nova.compute.manager [instance: 53d3e2f2-f63d-45c7-a5d1-1fa9a9f0ac0f] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 596.876259] env[61806]: ERROR nova.compute.manager [instance: 53d3e2f2-f63d-45c7-a5d1-1fa9a9f0ac0f] created_port_ids = self._update_ports_for_instance( [ 596.876671] env[61806]: ERROR nova.compute.manager [instance: 53d3e2f2-f63d-45c7-a5d1-1fa9a9f0ac0f] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 596.876671] env[61806]: ERROR nova.compute.manager [instance: 53d3e2f2-f63d-45c7-a5d1-1fa9a9f0ac0f] with excutils.save_and_reraise_exception(): [ 596.876671] env[61806]: ERROR nova.compute.manager [instance: 53d3e2f2-f63d-45c7-a5d1-1fa9a9f0ac0f] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 596.876671] env[61806]: ERROR nova.compute.manager [instance: 53d3e2f2-f63d-45c7-a5d1-1fa9a9f0ac0f] self.force_reraise() [ 596.876671] env[61806]: ERROR nova.compute.manager [instance: 53d3e2f2-f63d-45c7-a5d1-1fa9a9f0ac0f] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 596.876671] env[61806]: ERROR nova.compute.manager [instance: 53d3e2f2-f63d-45c7-a5d1-1fa9a9f0ac0f] raise self.value [ 596.876671] env[61806]: ERROR nova.compute.manager [instance: 53d3e2f2-f63d-45c7-a5d1-1fa9a9f0ac0f] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 596.876671] env[61806]: ERROR nova.compute.manager [instance: 53d3e2f2-f63d-45c7-a5d1-1fa9a9f0ac0f] updated_port = self._update_port( [ 596.876671] env[61806]: ERROR nova.compute.manager [instance: 53d3e2f2-f63d-45c7-a5d1-1fa9a9f0ac0f] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 596.876671] env[61806]: ERROR nova.compute.manager [instance: 53d3e2f2-f63d-45c7-a5d1-1fa9a9f0ac0f] _ensure_no_port_binding_failure(port) [ 596.876671] env[61806]: ERROR nova.compute.manager [instance: 53d3e2f2-f63d-45c7-a5d1-1fa9a9f0ac0f] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 596.876671] env[61806]: ERROR nova.compute.manager [instance: 53d3e2f2-f63d-45c7-a5d1-1fa9a9f0ac0f] raise exception.PortBindingFailed(port_id=port['id']) [ 596.877125] env[61806]: ERROR nova.compute.manager [instance: 53d3e2f2-f63d-45c7-a5d1-1fa9a9f0ac0f] nova.exception.PortBindingFailed: Binding failed for port f650d966-513e-4686-9faa-9dd139c8c888, please check neutron logs for more information. [ 596.877125] env[61806]: ERROR nova.compute.manager [instance: 53d3e2f2-f63d-45c7-a5d1-1fa9a9f0ac0f] [ 596.877125] env[61806]: INFO nova.compute.manager [None req-962d5603-e71b-4a26-895d-6dce1320b70e tempest-ListImageFiltersTestJSON-680116525 tempest-ListImageFiltersTestJSON-680116525-project-member] [instance: 53d3e2f2-f63d-45c7-a5d1-1fa9a9f0ac0f] Terminating instance [ 596.877224] env[61806]: DEBUG oslo_concurrency.lockutils [None req-962d5603-e71b-4a26-895d-6dce1320b70e tempest-ListImageFiltersTestJSON-680116525 tempest-ListImageFiltersTestJSON-680116525-project-member] Acquiring lock "refresh_cache-53d3e2f2-f63d-45c7-a5d1-1fa9a9f0ac0f" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 596.877311] env[61806]: DEBUG oslo_concurrency.lockutils [None req-962d5603-e71b-4a26-895d-6dce1320b70e tempest-ListImageFiltersTestJSON-680116525 tempest-ListImageFiltersTestJSON-680116525-project-member] Acquired lock "refresh_cache-53d3e2f2-f63d-45c7-a5d1-1fa9a9f0ac0f" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 596.877476] env[61806]: DEBUG nova.network.neutron [None req-962d5603-e71b-4a26-895d-6dce1320b70e tempest-ListImageFiltersTestJSON-680116525 tempest-ListImageFiltersTestJSON-680116525-project-member] [instance: 53d3e2f2-f63d-45c7-a5d1-1fa9a9f0ac0f] Building network info cache for instance {{(pid=61806) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 597.018374] env[61806]: INFO nova.compute.manager [-] [instance: 7e730579-8d0c-4226-86bc-292db16c49ee] Took 1.02 seconds to deallocate network for instance. [ 597.021274] env[61806]: DEBUG nova.compute.claims [None req-9101cb9e-b191-4241-8bc1-9b126210ca50 tempest-ListImageFiltersTestJSON-680116525 tempest-ListImageFiltersTestJSON-680116525-project-member] [instance: 7e730579-8d0c-4226-86bc-292db16c49ee] Aborting claim: {{(pid=61806) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 597.021582] env[61806]: DEBUG oslo_concurrency.lockutils [None req-9101cb9e-b191-4241-8bc1-9b126210ca50 tempest-ListImageFiltersTestJSON-680116525 tempest-ListImageFiltersTestJSON-680116525-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 597.142496] env[61806]: DEBUG oslo_concurrency.lockutils [None req-11b5d1a0-518c-42d9-940e-4bf652a6d358 tempest-DeleteServersAdminTestJSON-1553667494 tempest-DeleteServersAdminTestJSON-1553667494-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.586s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 597.143062] env[61806]: DEBUG nova.compute.manager [None req-11b5d1a0-518c-42d9-940e-4bf652a6d358 tempest-DeleteServersAdminTestJSON-1553667494 tempest-DeleteServersAdminTestJSON-1553667494-project-member] [instance: ddc801a0-0782-4c6d-bd72-95d0afd01339] Start building networks asynchronously for instance. {{(pid=61806) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 597.145562] env[61806]: DEBUG oslo_concurrency.lockutils [None req-38c4eaec-22d8-4bf0-bd45-321ef43329a0 tempest-ServersAdminTestJSON-14199920 tempest-ServersAdminTestJSON-14199920-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 15.867s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 597.396387] env[61806]: DEBUG nova.network.neutron [None req-962d5603-e71b-4a26-895d-6dce1320b70e tempest-ListImageFiltersTestJSON-680116525 tempest-ListImageFiltersTestJSON-680116525-project-member] [instance: 53d3e2f2-f63d-45c7-a5d1-1fa9a9f0ac0f] Instance cache missing network info. {{(pid=61806) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 597.443347] env[61806]: DEBUG nova.network.neutron [None req-962d5603-e71b-4a26-895d-6dce1320b70e tempest-ListImageFiltersTestJSON-680116525 tempest-ListImageFiltersTestJSON-680116525-project-member] [instance: 53d3e2f2-f63d-45c7-a5d1-1fa9a9f0ac0f] Updating instance_info_cache with network_info: [] {{(pid=61806) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 597.656995] env[61806]: DEBUG nova.compute.utils [None req-11b5d1a0-518c-42d9-940e-4bf652a6d358 tempest-DeleteServersAdminTestJSON-1553667494 tempest-DeleteServersAdminTestJSON-1553667494-project-member] Using /dev/sd instead of None {{(pid=61806) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 597.659216] env[61806]: DEBUG nova.compute.manager [None req-11b5d1a0-518c-42d9-940e-4bf652a6d358 tempest-DeleteServersAdminTestJSON-1553667494 tempest-DeleteServersAdminTestJSON-1553667494-project-member] [instance: ddc801a0-0782-4c6d-bd72-95d0afd01339] Allocating IP information in the background. {{(pid=61806) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 597.659449] env[61806]: DEBUG nova.network.neutron [None req-11b5d1a0-518c-42d9-940e-4bf652a6d358 tempest-DeleteServersAdminTestJSON-1553667494 tempest-DeleteServersAdminTestJSON-1553667494-project-member] [instance: ddc801a0-0782-4c6d-bd72-95d0afd01339] allocate_for_instance() {{(pid=61806) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 597.761041] env[61806]: DEBUG nova.policy [None req-11b5d1a0-518c-42d9-940e-4bf652a6d358 tempest-DeleteServersAdminTestJSON-1553667494 tempest-DeleteServersAdminTestJSON-1553667494-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '19ec5bc21d684c88a8004e53b9939ed1', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '45024eb0d0fe4ccdbb80b05fda894ce7', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61806) authorize /opt/stack/nova/nova/policy.py:201}} [ 597.908012] env[61806]: DEBUG nova.compute.manager [req-e5115be9-5f56-4afc-acf5-c862d5a1fed4 req-566b841b-0733-4bcd-9b61-6a62adab23dd service nova] [instance: 53d3e2f2-f63d-45c7-a5d1-1fa9a9f0ac0f] Received event network-changed-f650d966-513e-4686-9faa-9dd139c8c888 {{(pid=61806) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 597.908012] env[61806]: DEBUG nova.compute.manager [req-e5115be9-5f56-4afc-acf5-c862d5a1fed4 req-566b841b-0733-4bcd-9b61-6a62adab23dd service nova] [instance: 53d3e2f2-f63d-45c7-a5d1-1fa9a9f0ac0f] Refreshing instance network info cache due to event network-changed-f650d966-513e-4686-9faa-9dd139c8c888. {{(pid=61806) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 597.908012] env[61806]: DEBUG oslo_concurrency.lockutils [req-e5115be9-5f56-4afc-acf5-c862d5a1fed4 req-566b841b-0733-4bcd-9b61-6a62adab23dd service nova] Acquiring lock "refresh_cache-53d3e2f2-f63d-45c7-a5d1-1fa9a9f0ac0f" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 597.946313] env[61806]: DEBUG oslo_concurrency.lockutils [None req-962d5603-e71b-4a26-895d-6dce1320b70e tempest-ListImageFiltersTestJSON-680116525 tempest-ListImageFiltersTestJSON-680116525-project-member] Releasing lock "refresh_cache-53d3e2f2-f63d-45c7-a5d1-1fa9a9f0ac0f" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 597.947482] env[61806]: DEBUG nova.compute.manager [None req-962d5603-e71b-4a26-895d-6dce1320b70e tempest-ListImageFiltersTestJSON-680116525 tempest-ListImageFiltersTestJSON-680116525-project-member] [instance: 53d3e2f2-f63d-45c7-a5d1-1fa9a9f0ac0f] Start destroying the instance on the hypervisor. {{(pid=61806) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 597.947667] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-962d5603-e71b-4a26-895d-6dce1320b70e tempest-ListImageFiltersTestJSON-680116525 tempest-ListImageFiltersTestJSON-680116525-project-member] [instance: 53d3e2f2-f63d-45c7-a5d1-1fa9a9f0ac0f] Destroying instance {{(pid=61806) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 597.950792] env[61806]: DEBUG oslo_concurrency.lockutils [req-e5115be9-5f56-4afc-acf5-c862d5a1fed4 req-566b841b-0733-4bcd-9b61-6a62adab23dd service nova] Acquired lock "refresh_cache-53d3e2f2-f63d-45c7-a5d1-1fa9a9f0ac0f" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 597.950990] env[61806]: DEBUG nova.network.neutron [req-e5115be9-5f56-4afc-acf5-c862d5a1fed4 req-566b841b-0733-4bcd-9b61-6a62adab23dd service nova] [instance: 53d3e2f2-f63d-45c7-a5d1-1fa9a9f0ac0f] Refreshing network info cache for port f650d966-513e-4686-9faa-9dd139c8c888 {{(pid=61806) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 597.952103] env[61806]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-119202bd-d2fc-4582-9e7d-0f20b73dfef6 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 597.961148] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9d728ec5-0863-4f53-9fd6-954dc959b3a3 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 597.992097] env[61806]: WARNING nova.virt.vmwareapi.vmops [None req-962d5603-e71b-4a26-895d-6dce1320b70e tempest-ListImageFiltersTestJSON-680116525 tempest-ListImageFiltersTestJSON-680116525-project-member] [instance: 53d3e2f2-f63d-45c7-a5d1-1fa9a9f0ac0f] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 53d3e2f2-f63d-45c7-a5d1-1fa9a9f0ac0f could not be found. [ 597.992374] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-962d5603-e71b-4a26-895d-6dce1320b70e tempest-ListImageFiltersTestJSON-680116525 tempest-ListImageFiltersTestJSON-680116525-project-member] [instance: 53d3e2f2-f63d-45c7-a5d1-1fa9a9f0ac0f] Instance destroyed {{(pid=61806) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 597.992561] env[61806]: INFO nova.compute.manager [None req-962d5603-e71b-4a26-895d-6dce1320b70e tempest-ListImageFiltersTestJSON-680116525 tempest-ListImageFiltersTestJSON-680116525-project-member] [instance: 53d3e2f2-f63d-45c7-a5d1-1fa9a9f0ac0f] Took 0.05 seconds to destroy the instance on the hypervisor. [ 597.992801] env[61806]: DEBUG oslo.service.loopingcall [None req-962d5603-e71b-4a26-895d-6dce1320b70e tempest-ListImageFiltersTestJSON-680116525 tempest-ListImageFiltersTestJSON-680116525-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61806) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 597.995894] env[61806]: DEBUG nova.compute.manager [-] [instance: 53d3e2f2-f63d-45c7-a5d1-1fa9a9f0ac0f] Deallocating network for instance {{(pid=61806) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 597.995989] env[61806]: DEBUG nova.network.neutron [-] [instance: 53d3e2f2-f63d-45c7-a5d1-1fa9a9f0ac0f] deallocate_for_instance() {{(pid=61806) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 598.009730] env[61806]: DEBUG nova.network.neutron [-] [instance: 53d3e2f2-f63d-45c7-a5d1-1fa9a9f0ac0f] Instance cache missing network info. {{(pid=61806) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 598.133018] env[61806]: DEBUG nova.network.neutron [None req-11b5d1a0-518c-42d9-940e-4bf652a6d358 tempest-DeleteServersAdminTestJSON-1553667494 tempest-DeleteServersAdminTestJSON-1553667494-project-member] [instance: ddc801a0-0782-4c6d-bd72-95d0afd01339] Successfully created port: f65985e6-72fe-4194-a8d3-fc371e6fd74f {{(pid=61806) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 598.158259] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b7d5c707-fc16-4ff0-9970-3b148b4190ab {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 598.162957] env[61806]: DEBUG nova.compute.manager [None req-11b5d1a0-518c-42d9-940e-4bf652a6d358 tempest-DeleteServersAdminTestJSON-1553667494 tempest-DeleteServersAdminTestJSON-1553667494-project-member] [instance: ddc801a0-0782-4c6d-bd72-95d0afd01339] Start building block device mappings for instance. {{(pid=61806) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 598.169966] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-27d2e8c8-4882-43e1-839b-6bf15b498858 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 598.209290] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-438ebe5e-47c1-4e45-9f9a-ae320aac182a {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 598.218788] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-72d0383a-9454-4ff1-b0bf-236fc213e7bd {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 598.233455] env[61806]: DEBUG nova.compute.provider_tree [None req-38c4eaec-22d8-4bf0-bd45-321ef43329a0 tempest-ServersAdminTestJSON-14199920 tempest-ServersAdminTestJSON-14199920-project-member] Inventory has not changed in ProviderTree for provider: a2858be1-fd22-4e08-979e-87ad25293407 {{(pid=61806) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 598.475240] env[61806]: DEBUG nova.network.neutron [req-e5115be9-5f56-4afc-acf5-c862d5a1fed4 req-566b841b-0733-4bcd-9b61-6a62adab23dd service nova] [instance: 53d3e2f2-f63d-45c7-a5d1-1fa9a9f0ac0f] Instance cache missing network info. {{(pid=61806) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 598.517732] env[61806]: DEBUG nova.network.neutron [-] [instance: 53d3e2f2-f63d-45c7-a5d1-1fa9a9f0ac0f] Updating instance_info_cache with network_info: [] {{(pid=61806) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 598.547843] env[61806]: DEBUG nova.network.neutron [req-e5115be9-5f56-4afc-acf5-c862d5a1fed4 req-566b841b-0733-4bcd-9b61-6a62adab23dd service nova] [instance: 53d3e2f2-f63d-45c7-a5d1-1fa9a9f0ac0f] Updating instance_info_cache with network_info: [] {{(pid=61806) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 598.748778] env[61806]: DEBUG nova.scheduler.client.report [None req-38c4eaec-22d8-4bf0-bd45-321ef43329a0 tempest-ServersAdminTestJSON-14199920 tempest-ServersAdminTestJSON-14199920-project-member] Inventory has not changed for provider a2858be1-fd22-4e08-979e-87ad25293407 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 139, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61806) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 599.023674] env[61806]: INFO nova.compute.manager [-] [instance: 53d3e2f2-f63d-45c7-a5d1-1fa9a9f0ac0f] Took 1.03 seconds to deallocate network for instance. [ 599.026251] env[61806]: DEBUG nova.compute.claims [None req-962d5603-e71b-4a26-895d-6dce1320b70e tempest-ListImageFiltersTestJSON-680116525 tempest-ListImageFiltersTestJSON-680116525-project-member] [instance: 53d3e2f2-f63d-45c7-a5d1-1fa9a9f0ac0f] Aborting claim: {{(pid=61806) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 599.026441] env[61806]: DEBUG oslo_concurrency.lockutils [None req-962d5603-e71b-4a26-895d-6dce1320b70e tempest-ListImageFiltersTestJSON-680116525 tempest-ListImageFiltersTestJSON-680116525-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 599.056855] env[61806]: DEBUG oslo_concurrency.lockutils [req-e5115be9-5f56-4afc-acf5-c862d5a1fed4 req-566b841b-0733-4bcd-9b61-6a62adab23dd service nova] Releasing lock "refresh_cache-53d3e2f2-f63d-45c7-a5d1-1fa9a9f0ac0f" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 599.056855] env[61806]: DEBUG nova.compute.manager [req-e5115be9-5f56-4afc-acf5-c862d5a1fed4 req-566b841b-0733-4bcd-9b61-6a62adab23dd service nova] [instance: 53d3e2f2-f63d-45c7-a5d1-1fa9a9f0ac0f] Received event network-vif-deleted-f650d966-513e-4686-9faa-9dd139c8c888 {{(pid=61806) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 599.178020] env[61806]: DEBUG nova.compute.manager [None req-11b5d1a0-518c-42d9-940e-4bf652a6d358 tempest-DeleteServersAdminTestJSON-1553667494 tempest-DeleteServersAdminTestJSON-1553667494-project-member] [instance: ddc801a0-0782-4c6d-bd72-95d0afd01339] Start spawning the instance on the hypervisor. {{(pid=61806) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 599.208410] env[61806]: DEBUG nova.virt.hardware [None req-11b5d1a0-518c-42d9-940e-4bf652a6d358 tempest-DeleteServersAdminTestJSON-1553667494 tempest-DeleteServersAdminTestJSON-1553667494-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-15T18:05:10Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-15T18:04:52Z,direct_url=,disk_format='vmdk',id=ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='84ba83fce2c349a6988173c5d6fc3b07',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-15T18:04:53Z,virtual_size=,visibility=), allow threads: False {{(pid=61806) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 599.209651] env[61806]: DEBUG nova.virt.hardware [None req-11b5d1a0-518c-42d9-940e-4bf652a6d358 tempest-DeleteServersAdminTestJSON-1553667494 tempest-DeleteServersAdminTestJSON-1553667494-project-member] Flavor limits 0:0:0 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 599.209651] env[61806]: DEBUG nova.virt.hardware [None req-11b5d1a0-518c-42d9-940e-4bf652a6d358 tempest-DeleteServersAdminTestJSON-1553667494 tempest-DeleteServersAdminTestJSON-1553667494-project-member] Image limits 0:0:0 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 599.209651] env[61806]: DEBUG nova.virt.hardware [None req-11b5d1a0-518c-42d9-940e-4bf652a6d358 tempest-DeleteServersAdminTestJSON-1553667494 tempest-DeleteServersAdminTestJSON-1553667494-project-member] Flavor pref 0:0:0 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 599.209651] env[61806]: DEBUG nova.virt.hardware [None req-11b5d1a0-518c-42d9-940e-4bf652a6d358 tempest-DeleteServersAdminTestJSON-1553667494 tempest-DeleteServersAdminTestJSON-1553667494-project-member] Image pref 0:0:0 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 599.209651] env[61806]: DEBUG nova.virt.hardware [None req-11b5d1a0-518c-42d9-940e-4bf652a6d358 tempest-DeleteServersAdminTestJSON-1553667494 tempest-DeleteServersAdminTestJSON-1553667494-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 599.209866] env[61806]: DEBUG nova.virt.hardware [None req-11b5d1a0-518c-42d9-940e-4bf652a6d358 tempest-DeleteServersAdminTestJSON-1553667494 tempest-DeleteServersAdminTestJSON-1553667494-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61806) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 599.209866] env[61806]: DEBUG nova.virt.hardware [None req-11b5d1a0-518c-42d9-940e-4bf652a6d358 tempest-DeleteServersAdminTestJSON-1553667494 tempest-DeleteServersAdminTestJSON-1553667494-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61806) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 599.210045] env[61806]: DEBUG nova.virt.hardware [None req-11b5d1a0-518c-42d9-940e-4bf652a6d358 tempest-DeleteServersAdminTestJSON-1553667494 tempest-DeleteServersAdminTestJSON-1553667494-project-member] Got 1 possible topologies {{(pid=61806) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 599.210889] env[61806]: DEBUG nova.virt.hardware [None req-11b5d1a0-518c-42d9-940e-4bf652a6d358 tempest-DeleteServersAdminTestJSON-1553667494 tempest-DeleteServersAdminTestJSON-1553667494-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61806) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 599.210889] env[61806]: DEBUG nova.virt.hardware [None req-11b5d1a0-518c-42d9-940e-4bf652a6d358 tempest-DeleteServersAdminTestJSON-1553667494 tempest-DeleteServersAdminTestJSON-1553667494-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61806) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 599.211823] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bccfaf0f-f97a-471e-8248-c8d7d3698541 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 599.222499] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6026ced2-5863-489f-9740-6abb8561e536 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 599.243645] env[61806]: DEBUG oslo_concurrency.lockutils [None req-38c4eaec-22d8-4bf0-bd45-321ef43329a0 tempest-ServersAdminTestJSON-14199920 tempest-ServersAdminTestJSON-14199920-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.098s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 599.244194] env[61806]: ERROR nova.compute.manager [None req-38c4eaec-22d8-4bf0-bd45-321ef43329a0 tempest-ServersAdminTestJSON-14199920 tempest-ServersAdminTestJSON-14199920-project-member] [instance: 8905f20f-28a6-49f4-88bf-e52177c6a4b6] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 6c13d932-a28d-4aa7-99c6-9cda907b02db, please check neutron logs for more information. [ 599.244194] env[61806]: ERROR nova.compute.manager [instance: 8905f20f-28a6-49f4-88bf-e52177c6a4b6] Traceback (most recent call last): [ 599.244194] env[61806]: ERROR nova.compute.manager [instance: 8905f20f-28a6-49f4-88bf-e52177c6a4b6] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 599.244194] env[61806]: ERROR nova.compute.manager [instance: 8905f20f-28a6-49f4-88bf-e52177c6a4b6] self.driver.spawn(context, instance, image_meta, [ 599.244194] env[61806]: ERROR nova.compute.manager [instance: 8905f20f-28a6-49f4-88bf-e52177c6a4b6] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 599.244194] env[61806]: ERROR nova.compute.manager [instance: 8905f20f-28a6-49f4-88bf-e52177c6a4b6] self._vmops.spawn(context, instance, image_meta, injected_files, [ 599.244194] env[61806]: ERROR nova.compute.manager [instance: 8905f20f-28a6-49f4-88bf-e52177c6a4b6] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 599.244194] env[61806]: ERROR nova.compute.manager [instance: 8905f20f-28a6-49f4-88bf-e52177c6a4b6] vm_ref = self.build_virtual_machine(instance, [ 599.244194] env[61806]: ERROR nova.compute.manager [instance: 8905f20f-28a6-49f4-88bf-e52177c6a4b6] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 599.244194] env[61806]: ERROR nova.compute.manager [instance: 8905f20f-28a6-49f4-88bf-e52177c6a4b6] vif_infos = vmwarevif.get_vif_info(self._session, [ 599.244194] env[61806]: ERROR nova.compute.manager [instance: 8905f20f-28a6-49f4-88bf-e52177c6a4b6] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 599.244806] env[61806]: ERROR nova.compute.manager [instance: 8905f20f-28a6-49f4-88bf-e52177c6a4b6] for vif in network_info: [ 599.244806] env[61806]: ERROR nova.compute.manager [instance: 8905f20f-28a6-49f4-88bf-e52177c6a4b6] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 599.244806] env[61806]: ERROR nova.compute.manager [instance: 8905f20f-28a6-49f4-88bf-e52177c6a4b6] return self._sync_wrapper(fn, *args, **kwargs) [ 599.244806] env[61806]: ERROR nova.compute.manager [instance: 8905f20f-28a6-49f4-88bf-e52177c6a4b6] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 599.244806] env[61806]: ERROR nova.compute.manager [instance: 8905f20f-28a6-49f4-88bf-e52177c6a4b6] self.wait() [ 599.244806] env[61806]: ERROR nova.compute.manager [instance: 8905f20f-28a6-49f4-88bf-e52177c6a4b6] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 599.244806] env[61806]: ERROR nova.compute.manager [instance: 8905f20f-28a6-49f4-88bf-e52177c6a4b6] self[:] = self._gt.wait() [ 599.244806] env[61806]: ERROR nova.compute.manager [instance: 8905f20f-28a6-49f4-88bf-e52177c6a4b6] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 599.244806] env[61806]: ERROR nova.compute.manager [instance: 8905f20f-28a6-49f4-88bf-e52177c6a4b6] return self._exit_event.wait() [ 599.244806] env[61806]: ERROR nova.compute.manager [instance: 8905f20f-28a6-49f4-88bf-e52177c6a4b6] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 599.244806] env[61806]: ERROR nova.compute.manager [instance: 8905f20f-28a6-49f4-88bf-e52177c6a4b6] current.throw(*self._exc) [ 599.244806] env[61806]: ERROR nova.compute.manager [instance: 8905f20f-28a6-49f4-88bf-e52177c6a4b6] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 599.244806] env[61806]: ERROR nova.compute.manager [instance: 8905f20f-28a6-49f4-88bf-e52177c6a4b6] result = function(*args, **kwargs) [ 599.245264] env[61806]: ERROR nova.compute.manager [instance: 8905f20f-28a6-49f4-88bf-e52177c6a4b6] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 599.245264] env[61806]: ERROR nova.compute.manager [instance: 8905f20f-28a6-49f4-88bf-e52177c6a4b6] return func(*args, **kwargs) [ 599.245264] env[61806]: ERROR nova.compute.manager [instance: 8905f20f-28a6-49f4-88bf-e52177c6a4b6] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 599.245264] env[61806]: ERROR nova.compute.manager [instance: 8905f20f-28a6-49f4-88bf-e52177c6a4b6] raise e [ 599.245264] env[61806]: ERROR nova.compute.manager [instance: 8905f20f-28a6-49f4-88bf-e52177c6a4b6] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 599.245264] env[61806]: ERROR nova.compute.manager [instance: 8905f20f-28a6-49f4-88bf-e52177c6a4b6] nwinfo = self.network_api.allocate_for_instance( [ 599.245264] env[61806]: ERROR nova.compute.manager [instance: 8905f20f-28a6-49f4-88bf-e52177c6a4b6] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 599.245264] env[61806]: ERROR nova.compute.manager [instance: 8905f20f-28a6-49f4-88bf-e52177c6a4b6] created_port_ids = self._update_ports_for_instance( [ 599.245264] env[61806]: ERROR nova.compute.manager [instance: 8905f20f-28a6-49f4-88bf-e52177c6a4b6] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 599.245264] env[61806]: ERROR nova.compute.manager [instance: 8905f20f-28a6-49f4-88bf-e52177c6a4b6] with excutils.save_and_reraise_exception(): [ 599.245264] env[61806]: ERROR nova.compute.manager [instance: 8905f20f-28a6-49f4-88bf-e52177c6a4b6] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 599.245264] env[61806]: ERROR nova.compute.manager [instance: 8905f20f-28a6-49f4-88bf-e52177c6a4b6] self.force_reraise() [ 599.245264] env[61806]: ERROR nova.compute.manager [instance: 8905f20f-28a6-49f4-88bf-e52177c6a4b6] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 599.245533] env[61806]: ERROR nova.compute.manager [instance: 8905f20f-28a6-49f4-88bf-e52177c6a4b6] raise self.value [ 599.245533] env[61806]: ERROR nova.compute.manager [instance: 8905f20f-28a6-49f4-88bf-e52177c6a4b6] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 599.245533] env[61806]: ERROR nova.compute.manager [instance: 8905f20f-28a6-49f4-88bf-e52177c6a4b6] updated_port = self._update_port( [ 599.245533] env[61806]: ERROR nova.compute.manager [instance: 8905f20f-28a6-49f4-88bf-e52177c6a4b6] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 599.245533] env[61806]: ERROR nova.compute.manager [instance: 8905f20f-28a6-49f4-88bf-e52177c6a4b6] _ensure_no_port_binding_failure(port) [ 599.245533] env[61806]: ERROR nova.compute.manager [instance: 8905f20f-28a6-49f4-88bf-e52177c6a4b6] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 599.245533] env[61806]: ERROR nova.compute.manager [instance: 8905f20f-28a6-49f4-88bf-e52177c6a4b6] raise exception.PortBindingFailed(port_id=port['id']) [ 599.245533] env[61806]: ERROR nova.compute.manager [instance: 8905f20f-28a6-49f4-88bf-e52177c6a4b6] nova.exception.PortBindingFailed: Binding failed for port 6c13d932-a28d-4aa7-99c6-9cda907b02db, please check neutron logs for more information. [ 599.245533] env[61806]: ERROR nova.compute.manager [instance: 8905f20f-28a6-49f4-88bf-e52177c6a4b6] [ 599.245533] env[61806]: DEBUG nova.compute.utils [None req-38c4eaec-22d8-4bf0-bd45-321ef43329a0 tempest-ServersAdminTestJSON-14199920 tempest-ServersAdminTestJSON-14199920-project-member] [instance: 8905f20f-28a6-49f4-88bf-e52177c6a4b6] Binding failed for port 6c13d932-a28d-4aa7-99c6-9cda907b02db, please check neutron logs for more information. {{(pid=61806) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 599.247219] env[61806]: ERROR nova.compute.manager [None req-11b5d1a0-518c-42d9-940e-4bf652a6d358 tempest-DeleteServersAdminTestJSON-1553667494 tempest-DeleteServersAdminTestJSON-1553667494-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port f65985e6-72fe-4194-a8d3-fc371e6fd74f, please check neutron logs for more information. [ 599.247219] env[61806]: ERROR nova.compute.manager Traceback (most recent call last): [ 599.247219] env[61806]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 599.247219] env[61806]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 599.247219] env[61806]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 599.247219] env[61806]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 599.247219] env[61806]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 599.247219] env[61806]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 599.247219] env[61806]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 599.247219] env[61806]: ERROR nova.compute.manager self.force_reraise() [ 599.247219] env[61806]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 599.247219] env[61806]: ERROR nova.compute.manager raise self.value [ 599.247219] env[61806]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 599.247219] env[61806]: ERROR nova.compute.manager updated_port = self._update_port( [ 599.247219] env[61806]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 599.247219] env[61806]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 599.247819] env[61806]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 599.247819] env[61806]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 599.247819] env[61806]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port f65985e6-72fe-4194-a8d3-fc371e6fd74f, please check neutron logs for more information. [ 599.247819] env[61806]: ERROR nova.compute.manager [ 599.247819] env[61806]: Traceback (most recent call last): [ 599.247819] env[61806]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 599.247819] env[61806]: listener.cb(fileno) [ 599.247819] env[61806]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 599.247819] env[61806]: result = function(*args, **kwargs) [ 599.247819] env[61806]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 599.247819] env[61806]: return func(*args, **kwargs) [ 599.247819] env[61806]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 599.247819] env[61806]: raise e [ 599.247819] env[61806]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 599.247819] env[61806]: nwinfo = self.network_api.allocate_for_instance( [ 599.247819] env[61806]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 599.247819] env[61806]: created_port_ids = self._update_ports_for_instance( [ 599.247819] env[61806]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 599.247819] env[61806]: with excutils.save_and_reraise_exception(): [ 599.247819] env[61806]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 599.247819] env[61806]: self.force_reraise() [ 599.247819] env[61806]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 599.247819] env[61806]: raise self.value [ 599.247819] env[61806]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 599.247819] env[61806]: updated_port = self._update_port( [ 599.247819] env[61806]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 599.247819] env[61806]: _ensure_no_port_binding_failure(port) [ 599.247819] env[61806]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 599.247819] env[61806]: raise exception.PortBindingFailed(port_id=port['id']) [ 599.248409] env[61806]: nova.exception.PortBindingFailed: Binding failed for port f65985e6-72fe-4194-a8d3-fc371e6fd74f, please check neutron logs for more information. [ 599.248409] env[61806]: Removing descriptor: 17 [ 599.248409] env[61806]: DEBUG oslo_concurrency.lockutils [None req-37e71da7-cb5e-431d-ba3c-95daaf8af7f3 tempest-ServersAdminNegativeTestJSON-2128564044 tempest-ServersAdminNegativeTestJSON-2128564044-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 16.301s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 599.252121] env[61806]: ERROR nova.compute.manager [None req-11b5d1a0-518c-42d9-940e-4bf652a6d358 tempest-DeleteServersAdminTestJSON-1553667494 tempest-DeleteServersAdminTestJSON-1553667494-project-member] [instance: ddc801a0-0782-4c6d-bd72-95d0afd01339] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port f65985e6-72fe-4194-a8d3-fc371e6fd74f, please check neutron logs for more information. [ 599.252121] env[61806]: ERROR nova.compute.manager [instance: ddc801a0-0782-4c6d-bd72-95d0afd01339] Traceback (most recent call last): [ 599.252121] env[61806]: ERROR nova.compute.manager [instance: ddc801a0-0782-4c6d-bd72-95d0afd01339] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 599.252121] env[61806]: ERROR nova.compute.manager [instance: ddc801a0-0782-4c6d-bd72-95d0afd01339] yield resources [ 599.252121] env[61806]: ERROR nova.compute.manager [instance: ddc801a0-0782-4c6d-bd72-95d0afd01339] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 599.252121] env[61806]: ERROR nova.compute.manager [instance: ddc801a0-0782-4c6d-bd72-95d0afd01339] self.driver.spawn(context, instance, image_meta, [ 599.252121] env[61806]: ERROR nova.compute.manager [instance: ddc801a0-0782-4c6d-bd72-95d0afd01339] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 599.252121] env[61806]: ERROR nova.compute.manager [instance: ddc801a0-0782-4c6d-bd72-95d0afd01339] self._vmops.spawn(context, instance, image_meta, injected_files, [ 599.252121] env[61806]: ERROR nova.compute.manager [instance: ddc801a0-0782-4c6d-bd72-95d0afd01339] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 599.252121] env[61806]: ERROR nova.compute.manager [instance: ddc801a0-0782-4c6d-bd72-95d0afd01339] vm_ref = self.build_virtual_machine(instance, [ 599.252121] env[61806]: ERROR nova.compute.manager [instance: ddc801a0-0782-4c6d-bd72-95d0afd01339] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 599.252784] env[61806]: ERROR nova.compute.manager [instance: ddc801a0-0782-4c6d-bd72-95d0afd01339] vif_infos = vmwarevif.get_vif_info(self._session, [ 599.252784] env[61806]: ERROR nova.compute.manager [instance: ddc801a0-0782-4c6d-bd72-95d0afd01339] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 599.252784] env[61806]: ERROR nova.compute.manager [instance: ddc801a0-0782-4c6d-bd72-95d0afd01339] for vif in network_info: [ 599.252784] env[61806]: ERROR nova.compute.manager [instance: ddc801a0-0782-4c6d-bd72-95d0afd01339] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 599.252784] env[61806]: ERROR nova.compute.manager [instance: ddc801a0-0782-4c6d-bd72-95d0afd01339] return self._sync_wrapper(fn, *args, **kwargs) [ 599.252784] env[61806]: ERROR nova.compute.manager [instance: ddc801a0-0782-4c6d-bd72-95d0afd01339] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 599.252784] env[61806]: ERROR nova.compute.manager [instance: ddc801a0-0782-4c6d-bd72-95d0afd01339] self.wait() [ 599.252784] env[61806]: ERROR nova.compute.manager [instance: ddc801a0-0782-4c6d-bd72-95d0afd01339] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 599.252784] env[61806]: ERROR nova.compute.manager [instance: ddc801a0-0782-4c6d-bd72-95d0afd01339] self[:] = self._gt.wait() [ 599.252784] env[61806]: ERROR nova.compute.manager [instance: ddc801a0-0782-4c6d-bd72-95d0afd01339] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 599.252784] env[61806]: ERROR nova.compute.manager [instance: ddc801a0-0782-4c6d-bd72-95d0afd01339] return self._exit_event.wait() [ 599.252784] env[61806]: ERROR nova.compute.manager [instance: ddc801a0-0782-4c6d-bd72-95d0afd01339] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 599.252784] env[61806]: ERROR nova.compute.manager [instance: ddc801a0-0782-4c6d-bd72-95d0afd01339] result = hub.switch() [ 599.253471] env[61806]: ERROR nova.compute.manager [instance: ddc801a0-0782-4c6d-bd72-95d0afd01339] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 599.253471] env[61806]: ERROR nova.compute.manager [instance: ddc801a0-0782-4c6d-bd72-95d0afd01339] return self.greenlet.switch() [ 599.253471] env[61806]: ERROR nova.compute.manager [instance: ddc801a0-0782-4c6d-bd72-95d0afd01339] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 599.253471] env[61806]: ERROR nova.compute.manager [instance: ddc801a0-0782-4c6d-bd72-95d0afd01339] result = function(*args, **kwargs) [ 599.253471] env[61806]: ERROR nova.compute.manager [instance: ddc801a0-0782-4c6d-bd72-95d0afd01339] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 599.253471] env[61806]: ERROR nova.compute.manager [instance: ddc801a0-0782-4c6d-bd72-95d0afd01339] return func(*args, **kwargs) [ 599.253471] env[61806]: ERROR nova.compute.manager [instance: ddc801a0-0782-4c6d-bd72-95d0afd01339] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 599.253471] env[61806]: ERROR nova.compute.manager [instance: ddc801a0-0782-4c6d-bd72-95d0afd01339] raise e [ 599.253471] env[61806]: ERROR nova.compute.manager [instance: ddc801a0-0782-4c6d-bd72-95d0afd01339] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 599.253471] env[61806]: ERROR nova.compute.manager [instance: ddc801a0-0782-4c6d-bd72-95d0afd01339] nwinfo = self.network_api.allocate_for_instance( [ 599.253471] env[61806]: ERROR nova.compute.manager [instance: ddc801a0-0782-4c6d-bd72-95d0afd01339] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 599.253471] env[61806]: ERROR nova.compute.manager [instance: ddc801a0-0782-4c6d-bd72-95d0afd01339] created_port_ids = self._update_ports_for_instance( [ 599.253471] env[61806]: ERROR nova.compute.manager [instance: ddc801a0-0782-4c6d-bd72-95d0afd01339] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 599.253919] env[61806]: ERROR nova.compute.manager [instance: ddc801a0-0782-4c6d-bd72-95d0afd01339] with excutils.save_and_reraise_exception(): [ 599.253919] env[61806]: ERROR nova.compute.manager [instance: ddc801a0-0782-4c6d-bd72-95d0afd01339] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 599.253919] env[61806]: ERROR nova.compute.manager [instance: ddc801a0-0782-4c6d-bd72-95d0afd01339] self.force_reraise() [ 599.253919] env[61806]: ERROR nova.compute.manager [instance: ddc801a0-0782-4c6d-bd72-95d0afd01339] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 599.253919] env[61806]: ERROR nova.compute.manager [instance: ddc801a0-0782-4c6d-bd72-95d0afd01339] raise self.value [ 599.253919] env[61806]: ERROR nova.compute.manager [instance: ddc801a0-0782-4c6d-bd72-95d0afd01339] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 599.253919] env[61806]: ERROR nova.compute.manager [instance: ddc801a0-0782-4c6d-bd72-95d0afd01339] updated_port = self._update_port( [ 599.253919] env[61806]: ERROR nova.compute.manager [instance: ddc801a0-0782-4c6d-bd72-95d0afd01339] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 599.253919] env[61806]: ERROR nova.compute.manager [instance: ddc801a0-0782-4c6d-bd72-95d0afd01339] _ensure_no_port_binding_failure(port) [ 599.253919] env[61806]: ERROR nova.compute.manager [instance: ddc801a0-0782-4c6d-bd72-95d0afd01339] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 599.253919] env[61806]: ERROR nova.compute.manager [instance: ddc801a0-0782-4c6d-bd72-95d0afd01339] raise exception.PortBindingFailed(port_id=port['id']) [ 599.253919] env[61806]: ERROR nova.compute.manager [instance: ddc801a0-0782-4c6d-bd72-95d0afd01339] nova.exception.PortBindingFailed: Binding failed for port f65985e6-72fe-4194-a8d3-fc371e6fd74f, please check neutron logs for more information. [ 599.253919] env[61806]: ERROR nova.compute.manager [instance: ddc801a0-0782-4c6d-bd72-95d0afd01339] [ 599.254413] env[61806]: INFO nova.compute.manager [None req-11b5d1a0-518c-42d9-940e-4bf652a6d358 tempest-DeleteServersAdminTestJSON-1553667494 tempest-DeleteServersAdminTestJSON-1553667494-project-member] [instance: ddc801a0-0782-4c6d-bd72-95d0afd01339] Terminating instance [ 599.255571] env[61806]: DEBUG nova.compute.manager [None req-38c4eaec-22d8-4bf0-bd45-321ef43329a0 tempest-ServersAdminTestJSON-14199920 tempest-ServersAdminTestJSON-14199920-project-member] [instance: 8905f20f-28a6-49f4-88bf-e52177c6a4b6] Build of instance 8905f20f-28a6-49f4-88bf-e52177c6a4b6 was re-scheduled: Binding failed for port 6c13d932-a28d-4aa7-99c6-9cda907b02db, please check neutron logs for more information. {{(pid=61806) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 599.256210] env[61806]: DEBUG nova.compute.manager [None req-38c4eaec-22d8-4bf0-bd45-321ef43329a0 tempest-ServersAdminTestJSON-14199920 tempest-ServersAdminTestJSON-14199920-project-member] [instance: 8905f20f-28a6-49f4-88bf-e52177c6a4b6] Unplugging VIFs for instance {{(pid=61806) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 599.256469] env[61806]: DEBUG oslo_concurrency.lockutils [None req-38c4eaec-22d8-4bf0-bd45-321ef43329a0 tempest-ServersAdminTestJSON-14199920 tempest-ServersAdminTestJSON-14199920-project-member] Acquiring lock "refresh_cache-8905f20f-28a6-49f4-88bf-e52177c6a4b6" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 599.256651] env[61806]: DEBUG oslo_concurrency.lockutils [None req-38c4eaec-22d8-4bf0-bd45-321ef43329a0 tempest-ServersAdminTestJSON-14199920 tempest-ServersAdminTestJSON-14199920-project-member] Acquired lock "refresh_cache-8905f20f-28a6-49f4-88bf-e52177c6a4b6" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 599.257050] env[61806]: DEBUG nova.network.neutron [None req-38c4eaec-22d8-4bf0-bd45-321ef43329a0 tempest-ServersAdminTestJSON-14199920 tempest-ServersAdminTestJSON-14199920-project-member] [instance: 8905f20f-28a6-49f4-88bf-e52177c6a4b6] Building network info cache for instance {{(pid=61806) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 599.273154] env[61806]: DEBUG oslo_concurrency.lockutils [None req-11b5d1a0-518c-42d9-940e-4bf652a6d358 tempest-DeleteServersAdminTestJSON-1553667494 tempest-DeleteServersAdminTestJSON-1553667494-project-member] Acquiring lock "refresh_cache-ddc801a0-0782-4c6d-bd72-95d0afd01339" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 599.273243] env[61806]: DEBUG oslo_concurrency.lockutils [None req-11b5d1a0-518c-42d9-940e-4bf652a6d358 tempest-DeleteServersAdminTestJSON-1553667494 tempest-DeleteServersAdminTestJSON-1553667494-project-member] Acquired lock "refresh_cache-ddc801a0-0782-4c6d-bd72-95d0afd01339" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 599.273428] env[61806]: DEBUG nova.network.neutron [None req-11b5d1a0-518c-42d9-940e-4bf652a6d358 tempest-DeleteServersAdminTestJSON-1553667494 tempest-DeleteServersAdminTestJSON-1553667494-project-member] [instance: ddc801a0-0782-4c6d-bd72-95d0afd01339] Building network info cache for instance {{(pid=61806) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 599.781501] env[61806]: DEBUG nova.network.neutron [None req-38c4eaec-22d8-4bf0-bd45-321ef43329a0 tempest-ServersAdminTestJSON-14199920 tempest-ServersAdminTestJSON-14199920-project-member] [instance: 8905f20f-28a6-49f4-88bf-e52177c6a4b6] Instance cache missing network info. {{(pid=61806) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 599.795975] env[61806]: DEBUG nova.network.neutron [None req-11b5d1a0-518c-42d9-940e-4bf652a6d358 tempest-DeleteServersAdminTestJSON-1553667494 tempest-DeleteServersAdminTestJSON-1553667494-project-member] [instance: ddc801a0-0782-4c6d-bd72-95d0afd01339] Instance cache missing network info. {{(pid=61806) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 599.860984] env[61806]: DEBUG nova.network.neutron [None req-11b5d1a0-518c-42d9-940e-4bf652a6d358 tempest-DeleteServersAdminTestJSON-1553667494 tempest-DeleteServersAdminTestJSON-1553667494-project-member] [instance: ddc801a0-0782-4c6d-bd72-95d0afd01339] Updating instance_info_cache with network_info: [] {{(pid=61806) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 599.880222] env[61806]: DEBUG nova.network.neutron [None req-38c4eaec-22d8-4bf0-bd45-321ef43329a0 tempest-ServersAdminTestJSON-14199920 tempest-ServersAdminTestJSON-14199920-project-member] [instance: 8905f20f-28a6-49f4-88bf-e52177c6a4b6] Updating instance_info_cache with network_info: [] {{(pid=61806) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 599.959182] env[61806]: DEBUG nova.compute.manager [req-d5b4409d-e352-4514-bc78-dfdb66582c18 req-4b43592a-4716-4c9d-adaf-18dd9ebfad1a service nova] [instance: ddc801a0-0782-4c6d-bd72-95d0afd01339] Received event network-changed-f65985e6-72fe-4194-a8d3-fc371e6fd74f {{(pid=61806) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 599.959182] env[61806]: DEBUG nova.compute.manager [req-d5b4409d-e352-4514-bc78-dfdb66582c18 req-4b43592a-4716-4c9d-adaf-18dd9ebfad1a service nova] [instance: ddc801a0-0782-4c6d-bd72-95d0afd01339] Refreshing instance network info cache due to event network-changed-f65985e6-72fe-4194-a8d3-fc371e6fd74f. {{(pid=61806) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 599.959182] env[61806]: DEBUG oslo_concurrency.lockutils [req-d5b4409d-e352-4514-bc78-dfdb66582c18 req-4b43592a-4716-4c9d-adaf-18dd9ebfad1a service nova] Acquiring lock "refresh_cache-ddc801a0-0782-4c6d-bd72-95d0afd01339" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 600.207344] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f639ae51-5c0e-4839-8dfb-3fceac398607 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 600.216816] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a03c10ea-3a4a-477d-b6f9-e2876589f8e4 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 600.254846] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7379e160-5d45-4661-a54a-1817237b347b {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 600.265528] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2344f610-22e0-4a0d-b6c5-1e682b0fcb4a {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 600.280934] env[61806]: DEBUG nova.compute.provider_tree [None req-37e71da7-cb5e-431d-ba3c-95daaf8af7f3 tempest-ServersAdminNegativeTestJSON-2128564044 tempest-ServersAdminNegativeTestJSON-2128564044-project-member] Inventory has not changed in ProviderTree for provider: a2858be1-fd22-4e08-979e-87ad25293407 {{(pid=61806) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 600.370022] env[61806]: DEBUG oslo_concurrency.lockutils [None req-11b5d1a0-518c-42d9-940e-4bf652a6d358 tempest-DeleteServersAdminTestJSON-1553667494 tempest-DeleteServersAdminTestJSON-1553667494-project-member] Releasing lock "refresh_cache-ddc801a0-0782-4c6d-bd72-95d0afd01339" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 600.370022] env[61806]: DEBUG nova.compute.manager [None req-11b5d1a0-518c-42d9-940e-4bf652a6d358 tempest-DeleteServersAdminTestJSON-1553667494 tempest-DeleteServersAdminTestJSON-1553667494-project-member] [instance: ddc801a0-0782-4c6d-bd72-95d0afd01339] Start destroying the instance on the hypervisor. {{(pid=61806) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 600.370022] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-11b5d1a0-518c-42d9-940e-4bf652a6d358 tempest-DeleteServersAdminTestJSON-1553667494 tempest-DeleteServersAdminTestJSON-1553667494-project-member] [instance: ddc801a0-0782-4c6d-bd72-95d0afd01339] Destroying instance {{(pid=61806) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 600.370022] env[61806]: DEBUG oslo_concurrency.lockutils [req-d5b4409d-e352-4514-bc78-dfdb66582c18 req-4b43592a-4716-4c9d-adaf-18dd9ebfad1a service nova] Acquired lock "refresh_cache-ddc801a0-0782-4c6d-bd72-95d0afd01339" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 600.370022] env[61806]: DEBUG nova.network.neutron [req-d5b4409d-e352-4514-bc78-dfdb66582c18 req-4b43592a-4716-4c9d-adaf-18dd9ebfad1a service nova] [instance: ddc801a0-0782-4c6d-bd72-95d0afd01339] Refreshing network info cache for port f65985e6-72fe-4194-a8d3-fc371e6fd74f {{(pid=61806) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 600.370459] env[61806]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-0741b108-061d-4a68-80dd-17bb71607ae0 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 600.382741] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c196e44e-9577-4e88-b653-e6f29977ccce {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 600.398821] env[61806]: DEBUG oslo_concurrency.lockutils [None req-38c4eaec-22d8-4bf0-bd45-321ef43329a0 tempest-ServersAdminTestJSON-14199920 tempest-ServersAdminTestJSON-14199920-project-member] Releasing lock "refresh_cache-8905f20f-28a6-49f4-88bf-e52177c6a4b6" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 600.399100] env[61806]: DEBUG nova.compute.manager [None req-38c4eaec-22d8-4bf0-bd45-321ef43329a0 tempest-ServersAdminTestJSON-14199920 tempest-ServersAdminTestJSON-14199920-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61806) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 600.399224] env[61806]: DEBUG nova.compute.manager [None req-38c4eaec-22d8-4bf0-bd45-321ef43329a0 tempest-ServersAdminTestJSON-14199920 tempest-ServersAdminTestJSON-14199920-project-member] [instance: 8905f20f-28a6-49f4-88bf-e52177c6a4b6] Deallocating network for instance {{(pid=61806) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 600.399381] env[61806]: DEBUG nova.network.neutron [None req-38c4eaec-22d8-4bf0-bd45-321ef43329a0 tempest-ServersAdminTestJSON-14199920 tempest-ServersAdminTestJSON-14199920-project-member] [instance: 8905f20f-28a6-49f4-88bf-e52177c6a4b6] deallocate_for_instance() {{(pid=61806) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 600.412783] env[61806]: WARNING nova.virt.vmwareapi.vmops [None req-11b5d1a0-518c-42d9-940e-4bf652a6d358 tempest-DeleteServersAdminTestJSON-1553667494 tempest-DeleteServersAdminTestJSON-1553667494-project-member] [instance: ddc801a0-0782-4c6d-bd72-95d0afd01339] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance ddc801a0-0782-4c6d-bd72-95d0afd01339 could not be found. [ 600.412872] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-11b5d1a0-518c-42d9-940e-4bf652a6d358 tempest-DeleteServersAdminTestJSON-1553667494 tempest-DeleteServersAdminTestJSON-1553667494-project-member] [instance: ddc801a0-0782-4c6d-bd72-95d0afd01339] Instance destroyed {{(pid=61806) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 600.412978] env[61806]: INFO nova.compute.manager [None req-11b5d1a0-518c-42d9-940e-4bf652a6d358 tempest-DeleteServersAdminTestJSON-1553667494 tempest-DeleteServersAdminTestJSON-1553667494-project-member] [instance: ddc801a0-0782-4c6d-bd72-95d0afd01339] Took 0.05 seconds to destroy the instance on the hypervisor. [ 600.413293] env[61806]: DEBUG oslo.service.loopingcall [None req-11b5d1a0-518c-42d9-940e-4bf652a6d358 tempest-DeleteServersAdminTestJSON-1553667494 tempest-DeleteServersAdminTestJSON-1553667494-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61806) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 600.413662] env[61806]: DEBUG nova.compute.manager [-] [instance: ddc801a0-0782-4c6d-bd72-95d0afd01339] Deallocating network for instance {{(pid=61806) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 600.413662] env[61806]: DEBUG nova.network.neutron [-] [instance: ddc801a0-0782-4c6d-bd72-95d0afd01339] deallocate_for_instance() {{(pid=61806) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 600.429815] env[61806]: DEBUG nova.network.neutron [-] [instance: ddc801a0-0782-4c6d-bd72-95d0afd01339] Instance cache missing network info. {{(pid=61806) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 600.596686] env[61806]: DEBUG nova.network.neutron [None req-38c4eaec-22d8-4bf0-bd45-321ef43329a0 tempest-ServersAdminTestJSON-14199920 tempest-ServersAdminTestJSON-14199920-project-member] [instance: 8905f20f-28a6-49f4-88bf-e52177c6a4b6] Instance cache missing network info. {{(pid=61806) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 600.785535] env[61806]: DEBUG nova.scheduler.client.report [None req-37e71da7-cb5e-431d-ba3c-95daaf8af7f3 tempest-ServersAdminNegativeTestJSON-2128564044 tempest-ServersAdminNegativeTestJSON-2128564044-project-member] Inventory has not changed for provider a2858be1-fd22-4e08-979e-87ad25293407 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 139, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61806) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 600.804541] env[61806]: DEBUG oslo_concurrency.lockutils [None req-d17e7876-7c69-44af-acb2-32ed225ce945 tempest-ServerActionsTestOtherA-1637843849 tempest-ServerActionsTestOtherA-1637843849-project-member] Acquiring lock "15476ac0-289a-4e04-aa9d-4244c658e962" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 600.804784] env[61806]: DEBUG oslo_concurrency.lockutils [None req-d17e7876-7c69-44af-acb2-32ed225ce945 tempest-ServerActionsTestOtherA-1637843849 tempest-ServerActionsTestOtherA-1637843849-project-member] Lock "15476ac0-289a-4e04-aa9d-4244c658e962" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 600.892462] env[61806]: DEBUG nova.network.neutron [req-d5b4409d-e352-4514-bc78-dfdb66582c18 req-4b43592a-4716-4c9d-adaf-18dd9ebfad1a service nova] [instance: ddc801a0-0782-4c6d-bd72-95d0afd01339] Instance cache missing network info. {{(pid=61806) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 600.932403] env[61806]: DEBUG nova.network.neutron [-] [instance: ddc801a0-0782-4c6d-bd72-95d0afd01339] Updating instance_info_cache with network_info: [] {{(pid=61806) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 600.955963] env[61806]: DEBUG nova.network.neutron [req-d5b4409d-e352-4514-bc78-dfdb66582c18 req-4b43592a-4716-4c9d-adaf-18dd9ebfad1a service nova] [instance: ddc801a0-0782-4c6d-bd72-95d0afd01339] Updating instance_info_cache with network_info: [] {{(pid=61806) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 601.101273] env[61806]: DEBUG nova.network.neutron [None req-38c4eaec-22d8-4bf0-bd45-321ef43329a0 tempest-ServersAdminTestJSON-14199920 tempest-ServersAdminTestJSON-14199920-project-member] [instance: 8905f20f-28a6-49f4-88bf-e52177c6a4b6] Updating instance_info_cache with network_info: [] {{(pid=61806) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 601.290853] env[61806]: DEBUG oslo_concurrency.lockutils [None req-37e71da7-cb5e-431d-ba3c-95daaf8af7f3 tempest-ServersAdminNegativeTestJSON-2128564044 tempest-ServersAdminNegativeTestJSON-2128564044-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.043s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 601.291536] env[61806]: ERROR nova.compute.manager [None req-37e71da7-cb5e-431d-ba3c-95daaf8af7f3 tempest-ServersAdminNegativeTestJSON-2128564044 tempest-ServersAdminNegativeTestJSON-2128564044-project-member] [instance: 89ce0fd8-6a22-4b1c-bdff-d1584abd538a] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 63122136-e392-4570-8f9c-7785b771ac82, please check neutron logs for more information. [ 601.291536] env[61806]: ERROR nova.compute.manager [instance: 89ce0fd8-6a22-4b1c-bdff-d1584abd538a] Traceback (most recent call last): [ 601.291536] env[61806]: ERROR nova.compute.manager [instance: 89ce0fd8-6a22-4b1c-bdff-d1584abd538a] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 601.291536] env[61806]: ERROR nova.compute.manager [instance: 89ce0fd8-6a22-4b1c-bdff-d1584abd538a] self.driver.spawn(context, instance, image_meta, [ 601.291536] env[61806]: ERROR nova.compute.manager [instance: 89ce0fd8-6a22-4b1c-bdff-d1584abd538a] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 601.291536] env[61806]: ERROR nova.compute.manager [instance: 89ce0fd8-6a22-4b1c-bdff-d1584abd538a] self._vmops.spawn(context, instance, image_meta, injected_files, [ 601.291536] env[61806]: ERROR nova.compute.manager [instance: 89ce0fd8-6a22-4b1c-bdff-d1584abd538a] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 601.291536] env[61806]: ERROR nova.compute.manager [instance: 89ce0fd8-6a22-4b1c-bdff-d1584abd538a] vm_ref = self.build_virtual_machine(instance, [ 601.291536] env[61806]: ERROR nova.compute.manager [instance: 89ce0fd8-6a22-4b1c-bdff-d1584abd538a] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 601.291536] env[61806]: ERROR nova.compute.manager [instance: 89ce0fd8-6a22-4b1c-bdff-d1584abd538a] vif_infos = vmwarevif.get_vif_info(self._session, [ 601.291536] env[61806]: ERROR nova.compute.manager [instance: 89ce0fd8-6a22-4b1c-bdff-d1584abd538a] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 601.291794] env[61806]: ERROR nova.compute.manager [instance: 89ce0fd8-6a22-4b1c-bdff-d1584abd538a] for vif in network_info: [ 601.291794] env[61806]: ERROR nova.compute.manager [instance: 89ce0fd8-6a22-4b1c-bdff-d1584abd538a] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 601.291794] env[61806]: ERROR nova.compute.manager [instance: 89ce0fd8-6a22-4b1c-bdff-d1584abd538a] return self._sync_wrapper(fn, *args, **kwargs) [ 601.291794] env[61806]: ERROR nova.compute.manager [instance: 89ce0fd8-6a22-4b1c-bdff-d1584abd538a] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 601.291794] env[61806]: ERROR nova.compute.manager [instance: 89ce0fd8-6a22-4b1c-bdff-d1584abd538a] self.wait() [ 601.291794] env[61806]: ERROR nova.compute.manager [instance: 89ce0fd8-6a22-4b1c-bdff-d1584abd538a] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 601.291794] env[61806]: ERROR nova.compute.manager [instance: 89ce0fd8-6a22-4b1c-bdff-d1584abd538a] self[:] = self._gt.wait() [ 601.291794] env[61806]: ERROR nova.compute.manager [instance: 89ce0fd8-6a22-4b1c-bdff-d1584abd538a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 601.291794] env[61806]: ERROR nova.compute.manager [instance: 89ce0fd8-6a22-4b1c-bdff-d1584abd538a] return self._exit_event.wait() [ 601.291794] env[61806]: ERROR nova.compute.manager [instance: 89ce0fd8-6a22-4b1c-bdff-d1584abd538a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 601.291794] env[61806]: ERROR nova.compute.manager [instance: 89ce0fd8-6a22-4b1c-bdff-d1584abd538a] result = hub.switch() [ 601.291794] env[61806]: ERROR nova.compute.manager [instance: 89ce0fd8-6a22-4b1c-bdff-d1584abd538a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 601.291794] env[61806]: ERROR nova.compute.manager [instance: 89ce0fd8-6a22-4b1c-bdff-d1584abd538a] return self.greenlet.switch() [ 601.292142] env[61806]: ERROR nova.compute.manager [instance: 89ce0fd8-6a22-4b1c-bdff-d1584abd538a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 601.292142] env[61806]: ERROR nova.compute.manager [instance: 89ce0fd8-6a22-4b1c-bdff-d1584abd538a] result = function(*args, **kwargs) [ 601.292142] env[61806]: ERROR nova.compute.manager [instance: 89ce0fd8-6a22-4b1c-bdff-d1584abd538a] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 601.292142] env[61806]: ERROR nova.compute.manager [instance: 89ce0fd8-6a22-4b1c-bdff-d1584abd538a] return func(*args, **kwargs) [ 601.292142] env[61806]: ERROR nova.compute.manager [instance: 89ce0fd8-6a22-4b1c-bdff-d1584abd538a] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 601.292142] env[61806]: ERROR nova.compute.manager [instance: 89ce0fd8-6a22-4b1c-bdff-d1584abd538a] raise e [ 601.292142] env[61806]: ERROR nova.compute.manager [instance: 89ce0fd8-6a22-4b1c-bdff-d1584abd538a] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 601.292142] env[61806]: ERROR nova.compute.manager [instance: 89ce0fd8-6a22-4b1c-bdff-d1584abd538a] nwinfo = self.network_api.allocate_for_instance( [ 601.292142] env[61806]: ERROR nova.compute.manager [instance: 89ce0fd8-6a22-4b1c-bdff-d1584abd538a] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 601.292142] env[61806]: ERROR nova.compute.manager [instance: 89ce0fd8-6a22-4b1c-bdff-d1584abd538a] created_port_ids = self._update_ports_for_instance( [ 601.292142] env[61806]: ERROR nova.compute.manager [instance: 89ce0fd8-6a22-4b1c-bdff-d1584abd538a] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 601.292142] env[61806]: ERROR nova.compute.manager [instance: 89ce0fd8-6a22-4b1c-bdff-d1584abd538a] with excutils.save_and_reraise_exception(): [ 601.292142] env[61806]: ERROR nova.compute.manager [instance: 89ce0fd8-6a22-4b1c-bdff-d1584abd538a] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 601.292410] env[61806]: ERROR nova.compute.manager [instance: 89ce0fd8-6a22-4b1c-bdff-d1584abd538a] self.force_reraise() [ 601.292410] env[61806]: ERROR nova.compute.manager [instance: 89ce0fd8-6a22-4b1c-bdff-d1584abd538a] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 601.292410] env[61806]: ERROR nova.compute.manager [instance: 89ce0fd8-6a22-4b1c-bdff-d1584abd538a] raise self.value [ 601.292410] env[61806]: ERROR nova.compute.manager [instance: 89ce0fd8-6a22-4b1c-bdff-d1584abd538a] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 601.292410] env[61806]: ERROR nova.compute.manager [instance: 89ce0fd8-6a22-4b1c-bdff-d1584abd538a] updated_port = self._update_port( [ 601.292410] env[61806]: ERROR nova.compute.manager [instance: 89ce0fd8-6a22-4b1c-bdff-d1584abd538a] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 601.292410] env[61806]: ERROR nova.compute.manager [instance: 89ce0fd8-6a22-4b1c-bdff-d1584abd538a] _ensure_no_port_binding_failure(port) [ 601.292410] env[61806]: ERROR nova.compute.manager [instance: 89ce0fd8-6a22-4b1c-bdff-d1584abd538a] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 601.292410] env[61806]: ERROR nova.compute.manager [instance: 89ce0fd8-6a22-4b1c-bdff-d1584abd538a] raise exception.PortBindingFailed(port_id=port['id']) [ 601.292410] env[61806]: ERROR nova.compute.manager [instance: 89ce0fd8-6a22-4b1c-bdff-d1584abd538a] nova.exception.PortBindingFailed: Binding failed for port 63122136-e392-4570-8f9c-7785b771ac82, please check neutron logs for more information. [ 601.292410] env[61806]: ERROR nova.compute.manager [instance: 89ce0fd8-6a22-4b1c-bdff-d1584abd538a] [ 601.292687] env[61806]: DEBUG nova.compute.utils [None req-37e71da7-cb5e-431d-ba3c-95daaf8af7f3 tempest-ServersAdminNegativeTestJSON-2128564044 tempest-ServersAdminNegativeTestJSON-2128564044-project-member] [instance: 89ce0fd8-6a22-4b1c-bdff-d1584abd538a] Binding failed for port 63122136-e392-4570-8f9c-7785b771ac82, please check neutron logs for more information. {{(pid=61806) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 601.299033] env[61806]: DEBUG nova.compute.manager [None req-37e71da7-cb5e-431d-ba3c-95daaf8af7f3 tempest-ServersAdminNegativeTestJSON-2128564044 tempest-ServersAdminNegativeTestJSON-2128564044-project-member] [instance: 89ce0fd8-6a22-4b1c-bdff-d1584abd538a] Build of instance 89ce0fd8-6a22-4b1c-bdff-d1584abd538a was re-scheduled: Binding failed for port 63122136-e392-4570-8f9c-7785b771ac82, please check neutron logs for more information. {{(pid=61806) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 601.299033] env[61806]: DEBUG nova.compute.manager [None req-37e71da7-cb5e-431d-ba3c-95daaf8af7f3 tempest-ServersAdminNegativeTestJSON-2128564044 tempest-ServersAdminNegativeTestJSON-2128564044-project-member] [instance: 89ce0fd8-6a22-4b1c-bdff-d1584abd538a] Unplugging VIFs for instance {{(pid=61806) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 601.299033] env[61806]: DEBUG oslo_concurrency.lockutils [None req-37e71da7-cb5e-431d-ba3c-95daaf8af7f3 tempest-ServersAdminNegativeTestJSON-2128564044 tempest-ServersAdminNegativeTestJSON-2128564044-project-member] Acquiring lock "refresh_cache-89ce0fd8-6a22-4b1c-bdff-d1584abd538a" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 601.299033] env[61806]: DEBUG oslo_concurrency.lockutils [None req-37e71da7-cb5e-431d-ba3c-95daaf8af7f3 tempest-ServersAdminNegativeTestJSON-2128564044 tempest-ServersAdminNegativeTestJSON-2128564044-project-member] Acquired lock "refresh_cache-89ce0fd8-6a22-4b1c-bdff-d1584abd538a" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 601.299236] env[61806]: DEBUG nova.network.neutron [None req-37e71da7-cb5e-431d-ba3c-95daaf8af7f3 tempest-ServersAdminNegativeTestJSON-2128564044 tempest-ServersAdminNegativeTestJSON-2128564044-project-member] [instance: 89ce0fd8-6a22-4b1c-bdff-d1584abd538a] Building network info cache for instance {{(pid=61806) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 601.299236] env[61806]: DEBUG oslo_concurrency.lockutils [None req-ebff8349-d5ea-4a95-aee1-e4c9e8b6d3a4 tempest-ServerDiagnosticsTest-195484958 tempest-ServerDiagnosticsTest-195484958-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 15.979s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 601.300456] env[61806]: INFO nova.compute.claims [None req-ebff8349-d5ea-4a95-aee1-e4c9e8b6d3a4 tempest-ServerDiagnosticsTest-195484958 tempest-ServerDiagnosticsTest-195484958-project-member] [instance: 79fc9ea2-c427-47ce-acdb-1002ee5d78c9] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 601.435562] env[61806]: INFO nova.compute.manager [-] [instance: ddc801a0-0782-4c6d-bd72-95d0afd01339] Took 1.02 seconds to deallocate network for instance. [ 601.438165] env[61806]: DEBUG nova.compute.claims [None req-11b5d1a0-518c-42d9-940e-4bf652a6d358 tempest-DeleteServersAdminTestJSON-1553667494 tempest-DeleteServersAdminTestJSON-1553667494-project-member] [instance: ddc801a0-0782-4c6d-bd72-95d0afd01339] Aborting claim: {{(pid=61806) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 601.438980] env[61806]: DEBUG oslo_concurrency.lockutils [None req-11b5d1a0-518c-42d9-940e-4bf652a6d358 tempest-DeleteServersAdminTestJSON-1553667494 tempest-DeleteServersAdminTestJSON-1553667494-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 601.459645] env[61806]: DEBUG oslo_concurrency.lockutils [req-d5b4409d-e352-4514-bc78-dfdb66582c18 req-4b43592a-4716-4c9d-adaf-18dd9ebfad1a service nova] Releasing lock "refresh_cache-ddc801a0-0782-4c6d-bd72-95d0afd01339" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 601.459645] env[61806]: DEBUG nova.compute.manager [req-d5b4409d-e352-4514-bc78-dfdb66582c18 req-4b43592a-4716-4c9d-adaf-18dd9ebfad1a service nova] [instance: ddc801a0-0782-4c6d-bd72-95d0afd01339] Received event network-vif-deleted-f65985e6-72fe-4194-a8d3-fc371e6fd74f {{(pid=61806) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 601.601759] env[61806]: INFO nova.compute.manager [None req-38c4eaec-22d8-4bf0-bd45-321ef43329a0 tempest-ServersAdminTestJSON-14199920 tempest-ServersAdminTestJSON-14199920-project-member] [instance: 8905f20f-28a6-49f4-88bf-e52177c6a4b6] Took 1.20 seconds to deallocate network for instance. [ 601.823078] env[61806]: DEBUG nova.network.neutron [None req-37e71da7-cb5e-431d-ba3c-95daaf8af7f3 tempest-ServersAdminNegativeTestJSON-2128564044 tempest-ServersAdminNegativeTestJSON-2128564044-project-member] [instance: 89ce0fd8-6a22-4b1c-bdff-d1584abd538a] Instance cache missing network info. {{(pid=61806) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 601.885706] env[61806]: DEBUG nova.network.neutron [None req-37e71da7-cb5e-431d-ba3c-95daaf8af7f3 tempest-ServersAdminNegativeTestJSON-2128564044 tempest-ServersAdminNegativeTestJSON-2128564044-project-member] [instance: 89ce0fd8-6a22-4b1c-bdff-d1584abd538a] Updating instance_info_cache with network_info: [] {{(pid=61806) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 602.388035] env[61806]: DEBUG oslo_concurrency.lockutils [None req-37e71da7-cb5e-431d-ba3c-95daaf8af7f3 tempest-ServersAdminNegativeTestJSON-2128564044 tempest-ServersAdminNegativeTestJSON-2128564044-project-member] Releasing lock "refresh_cache-89ce0fd8-6a22-4b1c-bdff-d1584abd538a" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 602.388435] env[61806]: DEBUG nova.compute.manager [None req-37e71da7-cb5e-431d-ba3c-95daaf8af7f3 tempest-ServersAdminNegativeTestJSON-2128564044 tempest-ServersAdminNegativeTestJSON-2128564044-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61806) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 602.388435] env[61806]: DEBUG nova.compute.manager [None req-37e71da7-cb5e-431d-ba3c-95daaf8af7f3 tempest-ServersAdminNegativeTestJSON-2128564044 tempest-ServersAdminNegativeTestJSON-2128564044-project-member] [instance: 89ce0fd8-6a22-4b1c-bdff-d1584abd538a] Deallocating network for instance {{(pid=61806) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 602.388598] env[61806]: DEBUG nova.network.neutron [None req-37e71da7-cb5e-431d-ba3c-95daaf8af7f3 tempest-ServersAdminNegativeTestJSON-2128564044 tempest-ServersAdminNegativeTestJSON-2128564044-project-member] [instance: 89ce0fd8-6a22-4b1c-bdff-d1584abd538a] deallocate_for_instance() {{(pid=61806) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 602.408093] env[61806]: DEBUG nova.network.neutron [None req-37e71da7-cb5e-431d-ba3c-95daaf8af7f3 tempest-ServersAdminNegativeTestJSON-2128564044 tempest-ServersAdminNegativeTestJSON-2128564044-project-member] [instance: 89ce0fd8-6a22-4b1c-bdff-d1584abd538a] Instance cache missing network info. {{(pid=61806) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 602.624954] env[61806]: INFO nova.scheduler.client.report [None req-38c4eaec-22d8-4bf0-bd45-321ef43329a0 tempest-ServersAdminTestJSON-14199920 tempest-ServersAdminTestJSON-14199920-project-member] Deleted allocations for instance 8905f20f-28a6-49f4-88bf-e52177c6a4b6 [ 602.676327] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ee0c22af-b467-46c4-a434-c329d5e15666 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 602.684484] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2a9d92e4-2ed9-4221-91f2-1020c4d69906 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 602.716417] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-22c22173-76e0-4421-b01f-9ea4f516d6e9 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 602.723556] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3ad827f8-bfbd-472a-ac7b-f89626cd4e63 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 602.736856] env[61806]: DEBUG nova.compute.provider_tree [None req-ebff8349-d5ea-4a95-aee1-e4c9e8b6d3a4 tempest-ServerDiagnosticsTest-195484958 tempest-ServerDiagnosticsTest-195484958-project-member] Inventory has not changed in ProviderTree for provider: a2858be1-fd22-4e08-979e-87ad25293407 {{(pid=61806) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 602.911080] env[61806]: DEBUG nova.network.neutron [None req-37e71da7-cb5e-431d-ba3c-95daaf8af7f3 tempest-ServersAdminNegativeTestJSON-2128564044 tempest-ServersAdminNegativeTestJSON-2128564044-project-member] [instance: 89ce0fd8-6a22-4b1c-bdff-d1584abd538a] Updating instance_info_cache with network_info: [] {{(pid=61806) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 603.135033] env[61806]: DEBUG oslo_concurrency.lockutils [None req-38c4eaec-22d8-4bf0-bd45-321ef43329a0 tempest-ServersAdminTestJSON-14199920 tempest-ServersAdminTestJSON-14199920-project-member] Lock "8905f20f-28a6-49f4-88bf-e52177c6a4b6" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 67.970s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 603.240524] env[61806]: DEBUG nova.scheduler.client.report [None req-ebff8349-d5ea-4a95-aee1-e4c9e8b6d3a4 tempest-ServerDiagnosticsTest-195484958 tempest-ServerDiagnosticsTest-195484958-project-member] Inventory has not changed for provider a2858be1-fd22-4e08-979e-87ad25293407 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 139, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61806) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 603.414277] env[61806]: INFO nova.compute.manager [None req-37e71da7-cb5e-431d-ba3c-95daaf8af7f3 tempest-ServersAdminNegativeTestJSON-2128564044 tempest-ServersAdminNegativeTestJSON-2128564044-project-member] [instance: 89ce0fd8-6a22-4b1c-bdff-d1584abd538a] Took 1.03 seconds to deallocate network for instance. [ 603.640104] env[61806]: DEBUG nova.compute.manager [None req-a6b8da3a-7829-486a-884d-0f28be02e38c tempest-ServersTestManualDisk-151630353 tempest-ServersTestManualDisk-151630353-project-member] [instance: 38a79453-939d-4649-9419-a2799636180d] Starting instance... {{(pid=61806) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 603.746344] env[61806]: DEBUG oslo_concurrency.lockutils [None req-ebff8349-d5ea-4a95-aee1-e4c9e8b6d3a4 tempest-ServerDiagnosticsTest-195484958 tempest-ServerDiagnosticsTest-195484958-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.448s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 603.746897] env[61806]: DEBUG nova.compute.manager [None req-ebff8349-d5ea-4a95-aee1-e4c9e8b6d3a4 tempest-ServerDiagnosticsTest-195484958 tempest-ServerDiagnosticsTest-195484958-project-member] [instance: 79fc9ea2-c427-47ce-acdb-1002ee5d78c9] Start building networks asynchronously for instance. {{(pid=61806) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 603.750558] env[61806]: DEBUG oslo_concurrency.lockutils [None req-24088d08-d3d7-4f62-ad63-91029bd9b22d tempest-ServersTestMultiNic-1480543082 tempest-ServersTestMultiNic-1480543082-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 16.483s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 603.751932] env[61806]: INFO nova.compute.claims [None req-24088d08-d3d7-4f62-ad63-91029bd9b22d tempest-ServersTestMultiNic-1480543082 tempest-ServersTestMultiNic-1480543082-project-member] [instance: c2d1c9f6-f458-4843-982c-afdb2a90edae] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 604.163021] env[61806]: DEBUG oslo_concurrency.lockutils [None req-a6b8da3a-7829-486a-884d-0f28be02e38c tempest-ServersTestManualDisk-151630353 tempest-ServersTestManualDisk-151630353-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 604.256149] env[61806]: DEBUG nova.compute.utils [None req-ebff8349-d5ea-4a95-aee1-e4c9e8b6d3a4 tempest-ServerDiagnosticsTest-195484958 tempest-ServerDiagnosticsTest-195484958-project-member] Using /dev/sd instead of None {{(pid=61806) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 604.260982] env[61806]: DEBUG nova.compute.manager [None req-ebff8349-d5ea-4a95-aee1-e4c9e8b6d3a4 tempest-ServerDiagnosticsTest-195484958 tempest-ServerDiagnosticsTest-195484958-project-member] [instance: 79fc9ea2-c427-47ce-acdb-1002ee5d78c9] Allocating IP information in the background. {{(pid=61806) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 604.260982] env[61806]: DEBUG nova.network.neutron [None req-ebff8349-d5ea-4a95-aee1-e4c9e8b6d3a4 tempest-ServerDiagnosticsTest-195484958 tempest-ServerDiagnosticsTest-195484958-project-member] [instance: 79fc9ea2-c427-47ce-acdb-1002ee5d78c9] allocate_for_instance() {{(pid=61806) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 604.363859] env[61806]: DEBUG nova.policy [None req-ebff8349-d5ea-4a95-aee1-e4c9e8b6d3a4 tempest-ServerDiagnosticsTest-195484958 tempest-ServerDiagnosticsTest-195484958-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '2b5952aa6e3e416ea37aab443475648f', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '2ef18c5a54eb4dc18e0b9ec57601322b', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61806) authorize /opt/stack/nova/nova/policy.py:201}} [ 604.462390] env[61806]: INFO nova.scheduler.client.report [None req-37e71da7-cb5e-431d-ba3c-95daaf8af7f3 tempest-ServersAdminNegativeTestJSON-2128564044 tempest-ServersAdminNegativeTestJSON-2128564044-project-member] Deleted allocations for instance 89ce0fd8-6a22-4b1c-bdff-d1584abd538a [ 604.760298] env[61806]: DEBUG nova.compute.manager [None req-ebff8349-d5ea-4a95-aee1-e4c9e8b6d3a4 tempest-ServerDiagnosticsTest-195484958 tempest-ServerDiagnosticsTest-195484958-project-member] [instance: 79fc9ea2-c427-47ce-acdb-1002ee5d78c9] Start building block device mappings for instance. {{(pid=61806) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 604.899850] env[61806]: DEBUG nova.network.neutron [None req-ebff8349-d5ea-4a95-aee1-e4c9e8b6d3a4 tempest-ServerDiagnosticsTest-195484958 tempest-ServerDiagnosticsTest-195484958-project-member] [instance: 79fc9ea2-c427-47ce-acdb-1002ee5d78c9] Successfully created port: f3ed84a6-47ee-4b8a-875c-72f24cdfd5cb {{(pid=61806) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 604.972172] env[61806]: DEBUG oslo_concurrency.lockutils [None req-37e71da7-cb5e-431d-ba3c-95daaf8af7f3 tempest-ServersAdminNegativeTestJSON-2128564044 tempest-ServersAdminNegativeTestJSON-2128564044-project-member] Lock "89ce0fd8-6a22-4b1c-bdff-d1584abd538a" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 69.463s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 605.184550] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b80ce33e-7029-4067-8e71-88e3bb3c834f {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 605.192244] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3418fa2e-7d74-46dc-8756-1a92c94eee28 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 605.227671] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-228a57cf-0f7b-4b32-ab9d-9438dfef8547 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 605.235303] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9816d5d3-35ac-4f5e-b9a4-da076a5967a1 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 605.251314] env[61806]: DEBUG nova.compute.provider_tree [None req-24088d08-d3d7-4f62-ad63-91029bd9b22d tempest-ServersTestMultiNic-1480543082 tempest-ServersTestMultiNic-1480543082-project-member] Inventory has not changed in ProviderTree for provider: a2858be1-fd22-4e08-979e-87ad25293407 {{(pid=61806) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 605.476975] env[61806]: DEBUG nova.compute.manager [None req-98d962a0-2632-4a9a-8d12-70ce4acc382b tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] [instance: b8249dc5-5561-454a-aa01-477f9f1fc0e1] Starting instance... {{(pid=61806) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 605.755916] env[61806]: DEBUG nova.scheduler.client.report [None req-24088d08-d3d7-4f62-ad63-91029bd9b22d tempest-ServersTestMultiNic-1480543082 tempest-ServersTestMultiNic-1480543082-project-member] Inventory has not changed for provider a2858be1-fd22-4e08-979e-87ad25293407 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 139, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61806) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 605.776801] env[61806]: DEBUG nova.compute.manager [None req-ebff8349-d5ea-4a95-aee1-e4c9e8b6d3a4 tempest-ServerDiagnosticsTest-195484958 tempest-ServerDiagnosticsTest-195484958-project-member] [instance: 79fc9ea2-c427-47ce-acdb-1002ee5d78c9] Start spawning the instance on the hypervisor. {{(pid=61806) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 605.810018] env[61806]: DEBUG nova.virt.hardware [None req-ebff8349-d5ea-4a95-aee1-e4c9e8b6d3a4 tempest-ServerDiagnosticsTest-195484958 tempest-ServerDiagnosticsTest-195484958-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-15T18:05:10Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-15T18:04:52Z,direct_url=,disk_format='vmdk',id=ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='84ba83fce2c349a6988173c5d6fc3b07',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-15T18:04:53Z,virtual_size=,visibility=), allow threads: False {{(pid=61806) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 605.810018] env[61806]: DEBUG nova.virt.hardware [None req-ebff8349-d5ea-4a95-aee1-e4c9e8b6d3a4 tempest-ServerDiagnosticsTest-195484958 tempest-ServerDiagnosticsTest-195484958-project-member] Flavor limits 0:0:0 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 605.810018] env[61806]: DEBUG nova.virt.hardware [None req-ebff8349-d5ea-4a95-aee1-e4c9e8b6d3a4 tempest-ServerDiagnosticsTest-195484958 tempest-ServerDiagnosticsTest-195484958-project-member] Image limits 0:0:0 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 605.810389] env[61806]: DEBUG nova.virt.hardware [None req-ebff8349-d5ea-4a95-aee1-e4c9e8b6d3a4 tempest-ServerDiagnosticsTest-195484958 tempest-ServerDiagnosticsTest-195484958-project-member] Flavor pref 0:0:0 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 605.810389] env[61806]: DEBUG nova.virt.hardware [None req-ebff8349-d5ea-4a95-aee1-e4c9e8b6d3a4 tempest-ServerDiagnosticsTest-195484958 tempest-ServerDiagnosticsTest-195484958-project-member] Image pref 0:0:0 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 605.810389] env[61806]: DEBUG nova.virt.hardware [None req-ebff8349-d5ea-4a95-aee1-e4c9e8b6d3a4 tempest-ServerDiagnosticsTest-195484958 tempest-ServerDiagnosticsTest-195484958-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 605.810389] env[61806]: DEBUG nova.virt.hardware [None req-ebff8349-d5ea-4a95-aee1-e4c9e8b6d3a4 tempest-ServerDiagnosticsTest-195484958 tempest-ServerDiagnosticsTest-195484958-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61806) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 605.810389] env[61806]: DEBUG nova.virt.hardware [None req-ebff8349-d5ea-4a95-aee1-e4c9e8b6d3a4 tempest-ServerDiagnosticsTest-195484958 tempest-ServerDiagnosticsTest-195484958-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61806) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 605.810532] env[61806]: DEBUG nova.virt.hardware [None req-ebff8349-d5ea-4a95-aee1-e4c9e8b6d3a4 tempest-ServerDiagnosticsTest-195484958 tempest-ServerDiagnosticsTest-195484958-project-member] Got 1 possible topologies {{(pid=61806) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 605.810532] env[61806]: DEBUG nova.virt.hardware [None req-ebff8349-d5ea-4a95-aee1-e4c9e8b6d3a4 tempest-ServerDiagnosticsTest-195484958 tempest-ServerDiagnosticsTest-195484958-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61806) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 605.810532] env[61806]: DEBUG nova.virt.hardware [None req-ebff8349-d5ea-4a95-aee1-e4c9e8b6d3a4 tempest-ServerDiagnosticsTest-195484958 tempest-ServerDiagnosticsTest-195484958-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61806) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 605.811814] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-18e17fbd-3b49-4417-8208-4444553d80e4 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 605.823808] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0a562b69-3886-4605-b6d1-281abf107be6 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 605.877798] env[61806]: DEBUG nova.compute.manager [req-f04a2b7a-881c-4f35-9918-d733f6ae174d req-8ae883f1-bea5-459f-a79a-b7805cb3c861 service nova] [instance: 79fc9ea2-c427-47ce-acdb-1002ee5d78c9] Received event network-changed-f3ed84a6-47ee-4b8a-875c-72f24cdfd5cb {{(pid=61806) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 605.877798] env[61806]: DEBUG nova.compute.manager [req-f04a2b7a-881c-4f35-9918-d733f6ae174d req-8ae883f1-bea5-459f-a79a-b7805cb3c861 service nova] [instance: 79fc9ea2-c427-47ce-acdb-1002ee5d78c9] Refreshing instance network info cache due to event network-changed-f3ed84a6-47ee-4b8a-875c-72f24cdfd5cb. {{(pid=61806) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 605.877798] env[61806]: DEBUG oslo_concurrency.lockutils [req-f04a2b7a-881c-4f35-9918-d733f6ae174d req-8ae883f1-bea5-459f-a79a-b7805cb3c861 service nova] Acquiring lock "refresh_cache-79fc9ea2-c427-47ce-acdb-1002ee5d78c9" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 605.877798] env[61806]: DEBUG oslo_concurrency.lockutils [req-f04a2b7a-881c-4f35-9918-d733f6ae174d req-8ae883f1-bea5-459f-a79a-b7805cb3c861 service nova] Acquired lock "refresh_cache-79fc9ea2-c427-47ce-acdb-1002ee5d78c9" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 605.877798] env[61806]: DEBUG nova.network.neutron [req-f04a2b7a-881c-4f35-9918-d733f6ae174d req-8ae883f1-bea5-459f-a79a-b7805cb3c861 service nova] [instance: 79fc9ea2-c427-47ce-acdb-1002ee5d78c9] Refreshing network info cache for port f3ed84a6-47ee-4b8a-875c-72f24cdfd5cb {{(pid=61806) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 606.008099] env[61806]: DEBUG oslo_concurrency.lockutils [None req-98d962a0-2632-4a9a-8d12-70ce4acc382b tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 606.088363] env[61806]: ERROR nova.compute.manager [None req-ebff8349-d5ea-4a95-aee1-e4c9e8b6d3a4 tempest-ServerDiagnosticsTest-195484958 tempest-ServerDiagnosticsTest-195484958-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port f3ed84a6-47ee-4b8a-875c-72f24cdfd5cb, please check neutron logs for more information. [ 606.088363] env[61806]: ERROR nova.compute.manager Traceback (most recent call last): [ 606.088363] env[61806]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 606.088363] env[61806]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 606.088363] env[61806]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 606.088363] env[61806]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 606.088363] env[61806]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 606.088363] env[61806]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 606.088363] env[61806]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 606.088363] env[61806]: ERROR nova.compute.manager self.force_reraise() [ 606.088363] env[61806]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 606.088363] env[61806]: ERROR nova.compute.manager raise self.value [ 606.088363] env[61806]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 606.088363] env[61806]: ERROR nova.compute.manager updated_port = self._update_port( [ 606.088363] env[61806]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 606.088363] env[61806]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 606.088730] env[61806]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 606.088730] env[61806]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 606.088730] env[61806]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port f3ed84a6-47ee-4b8a-875c-72f24cdfd5cb, please check neutron logs for more information. [ 606.088730] env[61806]: ERROR nova.compute.manager [ 606.088730] env[61806]: Traceback (most recent call last): [ 606.088730] env[61806]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 606.088730] env[61806]: listener.cb(fileno) [ 606.088730] env[61806]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 606.088730] env[61806]: result = function(*args, **kwargs) [ 606.088730] env[61806]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 606.088730] env[61806]: return func(*args, **kwargs) [ 606.088730] env[61806]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 606.088730] env[61806]: raise e [ 606.088730] env[61806]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 606.088730] env[61806]: nwinfo = self.network_api.allocate_for_instance( [ 606.088730] env[61806]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 606.088730] env[61806]: created_port_ids = self._update_ports_for_instance( [ 606.088730] env[61806]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 606.088730] env[61806]: with excutils.save_and_reraise_exception(): [ 606.088730] env[61806]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 606.088730] env[61806]: self.force_reraise() [ 606.088730] env[61806]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 606.088730] env[61806]: raise self.value [ 606.088730] env[61806]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 606.088730] env[61806]: updated_port = self._update_port( [ 606.088730] env[61806]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 606.088730] env[61806]: _ensure_no_port_binding_failure(port) [ 606.088730] env[61806]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 606.088730] env[61806]: raise exception.PortBindingFailed(port_id=port['id']) [ 606.089398] env[61806]: nova.exception.PortBindingFailed: Binding failed for port f3ed84a6-47ee-4b8a-875c-72f24cdfd5cb, please check neutron logs for more information. [ 606.089398] env[61806]: Removing descriptor: 17 [ 606.089398] env[61806]: ERROR nova.compute.manager [None req-ebff8349-d5ea-4a95-aee1-e4c9e8b6d3a4 tempest-ServerDiagnosticsTest-195484958 tempest-ServerDiagnosticsTest-195484958-project-member] [instance: 79fc9ea2-c427-47ce-acdb-1002ee5d78c9] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port f3ed84a6-47ee-4b8a-875c-72f24cdfd5cb, please check neutron logs for more information. [ 606.089398] env[61806]: ERROR nova.compute.manager [instance: 79fc9ea2-c427-47ce-acdb-1002ee5d78c9] Traceback (most recent call last): [ 606.089398] env[61806]: ERROR nova.compute.manager [instance: 79fc9ea2-c427-47ce-acdb-1002ee5d78c9] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 606.089398] env[61806]: ERROR nova.compute.manager [instance: 79fc9ea2-c427-47ce-acdb-1002ee5d78c9] yield resources [ 606.089398] env[61806]: ERROR nova.compute.manager [instance: 79fc9ea2-c427-47ce-acdb-1002ee5d78c9] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 606.089398] env[61806]: ERROR nova.compute.manager [instance: 79fc9ea2-c427-47ce-acdb-1002ee5d78c9] self.driver.spawn(context, instance, image_meta, [ 606.089398] env[61806]: ERROR nova.compute.manager [instance: 79fc9ea2-c427-47ce-acdb-1002ee5d78c9] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 606.089398] env[61806]: ERROR nova.compute.manager [instance: 79fc9ea2-c427-47ce-acdb-1002ee5d78c9] self._vmops.spawn(context, instance, image_meta, injected_files, [ 606.089398] env[61806]: ERROR nova.compute.manager [instance: 79fc9ea2-c427-47ce-acdb-1002ee5d78c9] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 606.089398] env[61806]: ERROR nova.compute.manager [instance: 79fc9ea2-c427-47ce-acdb-1002ee5d78c9] vm_ref = self.build_virtual_machine(instance, [ 606.089679] env[61806]: ERROR nova.compute.manager [instance: 79fc9ea2-c427-47ce-acdb-1002ee5d78c9] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 606.089679] env[61806]: ERROR nova.compute.manager [instance: 79fc9ea2-c427-47ce-acdb-1002ee5d78c9] vif_infos = vmwarevif.get_vif_info(self._session, [ 606.089679] env[61806]: ERROR nova.compute.manager [instance: 79fc9ea2-c427-47ce-acdb-1002ee5d78c9] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 606.089679] env[61806]: ERROR nova.compute.manager [instance: 79fc9ea2-c427-47ce-acdb-1002ee5d78c9] for vif in network_info: [ 606.089679] env[61806]: ERROR nova.compute.manager [instance: 79fc9ea2-c427-47ce-acdb-1002ee5d78c9] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 606.089679] env[61806]: ERROR nova.compute.manager [instance: 79fc9ea2-c427-47ce-acdb-1002ee5d78c9] return self._sync_wrapper(fn, *args, **kwargs) [ 606.089679] env[61806]: ERROR nova.compute.manager [instance: 79fc9ea2-c427-47ce-acdb-1002ee5d78c9] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 606.089679] env[61806]: ERROR nova.compute.manager [instance: 79fc9ea2-c427-47ce-acdb-1002ee5d78c9] self.wait() [ 606.089679] env[61806]: ERROR nova.compute.manager [instance: 79fc9ea2-c427-47ce-acdb-1002ee5d78c9] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 606.089679] env[61806]: ERROR nova.compute.manager [instance: 79fc9ea2-c427-47ce-acdb-1002ee5d78c9] self[:] = self._gt.wait() [ 606.089679] env[61806]: ERROR nova.compute.manager [instance: 79fc9ea2-c427-47ce-acdb-1002ee5d78c9] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 606.089679] env[61806]: ERROR nova.compute.manager [instance: 79fc9ea2-c427-47ce-acdb-1002ee5d78c9] return self._exit_event.wait() [ 606.089679] env[61806]: ERROR nova.compute.manager [instance: 79fc9ea2-c427-47ce-acdb-1002ee5d78c9] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 606.089987] env[61806]: ERROR nova.compute.manager [instance: 79fc9ea2-c427-47ce-acdb-1002ee5d78c9] result = hub.switch() [ 606.089987] env[61806]: ERROR nova.compute.manager [instance: 79fc9ea2-c427-47ce-acdb-1002ee5d78c9] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 606.089987] env[61806]: ERROR nova.compute.manager [instance: 79fc9ea2-c427-47ce-acdb-1002ee5d78c9] return self.greenlet.switch() [ 606.089987] env[61806]: ERROR nova.compute.manager [instance: 79fc9ea2-c427-47ce-acdb-1002ee5d78c9] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 606.089987] env[61806]: ERROR nova.compute.manager [instance: 79fc9ea2-c427-47ce-acdb-1002ee5d78c9] result = function(*args, **kwargs) [ 606.089987] env[61806]: ERROR nova.compute.manager [instance: 79fc9ea2-c427-47ce-acdb-1002ee5d78c9] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 606.089987] env[61806]: ERROR nova.compute.manager [instance: 79fc9ea2-c427-47ce-acdb-1002ee5d78c9] return func(*args, **kwargs) [ 606.089987] env[61806]: ERROR nova.compute.manager [instance: 79fc9ea2-c427-47ce-acdb-1002ee5d78c9] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 606.089987] env[61806]: ERROR nova.compute.manager [instance: 79fc9ea2-c427-47ce-acdb-1002ee5d78c9] raise e [ 606.089987] env[61806]: ERROR nova.compute.manager [instance: 79fc9ea2-c427-47ce-acdb-1002ee5d78c9] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 606.089987] env[61806]: ERROR nova.compute.manager [instance: 79fc9ea2-c427-47ce-acdb-1002ee5d78c9] nwinfo = self.network_api.allocate_for_instance( [ 606.089987] env[61806]: ERROR nova.compute.manager [instance: 79fc9ea2-c427-47ce-acdb-1002ee5d78c9] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 606.089987] env[61806]: ERROR nova.compute.manager [instance: 79fc9ea2-c427-47ce-acdb-1002ee5d78c9] created_port_ids = self._update_ports_for_instance( [ 606.090264] env[61806]: ERROR nova.compute.manager [instance: 79fc9ea2-c427-47ce-acdb-1002ee5d78c9] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 606.090264] env[61806]: ERROR nova.compute.manager [instance: 79fc9ea2-c427-47ce-acdb-1002ee5d78c9] with excutils.save_and_reraise_exception(): [ 606.090264] env[61806]: ERROR nova.compute.manager [instance: 79fc9ea2-c427-47ce-acdb-1002ee5d78c9] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 606.090264] env[61806]: ERROR nova.compute.manager [instance: 79fc9ea2-c427-47ce-acdb-1002ee5d78c9] self.force_reraise() [ 606.090264] env[61806]: ERROR nova.compute.manager [instance: 79fc9ea2-c427-47ce-acdb-1002ee5d78c9] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 606.090264] env[61806]: ERROR nova.compute.manager [instance: 79fc9ea2-c427-47ce-acdb-1002ee5d78c9] raise self.value [ 606.090264] env[61806]: ERROR nova.compute.manager [instance: 79fc9ea2-c427-47ce-acdb-1002ee5d78c9] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 606.090264] env[61806]: ERROR nova.compute.manager [instance: 79fc9ea2-c427-47ce-acdb-1002ee5d78c9] updated_port = self._update_port( [ 606.090264] env[61806]: ERROR nova.compute.manager [instance: 79fc9ea2-c427-47ce-acdb-1002ee5d78c9] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 606.090264] env[61806]: ERROR nova.compute.manager [instance: 79fc9ea2-c427-47ce-acdb-1002ee5d78c9] _ensure_no_port_binding_failure(port) [ 606.090264] env[61806]: ERROR nova.compute.manager [instance: 79fc9ea2-c427-47ce-acdb-1002ee5d78c9] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 606.090264] env[61806]: ERROR nova.compute.manager [instance: 79fc9ea2-c427-47ce-acdb-1002ee5d78c9] raise exception.PortBindingFailed(port_id=port['id']) [ 606.090517] env[61806]: ERROR nova.compute.manager [instance: 79fc9ea2-c427-47ce-acdb-1002ee5d78c9] nova.exception.PortBindingFailed: Binding failed for port f3ed84a6-47ee-4b8a-875c-72f24cdfd5cb, please check neutron logs for more information. [ 606.090517] env[61806]: ERROR nova.compute.manager [instance: 79fc9ea2-c427-47ce-acdb-1002ee5d78c9] [ 606.090517] env[61806]: INFO nova.compute.manager [None req-ebff8349-d5ea-4a95-aee1-e4c9e8b6d3a4 tempest-ServerDiagnosticsTest-195484958 tempest-ServerDiagnosticsTest-195484958-project-member] [instance: 79fc9ea2-c427-47ce-acdb-1002ee5d78c9] Terminating instance [ 606.091614] env[61806]: DEBUG oslo_concurrency.lockutils [None req-ebff8349-d5ea-4a95-aee1-e4c9e8b6d3a4 tempest-ServerDiagnosticsTest-195484958 tempest-ServerDiagnosticsTest-195484958-project-member] Acquiring lock "refresh_cache-79fc9ea2-c427-47ce-acdb-1002ee5d78c9" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 606.267849] env[61806]: DEBUG oslo_concurrency.lockutils [None req-24088d08-d3d7-4f62-ad63-91029bd9b22d tempest-ServersTestMultiNic-1480543082 tempest-ServersTestMultiNic-1480543082-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.517s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 606.268443] env[61806]: DEBUG nova.compute.manager [None req-24088d08-d3d7-4f62-ad63-91029bd9b22d tempest-ServersTestMultiNic-1480543082 tempest-ServersTestMultiNic-1480543082-project-member] [instance: c2d1c9f6-f458-4843-982c-afdb2a90edae] Start building networks asynchronously for instance. {{(pid=61806) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 606.272038] env[61806]: DEBUG oslo_concurrency.lockutils [None req-5c11e673-bd02-4846-9be6-b5679ed1117d tempest-VolumesAssistedSnapshotsTest-2037321669 tempest-VolumesAssistedSnapshotsTest-2037321669-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 15.852s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 606.396934] env[61806]: DEBUG nova.network.neutron [req-f04a2b7a-881c-4f35-9918-d733f6ae174d req-8ae883f1-bea5-459f-a79a-b7805cb3c861 service nova] [instance: 79fc9ea2-c427-47ce-acdb-1002ee5d78c9] Instance cache missing network info. {{(pid=61806) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 606.456688] env[61806]: DEBUG nova.network.neutron [req-f04a2b7a-881c-4f35-9918-d733f6ae174d req-8ae883f1-bea5-459f-a79a-b7805cb3c861 service nova] [instance: 79fc9ea2-c427-47ce-acdb-1002ee5d78c9] Updating instance_info_cache with network_info: [] {{(pid=61806) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 606.780165] env[61806]: DEBUG nova.compute.utils [None req-24088d08-d3d7-4f62-ad63-91029bd9b22d tempest-ServersTestMultiNic-1480543082 tempest-ServersTestMultiNic-1480543082-project-member] Using /dev/sd instead of None {{(pid=61806) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 606.781721] env[61806]: DEBUG nova.compute.manager [None req-24088d08-d3d7-4f62-ad63-91029bd9b22d tempest-ServersTestMultiNic-1480543082 tempest-ServersTestMultiNic-1480543082-project-member] [instance: c2d1c9f6-f458-4843-982c-afdb2a90edae] Allocating IP information in the background. {{(pid=61806) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 606.781810] env[61806]: DEBUG nova.network.neutron [None req-24088d08-d3d7-4f62-ad63-91029bd9b22d tempest-ServersTestMultiNic-1480543082 tempest-ServersTestMultiNic-1480543082-project-member] [instance: c2d1c9f6-f458-4843-982c-afdb2a90edae] allocate_for_instance() {{(pid=61806) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 606.842092] env[61806]: DEBUG nova.policy [None req-24088d08-d3d7-4f62-ad63-91029bd9b22d tempest-ServersTestMultiNic-1480543082 tempest-ServersTestMultiNic-1480543082-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '2e72eab1cc5c4fbba3e2d7dbfa0035ec', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '4abf9a81bb9649db910a44e5562dd342', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61806) authorize /opt/stack/nova/nova/policy.py:201}} [ 606.963632] env[61806]: DEBUG oslo_concurrency.lockutils [req-f04a2b7a-881c-4f35-9918-d733f6ae174d req-8ae883f1-bea5-459f-a79a-b7805cb3c861 service nova] Releasing lock "refresh_cache-79fc9ea2-c427-47ce-acdb-1002ee5d78c9" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 606.964060] env[61806]: DEBUG oslo_concurrency.lockutils [None req-ebff8349-d5ea-4a95-aee1-e4c9e8b6d3a4 tempest-ServerDiagnosticsTest-195484958 tempest-ServerDiagnosticsTest-195484958-project-member] Acquired lock "refresh_cache-79fc9ea2-c427-47ce-acdb-1002ee5d78c9" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 606.964252] env[61806]: DEBUG nova.network.neutron [None req-ebff8349-d5ea-4a95-aee1-e4c9e8b6d3a4 tempest-ServerDiagnosticsTest-195484958 tempest-ServerDiagnosticsTest-195484958-project-member] [instance: 79fc9ea2-c427-47ce-acdb-1002ee5d78c9] Building network info cache for instance {{(pid=61806) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 607.177661] env[61806]: DEBUG nova.network.neutron [None req-24088d08-d3d7-4f62-ad63-91029bd9b22d tempest-ServersTestMultiNic-1480543082 tempest-ServersTestMultiNic-1480543082-project-member] [instance: c2d1c9f6-f458-4843-982c-afdb2a90edae] Successfully created port: 3e148a81-f55b-4f13-adfc-ca47776ae7c5 {{(pid=61806) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 607.200272] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-71f7fc85-7883-4a45-adbe-0df3e9d6f772 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 607.208238] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-00cccccd-3240-47d6-9e7d-42449e2ae388 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 607.239335] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8222400a-4484-4429-a658-dea9a258c025 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 607.246536] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-52b4fc11-1a8d-4e6a-9a71-36fadee3d7c2 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 607.262386] env[61806]: DEBUG nova.compute.provider_tree [None req-5c11e673-bd02-4846-9be6-b5679ed1117d tempest-VolumesAssistedSnapshotsTest-2037321669 tempest-VolumesAssistedSnapshotsTest-2037321669-project-member] Inventory has not changed in ProviderTree for provider: a2858be1-fd22-4e08-979e-87ad25293407 {{(pid=61806) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 607.286732] env[61806]: DEBUG nova.compute.manager [None req-24088d08-d3d7-4f62-ad63-91029bd9b22d tempest-ServersTestMultiNic-1480543082 tempest-ServersTestMultiNic-1480543082-project-member] [instance: c2d1c9f6-f458-4843-982c-afdb2a90edae] Start building block device mappings for instance. {{(pid=61806) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 607.486193] env[61806]: DEBUG nova.network.neutron [None req-ebff8349-d5ea-4a95-aee1-e4c9e8b6d3a4 tempest-ServerDiagnosticsTest-195484958 tempest-ServerDiagnosticsTest-195484958-project-member] [instance: 79fc9ea2-c427-47ce-acdb-1002ee5d78c9] Instance cache missing network info. {{(pid=61806) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 607.555227] env[61806]: DEBUG nova.network.neutron [None req-24088d08-d3d7-4f62-ad63-91029bd9b22d tempest-ServersTestMultiNic-1480543082 tempest-ServersTestMultiNic-1480543082-project-member] [instance: c2d1c9f6-f458-4843-982c-afdb2a90edae] Successfully created port: f1f8ff4b-1b6e-4512-a569-032b54010b8c {{(pid=61806) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 607.557277] env[61806]: DEBUG nova.network.neutron [None req-ebff8349-d5ea-4a95-aee1-e4c9e8b6d3a4 tempest-ServerDiagnosticsTest-195484958 tempest-ServerDiagnosticsTest-195484958-project-member] [instance: 79fc9ea2-c427-47ce-acdb-1002ee5d78c9] Updating instance_info_cache with network_info: [] {{(pid=61806) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 607.765793] env[61806]: DEBUG nova.scheduler.client.report [None req-5c11e673-bd02-4846-9be6-b5679ed1117d tempest-VolumesAssistedSnapshotsTest-2037321669 tempest-VolumesAssistedSnapshotsTest-2037321669-project-member] Inventory has not changed for provider a2858be1-fd22-4e08-979e-87ad25293407 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 139, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61806) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 607.847261] env[61806]: DEBUG nova.network.neutron [None req-24088d08-d3d7-4f62-ad63-91029bd9b22d tempest-ServersTestMultiNic-1480543082 tempest-ServersTestMultiNic-1480543082-project-member] [instance: c2d1c9f6-f458-4843-982c-afdb2a90edae] Successfully created port: 3f63a718-d3f8-4933-9f1f-5549fc0960cf {{(pid=61806) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 607.902234] env[61806]: DEBUG nova.compute.manager [req-88b00f29-408e-4aa5-be5e-911fcde5660b req-1b360441-d20a-48fc-888b-dc1fcb511da9 service nova] [instance: 79fc9ea2-c427-47ce-acdb-1002ee5d78c9] Received event network-vif-deleted-f3ed84a6-47ee-4b8a-875c-72f24cdfd5cb {{(pid=61806) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 608.061052] env[61806]: DEBUG oslo_concurrency.lockutils [None req-ebff8349-d5ea-4a95-aee1-e4c9e8b6d3a4 tempest-ServerDiagnosticsTest-195484958 tempest-ServerDiagnosticsTest-195484958-project-member] Releasing lock "refresh_cache-79fc9ea2-c427-47ce-acdb-1002ee5d78c9" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 608.061945] env[61806]: DEBUG nova.compute.manager [None req-ebff8349-d5ea-4a95-aee1-e4c9e8b6d3a4 tempest-ServerDiagnosticsTest-195484958 tempest-ServerDiagnosticsTest-195484958-project-member] [instance: 79fc9ea2-c427-47ce-acdb-1002ee5d78c9] Start destroying the instance on the hypervisor. {{(pid=61806) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 608.062573] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-ebff8349-d5ea-4a95-aee1-e4c9e8b6d3a4 tempest-ServerDiagnosticsTest-195484958 tempest-ServerDiagnosticsTest-195484958-project-member] [instance: 79fc9ea2-c427-47ce-acdb-1002ee5d78c9] Destroying instance {{(pid=61806) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 608.062758] env[61806]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-a8384370-42a7-4408-976d-fa62a295449c {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 608.075397] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fe06e08e-3855-48d2-8f38-07ed7c8362a0 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 608.098724] env[61806]: WARNING nova.virt.vmwareapi.vmops [None req-ebff8349-d5ea-4a95-aee1-e4c9e8b6d3a4 tempest-ServerDiagnosticsTest-195484958 tempest-ServerDiagnosticsTest-195484958-project-member] [instance: 79fc9ea2-c427-47ce-acdb-1002ee5d78c9] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 79fc9ea2-c427-47ce-acdb-1002ee5d78c9 could not be found. [ 608.099032] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-ebff8349-d5ea-4a95-aee1-e4c9e8b6d3a4 tempest-ServerDiagnosticsTest-195484958 tempest-ServerDiagnosticsTest-195484958-project-member] [instance: 79fc9ea2-c427-47ce-acdb-1002ee5d78c9] Instance destroyed {{(pid=61806) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 608.099266] env[61806]: INFO nova.compute.manager [None req-ebff8349-d5ea-4a95-aee1-e4c9e8b6d3a4 tempest-ServerDiagnosticsTest-195484958 tempest-ServerDiagnosticsTest-195484958-project-member] [instance: 79fc9ea2-c427-47ce-acdb-1002ee5d78c9] Took 0.04 seconds to destroy the instance on the hypervisor. [ 608.099941] env[61806]: DEBUG oslo.service.loopingcall [None req-ebff8349-d5ea-4a95-aee1-e4c9e8b6d3a4 tempest-ServerDiagnosticsTest-195484958 tempest-ServerDiagnosticsTest-195484958-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61806) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 608.100411] env[61806]: DEBUG nova.compute.manager [-] [instance: 79fc9ea2-c427-47ce-acdb-1002ee5d78c9] Deallocating network for instance {{(pid=61806) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 608.100526] env[61806]: DEBUG nova.network.neutron [-] [instance: 79fc9ea2-c427-47ce-acdb-1002ee5d78c9] deallocate_for_instance() {{(pid=61806) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 608.118303] env[61806]: DEBUG nova.network.neutron [-] [instance: 79fc9ea2-c427-47ce-acdb-1002ee5d78c9] Instance cache missing network info. {{(pid=61806) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 608.274155] env[61806]: DEBUG oslo_concurrency.lockutils [None req-5c11e673-bd02-4846-9be6-b5679ed1117d tempest-VolumesAssistedSnapshotsTest-2037321669 tempest-VolumesAssistedSnapshotsTest-2037321669-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.002s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 608.274811] env[61806]: ERROR nova.compute.manager [None req-5c11e673-bd02-4846-9be6-b5679ed1117d tempest-VolumesAssistedSnapshotsTest-2037321669 tempest-VolumesAssistedSnapshotsTest-2037321669-project-member] [instance: 76722306-a9ad-4ef9-b579-d958c36cf6bc] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 0094ad04-66d2-46ea-8ce7-09664ac7c614, please check neutron logs for more information. [ 608.274811] env[61806]: ERROR nova.compute.manager [instance: 76722306-a9ad-4ef9-b579-d958c36cf6bc] Traceback (most recent call last): [ 608.274811] env[61806]: ERROR nova.compute.manager [instance: 76722306-a9ad-4ef9-b579-d958c36cf6bc] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 608.274811] env[61806]: ERROR nova.compute.manager [instance: 76722306-a9ad-4ef9-b579-d958c36cf6bc] self.driver.spawn(context, instance, image_meta, [ 608.274811] env[61806]: ERROR nova.compute.manager [instance: 76722306-a9ad-4ef9-b579-d958c36cf6bc] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 608.274811] env[61806]: ERROR nova.compute.manager [instance: 76722306-a9ad-4ef9-b579-d958c36cf6bc] self._vmops.spawn(context, instance, image_meta, injected_files, [ 608.274811] env[61806]: ERROR nova.compute.manager [instance: 76722306-a9ad-4ef9-b579-d958c36cf6bc] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 608.274811] env[61806]: ERROR nova.compute.manager [instance: 76722306-a9ad-4ef9-b579-d958c36cf6bc] vm_ref = self.build_virtual_machine(instance, [ 608.274811] env[61806]: ERROR nova.compute.manager [instance: 76722306-a9ad-4ef9-b579-d958c36cf6bc] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 608.274811] env[61806]: ERROR nova.compute.manager [instance: 76722306-a9ad-4ef9-b579-d958c36cf6bc] vif_infos = vmwarevif.get_vif_info(self._session, [ 608.274811] env[61806]: ERROR nova.compute.manager [instance: 76722306-a9ad-4ef9-b579-d958c36cf6bc] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 608.275075] env[61806]: ERROR nova.compute.manager [instance: 76722306-a9ad-4ef9-b579-d958c36cf6bc] for vif in network_info: [ 608.275075] env[61806]: ERROR nova.compute.manager [instance: 76722306-a9ad-4ef9-b579-d958c36cf6bc] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 608.275075] env[61806]: ERROR nova.compute.manager [instance: 76722306-a9ad-4ef9-b579-d958c36cf6bc] return self._sync_wrapper(fn, *args, **kwargs) [ 608.275075] env[61806]: ERROR nova.compute.manager [instance: 76722306-a9ad-4ef9-b579-d958c36cf6bc] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 608.275075] env[61806]: ERROR nova.compute.manager [instance: 76722306-a9ad-4ef9-b579-d958c36cf6bc] self.wait() [ 608.275075] env[61806]: ERROR nova.compute.manager [instance: 76722306-a9ad-4ef9-b579-d958c36cf6bc] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 608.275075] env[61806]: ERROR nova.compute.manager [instance: 76722306-a9ad-4ef9-b579-d958c36cf6bc] self[:] = self._gt.wait() [ 608.275075] env[61806]: ERROR nova.compute.manager [instance: 76722306-a9ad-4ef9-b579-d958c36cf6bc] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 608.275075] env[61806]: ERROR nova.compute.manager [instance: 76722306-a9ad-4ef9-b579-d958c36cf6bc] return self._exit_event.wait() [ 608.275075] env[61806]: ERROR nova.compute.manager [instance: 76722306-a9ad-4ef9-b579-d958c36cf6bc] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 608.275075] env[61806]: ERROR nova.compute.manager [instance: 76722306-a9ad-4ef9-b579-d958c36cf6bc] result = hub.switch() [ 608.275075] env[61806]: ERROR nova.compute.manager [instance: 76722306-a9ad-4ef9-b579-d958c36cf6bc] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 608.275075] env[61806]: ERROR nova.compute.manager [instance: 76722306-a9ad-4ef9-b579-d958c36cf6bc] return self.greenlet.switch() [ 608.275344] env[61806]: ERROR nova.compute.manager [instance: 76722306-a9ad-4ef9-b579-d958c36cf6bc] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 608.275344] env[61806]: ERROR nova.compute.manager [instance: 76722306-a9ad-4ef9-b579-d958c36cf6bc] result = function(*args, **kwargs) [ 608.275344] env[61806]: ERROR nova.compute.manager [instance: 76722306-a9ad-4ef9-b579-d958c36cf6bc] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 608.275344] env[61806]: ERROR nova.compute.manager [instance: 76722306-a9ad-4ef9-b579-d958c36cf6bc] return func(*args, **kwargs) [ 608.275344] env[61806]: ERROR nova.compute.manager [instance: 76722306-a9ad-4ef9-b579-d958c36cf6bc] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 608.275344] env[61806]: ERROR nova.compute.manager [instance: 76722306-a9ad-4ef9-b579-d958c36cf6bc] raise e [ 608.275344] env[61806]: ERROR nova.compute.manager [instance: 76722306-a9ad-4ef9-b579-d958c36cf6bc] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 608.275344] env[61806]: ERROR nova.compute.manager [instance: 76722306-a9ad-4ef9-b579-d958c36cf6bc] nwinfo = self.network_api.allocate_for_instance( [ 608.275344] env[61806]: ERROR nova.compute.manager [instance: 76722306-a9ad-4ef9-b579-d958c36cf6bc] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 608.275344] env[61806]: ERROR nova.compute.manager [instance: 76722306-a9ad-4ef9-b579-d958c36cf6bc] created_port_ids = self._update_ports_for_instance( [ 608.275344] env[61806]: ERROR nova.compute.manager [instance: 76722306-a9ad-4ef9-b579-d958c36cf6bc] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 608.275344] env[61806]: ERROR nova.compute.manager [instance: 76722306-a9ad-4ef9-b579-d958c36cf6bc] with excutils.save_and_reraise_exception(): [ 608.275344] env[61806]: ERROR nova.compute.manager [instance: 76722306-a9ad-4ef9-b579-d958c36cf6bc] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 608.275674] env[61806]: ERROR nova.compute.manager [instance: 76722306-a9ad-4ef9-b579-d958c36cf6bc] self.force_reraise() [ 608.275674] env[61806]: ERROR nova.compute.manager [instance: 76722306-a9ad-4ef9-b579-d958c36cf6bc] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 608.275674] env[61806]: ERROR nova.compute.manager [instance: 76722306-a9ad-4ef9-b579-d958c36cf6bc] raise self.value [ 608.275674] env[61806]: ERROR nova.compute.manager [instance: 76722306-a9ad-4ef9-b579-d958c36cf6bc] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 608.275674] env[61806]: ERROR nova.compute.manager [instance: 76722306-a9ad-4ef9-b579-d958c36cf6bc] updated_port = self._update_port( [ 608.275674] env[61806]: ERROR nova.compute.manager [instance: 76722306-a9ad-4ef9-b579-d958c36cf6bc] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 608.275674] env[61806]: ERROR nova.compute.manager [instance: 76722306-a9ad-4ef9-b579-d958c36cf6bc] _ensure_no_port_binding_failure(port) [ 608.275674] env[61806]: ERROR nova.compute.manager [instance: 76722306-a9ad-4ef9-b579-d958c36cf6bc] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 608.275674] env[61806]: ERROR nova.compute.manager [instance: 76722306-a9ad-4ef9-b579-d958c36cf6bc] raise exception.PortBindingFailed(port_id=port['id']) [ 608.275674] env[61806]: ERROR nova.compute.manager [instance: 76722306-a9ad-4ef9-b579-d958c36cf6bc] nova.exception.PortBindingFailed: Binding failed for port 0094ad04-66d2-46ea-8ce7-09664ac7c614, please check neutron logs for more information. [ 608.275674] env[61806]: ERROR nova.compute.manager [instance: 76722306-a9ad-4ef9-b579-d958c36cf6bc] [ 608.276630] env[61806]: DEBUG nova.compute.utils [None req-5c11e673-bd02-4846-9be6-b5679ed1117d tempest-VolumesAssistedSnapshotsTest-2037321669 tempest-VolumesAssistedSnapshotsTest-2037321669-project-member] [instance: 76722306-a9ad-4ef9-b579-d958c36cf6bc] Binding failed for port 0094ad04-66d2-46ea-8ce7-09664ac7c614, please check neutron logs for more information. {{(pid=61806) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 608.282132] env[61806]: DEBUG oslo_concurrency.lockutils [None req-b50d0802-a025-4c5c-b8e9-95dd624e7ff5 tempest-ServersWithSpecificFlavorTestJSON-815343160 tempest-ServersWithSpecificFlavorTestJSON-815343160-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 16.426s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 608.288941] env[61806]: DEBUG nova.compute.manager [None req-5c11e673-bd02-4846-9be6-b5679ed1117d tempest-VolumesAssistedSnapshotsTest-2037321669 tempest-VolumesAssistedSnapshotsTest-2037321669-project-member] [instance: 76722306-a9ad-4ef9-b579-d958c36cf6bc] Build of instance 76722306-a9ad-4ef9-b579-d958c36cf6bc was re-scheduled: Binding failed for port 0094ad04-66d2-46ea-8ce7-09664ac7c614, please check neutron logs for more information. {{(pid=61806) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 608.289515] env[61806]: DEBUG nova.compute.manager [None req-5c11e673-bd02-4846-9be6-b5679ed1117d tempest-VolumesAssistedSnapshotsTest-2037321669 tempest-VolumesAssistedSnapshotsTest-2037321669-project-member] [instance: 76722306-a9ad-4ef9-b579-d958c36cf6bc] Unplugging VIFs for instance {{(pid=61806) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 608.289515] env[61806]: DEBUG oslo_concurrency.lockutils [None req-5c11e673-bd02-4846-9be6-b5679ed1117d tempest-VolumesAssistedSnapshotsTest-2037321669 tempest-VolumesAssistedSnapshotsTest-2037321669-project-member] Acquiring lock "refresh_cache-76722306-a9ad-4ef9-b579-d958c36cf6bc" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 608.289638] env[61806]: DEBUG oslo_concurrency.lockutils [None req-5c11e673-bd02-4846-9be6-b5679ed1117d tempest-VolumesAssistedSnapshotsTest-2037321669 tempest-VolumesAssistedSnapshotsTest-2037321669-project-member] Acquired lock "refresh_cache-76722306-a9ad-4ef9-b579-d958c36cf6bc" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 608.289815] env[61806]: DEBUG nova.network.neutron [None req-5c11e673-bd02-4846-9be6-b5679ed1117d tempest-VolumesAssistedSnapshotsTest-2037321669 tempest-VolumesAssistedSnapshotsTest-2037321669-project-member] [instance: 76722306-a9ad-4ef9-b579-d958c36cf6bc] Building network info cache for instance {{(pid=61806) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 608.292510] env[61806]: DEBUG nova.compute.manager [None req-24088d08-d3d7-4f62-ad63-91029bd9b22d tempest-ServersTestMultiNic-1480543082 tempest-ServersTestMultiNic-1480543082-project-member] [instance: c2d1c9f6-f458-4843-982c-afdb2a90edae] Start spawning the instance on the hypervisor. {{(pid=61806) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 608.327935] env[61806]: DEBUG nova.virt.hardware [None req-24088d08-d3d7-4f62-ad63-91029bd9b22d tempest-ServersTestMultiNic-1480543082 tempest-ServersTestMultiNic-1480543082-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-15T18:05:10Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-15T18:04:52Z,direct_url=,disk_format='vmdk',id=ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='84ba83fce2c349a6988173c5d6fc3b07',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-15T18:04:53Z,virtual_size=,visibility=), allow threads: False {{(pid=61806) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 608.328197] env[61806]: DEBUG nova.virt.hardware [None req-24088d08-d3d7-4f62-ad63-91029bd9b22d tempest-ServersTestMultiNic-1480543082 tempest-ServersTestMultiNic-1480543082-project-member] Flavor limits 0:0:0 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 608.328371] env[61806]: DEBUG nova.virt.hardware [None req-24088d08-d3d7-4f62-ad63-91029bd9b22d tempest-ServersTestMultiNic-1480543082 tempest-ServersTestMultiNic-1480543082-project-member] Image limits 0:0:0 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 608.328556] env[61806]: DEBUG nova.virt.hardware [None req-24088d08-d3d7-4f62-ad63-91029bd9b22d tempest-ServersTestMultiNic-1480543082 tempest-ServersTestMultiNic-1480543082-project-member] Flavor pref 0:0:0 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 608.328727] env[61806]: DEBUG nova.virt.hardware [None req-24088d08-d3d7-4f62-ad63-91029bd9b22d tempest-ServersTestMultiNic-1480543082 tempest-ServersTestMultiNic-1480543082-project-member] Image pref 0:0:0 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 608.328863] env[61806]: DEBUG nova.virt.hardware [None req-24088d08-d3d7-4f62-ad63-91029bd9b22d tempest-ServersTestMultiNic-1480543082 tempest-ServersTestMultiNic-1480543082-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 608.329106] env[61806]: DEBUG nova.virt.hardware [None req-24088d08-d3d7-4f62-ad63-91029bd9b22d tempest-ServersTestMultiNic-1480543082 tempest-ServersTestMultiNic-1480543082-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61806) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 608.329276] env[61806]: DEBUG nova.virt.hardware [None req-24088d08-d3d7-4f62-ad63-91029bd9b22d tempest-ServersTestMultiNic-1480543082 tempest-ServersTestMultiNic-1480543082-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61806) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 608.329443] env[61806]: DEBUG nova.virt.hardware [None req-24088d08-d3d7-4f62-ad63-91029bd9b22d tempest-ServersTestMultiNic-1480543082 tempest-ServersTestMultiNic-1480543082-project-member] Got 1 possible topologies {{(pid=61806) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 608.329607] env[61806]: DEBUG nova.virt.hardware [None req-24088d08-d3d7-4f62-ad63-91029bd9b22d tempest-ServersTestMultiNic-1480543082 tempest-ServersTestMultiNic-1480543082-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61806) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 608.329776] env[61806]: DEBUG nova.virt.hardware [None req-24088d08-d3d7-4f62-ad63-91029bd9b22d tempest-ServersTestMultiNic-1480543082 tempest-ServersTestMultiNic-1480543082-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61806) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 608.330934] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-15c922b8-9b94-439b-b976-d64c968c4a0d {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 608.339150] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-34783076-be30-4b96-b386-5be4ddfb443d {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 608.620948] env[61806]: DEBUG nova.network.neutron [-] [instance: 79fc9ea2-c427-47ce-acdb-1002ee5d78c9] Updating instance_info_cache with network_info: [] {{(pid=61806) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 608.811986] env[61806]: DEBUG nova.network.neutron [None req-5c11e673-bd02-4846-9be6-b5679ed1117d tempest-VolumesAssistedSnapshotsTest-2037321669 tempest-VolumesAssistedSnapshotsTest-2037321669-project-member] [instance: 76722306-a9ad-4ef9-b579-d958c36cf6bc] Instance cache missing network info. {{(pid=61806) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 608.962739] env[61806]: DEBUG nova.network.neutron [None req-5c11e673-bd02-4846-9be6-b5679ed1117d tempest-VolumesAssistedSnapshotsTest-2037321669 tempest-VolumesAssistedSnapshotsTest-2037321669-project-member] [instance: 76722306-a9ad-4ef9-b579-d958c36cf6bc] Updating instance_info_cache with network_info: [] {{(pid=61806) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 609.001686] env[61806]: ERROR nova.compute.manager [None req-24088d08-d3d7-4f62-ad63-91029bd9b22d tempest-ServersTestMultiNic-1480543082 tempest-ServersTestMultiNic-1480543082-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 3e148a81-f55b-4f13-adfc-ca47776ae7c5, please check neutron logs for more information. [ 609.001686] env[61806]: ERROR nova.compute.manager Traceback (most recent call last): [ 609.001686] env[61806]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 609.001686] env[61806]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 609.001686] env[61806]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 609.001686] env[61806]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 609.001686] env[61806]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 609.001686] env[61806]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 609.001686] env[61806]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 609.001686] env[61806]: ERROR nova.compute.manager self.force_reraise() [ 609.001686] env[61806]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 609.001686] env[61806]: ERROR nova.compute.manager raise self.value [ 609.001686] env[61806]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 609.001686] env[61806]: ERROR nova.compute.manager updated_port = self._update_port( [ 609.001686] env[61806]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 609.001686] env[61806]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 609.002461] env[61806]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 609.002461] env[61806]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 609.002461] env[61806]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 3e148a81-f55b-4f13-adfc-ca47776ae7c5, please check neutron logs for more information. [ 609.002461] env[61806]: ERROR nova.compute.manager [ 609.002461] env[61806]: Traceback (most recent call last): [ 609.002461] env[61806]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 609.002461] env[61806]: listener.cb(fileno) [ 609.002461] env[61806]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 609.002461] env[61806]: result = function(*args, **kwargs) [ 609.002461] env[61806]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 609.002461] env[61806]: return func(*args, **kwargs) [ 609.002461] env[61806]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 609.002461] env[61806]: raise e [ 609.002461] env[61806]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 609.002461] env[61806]: nwinfo = self.network_api.allocate_for_instance( [ 609.002461] env[61806]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 609.002461] env[61806]: created_port_ids = self._update_ports_for_instance( [ 609.002461] env[61806]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 609.002461] env[61806]: with excutils.save_and_reraise_exception(): [ 609.002461] env[61806]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 609.002461] env[61806]: self.force_reraise() [ 609.002461] env[61806]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 609.002461] env[61806]: raise self.value [ 609.002461] env[61806]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 609.002461] env[61806]: updated_port = self._update_port( [ 609.002461] env[61806]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 609.002461] env[61806]: _ensure_no_port_binding_failure(port) [ 609.002461] env[61806]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 609.002461] env[61806]: raise exception.PortBindingFailed(port_id=port['id']) [ 609.003425] env[61806]: nova.exception.PortBindingFailed: Binding failed for port 3e148a81-f55b-4f13-adfc-ca47776ae7c5, please check neutron logs for more information. [ 609.003425] env[61806]: Removing descriptor: 17 [ 609.003425] env[61806]: ERROR nova.compute.manager [None req-24088d08-d3d7-4f62-ad63-91029bd9b22d tempest-ServersTestMultiNic-1480543082 tempest-ServersTestMultiNic-1480543082-project-member] [instance: c2d1c9f6-f458-4843-982c-afdb2a90edae] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 3e148a81-f55b-4f13-adfc-ca47776ae7c5, please check neutron logs for more information. [ 609.003425] env[61806]: ERROR nova.compute.manager [instance: c2d1c9f6-f458-4843-982c-afdb2a90edae] Traceback (most recent call last): [ 609.003425] env[61806]: ERROR nova.compute.manager [instance: c2d1c9f6-f458-4843-982c-afdb2a90edae] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 609.003425] env[61806]: ERROR nova.compute.manager [instance: c2d1c9f6-f458-4843-982c-afdb2a90edae] yield resources [ 609.003425] env[61806]: ERROR nova.compute.manager [instance: c2d1c9f6-f458-4843-982c-afdb2a90edae] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 609.003425] env[61806]: ERROR nova.compute.manager [instance: c2d1c9f6-f458-4843-982c-afdb2a90edae] self.driver.spawn(context, instance, image_meta, [ 609.003425] env[61806]: ERROR nova.compute.manager [instance: c2d1c9f6-f458-4843-982c-afdb2a90edae] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 609.003425] env[61806]: ERROR nova.compute.manager [instance: c2d1c9f6-f458-4843-982c-afdb2a90edae] self._vmops.spawn(context, instance, image_meta, injected_files, [ 609.003425] env[61806]: ERROR nova.compute.manager [instance: c2d1c9f6-f458-4843-982c-afdb2a90edae] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 609.003425] env[61806]: ERROR nova.compute.manager [instance: c2d1c9f6-f458-4843-982c-afdb2a90edae] vm_ref = self.build_virtual_machine(instance, [ 609.003682] env[61806]: ERROR nova.compute.manager [instance: c2d1c9f6-f458-4843-982c-afdb2a90edae] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 609.003682] env[61806]: ERROR nova.compute.manager [instance: c2d1c9f6-f458-4843-982c-afdb2a90edae] vif_infos = vmwarevif.get_vif_info(self._session, [ 609.003682] env[61806]: ERROR nova.compute.manager [instance: c2d1c9f6-f458-4843-982c-afdb2a90edae] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 609.003682] env[61806]: ERROR nova.compute.manager [instance: c2d1c9f6-f458-4843-982c-afdb2a90edae] for vif in network_info: [ 609.003682] env[61806]: ERROR nova.compute.manager [instance: c2d1c9f6-f458-4843-982c-afdb2a90edae] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 609.003682] env[61806]: ERROR nova.compute.manager [instance: c2d1c9f6-f458-4843-982c-afdb2a90edae] return self._sync_wrapper(fn, *args, **kwargs) [ 609.003682] env[61806]: ERROR nova.compute.manager [instance: c2d1c9f6-f458-4843-982c-afdb2a90edae] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 609.003682] env[61806]: ERROR nova.compute.manager [instance: c2d1c9f6-f458-4843-982c-afdb2a90edae] self.wait() [ 609.003682] env[61806]: ERROR nova.compute.manager [instance: c2d1c9f6-f458-4843-982c-afdb2a90edae] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 609.003682] env[61806]: ERROR nova.compute.manager [instance: c2d1c9f6-f458-4843-982c-afdb2a90edae] self[:] = self._gt.wait() [ 609.003682] env[61806]: ERROR nova.compute.manager [instance: c2d1c9f6-f458-4843-982c-afdb2a90edae] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 609.003682] env[61806]: ERROR nova.compute.manager [instance: c2d1c9f6-f458-4843-982c-afdb2a90edae] return self._exit_event.wait() [ 609.003682] env[61806]: ERROR nova.compute.manager [instance: c2d1c9f6-f458-4843-982c-afdb2a90edae] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 609.003950] env[61806]: ERROR nova.compute.manager [instance: c2d1c9f6-f458-4843-982c-afdb2a90edae] result = hub.switch() [ 609.003950] env[61806]: ERROR nova.compute.manager [instance: c2d1c9f6-f458-4843-982c-afdb2a90edae] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 609.003950] env[61806]: ERROR nova.compute.manager [instance: c2d1c9f6-f458-4843-982c-afdb2a90edae] return self.greenlet.switch() [ 609.003950] env[61806]: ERROR nova.compute.manager [instance: c2d1c9f6-f458-4843-982c-afdb2a90edae] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 609.003950] env[61806]: ERROR nova.compute.manager [instance: c2d1c9f6-f458-4843-982c-afdb2a90edae] result = function(*args, **kwargs) [ 609.003950] env[61806]: ERROR nova.compute.manager [instance: c2d1c9f6-f458-4843-982c-afdb2a90edae] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 609.003950] env[61806]: ERROR nova.compute.manager [instance: c2d1c9f6-f458-4843-982c-afdb2a90edae] return func(*args, **kwargs) [ 609.003950] env[61806]: ERROR nova.compute.manager [instance: c2d1c9f6-f458-4843-982c-afdb2a90edae] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 609.003950] env[61806]: ERROR nova.compute.manager [instance: c2d1c9f6-f458-4843-982c-afdb2a90edae] raise e [ 609.003950] env[61806]: ERROR nova.compute.manager [instance: c2d1c9f6-f458-4843-982c-afdb2a90edae] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 609.003950] env[61806]: ERROR nova.compute.manager [instance: c2d1c9f6-f458-4843-982c-afdb2a90edae] nwinfo = self.network_api.allocate_for_instance( [ 609.003950] env[61806]: ERROR nova.compute.manager [instance: c2d1c9f6-f458-4843-982c-afdb2a90edae] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 609.003950] env[61806]: ERROR nova.compute.manager [instance: c2d1c9f6-f458-4843-982c-afdb2a90edae] created_port_ids = self._update_ports_for_instance( [ 609.004230] env[61806]: ERROR nova.compute.manager [instance: c2d1c9f6-f458-4843-982c-afdb2a90edae] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 609.004230] env[61806]: ERROR nova.compute.manager [instance: c2d1c9f6-f458-4843-982c-afdb2a90edae] with excutils.save_and_reraise_exception(): [ 609.004230] env[61806]: ERROR nova.compute.manager [instance: c2d1c9f6-f458-4843-982c-afdb2a90edae] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 609.004230] env[61806]: ERROR nova.compute.manager [instance: c2d1c9f6-f458-4843-982c-afdb2a90edae] self.force_reraise() [ 609.004230] env[61806]: ERROR nova.compute.manager [instance: c2d1c9f6-f458-4843-982c-afdb2a90edae] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 609.004230] env[61806]: ERROR nova.compute.manager [instance: c2d1c9f6-f458-4843-982c-afdb2a90edae] raise self.value [ 609.004230] env[61806]: ERROR nova.compute.manager [instance: c2d1c9f6-f458-4843-982c-afdb2a90edae] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 609.004230] env[61806]: ERROR nova.compute.manager [instance: c2d1c9f6-f458-4843-982c-afdb2a90edae] updated_port = self._update_port( [ 609.004230] env[61806]: ERROR nova.compute.manager [instance: c2d1c9f6-f458-4843-982c-afdb2a90edae] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 609.004230] env[61806]: ERROR nova.compute.manager [instance: c2d1c9f6-f458-4843-982c-afdb2a90edae] _ensure_no_port_binding_failure(port) [ 609.004230] env[61806]: ERROR nova.compute.manager [instance: c2d1c9f6-f458-4843-982c-afdb2a90edae] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 609.004230] env[61806]: ERROR nova.compute.manager [instance: c2d1c9f6-f458-4843-982c-afdb2a90edae] raise exception.PortBindingFailed(port_id=port['id']) [ 609.004477] env[61806]: ERROR nova.compute.manager [instance: c2d1c9f6-f458-4843-982c-afdb2a90edae] nova.exception.PortBindingFailed: Binding failed for port 3e148a81-f55b-4f13-adfc-ca47776ae7c5, please check neutron logs for more information. [ 609.004477] env[61806]: ERROR nova.compute.manager [instance: c2d1c9f6-f458-4843-982c-afdb2a90edae] [ 609.004477] env[61806]: INFO nova.compute.manager [None req-24088d08-d3d7-4f62-ad63-91029bd9b22d tempest-ServersTestMultiNic-1480543082 tempest-ServersTestMultiNic-1480543082-project-member] [instance: c2d1c9f6-f458-4843-982c-afdb2a90edae] Terminating instance [ 609.005303] env[61806]: DEBUG oslo_concurrency.lockutils [None req-24088d08-d3d7-4f62-ad63-91029bd9b22d tempest-ServersTestMultiNic-1480543082 tempest-ServersTestMultiNic-1480543082-project-member] Acquiring lock "refresh_cache-c2d1c9f6-f458-4843-982c-afdb2a90edae" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 609.005468] env[61806]: DEBUG oslo_concurrency.lockutils [None req-24088d08-d3d7-4f62-ad63-91029bd9b22d tempest-ServersTestMultiNic-1480543082 tempest-ServersTestMultiNic-1480543082-project-member] Acquired lock "refresh_cache-c2d1c9f6-f458-4843-982c-afdb2a90edae" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 609.005633] env[61806]: DEBUG nova.network.neutron [None req-24088d08-d3d7-4f62-ad63-91029bd9b22d tempest-ServersTestMultiNic-1480543082 tempest-ServersTestMultiNic-1480543082-project-member] [instance: c2d1c9f6-f458-4843-982c-afdb2a90edae] Building network info cache for instance {{(pid=61806) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 609.125393] env[61806]: INFO nova.compute.manager [-] [instance: 79fc9ea2-c427-47ce-acdb-1002ee5d78c9] Took 1.02 seconds to deallocate network for instance. [ 609.127871] env[61806]: DEBUG nova.compute.claims [None req-ebff8349-d5ea-4a95-aee1-e4c9e8b6d3a4 tempest-ServerDiagnosticsTest-195484958 tempest-ServerDiagnosticsTest-195484958-project-member] [instance: 79fc9ea2-c427-47ce-acdb-1002ee5d78c9] Aborting claim: {{(pid=61806) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 609.128058] env[61806]: DEBUG oslo_concurrency.lockutils [None req-ebff8349-d5ea-4a95-aee1-e4c9e8b6d3a4 tempest-ServerDiagnosticsTest-195484958 tempest-ServerDiagnosticsTest-195484958-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 609.210882] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-07032fae-abc2-41a0-a20d-3fc26a2f49c8 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 609.219212] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6703dc57-6517-4e94-81cc-8057effe7fc0 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 609.254125] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3ee4095f-587d-47ec-a32a-b0c164b1b40a {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 609.262280] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0e9ca877-1211-45f3-a0bf-206f95e6ad8d {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 609.279082] env[61806]: DEBUG nova.compute.provider_tree [None req-b50d0802-a025-4c5c-b8e9-95dd624e7ff5 tempest-ServersWithSpecificFlavorTestJSON-815343160 tempest-ServersWithSpecificFlavorTestJSON-815343160-project-member] Inventory has not changed in ProviderTree for provider: a2858be1-fd22-4e08-979e-87ad25293407 {{(pid=61806) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 609.464164] env[61806]: DEBUG oslo_concurrency.lockutils [None req-5c11e673-bd02-4846-9be6-b5679ed1117d tempest-VolumesAssistedSnapshotsTest-2037321669 tempest-VolumesAssistedSnapshotsTest-2037321669-project-member] Releasing lock "refresh_cache-76722306-a9ad-4ef9-b579-d958c36cf6bc" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 609.464447] env[61806]: DEBUG nova.compute.manager [None req-5c11e673-bd02-4846-9be6-b5679ed1117d tempest-VolumesAssistedSnapshotsTest-2037321669 tempest-VolumesAssistedSnapshotsTest-2037321669-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61806) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 609.464624] env[61806]: DEBUG nova.compute.manager [None req-5c11e673-bd02-4846-9be6-b5679ed1117d tempest-VolumesAssistedSnapshotsTest-2037321669 tempest-VolumesAssistedSnapshotsTest-2037321669-project-member] [instance: 76722306-a9ad-4ef9-b579-d958c36cf6bc] Deallocating network for instance {{(pid=61806) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 609.464738] env[61806]: DEBUG nova.network.neutron [None req-5c11e673-bd02-4846-9be6-b5679ed1117d tempest-VolumesAssistedSnapshotsTest-2037321669 tempest-VolumesAssistedSnapshotsTest-2037321669-project-member] [instance: 76722306-a9ad-4ef9-b579-d958c36cf6bc] deallocate_for_instance() {{(pid=61806) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 609.486971] env[61806]: DEBUG nova.network.neutron [None req-5c11e673-bd02-4846-9be6-b5679ed1117d tempest-VolumesAssistedSnapshotsTest-2037321669 tempest-VolumesAssistedSnapshotsTest-2037321669-project-member] [instance: 76722306-a9ad-4ef9-b579-d958c36cf6bc] Instance cache missing network info. {{(pid=61806) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 609.571471] env[61806]: DEBUG nova.network.neutron [None req-24088d08-d3d7-4f62-ad63-91029bd9b22d tempest-ServersTestMultiNic-1480543082 tempest-ServersTestMultiNic-1480543082-project-member] [instance: c2d1c9f6-f458-4843-982c-afdb2a90edae] Instance cache missing network info. {{(pid=61806) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 609.751106] env[61806]: DEBUG nova.network.neutron [None req-24088d08-d3d7-4f62-ad63-91029bd9b22d tempest-ServersTestMultiNic-1480543082 tempest-ServersTestMultiNic-1480543082-project-member] [instance: c2d1c9f6-f458-4843-982c-afdb2a90edae] Updating instance_info_cache with network_info: [] {{(pid=61806) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 609.782647] env[61806]: DEBUG nova.scheduler.client.report [None req-b50d0802-a025-4c5c-b8e9-95dd624e7ff5 tempest-ServersWithSpecificFlavorTestJSON-815343160 tempest-ServersWithSpecificFlavorTestJSON-815343160-project-member] Inventory has not changed for provider a2858be1-fd22-4e08-979e-87ad25293407 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 139, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61806) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 609.940736] env[61806]: DEBUG nova.compute.manager [req-eb25ea80-c0d0-431f-aaee-a1e663ef0892 req-a6be85b4-eda5-4c9f-a4f7-41215c34c87e service nova] [instance: c2d1c9f6-f458-4843-982c-afdb2a90edae] Received event network-changed-3e148a81-f55b-4f13-adfc-ca47776ae7c5 {{(pid=61806) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 609.941024] env[61806]: DEBUG nova.compute.manager [req-eb25ea80-c0d0-431f-aaee-a1e663ef0892 req-a6be85b4-eda5-4c9f-a4f7-41215c34c87e service nova] [instance: c2d1c9f6-f458-4843-982c-afdb2a90edae] Refreshing instance network info cache due to event network-changed-3e148a81-f55b-4f13-adfc-ca47776ae7c5. {{(pid=61806) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 609.941253] env[61806]: DEBUG oslo_concurrency.lockutils [req-eb25ea80-c0d0-431f-aaee-a1e663ef0892 req-a6be85b4-eda5-4c9f-a4f7-41215c34c87e service nova] Acquiring lock "refresh_cache-c2d1c9f6-f458-4843-982c-afdb2a90edae" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 609.990773] env[61806]: DEBUG nova.network.neutron [None req-5c11e673-bd02-4846-9be6-b5679ed1117d tempest-VolumesAssistedSnapshotsTest-2037321669 tempest-VolumesAssistedSnapshotsTest-2037321669-project-member] [instance: 76722306-a9ad-4ef9-b579-d958c36cf6bc] Updating instance_info_cache with network_info: [] {{(pid=61806) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 610.253694] env[61806]: DEBUG oslo_concurrency.lockutils [None req-24088d08-d3d7-4f62-ad63-91029bd9b22d tempest-ServersTestMultiNic-1480543082 tempest-ServersTestMultiNic-1480543082-project-member] Releasing lock "refresh_cache-c2d1c9f6-f458-4843-982c-afdb2a90edae" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 610.254228] env[61806]: DEBUG nova.compute.manager [None req-24088d08-d3d7-4f62-ad63-91029bd9b22d tempest-ServersTestMultiNic-1480543082 tempest-ServersTestMultiNic-1480543082-project-member] [instance: c2d1c9f6-f458-4843-982c-afdb2a90edae] Start destroying the instance on the hypervisor. {{(pid=61806) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 610.254458] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-24088d08-d3d7-4f62-ad63-91029bd9b22d tempest-ServersTestMultiNic-1480543082 tempest-ServersTestMultiNic-1480543082-project-member] [instance: c2d1c9f6-f458-4843-982c-afdb2a90edae] Destroying instance {{(pid=61806) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 610.254826] env[61806]: DEBUG oslo_concurrency.lockutils [req-eb25ea80-c0d0-431f-aaee-a1e663ef0892 req-a6be85b4-eda5-4c9f-a4f7-41215c34c87e service nova] Acquired lock "refresh_cache-c2d1c9f6-f458-4843-982c-afdb2a90edae" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 610.255095] env[61806]: DEBUG nova.network.neutron [req-eb25ea80-c0d0-431f-aaee-a1e663ef0892 req-a6be85b4-eda5-4c9f-a4f7-41215c34c87e service nova] [instance: c2d1c9f6-f458-4843-982c-afdb2a90edae] Refreshing network info cache for port 3e148a81-f55b-4f13-adfc-ca47776ae7c5 {{(pid=61806) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 610.256711] env[61806]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-eed94647-76f1-4761-9b53-7cb014cf3af7 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 610.266448] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d02040f6-04bd-412c-9ba3-3471b4f262e3 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 610.291442] env[61806]: DEBUG oslo_concurrency.lockutils [None req-b50d0802-a025-4c5c-b8e9-95dd624e7ff5 tempest-ServersWithSpecificFlavorTestJSON-815343160 tempest-ServersWithSpecificFlavorTestJSON-815343160-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.009s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 610.291784] env[61806]: ERROR nova.compute.manager [None req-b50d0802-a025-4c5c-b8e9-95dd624e7ff5 tempest-ServersWithSpecificFlavorTestJSON-815343160 tempest-ServersWithSpecificFlavorTestJSON-815343160-project-member] [instance: b68b509e-96dc-489d-9cf1-93ba91095af1] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port ec4cd098-60bf-4a9b-9f8a-2d85646b89ab, please check neutron logs for more information. [ 610.291784] env[61806]: ERROR nova.compute.manager [instance: b68b509e-96dc-489d-9cf1-93ba91095af1] Traceback (most recent call last): [ 610.291784] env[61806]: ERROR nova.compute.manager [instance: b68b509e-96dc-489d-9cf1-93ba91095af1] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 610.291784] env[61806]: ERROR nova.compute.manager [instance: b68b509e-96dc-489d-9cf1-93ba91095af1] self.driver.spawn(context, instance, image_meta, [ 610.291784] env[61806]: ERROR nova.compute.manager [instance: b68b509e-96dc-489d-9cf1-93ba91095af1] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 610.291784] env[61806]: ERROR nova.compute.manager [instance: b68b509e-96dc-489d-9cf1-93ba91095af1] self._vmops.spawn(context, instance, image_meta, injected_files, [ 610.291784] env[61806]: ERROR nova.compute.manager [instance: b68b509e-96dc-489d-9cf1-93ba91095af1] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 610.291784] env[61806]: ERROR nova.compute.manager [instance: b68b509e-96dc-489d-9cf1-93ba91095af1] vm_ref = self.build_virtual_machine(instance, [ 610.291784] env[61806]: ERROR nova.compute.manager [instance: b68b509e-96dc-489d-9cf1-93ba91095af1] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 610.291784] env[61806]: ERROR nova.compute.manager [instance: b68b509e-96dc-489d-9cf1-93ba91095af1] vif_infos = vmwarevif.get_vif_info(self._session, [ 610.291784] env[61806]: ERROR nova.compute.manager [instance: b68b509e-96dc-489d-9cf1-93ba91095af1] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 610.292215] env[61806]: ERROR nova.compute.manager [instance: b68b509e-96dc-489d-9cf1-93ba91095af1] for vif in network_info: [ 610.292215] env[61806]: ERROR nova.compute.manager [instance: b68b509e-96dc-489d-9cf1-93ba91095af1] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 610.292215] env[61806]: ERROR nova.compute.manager [instance: b68b509e-96dc-489d-9cf1-93ba91095af1] return self._sync_wrapper(fn, *args, **kwargs) [ 610.292215] env[61806]: ERROR nova.compute.manager [instance: b68b509e-96dc-489d-9cf1-93ba91095af1] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 610.292215] env[61806]: ERROR nova.compute.manager [instance: b68b509e-96dc-489d-9cf1-93ba91095af1] self.wait() [ 610.292215] env[61806]: ERROR nova.compute.manager [instance: b68b509e-96dc-489d-9cf1-93ba91095af1] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 610.292215] env[61806]: ERROR nova.compute.manager [instance: b68b509e-96dc-489d-9cf1-93ba91095af1] self[:] = self._gt.wait() [ 610.292215] env[61806]: ERROR nova.compute.manager [instance: b68b509e-96dc-489d-9cf1-93ba91095af1] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 610.292215] env[61806]: ERROR nova.compute.manager [instance: b68b509e-96dc-489d-9cf1-93ba91095af1] return self._exit_event.wait() [ 610.292215] env[61806]: ERROR nova.compute.manager [instance: b68b509e-96dc-489d-9cf1-93ba91095af1] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 610.292215] env[61806]: ERROR nova.compute.manager [instance: b68b509e-96dc-489d-9cf1-93ba91095af1] result = hub.switch() [ 610.292215] env[61806]: ERROR nova.compute.manager [instance: b68b509e-96dc-489d-9cf1-93ba91095af1] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 610.292215] env[61806]: ERROR nova.compute.manager [instance: b68b509e-96dc-489d-9cf1-93ba91095af1] return self.greenlet.switch() [ 610.292603] env[61806]: ERROR nova.compute.manager [instance: b68b509e-96dc-489d-9cf1-93ba91095af1] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 610.292603] env[61806]: ERROR nova.compute.manager [instance: b68b509e-96dc-489d-9cf1-93ba91095af1] result = function(*args, **kwargs) [ 610.292603] env[61806]: ERROR nova.compute.manager [instance: b68b509e-96dc-489d-9cf1-93ba91095af1] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 610.292603] env[61806]: ERROR nova.compute.manager [instance: b68b509e-96dc-489d-9cf1-93ba91095af1] return func(*args, **kwargs) [ 610.292603] env[61806]: ERROR nova.compute.manager [instance: b68b509e-96dc-489d-9cf1-93ba91095af1] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 610.292603] env[61806]: ERROR nova.compute.manager [instance: b68b509e-96dc-489d-9cf1-93ba91095af1] raise e [ 610.292603] env[61806]: ERROR nova.compute.manager [instance: b68b509e-96dc-489d-9cf1-93ba91095af1] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 610.292603] env[61806]: ERROR nova.compute.manager [instance: b68b509e-96dc-489d-9cf1-93ba91095af1] nwinfo = self.network_api.allocate_for_instance( [ 610.292603] env[61806]: ERROR nova.compute.manager [instance: b68b509e-96dc-489d-9cf1-93ba91095af1] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 610.292603] env[61806]: ERROR nova.compute.manager [instance: b68b509e-96dc-489d-9cf1-93ba91095af1] created_port_ids = self._update_ports_for_instance( [ 610.292603] env[61806]: ERROR nova.compute.manager [instance: b68b509e-96dc-489d-9cf1-93ba91095af1] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 610.292603] env[61806]: ERROR nova.compute.manager [instance: b68b509e-96dc-489d-9cf1-93ba91095af1] with excutils.save_and_reraise_exception(): [ 610.292603] env[61806]: ERROR nova.compute.manager [instance: b68b509e-96dc-489d-9cf1-93ba91095af1] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 610.293020] env[61806]: ERROR nova.compute.manager [instance: b68b509e-96dc-489d-9cf1-93ba91095af1] self.force_reraise() [ 610.293020] env[61806]: ERROR nova.compute.manager [instance: b68b509e-96dc-489d-9cf1-93ba91095af1] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 610.293020] env[61806]: ERROR nova.compute.manager [instance: b68b509e-96dc-489d-9cf1-93ba91095af1] raise self.value [ 610.293020] env[61806]: ERROR nova.compute.manager [instance: b68b509e-96dc-489d-9cf1-93ba91095af1] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 610.293020] env[61806]: ERROR nova.compute.manager [instance: b68b509e-96dc-489d-9cf1-93ba91095af1] updated_port = self._update_port( [ 610.293020] env[61806]: ERROR nova.compute.manager [instance: b68b509e-96dc-489d-9cf1-93ba91095af1] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 610.293020] env[61806]: ERROR nova.compute.manager [instance: b68b509e-96dc-489d-9cf1-93ba91095af1] _ensure_no_port_binding_failure(port) [ 610.293020] env[61806]: ERROR nova.compute.manager [instance: b68b509e-96dc-489d-9cf1-93ba91095af1] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 610.293020] env[61806]: ERROR nova.compute.manager [instance: b68b509e-96dc-489d-9cf1-93ba91095af1] raise exception.PortBindingFailed(port_id=port['id']) [ 610.293020] env[61806]: ERROR nova.compute.manager [instance: b68b509e-96dc-489d-9cf1-93ba91095af1] nova.exception.PortBindingFailed: Binding failed for port ec4cd098-60bf-4a9b-9f8a-2d85646b89ab, please check neutron logs for more information. [ 610.293020] env[61806]: ERROR nova.compute.manager [instance: b68b509e-96dc-489d-9cf1-93ba91095af1] [ 610.293264] env[61806]: DEBUG nova.compute.utils [None req-b50d0802-a025-4c5c-b8e9-95dd624e7ff5 tempest-ServersWithSpecificFlavorTestJSON-815343160 tempest-ServersWithSpecificFlavorTestJSON-815343160-project-member] [instance: b68b509e-96dc-489d-9cf1-93ba91095af1] Binding failed for port ec4cd098-60bf-4a9b-9f8a-2d85646b89ab, please check neutron logs for more information. {{(pid=61806) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 610.293939] env[61806]: WARNING nova.virt.vmwareapi.vmops [None req-24088d08-d3d7-4f62-ad63-91029bd9b22d tempest-ServersTestMultiNic-1480543082 tempest-ServersTestMultiNic-1480543082-project-member] [instance: c2d1c9f6-f458-4843-982c-afdb2a90edae] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance c2d1c9f6-f458-4843-982c-afdb2a90edae could not be found. [ 610.297430] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-24088d08-d3d7-4f62-ad63-91029bd9b22d tempest-ServersTestMultiNic-1480543082 tempest-ServersTestMultiNic-1480543082-project-member] [instance: c2d1c9f6-f458-4843-982c-afdb2a90edae] Instance destroyed {{(pid=61806) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 610.297430] env[61806]: INFO nova.compute.manager [None req-24088d08-d3d7-4f62-ad63-91029bd9b22d tempest-ServersTestMultiNic-1480543082 tempest-ServersTestMultiNic-1480543082-project-member] [instance: c2d1c9f6-f458-4843-982c-afdb2a90edae] Took 0.04 seconds to destroy the instance on the hypervisor. [ 610.297430] env[61806]: DEBUG oslo.service.loopingcall [None req-24088d08-d3d7-4f62-ad63-91029bd9b22d tempest-ServersTestMultiNic-1480543082 tempest-ServersTestMultiNic-1480543082-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61806) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 610.297430] env[61806]: DEBUG nova.compute.manager [None req-b50d0802-a025-4c5c-b8e9-95dd624e7ff5 tempest-ServersWithSpecificFlavorTestJSON-815343160 tempest-ServersWithSpecificFlavorTestJSON-815343160-project-member] [instance: b68b509e-96dc-489d-9cf1-93ba91095af1] Build of instance b68b509e-96dc-489d-9cf1-93ba91095af1 was re-scheduled: Binding failed for port ec4cd098-60bf-4a9b-9f8a-2d85646b89ab, please check neutron logs for more information. {{(pid=61806) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 610.297641] env[61806]: DEBUG nova.compute.manager [None req-b50d0802-a025-4c5c-b8e9-95dd624e7ff5 tempest-ServersWithSpecificFlavorTestJSON-815343160 tempest-ServersWithSpecificFlavorTestJSON-815343160-project-member] [instance: b68b509e-96dc-489d-9cf1-93ba91095af1] Unplugging VIFs for instance {{(pid=61806) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 610.297641] env[61806]: DEBUG oslo_concurrency.lockutils [None req-b50d0802-a025-4c5c-b8e9-95dd624e7ff5 tempest-ServersWithSpecificFlavorTestJSON-815343160 tempest-ServersWithSpecificFlavorTestJSON-815343160-project-member] Acquiring lock "refresh_cache-b68b509e-96dc-489d-9cf1-93ba91095af1" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 610.297641] env[61806]: DEBUG oslo_concurrency.lockutils [None req-b50d0802-a025-4c5c-b8e9-95dd624e7ff5 tempest-ServersWithSpecificFlavorTestJSON-815343160 tempest-ServersWithSpecificFlavorTestJSON-815343160-project-member] Acquired lock "refresh_cache-b68b509e-96dc-489d-9cf1-93ba91095af1" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 610.297641] env[61806]: DEBUG nova.network.neutron [None req-b50d0802-a025-4c5c-b8e9-95dd624e7ff5 tempest-ServersWithSpecificFlavorTestJSON-815343160 tempest-ServersWithSpecificFlavorTestJSON-815343160-project-member] [instance: b68b509e-96dc-489d-9cf1-93ba91095af1] Building network info cache for instance {{(pid=61806) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 610.297749] env[61806]: DEBUG oslo_concurrency.lockutils [None req-e34d2bcf-266f-4faa-8f31-6bd93f6ab176 tempest-ImagesTestJSON-618852556 tempest-ImagesTestJSON-618852556-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 16.065s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 610.298325] env[61806]: INFO nova.compute.claims [None req-e34d2bcf-266f-4faa-8f31-6bd93f6ab176 tempest-ImagesTestJSON-618852556 tempest-ImagesTestJSON-618852556-project-member] [instance: 65f3b2f3-6d25-44f0-95cb-f9a7e75f8e82] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 610.300705] env[61806]: DEBUG nova.compute.manager [-] [instance: c2d1c9f6-f458-4843-982c-afdb2a90edae] Deallocating network for instance {{(pid=61806) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 610.300832] env[61806]: DEBUG nova.network.neutron [-] [instance: c2d1c9f6-f458-4843-982c-afdb2a90edae] deallocate_for_instance() {{(pid=61806) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 610.356011] env[61806]: DEBUG nova.network.neutron [-] [instance: c2d1c9f6-f458-4843-982c-afdb2a90edae] Instance cache missing network info. {{(pid=61806) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 610.494561] env[61806]: INFO nova.compute.manager [None req-5c11e673-bd02-4846-9be6-b5679ed1117d tempest-VolumesAssistedSnapshotsTest-2037321669 tempest-VolumesAssistedSnapshotsTest-2037321669-project-member] [instance: 76722306-a9ad-4ef9-b579-d958c36cf6bc] Took 1.03 seconds to deallocate network for instance. [ 610.803848] env[61806]: DEBUG nova.network.neutron [req-eb25ea80-c0d0-431f-aaee-a1e663ef0892 req-a6be85b4-eda5-4c9f-a4f7-41215c34c87e service nova] [instance: c2d1c9f6-f458-4843-982c-afdb2a90edae] Instance cache missing network info. {{(pid=61806) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 610.846138] env[61806]: DEBUG nova.network.neutron [None req-b50d0802-a025-4c5c-b8e9-95dd624e7ff5 tempest-ServersWithSpecificFlavorTestJSON-815343160 tempest-ServersWithSpecificFlavorTestJSON-815343160-project-member] [instance: b68b509e-96dc-489d-9cf1-93ba91095af1] Instance cache missing network info. {{(pid=61806) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 610.995016] env[61806]: DEBUG nova.network.neutron [None req-b50d0802-a025-4c5c-b8e9-95dd624e7ff5 tempest-ServersWithSpecificFlavorTestJSON-815343160 tempest-ServersWithSpecificFlavorTestJSON-815343160-project-member] [instance: b68b509e-96dc-489d-9cf1-93ba91095af1] Updating instance_info_cache with network_info: [] {{(pid=61806) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 611.096132] env[61806]: DEBUG nova.network.neutron [req-eb25ea80-c0d0-431f-aaee-a1e663ef0892 req-a6be85b4-eda5-4c9f-a4f7-41215c34c87e service nova] [instance: c2d1c9f6-f458-4843-982c-afdb2a90edae] Updating instance_info_cache with network_info: [] {{(pid=61806) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 611.461418] env[61806]: DEBUG nova.network.neutron [-] [instance: c2d1c9f6-f458-4843-982c-afdb2a90edae] Updating instance_info_cache with network_info: [] {{(pid=61806) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 611.498478] env[61806]: DEBUG oslo_concurrency.lockutils [None req-b50d0802-a025-4c5c-b8e9-95dd624e7ff5 tempest-ServersWithSpecificFlavorTestJSON-815343160 tempest-ServersWithSpecificFlavorTestJSON-815343160-project-member] Releasing lock "refresh_cache-b68b509e-96dc-489d-9cf1-93ba91095af1" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 611.498732] env[61806]: DEBUG nova.compute.manager [None req-b50d0802-a025-4c5c-b8e9-95dd624e7ff5 tempest-ServersWithSpecificFlavorTestJSON-815343160 tempest-ServersWithSpecificFlavorTestJSON-815343160-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61806) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 611.498873] env[61806]: DEBUG nova.compute.manager [None req-b50d0802-a025-4c5c-b8e9-95dd624e7ff5 tempest-ServersWithSpecificFlavorTestJSON-815343160 tempest-ServersWithSpecificFlavorTestJSON-815343160-project-member] [instance: b68b509e-96dc-489d-9cf1-93ba91095af1] Deallocating network for instance {{(pid=61806) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 611.499035] env[61806]: DEBUG nova.network.neutron [None req-b50d0802-a025-4c5c-b8e9-95dd624e7ff5 tempest-ServersWithSpecificFlavorTestJSON-815343160 tempest-ServersWithSpecificFlavorTestJSON-815343160-project-member] [instance: b68b509e-96dc-489d-9cf1-93ba91095af1] deallocate_for_instance() {{(pid=61806) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 611.524056] env[61806]: DEBUG nova.network.neutron [None req-b50d0802-a025-4c5c-b8e9-95dd624e7ff5 tempest-ServersWithSpecificFlavorTestJSON-815343160 tempest-ServersWithSpecificFlavorTestJSON-815343160-project-member] [instance: b68b509e-96dc-489d-9cf1-93ba91095af1] Instance cache missing network info. {{(pid=61806) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 611.533914] env[61806]: INFO nova.scheduler.client.report [None req-5c11e673-bd02-4846-9be6-b5679ed1117d tempest-VolumesAssistedSnapshotsTest-2037321669 tempest-VolumesAssistedSnapshotsTest-2037321669-project-member] Deleted allocations for instance 76722306-a9ad-4ef9-b579-d958c36cf6bc [ 611.600555] env[61806]: DEBUG oslo_concurrency.lockutils [req-eb25ea80-c0d0-431f-aaee-a1e663ef0892 req-a6be85b4-eda5-4c9f-a4f7-41215c34c87e service nova] Releasing lock "refresh_cache-c2d1c9f6-f458-4843-982c-afdb2a90edae" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 611.600863] env[61806]: DEBUG nova.compute.manager [req-eb25ea80-c0d0-431f-aaee-a1e663ef0892 req-a6be85b4-eda5-4c9f-a4f7-41215c34c87e service nova] [instance: c2d1c9f6-f458-4843-982c-afdb2a90edae] Received event network-vif-deleted-3e148a81-f55b-4f13-adfc-ca47776ae7c5 {{(pid=61806) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 611.766726] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8a4f8878-a0f6-4b52-a948-a1e571c58fe7 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 611.776358] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-892a180d-cf45-413c-bda0-cc889e3939dd {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 611.812825] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-34842651-f547-4b73-9a68-f52f872f76a4 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 611.820802] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4e6cc954-e47b-4498-bde2-8ffbc3108992 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 611.835702] env[61806]: DEBUG nova.compute.provider_tree [None req-e34d2bcf-266f-4faa-8f31-6bd93f6ab176 tempest-ImagesTestJSON-618852556 tempest-ImagesTestJSON-618852556-project-member] Inventory has not changed in ProviderTree for provider: a2858be1-fd22-4e08-979e-87ad25293407 {{(pid=61806) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 611.965774] env[61806]: INFO nova.compute.manager [-] [instance: c2d1c9f6-f458-4843-982c-afdb2a90edae] Took 1.66 seconds to deallocate network for instance. [ 611.967095] env[61806]: DEBUG nova.compute.claims [None req-24088d08-d3d7-4f62-ad63-91029bd9b22d tempest-ServersTestMultiNic-1480543082 tempest-ServersTestMultiNic-1480543082-project-member] [instance: c2d1c9f6-f458-4843-982c-afdb2a90edae] Aborting claim: {{(pid=61806) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 611.967447] env[61806]: DEBUG oslo_concurrency.lockutils [None req-24088d08-d3d7-4f62-ad63-91029bd9b22d tempest-ServersTestMultiNic-1480543082 tempest-ServersTestMultiNic-1480543082-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 612.027528] env[61806]: DEBUG nova.network.neutron [None req-b50d0802-a025-4c5c-b8e9-95dd624e7ff5 tempest-ServersWithSpecificFlavorTestJSON-815343160 tempest-ServersWithSpecificFlavorTestJSON-815343160-project-member] [instance: b68b509e-96dc-489d-9cf1-93ba91095af1] Updating instance_info_cache with network_info: [] {{(pid=61806) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 612.044995] env[61806]: DEBUG oslo_concurrency.lockutils [None req-5c11e673-bd02-4846-9be6-b5679ed1117d tempest-VolumesAssistedSnapshotsTest-2037321669 tempest-VolumesAssistedSnapshotsTest-2037321669-project-member] Lock "76722306-a9ad-4ef9-b579-d958c36cf6bc" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 73.699s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 612.341590] env[61806]: DEBUG nova.scheduler.client.report [None req-e34d2bcf-266f-4faa-8f31-6bd93f6ab176 tempest-ImagesTestJSON-618852556 tempest-ImagesTestJSON-618852556-project-member] Inventory has not changed for provider a2858be1-fd22-4e08-979e-87ad25293407 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 139, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61806) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 612.530336] env[61806]: INFO nova.compute.manager [None req-b50d0802-a025-4c5c-b8e9-95dd624e7ff5 tempest-ServersWithSpecificFlavorTestJSON-815343160 tempest-ServersWithSpecificFlavorTestJSON-815343160-project-member] [instance: b68b509e-96dc-489d-9cf1-93ba91095af1] Took 1.03 seconds to deallocate network for instance. [ 612.548158] env[61806]: DEBUG nova.compute.manager [None req-3e50272a-5153-4e11-beee-d774a917f2df tempest-ServerPasswordTestJSON-1741413915 tempest-ServerPasswordTestJSON-1741413915-project-member] [instance: 0d757385-ff1c-49c9-825f-f9d542ae074f] Starting instance... {{(pid=61806) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 612.845108] env[61806]: DEBUG oslo_concurrency.lockutils [None req-e34d2bcf-266f-4faa-8f31-6bd93f6ab176 tempest-ImagesTestJSON-618852556 tempest-ImagesTestJSON-618852556-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.548s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 612.845770] env[61806]: DEBUG nova.compute.manager [None req-e34d2bcf-266f-4faa-8f31-6bd93f6ab176 tempest-ImagesTestJSON-618852556 tempest-ImagesTestJSON-618852556-project-member] [instance: 65f3b2f3-6d25-44f0-95cb-f9a7e75f8e82] Start building networks asynchronously for instance. {{(pid=61806) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 612.848391] env[61806]: DEBUG oslo_concurrency.lockutils [None req-9101cb9e-b191-4241-8bc1-9b126210ca50 tempest-ListImageFiltersTestJSON-680116525 tempest-ListImageFiltersTestJSON-680116525-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 15.827s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 613.068126] env[61806]: DEBUG oslo_concurrency.lockutils [None req-3e50272a-5153-4e11-beee-d774a917f2df tempest-ServerPasswordTestJSON-1741413915 tempest-ServerPasswordTestJSON-1741413915-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 613.356837] env[61806]: DEBUG nova.compute.utils [None req-e34d2bcf-266f-4faa-8f31-6bd93f6ab176 tempest-ImagesTestJSON-618852556 tempest-ImagesTestJSON-618852556-project-member] Using /dev/sd instead of None {{(pid=61806) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 613.358924] env[61806]: DEBUG nova.compute.manager [None req-e34d2bcf-266f-4faa-8f31-6bd93f6ab176 tempest-ImagesTestJSON-618852556 tempest-ImagesTestJSON-618852556-project-member] [instance: 65f3b2f3-6d25-44f0-95cb-f9a7e75f8e82] Allocating IP information in the background. {{(pid=61806) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 613.358924] env[61806]: DEBUG nova.network.neutron [None req-e34d2bcf-266f-4faa-8f31-6bd93f6ab176 tempest-ImagesTestJSON-618852556 tempest-ImagesTestJSON-618852556-project-member] [instance: 65f3b2f3-6d25-44f0-95cb-f9a7e75f8e82] allocate_for_instance() {{(pid=61806) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 613.421311] env[61806]: DEBUG nova.policy [None req-e34d2bcf-266f-4faa-8f31-6bd93f6ab176 tempest-ImagesTestJSON-618852556 tempest-ImagesTestJSON-618852556-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'e6fa11a9b63f4cd6b04baf3115431167', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '4b9aa91aff1d4008ac5096902b77f852', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61806) authorize /opt/stack/nova/nova/policy.py:201}} [ 613.564360] env[61806]: INFO nova.scheduler.client.report [None req-b50d0802-a025-4c5c-b8e9-95dd624e7ff5 tempest-ServersWithSpecificFlavorTestJSON-815343160 tempest-ServersWithSpecificFlavorTestJSON-815343160-project-member] Deleted allocations for instance b68b509e-96dc-489d-9cf1-93ba91095af1 [ 613.727472] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eb4addc2-7c1e-4278-8b3a-271cb6d5e602 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 613.735631] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7f0c304d-0285-48cf-a2da-7c76c33523dd {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 613.773290] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bc58619f-a2ed-4ce9-b840-49060d3a0220 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 613.783765] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8655e5b9-7184-4e84-8653-f1eb6e06eff0 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 613.796965] env[61806]: DEBUG nova.compute.provider_tree [None req-9101cb9e-b191-4241-8bc1-9b126210ca50 tempest-ListImageFiltersTestJSON-680116525 tempest-ListImageFiltersTestJSON-680116525-project-member] Inventory has not changed in ProviderTree for provider: a2858be1-fd22-4e08-979e-87ad25293407 {{(pid=61806) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 613.824439] env[61806]: DEBUG nova.network.neutron [None req-e34d2bcf-266f-4faa-8f31-6bd93f6ab176 tempest-ImagesTestJSON-618852556 tempest-ImagesTestJSON-618852556-project-member] [instance: 65f3b2f3-6d25-44f0-95cb-f9a7e75f8e82] Successfully created port: ddedeb2e-1da8-4774-925f-5f0e09c0cae8 {{(pid=61806) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 613.860894] env[61806]: DEBUG nova.compute.manager [None req-e34d2bcf-266f-4faa-8f31-6bd93f6ab176 tempest-ImagesTestJSON-618852556 tempest-ImagesTestJSON-618852556-project-member] [instance: 65f3b2f3-6d25-44f0-95cb-f9a7e75f8e82] Start building block device mappings for instance. {{(pid=61806) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 614.073563] env[61806]: DEBUG oslo_concurrency.lockutils [None req-b50d0802-a025-4c5c-b8e9-95dd624e7ff5 tempest-ServersWithSpecificFlavorTestJSON-815343160 tempest-ServersWithSpecificFlavorTestJSON-815343160-project-member] Lock "b68b509e-96dc-489d-9cf1-93ba91095af1" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 75.061s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 614.300368] env[61806]: DEBUG nova.scheduler.client.report [None req-9101cb9e-b191-4241-8bc1-9b126210ca50 tempest-ListImageFiltersTestJSON-680116525 tempest-ListImageFiltersTestJSON-680116525-project-member] Inventory has not changed for provider a2858be1-fd22-4e08-979e-87ad25293407 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 139, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61806) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 614.574844] env[61806]: DEBUG nova.compute.manager [None req-f8b2f38e-90f5-4b03-89bc-ca1926adb40c tempest-FloatingIPsAssociationTestJSON-35718468 tempest-FloatingIPsAssociationTestJSON-35718468-project-member] [instance: 88cb0961-ad34-4ffc-8fb2-6034f08ee740] Starting instance... {{(pid=61806) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 614.807028] env[61806]: DEBUG oslo_concurrency.lockutils [None req-9101cb9e-b191-4241-8bc1-9b126210ca50 tempest-ListImageFiltersTestJSON-680116525 tempest-ListImageFiltersTestJSON-680116525-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.958s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 614.807706] env[61806]: ERROR nova.compute.manager [None req-9101cb9e-b191-4241-8bc1-9b126210ca50 tempest-ListImageFiltersTestJSON-680116525 tempest-ListImageFiltersTestJSON-680116525-project-member] [instance: 7e730579-8d0c-4226-86bc-292db16c49ee] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 24a71d6d-5a30-4980-900b-c78cb8631d22, please check neutron logs for more information. [ 614.807706] env[61806]: ERROR nova.compute.manager [instance: 7e730579-8d0c-4226-86bc-292db16c49ee] Traceback (most recent call last): [ 614.807706] env[61806]: ERROR nova.compute.manager [instance: 7e730579-8d0c-4226-86bc-292db16c49ee] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 614.807706] env[61806]: ERROR nova.compute.manager [instance: 7e730579-8d0c-4226-86bc-292db16c49ee] self.driver.spawn(context, instance, image_meta, [ 614.807706] env[61806]: ERROR nova.compute.manager [instance: 7e730579-8d0c-4226-86bc-292db16c49ee] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 614.807706] env[61806]: ERROR nova.compute.manager [instance: 7e730579-8d0c-4226-86bc-292db16c49ee] self._vmops.spawn(context, instance, image_meta, injected_files, [ 614.807706] env[61806]: ERROR nova.compute.manager [instance: 7e730579-8d0c-4226-86bc-292db16c49ee] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 614.807706] env[61806]: ERROR nova.compute.manager [instance: 7e730579-8d0c-4226-86bc-292db16c49ee] vm_ref = self.build_virtual_machine(instance, [ 614.807706] env[61806]: ERROR nova.compute.manager [instance: 7e730579-8d0c-4226-86bc-292db16c49ee] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 614.807706] env[61806]: ERROR nova.compute.manager [instance: 7e730579-8d0c-4226-86bc-292db16c49ee] vif_infos = vmwarevif.get_vif_info(self._session, [ 614.807706] env[61806]: ERROR nova.compute.manager [instance: 7e730579-8d0c-4226-86bc-292db16c49ee] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 614.807971] env[61806]: ERROR nova.compute.manager [instance: 7e730579-8d0c-4226-86bc-292db16c49ee] for vif in network_info: [ 614.807971] env[61806]: ERROR nova.compute.manager [instance: 7e730579-8d0c-4226-86bc-292db16c49ee] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 614.807971] env[61806]: ERROR nova.compute.manager [instance: 7e730579-8d0c-4226-86bc-292db16c49ee] return self._sync_wrapper(fn, *args, **kwargs) [ 614.807971] env[61806]: ERROR nova.compute.manager [instance: 7e730579-8d0c-4226-86bc-292db16c49ee] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 614.807971] env[61806]: ERROR nova.compute.manager [instance: 7e730579-8d0c-4226-86bc-292db16c49ee] self.wait() [ 614.807971] env[61806]: ERROR nova.compute.manager [instance: 7e730579-8d0c-4226-86bc-292db16c49ee] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 614.807971] env[61806]: ERROR nova.compute.manager [instance: 7e730579-8d0c-4226-86bc-292db16c49ee] self[:] = self._gt.wait() [ 614.807971] env[61806]: ERROR nova.compute.manager [instance: 7e730579-8d0c-4226-86bc-292db16c49ee] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 614.807971] env[61806]: ERROR nova.compute.manager [instance: 7e730579-8d0c-4226-86bc-292db16c49ee] return self._exit_event.wait() [ 614.807971] env[61806]: ERROR nova.compute.manager [instance: 7e730579-8d0c-4226-86bc-292db16c49ee] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 614.807971] env[61806]: ERROR nova.compute.manager [instance: 7e730579-8d0c-4226-86bc-292db16c49ee] current.throw(*self._exc) [ 614.807971] env[61806]: ERROR nova.compute.manager [instance: 7e730579-8d0c-4226-86bc-292db16c49ee] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 614.807971] env[61806]: ERROR nova.compute.manager [instance: 7e730579-8d0c-4226-86bc-292db16c49ee] result = function(*args, **kwargs) [ 614.808255] env[61806]: ERROR nova.compute.manager [instance: 7e730579-8d0c-4226-86bc-292db16c49ee] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 614.808255] env[61806]: ERROR nova.compute.manager [instance: 7e730579-8d0c-4226-86bc-292db16c49ee] return func(*args, **kwargs) [ 614.808255] env[61806]: ERROR nova.compute.manager [instance: 7e730579-8d0c-4226-86bc-292db16c49ee] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 614.808255] env[61806]: ERROR nova.compute.manager [instance: 7e730579-8d0c-4226-86bc-292db16c49ee] raise e [ 614.808255] env[61806]: ERROR nova.compute.manager [instance: 7e730579-8d0c-4226-86bc-292db16c49ee] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 614.808255] env[61806]: ERROR nova.compute.manager [instance: 7e730579-8d0c-4226-86bc-292db16c49ee] nwinfo = self.network_api.allocate_for_instance( [ 614.808255] env[61806]: ERROR nova.compute.manager [instance: 7e730579-8d0c-4226-86bc-292db16c49ee] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 614.808255] env[61806]: ERROR nova.compute.manager [instance: 7e730579-8d0c-4226-86bc-292db16c49ee] created_port_ids = self._update_ports_for_instance( [ 614.808255] env[61806]: ERROR nova.compute.manager [instance: 7e730579-8d0c-4226-86bc-292db16c49ee] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 614.808255] env[61806]: ERROR nova.compute.manager [instance: 7e730579-8d0c-4226-86bc-292db16c49ee] with excutils.save_and_reraise_exception(): [ 614.808255] env[61806]: ERROR nova.compute.manager [instance: 7e730579-8d0c-4226-86bc-292db16c49ee] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 614.808255] env[61806]: ERROR nova.compute.manager [instance: 7e730579-8d0c-4226-86bc-292db16c49ee] self.force_reraise() [ 614.808255] env[61806]: ERROR nova.compute.manager [instance: 7e730579-8d0c-4226-86bc-292db16c49ee] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 614.808541] env[61806]: ERROR nova.compute.manager [instance: 7e730579-8d0c-4226-86bc-292db16c49ee] raise self.value [ 614.808541] env[61806]: ERROR nova.compute.manager [instance: 7e730579-8d0c-4226-86bc-292db16c49ee] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 614.808541] env[61806]: ERROR nova.compute.manager [instance: 7e730579-8d0c-4226-86bc-292db16c49ee] updated_port = self._update_port( [ 614.808541] env[61806]: ERROR nova.compute.manager [instance: 7e730579-8d0c-4226-86bc-292db16c49ee] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 614.808541] env[61806]: ERROR nova.compute.manager [instance: 7e730579-8d0c-4226-86bc-292db16c49ee] _ensure_no_port_binding_failure(port) [ 614.808541] env[61806]: ERROR nova.compute.manager [instance: 7e730579-8d0c-4226-86bc-292db16c49ee] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 614.808541] env[61806]: ERROR nova.compute.manager [instance: 7e730579-8d0c-4226-86bc-292db16c49ee] raise exception.PortBindingFailed(port_id=port['id']) [ 614.808541] env[61806]: ERROR nova.compute.manager [instance: 7e730579-8d0c-4226-86bc-292db16c49ee] nova.exception.PortBindingFailed: Binding failed for port 24a71d6d-5a30-4980-900b-c78cb8631d22, please check neutron logs for more information. [ 614.808541] env[61806]: ERROR nova.compute.manager [instance: 7e730579-8d0c-4226-86bc-292db16c49ee] [ 614.808738] env[61806]: DEBUG nova.compute.utils [None req-9101cb9e-b191-4241-8bc1-9b126210ca50 tempest-ListImageFiltersTestJSON-680116525 tempest-ListImageFiltersTestJSON-680116525-project-member] [instance: 7e730579-8d0c-4226-86bc-292db16c49ee] Binding failed for port 24a71d6d-5a30-4980-900b-c78cb8631d22, please check neutron logs for more information. {{(pid=61806) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 614.810686] env[61806]: DEBUG oslo_concurrency.lockutils [None req-962d5603-e71b-4a26-895d-6dce1320b70e tempest-ListImageFiltersTestJSON-680116525 tempest-ListImageFiltersTestJSON-680116525-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 15.784s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 614.813792] env[61806]: DEBUG nova.compute.manager [None req-9101cb9e-b191-4241-8bc1-9b126210ca50 tempest-ListImageFiltersTestJSON-680116525 tempest-ListImageFiltersTestJSON-680116525-project-member] [instance: 7e730579-8d0c-4226-86bc-292db16c49ee] Build of instance 7e730579-8d0c-4226-86bc-292db16c49ee was re-scheduled: Binding failed for port 24a71d6d-5a30-4980-900b-c78cb8631d22, please check neutron logs for more information. {{(pid=61806) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 614.814863] env[61806]: DEBUG nova.compute.manager [None req-9101cb9e-b191-4241-8bc1-9b126210ca50 tempest-ListImageFiltersTestJSON-680116525 tempest-ListImageFiltersTestJSON-680116525-project-member] [instance: 7e730579-8d0c-4226-86bc-292db16c49ee] Unplugging VIFs for instance {{(pid=61806) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 614.815172] env[61806]: DEBUG oslo_concurrency.lockutils [None req-9101cb9e-b191-4241-8bc1-9b126210ca50 tempest-ListImageFiltersTestJSON-680116525 tempest-ListImageFiltersTestJSON-680116525-project-member] Acquiring lock "refresh_cache-7e730579-8d0c-4226-86bc-292db16c49ee" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 614.815337] env[61806]: DEBUG oslo_concurrency.lockutils [None req-9101cb9e-b191-4241-8bc1-9b126210ca50 tempest-ListImageFiltersTestJSON-680116525 tempest-ListImageFiltersTestJSON-680116525-project-member] Acquired lock "refresh_cache-7e730579-8d0c-4226-86bc-292db16c49ee" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 614.815503] env[61806]: DEBUG nova.network.neutron [None req-9101cb9e-b191-4241-8bc1-9b126210ca50 tempest-ListImageFiltersTestJSON-680116525 tempest-ListImageFiltersTestJSON-680116525-project-member] [instance: 7e730579-8d0c-4226-86bc-292db16c49ee] Building network info cache for instance {{(pid=61806) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 614.872249] env[61806]: DEBUG nova.compute.manager [None req-e34d2bcf-266f-4faa-8f31-6bd93f6ab176 tempest-ImagesTestJSON-618852556 tempest-ImagesTestJSON-618852556-project-member] [instance: 65f3b2f3-6d25-44f0-95cb-f9a7e75f8e82] Start spawning the instance on the hypervisor. {{(pid=61806) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 614.901039] env[61806]: DEBUG nova.virt.hardware [None req-e34d2bcf-266f-4faa-8f31-6bd93f6ab176 tempest-ImagesTestJSON-618852556 tempest-ImagesTestJSON-618852556-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-15T18:05:10Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-15T18:04:52Z,direct_url=,disk_format='vmdk',id=ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='84ba83fce2c349a6988173c5d6fc3b07',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-15T18:04:53Z,virtual_size=,visibility=), allow threads: False {{(pid=61806) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 614.901331] env[61806]: DEBUG nova.virt.hardware [None req-e34d2bcf-266f-4faa-8f31-6bd93f6ab176 tempest-ImagesTestJSON-618852556 tempest-ImagesTestJSON-618852556-project-member] Flavor limits 0:0:0 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 614.901551] env[61806]: DEBUG nova.virt.hardware [None req-e34d2bcf-266f-4faa-8f31-6bd93f6ab176 tempest-ImagesTestJSON-618852556 tempest-ImagesTestJSON-618852556-project-member] Image limits 0:0:0 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 614.901732] env[61806]: DEBUG nova.virt.hardware [None req-e34d2bcf-266f-4faa-8f31-6bd93f6ab176 tempest-ImagesTestJSON-618852556 tempest-ImagesTestJSON-618852556-project-member] Flavor pref 0:0:0 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 614.901809] env[61806]: DEBUG nova.virt.hardware [None req-e34d2bcf-266f-4faa-8f31-6bd93f6ab176 tempest-ImagesTestJSON-618852556 tempest-ImagesTestJSON-618852556-project-member] Image pref 0:0:0 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 614.901954] env[61806]: DEBUG nova.virt.hardware [None req-e34d2bcf-266f-4faa-8f31-6bd93f6ab176 tempest-ImagesTestJSON-618852556 tempest-ImagesTestJSON-618852556-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 614.902185] env[61806]: DEBUG nova.virt.hardware [None req-e34d2bcf-266f-4faa-8f31-6bd93f6ab176 tempest-ImagesTestJSON-618852556 tempest-ImagesTestJSON-618852556-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61806) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 614.902361] env[61806]: DEBUG nova.virt.hardware [None req-e34d2bcf-266f-4faa-8f31-6bd93f6ab176 tempest-ImagesTestJSON-618852556 tempest-ImagesTestJSON-618852556-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61806) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 614.902513] env[61806]: DEBUG nova.virt.hardware [None req-e34d2bcf-266f-4faa-8f31-6bd93f6ab176 tempest-ImagesTestJSON-618852556 tempest-ImagesTestJSON-618852556-project-member] Got 1 possible topologies {{(pid=61806) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 614.902673] env[61806]: DEBUG nova.virt.hardware [None req-e34d2bcf-266f-4faa-8f31-6bd93f6ab176 tempest-ImagesTestJSON-618852556 tempest-ImagesTestJSON-618852556-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61806) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 614.902881] env[61806]: DEBUG nova.virt.hardware [None req-e34d2bcf-266f-4faa-8f31-6bd93f6ab176 tempest-ImagesTestJSON-618852556 tempest-ImagesTestJSON-618852556-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61806) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 614.903754] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-686b511b-7fe7-48ce-8301-d22351f0e108 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 614.911622] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8946f910-cb39-4b7d-8bb6-ee4e39e15619 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 615.109230] env[61806]: DEBUG oslo_concurrency.lockutils [None req-f8b2f38e-90f5-4b03-89bc-ca1926adb40c tempest-FloatingIPsAssociationTestJSON-35718468 tempest-FloatingIPsAssociationTestJSON-35718468-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 615.171194] env[61806]: ERROR nova.compute.manager [None req-e34d2bcf-266f-4faa-8f31-6bd93f6ab176 tempest-ImagesTestJSON-618852556 tempest-ImagesTestJSON-618852556-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port ddedeb2e-1da8-4774-925f-5f0e09c0cae8, please check neutron logs for more information. [ 615.171194] env[61806]: ERROR nova.compute.manager Traceback (most recent call last): [ 615.171194] env[61806]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 615.171194] env[61806]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 615.171194] env[61806]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 615.171194] env[61806]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 615.171194] env[61806]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 615.171194] env[61806]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 615.171194] env[61806]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 615.171194] env[61806]: ERROR nova.compute.manager self.force_reraise() [ 615.171194] env[61806]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 615.171194] env[61806]: ERROR nova.compute.manager raise self.value [ 615.171194] env[61806]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 615.171194] env[61806]: ERROR nova.compute.manager updated_port = self._update_port( [ 615.171194] env[61806]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 615.171194] env[61806]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 615.171599] env[61806]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 615.171599] env[61806]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 615.171599] env[61806]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port ddedeb2e-1da8-4774-925f-5f0e09c0cae8, please check neutron logs for more information. [ 615.171599] env[61806]: ERROR nova.compute.manager [ 615.171599] env[61806]: Traceback (most recent call last): [ 615.171599] env[61806]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 615.171599] env[61806]: listener.cb(fileno) [ 615.171599] env[61806]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 615.171599] env[61806]: result = function(*args, **kwargs) [ 615.171599] env[61806]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 615.171599] env[61806]: return func(*args, **kwargs) [ 615.171599] env[61806]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 615.171599] env[61806]: raise e [ 615.171599] env[61806]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 615.171599] env[61806]: nwinfo = self.network_api.allocate_for_instance( [ 615.171599] env[61806]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 615.171599] env[61806]: created_port_ids = self._update_ports_for_instance( [ 615.171599] env[61806]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 615.171599] env[61806]: with excutils.save_and_reraise_exception(): [ 615.171599] env[61806]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 615.171599] env[61806]: self.force_reraise() [ 615.171599] env[61806]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 615.171599] env[61806]: raise self.value [ 615.171599] env[61806]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 615.171599] env[61806]: updated_port = self._update_port( [ 615.171599] env[61806]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 615.171599] env[61806]: _ensure_no_port_binding_failure(port) [ 615.171599] env[61806]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 615.171599] env[61806]: raise exception.PortBindingFailed(port_id=port['id']) [ 615.172214] env[61806]: nova.exception.PortBindingFailed: Binding failed for port ddedeb2e-1da8-4774-925f-5f0e09c0cae8, please check neutron logs for more information. [ 615.172214] env[61806]: Removing descriptor: 16 [ 615.172214] env[61806]: ERROR nova.compute.manager [None req-e34d2bcf-266f-4faa-8f31-6bd93f6ab176 tempest-ImagesTestJSON-618852556 tempest-ImagesTestJSON-618852556-project-member] [instance: 65f3b2f3-6d25-44f0-95cb-f9a7e75f8e82] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port ddedeb2e-1da8-4774-925f-5f0e09c0cae8, please check neutron logs for more information. [ 615.172214] env[61806]: ERROR nova.compute.manager [instance: 65f3b2f3-6d25-44f0-95cb-f9a7e75f8e82] Traceback (most recent call last): [ 615.172214] env[61806]: ERROR nova.compute.manager [instance: 65f3b2f3-6d25-44f0-95cb-f9a7e75f8e82] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 615.172214] env[61806]: ERROR nova.compute.manager [instance: 65f3b2f3-6d25-44f0-95cb-f9a7e75f8e82] yield resources [ 615.172214] env[61806]: ERROR nova.compute.manager [instance: 65f3b2f3-6d25-44f0-95cb-f9a7e75f8e82] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 615.172214] env[61806]: ERROR nova.compute.manager [instance: 65f3b2f3-6d25-44f0-95cb-f9a7e75f8e82] self.driver.spawn(context, instance, image_meta, [ 615.172214] env[61806]: ERROR nova.compute.manager [instance: 65f3b2f3-6d25-44f0-95cb-f9a7e75f8e82] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 615.172214] env[61806]: ERROR nova.compute.manager [instance: 65f3b2f3-6d25-44f0-95cb-f9a7e75f8e82] self._vmops.spawn(context, instance, image_meta, injected_files, [ 615.172214] env[61806]: ERROR nova.compute.manager [instance: 65f3b2f3-6d25-44f0-95cb-f9a7e75f8e82] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 615.172214] env[61806]: ERROR nova.compute.manager [instance: 65f3b2f3-6d25-44f0-95cb-f9a7e75f8e82] vm_ref = self.build_virtual_machine(instance, [ 615.172586] env[61806]: ERROR nova.compute.manager [instance: 65f3b2f3-6d25-44f0-95cb-f9a7e75f8e82] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 615.172586] env[61806]: ERROR nova.compute.manager [instance: 65f3b2f3-6d25-44f0-95cb-f9a7e75f8e82] vif_infos = vmwarevif.get_vif_info(self._session, [ 615.172586] env[61806]: ERROR nova.compute.manager [instance: 65f3b2f3-6d25-44f0-95cb-f9a7e75f8e82] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 615.172586] env[61806]: ERROR nova.compute.manager [instance: 65f3b2f3-6d25-44f0-95cb-f9a7e75f8e82] for vif in network_info: [ 615.172586] env[61806]: ERROR nova.compute.manager [instance: 65f3b2f3-6d25-44f0-95cb-f9a7e75f8e82] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 615.172586] env[61806]: ERROR nova.compute.manager [instance: 65f3b2f3-6d25-44f0-95cb-f9a7e75f8e82] return self._sync_wrapper(fn, *args, **kwargs) [ 615.172586] env[61806]: ERROR nova.compute.manager [instance: 65f3b2f3-6d25-44f0-95cb-f9a7e75f8e82] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 615.172586] env[61806]: ERROR nova.compute.manager [instance: 65f3b2f3-6d25-44f0-95cb-f9a7e75f8e82] self.wait() [ 615.172586] env[61806]: ERROR nova.compute.manager [instance: 65f3b2f3-6d25-44f0-95cb-f9a7e75f8e82] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 615.172586] env[61806]: ERROR nova.compute.manager [instance: 65f3b2f3-6d25-44f0-95cb-f9a7e75f8e82] self[:] = self._gt.wait() [ 615.172586] env[61806]: ERROR nova.compute.manager [instance: 65f3b2f3-6d25-44f0-95cb-f9a7e75f8e82] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 615.172586] env[61806]: ERROR nova.compute.manager [instance: 65f3b2f3-6d25-44f0-95cb-f9a7e75f8e82] return self._exit_event.wait() [ 615.172586] env[61806]: ERROR nova.compute.manager [instance: 65f3b2f3-6d25-44f0-95cb-f9a7e75f8e82] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 615.172923] env[61806]: ERROR nova.compute.manager [instance: 65f3b2f3-6d25-44f0-95cb-f9a7e75f8e82] result = hub.switch() [ 615.172923] env[61806]: ERROR nova.compute.manager [instance: 65f3b2f3-6d25-44f0-95cb-f9a7e75f8e82] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 615.172923] env[61806]: ERROR nova.compute.manager [instance: 65f3b2f3-6d25-44f0-95cb-f9a7e75f8e82] return self.greenlet.switch() [ 615.172923] env[61806]: ERROR nova.compute.manager [instance: 65f3b2f3-6d25-44f0-95cb-f9a7e75f8e82] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 615.172923] env[61806]: ERROR nova.compute.manager [instance: 65f3b2f3-6d25-44f0-95cb-f9a7e75f8e82] result = function(*args, **kwargs) [ 615.172923] env[61806]: ERROR nova.compute.manager [instance: 65f3b2f3-6d25-44f0-95cb-f9a7e75f8e82] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 615.172923] env[61806]: ERROR nova.compute.manager [instance: 65f3b2f3-6d25-44f0-95cb-f9a7e75f8e82] return func(*args, **kwargs) [ 615.172923] env[61806]: ERROR nova.compute.manager [instance: 65f3b2f3-6d25-44f0-95cb-f9a7e75f8e82] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 615.172923] env[61806]: ERROR nova.compute.manager [instance: 65f3b2f3-6d25-44f0-95cb-f9a7e75f8e82] raise e [ 615.172923] env[61806]: ERROR nova.compute.manager [instance: 65f3b2f3-6d25-44f0-95cb-f9a7e75f8e82] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 615.172923] env[61806]: ERROR nova.compute.manager [instance: 65f3b2f3-6d25-44f0-95cb-f9a7e75f8e82] nwinfo = self.network_api.allocate_for_instance( [ 615.172923] env[61806]: ERROR nova.compute.manager [instance: 65f3b2f3-6d25-44f0-95cb-f9a7e75f8e82] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 615.172923] env[61806]: ERROR nova.compute.manager [instance: 65f3b2f3-6d25-44f0-95cb-f9a7e75f8e82] created_port_ids = self._update_ports_for_instance( [ 615.173264] env[61806]: ERROR nova.compute.manager [instance: 65f3b2f3-6d25-44f0-95cb-f9a7e75f8e82] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 615.173264] env[61806]: ERROR nova.compute.manager [instance: 65f3b2f3-6d25-44f0-95cb-f9a7e75f8e82] with excutils.save_and_reraise_exception(): [ 615.173264] env[61806]: ERROR nova.compute.manager [instance: 65f3b2f3-6d25-44f0-95cb-f9a7e75f8e82] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 615.173264] env[61806]: ERROR nova.compute.manager [instance: 65f3b2f3-6d25-44f0-95cb-f9a7e75f8e82] self.force_reraise() [ 615.173264] env[61806]: ERROR nova.compute.manager [instance: 65f3b2f3-6d25-44f0-95cb-f9a7e75f8e82] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 615.173264] env[61806]: ERROR nova.compute.manager [instance: 65f3b2f3-6d25-44f0-95cb-f9a7e75f8e82] raise self.value [ 615.173264] env[61806]: ERROR nova.compute.manager [instance: 65f3b2f3-6d25-44f0-95cb-f9a7e75f8e82] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 615.173264] env[61806]: ERROR nova.compute.manager [instance: 65f3b2f3-6d25-44f0-95cb-f9a7e75f8e82] updated_port = self._update_port( [ 615.173264] env[61806]: ERROR nova.compute.manager [instance: 65f3b2f3-6d25-44f0-95cb-f9a7e75f8e82] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 615.173264] env[61806]: ERROR nova.compute.manager [instance: 65f3b2f3-6d25-44f0-95cb-f9a7e75f8e82] _ensure_no_port_binding_failure(port) [ 615.173264] env[61806]: ERROR nova.compute.manager [instance: 65f3b2f3-6d25-44f0-95cb-f9a7e75f8e82] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 615.173264] env[61806]: ERROR nova.compute.manager [instance: 65f3b2f3-6d25-44f0-95cb-f9a7e75f8e82] raise exception.PortBindingFailed(port_id=port['id']) [ 615.173574] env[61806]: ERROR nova.compute.manager [instance: 65f3b2f3-6d25-44f0-95cb-f9a7e75f8e82] nova.exception.PortBindingFailed: Binding failed for port ddedeb2e-1da8-4774-925f-5f0e09c0cae8, please check neutron logs for more information. [ 615.173574] env[61806]: ERROR nova.compute.manager [instance: 65f3b2f3-6d25-44f0-95cb-f9a7e75f8e82] [ 615.173574] env[61806]: INFO nova.compute.manager [None req-e34d2bcf-266f-4faa-8f31-6bd93f6ab176 tempest-ImagesTestJSON-618852556 tempest-ImagesTestJSON-618852556-project-member] [instance: 65f3b2f3-6d25-44f0-95cb-f9a7e75f8e82] Terminating instance [ 615.174544] env[61806]: DEBUG oslo_concurrency.lockutils [None req-e34d2bcf-266f-4faa-8f31-6bd93f6ab176 tempest-ImagesTestJSON-618852556 tempest-ImagesTestJSON-618852556-project-member] Acquiring lock "refresh_cache-65f3b2f3-6d25-44f0-95cb-f9a7e75f8e82" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 615.174703] env[61806]: DEBUG oslo_concurrency.lockutils [None req-e34d2bcf-266f-4faa-8f31-6bd93f6ab176 tempest-ImagesTestJSON-618852556 tempest-ImagesTestJSON-618852556-project-member] Acquired lock "refresh_cache-65f3b2f3-6d25-44f0-95cb-f9a7e75f8e82" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 615.174869] env[61806]: DEBUG nova.network.neutron [None req-e34d2bcf-266f-4faa-8f31-6bd93f6ab176 tempest-ImagesTestJSON-618852556 tempest-ImagesTestJSON-618852556-project-member] [instance: 65f3b2f3-6d25-44f0-95cb-f9a7e75f8e82] Building network info cache for instance {{(pid=61806) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 615.214215] env[61806]: DEBUG nova.compute.manager [req-7feed7ee-b470-45d3-b2a6-c1bf40c7e0b9 req-15927e58-6983-48e4-a9f6-6f3f6b0c4f9c service nova] [instance: 65f3b2f3-6d25-44f0-95cb-f9a7e75f8e82] Received event network-changed-ddedeb2e-1da8-4774-925f-5f0e09c0cae8 {{(pid=61806) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 615.214478] env[61806]: DEBUG nova.compute.manager [req-7feed7ee-b470-45d3-b2a6-c1bf40c7e0b9 req-15927e58-6983-48e4-a9f6-6f3f6b0c4f9c service nova] [instance: 65f3b2f3-6d25-44f0-95cb-f9a7e75f8e82] Refreshing instance network info cache due to event network-changed-ddedeb2e-1da8-4774-925f-5f0e09c0cae8. {{(pid=61806) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 615.214734] env[61806]: DEBUG oslo_concurrency.lockutils [req-7feed7ee-b470-45d3-b2a6-c1bf40c7e0b9 req-15927e58-6983-48e4-a9f6-6f3f6b0c4f9c service nova] Acquiring lock "refresh_cache-65f3b2f3-6d25-44f0-95cb-f9a7e75f8e82" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 615.340428] env[61806]: DEBUG nova.network.neutron [None req-9101cb9e-b191-4241-8bc1-9b126210ca50 tempest-ListImageFiltersTestJSON-680116525 tempest-ListImageFiltersTestJSON-680116525-project-member] [instance: 7e730579-8d0c-4226-86bc-292db16c49ee] Instance cache missing network info. {{(pid=61806) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 615.444935] env[61806]: DEBUG nova.network.neutron [None req-9101cb9e-b191-4241-8bc1-9b126210ca50 tempest-ListImageFiltersTestJSON-680116525 tempest-ListImageFiltersTestJSON-680116525-project-member] [instance: 7e730579-8d0c-4226-86bc-292db16c49ee] Updating instance_info_cache with network_info: [] {{(pid=61806) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 615.702632] env[61806]: DEBUG nova.network.neutron [None req-e34d2bcf-266f-4faa-8f31-6bd93f6ab176 tempest-ImagesTestJSON-618852556 tempest-ImagesTestJSON-618852556-project-member] [instance: 65f3b2f3-6d25-44f0-95cb-f9a7e75f8e82] Instance cache missing network info. {{(pid=61806) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 615.745342] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2c9144e4-d991-46eb-88d0-d1f1c1a62e22 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 615.753223] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-19a2a1b9-9a06-4984-8382-9b251282674c {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 615.786692] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bd71c0d7-1f01-4198-af5b-2539148610f5 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 615.795405] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-af35b609-fd4f-416b-a328-0317ade2d63d {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 615.814098] env[61806]: DEBUG nova.compute.provider_tree [None req-962d5603-e71b-4a26-895d-6dce1320b70e tempest-ListImageFiltersTestJSON-680116525 tempest-ListImageFiltersTestJSON-680116525-project-member] Inventory has not changed in ProviderTree for provider: a2858be1-fd22-4e08-979e-87ad25293407 {{(pid=61806) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 615.866188] env[61806]: DEBUG nova.network.neutron [None req-e34d2bcf-266f-4faa-8f31-6bd93f6ab176 tempest-ImagesTestJSON-618852556 tempest-ImagesTestJSON-618852556-project-member] [instance: 65f3b2f3-6d25-44f0-95cb-f9a7e75f8e82] Updating instance_info_cache with network_info: [] {{(pid=61806) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 615.948196] env[61806]: DEBUG oslo_concurrency.lockutils [None req-9101cb9e-b191-4241-8bc1-9b126210ca50 tempest-ListImageFiltersTestJSON-680116525 tempest-ListImageFiltersTestJSON-680116525-project-member] Releasing lock "refresh_cache-7e730579-8d0c-4226-86bc-292db16c49ee" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 615.948504] env[61806]: DEBUG nova.compute.manager [None req-9101cb9e-b191-4241-8bc1-9b126210ca50 tempest-ListImageFiltersTestJSON-680116525 tempest-ListImageFiltersTestJSON-680116525-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61806) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 615.948630] env[61806]: DEBUG nova.compute.manager [None req-9101cb9e-b191-4241-8bc1-9b126210ca50 tempest-ListImageFiltersTestJSON-680116525 tempest-ListImageFiltersTestJSON-680116525-project-member] [instance: 7e730579-8d0c-4226-86bc-292db16c49ee] Deallocating network for instance {{(pid=61806) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 615.948840] env[61806]: DEBUG nova.network.neutron [None req-9101cb9e-b191-4241-8bc1-9b126210ca50 tempest-ListImageFiltersTestJSON-680116525 tempest-ListImageFiltersTestJSON-680116525-project-member] [instance: 7e730579-8d0c-4226-86bc-292db16c49ee] deallocate_for_instance() {{(pid=61806) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 615.974054] env[61806]: DEBUG nova.network.neutron [None req-9101cb9e-b191-4241-8bc1-9b126210ca50 tempest-ListImageFiltersTestJSON-680116525 tempest-ListImageFiltersTestJSON-680116525-project-member] [instance: 7e730579-8d0c-4226-86bc-292db16c49ee] Instance cache missing network info. {{(pid=61806) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 616.314550] env[61806]: DEBUG nova.scheduler.client.report [None req-962d5603-e71b-4a26-895d-6dce1320b70e tempest-ListImageFiltersTestJSON-680116525 tempest-ListImageFiltersTestJSON-680116525-project-member] Inventory has not changed for provider a2858be1-fd22-4e08-979e-87ad25293407 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 139, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61806) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 616.368501] env[61806]: DEBUG oslo_concurrency.lockutils [None req-e34d2bcf-266f-4faa-8f31-6bd93f6ab176 tempest-ImagesTestJSON-618852556 tempest-ImagesTestJSON-618852556-project-member] Releasing lock "refresh_cache-65f3b2f3-6d25-44f0-95cb-f9a7e75f8e82" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 616.368899] env[61806]: DEBUG nova.compute.manager [None req-e34d2bcf-266f-4faa-8f31-6bd93f6ab176 tempest-ImagesTestJSON-618852556 tempest-ImagesTestJSON-618852556-project-member] [instance: 65f3b2f3-6d25-44f0-95cb-f9a7e75f8e82] Start destroying the instance on the hypervisor. {{(pid=61806) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 616.369117] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-e34d2bcf-266f-4faa-8f31-6bd93f6ab176 tempest-ImagesTestJSON-618852556 tempest-ImagesTestJSON-618852556-project-member] [instance: 65f3b2f3-6d25-44f0-95cb-f9a7e75f8e82] Destroying instance {{(pid=61806) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 616.369671] env[61806]: DEBUG oslo_concurrency.lockutils [req-7feed7ee-b470-45d3-b2a6-c1bf40c7e0b9 req-15927e58-6983-48e4-a9f6-6f3f6b0c4f9c service nova] Acquired lock "refresh_cache-65f3b2f3-6d25-44f0-95cb-f9a7e75f8e82" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 616.370829] env[61806]: DEBUG nova.network.neutron [req-7feed7ee-b470-45d3-b2a6-c1bf40c7e0b9 req-15927e58-6983-48e4-a9f6-6f3f6b0c4f9c service nova] [instance: 65f3b2f3-6d25-44f0-95cb-f9a7e75f8e82] Refreshing network info cache for port ddedeb2e-1da8-4774-925f-5f0e09c0cae8 {{(pid=61806) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 616.374373] env[61806]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-951311c9-8ddc-46db-8076-edb652d2a08b {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 616.382734] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e92d9184-7328-4db1-b0cf-91bcba213360 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 616.407118] env[61806]: WARNING nova.virt.vmwareapi.vmops [None req-e34d2bcf-266f-4faa-8f31-6bd93f6ab176 tempest-ImagesTestJSON-618852556 tempest-ImagesTestJSON-618852556-project-member] [instance: 65f3b2f3-6d25-44f0-95cb-f9a7e75f8e82] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 65f3b2f3-6d25-44f0-95cb-f9a7e75f8e82 could not be found. [ 616.407387] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-e34d2bcf-266f-4faa-8f31-6bd93f6ab176 tempest-ImagesTestJSON-618852556 tempest-ImagesTestJSON-618852556-project-member] [instance: 65f3b2f3-6d25-44f0-95cb-f9a7e75f8e82] Instance destroyed {{(pid=61806) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 616.407589] env[61806]: INFO nova.compute.manager [None req-e34d2bcf-266f-4faa-8f31-6bd93f6ab176 tempest-ImagesTestJSON-618852556 tempest-ImagesTestJSON-618852556-project-member] [instance: 65f3b2f3-6d25-44f0-95cb-f9a7e75f8e82] Took 0.04 seconds to destroy the instance on the hypervisor. [ 616.407849] env[61806]: DEBUG oslo.service.loopingcall [None req-e34d2bcf-266f-4faa-8f31-6bd93f6ab176 tempest-ImagesTestJSON-618852556 tempest-ImagesTestJSON-618852556-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61806) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 616.408413] env[61806]: DEBUG nova.compute.manager [-] [instance: 65f3b2f3-6d25-44f0-95cb-f9a7e75f8e82] Deallocating network for instance {{(pid=61806) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 616.408541] env[61806]: DEBUG nova.network.neutron [-] [instance: 65f3b2f3-6d25-44f0-95cb-f9a7e75f8e82] deallocate_for_instance() {{(pid=61806) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 616.440621] env[61806]: DEBUG nova.network.neutron [-] [instance: 65f3b2f3-6d25-44f0-95cb-f9a7e75f8e82] Instance cache missing network info. {{(pid=61806) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 616.476302] env[61806]: DEBUG nova.network.neutron [None req-9101cb9e-b191-4241-8bc1-9b126210ca50 tempest-ListImageFiltersTestJSON-680116525 tempest-ListImageFiltersTestJSON-680116525-project-member] [instance: 7e730579-8d0c-4226-86bc-292db16c49ee] Updating instance_info_cache with network_info: [] {{(pid=61806) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 616.820118] env[61806]: DEBUG oslo_concurrency.lockutils [None req-962d5603-e71b-4a26-895d-6dce1320b70e tempest-ListImageFiltersTestJSON-680116525 tempest-ListImageFiltersTestJSON-680116525-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.009s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 616.820888] env[61806]: ERROR nova.compute.manager [None req-962d5603-e71b-4a26-895d-6dce1320b70e tempest-ListImageFiltersTestJSON-680116525 tempest-ListImageFiltersTestJSON-680116525-project-member] [instance: 53d3e2f2-f63d-45c7-a5d1-1fa9a9f0ac0f] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port f650d966-513e-4686-9faa-9dd139c8c888, please check neutron logs for more information. [ 616.820888] env[61806]: ERROR nova.compute.manager [instance: 53d3e2f2-f63d-45c7-a5d1-1fa9a9f0ac0f] Traceback (most recent call last): [ 616.820888] env[61806]: ERROR nova.compute.manager [instance: 53d3e2f2-f63d-45c7-a5d1-1fa9a9f0ac0f] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 616.820888] env[61806]: ERROR nova.compute.manager [instance: 53d3e2f2-f63d-45c7-a5d1-1fa9a9f0ac0f] self.driver.spawn(context, instance, image_meta, [ 616.820888] env[61806]: ERROR nova.compute.manager [instance: 53d3e2f2-f63d-45c7-a5d1-1fa9a9f0ac0f] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 616.820888] env[61806]: ERROR nova.compute.manager [instance: 53d3e2f2-f63d-45c7-a5d1-1fa9a9f0ac0f] self._vmops.spawn(context, instance, image_meta, injected_files, [ 616.820888] env[61806]: ERROR nova.compute.manager [instance: 53d3e2f2-f63d-45c7-a5d1-1fa9a9f0ac0f] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 616.820888] env[61806]: ERROR nova.compute.manager [instance: 53d3e2f2-f63d-45c7-a5d1-1fa9a9f0ac0f] vm_ref = self.build_virtual_machine(instance, [ 616.820888] env[61806]: ERROR nova.compute.manager [instance: 53d3e2f2-f63d-45c7-a5d1-1fa9a9f0ac0f] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 616.820888] env[61806]: ERROR nova.compute.manager [instance: 53d3e2f2-f63d-45c7-a5d1-1fa9a9f0ac0f] vif_infos = vmwarevif.get_vif_info(self._session, [ 616.820888] env[61806]: ERROR nova.compute.manager [instance: 53d3e2f2-f63d-45c7-a5d1-1fa9a9f0ac0f] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 616.821192] env[61806]: ERROR nova.compute.manager [instance: 53d3e2f2-f63d-45c7-a5d1-1fa9a9f0ac0f] for vif in network_info: [ 616.821192] env[61806]: ERROR nova.compute.manager [instance: 53d3e2f2-f63d-45c7-a5d1-1fa9a9f0ac0f] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 616.821192] env[61806]: ERROR nova.compute.manager [instance: 53d3e2f2-f63d-45c7-a5d1-1fa9a9f0ac0f] return self._sync_wrapper(fn, *args, **kwargs) [ 616.821192] env[61806]: ERROR nova.compute.manager [instance: 53d3e2f2-f63d-45c7-a5d1-1fa9a9f0ac0f] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 616.821192] env[61806]: ERROR nova.compute.manager [instance: 53d3e2f2-f63d-45c7-a5d1-1fa9a9f0ac0f] self.wait() [ 616.821192] env[61806]: ERROR nova.compute.manager [instance: 53d3e2f2-f63d-45c7-a5d1-1fa9a9f0ac0f] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 616.821192] env[61806]: ERROR nova.compute.manager [instance: 53d3e2f2-f63d-45c7-a5d1-1fa9a9f0ac0f] self[:] = self._gt.wait() [ 616.821192] env[61806]: ERROR nova.compute.manager [instance: 53d3e2f2-f63d-45c7-a5d1-1fa9a9f0ac0f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 616.821192] env[61806]: ERROR nova.compute.manager [instance: 53d3e2f2-f63d-45c7-a5d1-1fa9a9f0ac0f] return self._exit_event.wait() [ 616.821192] env[61806]: ERROR nova.compute.manager [instance: 53d3e2f2-f63d-45c7-a5d1-1fa9a9f0ac0f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 616.821192] env[61806]: ERROR nova.compute.manager [instance: 53d3e2f2-f63d-45c7-a5d1-1fa9a9f0ac0f] result = hub.switch() [ 616.821192] env[61806]: ERROR nova.compute.manager [instance: 53d3e2f2-f63d-45c7-a5d1-1fa9a9f0ac0f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 616.821192] env[61806]: ERROR nova.compute.manager [instance: 53d3e2f2-f63d-45c7-a5d1-1fa9a9f0ac0f] return self.greenlet.switch() [ 616.821539] env[61806]: ERROR nova.compute.manager [instance: 53d3e2f2-f63d-45c7-a5d1-1fa9a9f0ac0f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 616.821539] env[61806]: ERROR nova.compute.manager [instance: 53d3e2f2-f63d-45c7-a5d1-1fa9a9f0ac0f] result = function(*args, **kwargs) [ 616.821539] env[61806]: ERROR nova.compute.manager [instance: 53d3e2f2-f63d-45c7-a5d1-1fa9a9f0ac0f] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 616.821539] env[61806]: ERROR nova.compute.manager [instance: 53d3e2f2-f63d-45c7-a5d1-1fa9a9f0ac0f] return func(*args, **kwargs) [ 616.821539] env[61806]: ERROR nova.compute.manager [instance: 53d3e2f2-f63d-45c7-a5d1-1fa9a9f0ac0f] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 616.821539] env[61806]: ERROR nova.compute.manager [instance: 53d3e2f2-f63d-45c7-a5d1-1fa9a9f0ac0f] raise e [ 616.821539] env[61806]: ERROR nova.compute.manager [instance: 53d3e2f2-f63d-45c7-a5d1-1fa9a9f0ac0f] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 616.821539] env[61806]: ERROR nova.compute.manager [instance: 53d3e2f2-f63d-45c7-a5d1-1fa9a9f0ac0f] nwinfo = self.network_api.allocate_for_instance( [ 616.821539] env[61806]: ERROR nova.compute.manager [instance: 53d3e2f2-f63d-45c7-a5d1-1fa9a9f0ac0f] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 616.821539] env[61806]: ERROR nova.compute.manager [instance: 53d3e2f2-f63d-45c7-a5d1-1fa9a9f0ac0f] created_port_ids = self._update_ports_for_instance( [ 616.821539] env[61806]: ERROR nova.compute.manager [instance: 53d3e2f2-f63d-45c7-a5d1-1fa9a9f0ac0f] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 616.821539] env[61806]: ERROR nova.compute.manager [instance: 53d3e2f2-f63d-45c7-a5d1-1fa9a9f0ac0f] with excutils.save_and_reraise_exception(): [ 616.821539] env[61806]: ERROR nova.compute.manager [instance: 53d3e2f2-f63d-45c7-a5d1-1fa9a9f0ac0f] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 616.821877] env[61806]: ERROR nova.compute.manager [instance: 53d3e2f2-f63d-45c7-a5d1-1fa9a9f0ac0f] self.force_reraise() [ 616.821877] env[61806]: ERROR nova.compute.manager [instance: 53d3e2f2-f63d-45c7-a5d1-1fa9a9f0ac0f] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 616.821877] env[61806]: ERROR nova.compute.manager [instance: 53d3e2f2-f63d-45c7-a5d1-1fa9a9f0ac0f] raise self.value [ 616.821877] env[61806]: ERROR nova.compute.manager [instance: 53d3e2f2-f63d-45c7-a5d1-1fa9a9f0ac0f] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 616.821877] env[61806]: ERROR nova.compute.manager [instance: 53d3e2f2-f63d-45c7-a5d1-1fa9a9f0ac0f] updated_port = self._update_port( [ 616.821877] env[61806]: ERROR nova.compute.manager [instance: 53d3e2f2-f63d-45c7-a5d1-1fa9a9f0ac0f] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 616.821877] env[61806]: ERROR nova.compute.manager [instance: 53d3e2f2-f63d-45c7-a5d1-1fa9a9f0ac0f] _ensure_no_port_binding_failure(port) [ 616.821877] env[61806]: ERROR nova.compute.manager [instance: 53d3e2f2-f63d-45c7-a5d1-1fa9a9f0ac0f] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 616.821877] env[61806]: ERROR nova.compute.manager [instance: 53d3e2f2-f63d-45c7-a5d1-1fa9a9f0ac0f] raise exception.PortBindingFailed(port_id=port['id']) [ 616.821877] env[61806]: ERROR nova.compute.manager [instance: 53d3e2f2-f63d-45c7-a5d1-1fa9a9f0ac0f] nova.exception.PortBindingFailed: Binding failed for port f650d966-513e-4686-9faa-9dd139c8c888, please check neutron logs for more information. [ 616.821877] env[61806]: ERROR nova.compute.manager [instance: 53d3e2f2-f63d-45c7-a5d1-1fa9a9f0ac0f] [ 616.822187] env[61806]: DEBUG nova.compute.utils [None req-962d5603-e71b-4a26-895d-6dce1320b70e tempest-ListImageFiltersTestJSON-680116525 tempest-ListImageFiltersTestJSON-680116525-project-member] [instance: 53d3e2f2-f63d-45c7-a5d1-1fa9a9f0ac0f] Binding failed for port f650d966-513e-4686-9faa-9dd139c8c888, please check neutron logs for more information. {{(pid=61806) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 616.823817] env[61806]: DEBUG nova.compute.manager [None req-962d5603-e71b-4a26-895d-6dce1320b70e tempest-ListImageFiltersTestJSON-680116525 tempest-ListImageFiltersTestJSON-680116525-project-member] [instance: 53d3e2f2-f63d-45c7-a5d1-1fa9a9f0ac0f] Build of instance 53d3e2f2-f63d-45c7-a5d1-1fa9a9f0ac0f was re-scheduled: Binding failed for port f650d966-513e-4686-9faa-9dd139c8c888, please check neutron logs for more information. {{(pid=61806) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 616.824257] env[61806]: DEBUG nova.compute.manager [None req-962d5603-e71b-4a26-895d-6dce1320b70e tempest-ListImageFiltersTestJSON-680116525 tempest-ListImageFiltersTestJSON-680116525-project-member] [instance: 53d3e2f2-f63d-45c7-a5d1-1fa9a9f0ac0f] Unplugging VIFs for instance {{(pid=61806) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 616.824494] env[61806]: DEBUG oslo_concurrency.lockutils [None req-962d5603-e71b-4a26-895d-6dce1320b70e tempest-ListImageFiltersTestJSON-680116525 tempest-ListImageFiltersTestJSON-680116525-project-member] Acquiring lock "refresh_cache-53d3e2f2-f63d-45c7-a5d1-1fa9a9f0ac0f" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 616.824642] env[61806]: DEBUG oslo_concurrency.lockutils [None req-962d5603-e71b-4a26-895d-6dce1320b70e tempest-ListImageFiltersTestJSON-680116525 tempest-ListImageFiltersTestJSON-680116525-project-member] Acquired lock "refresh_cache-53d3e2f2-f63d-45c7-a5d1-1fa9a9f0ac0f" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 616.824800] env[61806]: DEBUG nova.network.neutron [None req-962d5603-e71b-4a26-895d-6dce1320b70e tempest-ListImageFiltersTestJSON-680116525 tempest-ListImageFiltersTestJSON-680116525-project-member] [instance: 53d3e2f2-f63d-45c7-a5d1-1fa9a9f0ac0f] Building network info cache for instance {{(pid=61806) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 616.828655] env[61806]: DEBUG oslo_concurrency.lockutils [None req-11b5d1a0-518c-42d9-940e-4bf652a6d358 tempest-DeleteServersAdminTestJSON-1553667494 tempest-DeleteServersAdminTestJSON-1553667494-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 15.388s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 616.911147] env[61806]: DEBUG nova.network.neutron [req-7feed7ee-b470-45d3-b2a6-c1bf40c7e0b9 req-15927e58-6983-48e4-a9f6-6f3f6b0c4f9c service nova] [instance: 65f3b2f3-6d25-44f0-95cb-f9a7e75f8e82] Instance cache missing network info. {{(pid=61806) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 616.944792] env[61806]: DEBUG nova.network.neutron [-] [instance: 65f3b2f3-6d25-44f0-95cb-f9a7e75f8e82] Updating instance_info_cache with network_info: [] {{(pid=61806) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 616.982195] env[61806]: INFO nova.compute.manager [None req-9101cb9e-b191-4241-8bc1-9b126210ca50 tempest-ListImageFiltersTestJSON-680116525 tempest-ListImageFiltersTestJSON-680116525-project-member] [instance: 7e730579-8d0c-4226-86bc-292db16c49ee] Took 1.03 seconds to deallocate network for instance. [ 617.131274] env[61806]: DEBUG nova.network.neutron [req-7feed7ee-b470-45d3-b2a6-c1bf40c7e0b9 req-15927e58-6983-48e4-a9f6-6f3f6b0c4f9c service nova] [instance: 65f3b2f3-6d25-44f0-95cb-f9a7e75f8e82] Updating instance_info_cache with network_info: [] {{(pid=61806) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 617.276620] env[61806]: DEBUG nova.compute.manager [req-188be06a-a67d-4d41-969a-801b81bc12ab req-eee33904-f238-496f-9350-9e71447cc8cc service nova] [instance: 65f3b2f3-6d25-44f0-95cb-f9a7e75f8e82] Received event network-vif-deleted-ddedeb2e-1da8-4774-925f-5f0e09c0cae8 {{(pid=61806) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 617.354057] env[61806]: DEBUG nova.network.neutron [None req-962d5603-e71b-4a26-895d-6dce1320b70e tempest-ListImageFiltersTestJSON-680116525 tempest-ListImageFiltersTestJSON-680116525-project-member] [instance: 53d3e2f2-f63d-45c7-a5d1-1fa9a9f0ac0f] Instance cache missing network info. {{(pid=61806) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 617.448139] env[61806]: INFO nova.compute.manager [-] [instance: 65f3b2f3-6d25-44f0-95cb-f9a7e75f8e82] Took 1.04 seconds to deallocate network for instance. [ 617.451038] env[61806]: DEBUG nova.compute.claims [None req-e34d2bcf-266f-4faa-8f31-6bd93f6ab176 tempest-ImagesTestJSON-618852556 tempest-ImagesTestJSON-618852556-project-member] [instance: 65f3b2f3-6d25-44f0-95cb-f9a7e75f8e82] Aborting claim: {{(pid=61806) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 617.451218] env[61806]: DEBUG oslo_concurrency.lockutils [None req-e34d2bcf-266f-4faa-8f31-6bd93f6ab176 tempest-ImagesTestJSON-618852556 tempest-ImagesTestJSON-618852556-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 617.517867] env[61806]: DEBUG nova.network.neutron [None req-962d5603-e71b-4a26-895d-6dce1320b70e tempest-ListImageFiltersTestJSON-680116525 tempest-ListImageFiltersTestJSON-680116525-project-member] [instance: 53d3e2f2-f63d-45c7-a5d1-1fa9a9f0ac0f] Updating instance_info_cache with network_info: [] {{(pid=61806) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 617.637680] env[61806]: DEBUG oslo_concurrency.lockutils [req-7feed7ee-b470-45d3-b2a6-c1bf40c7e0b9 req-15927e58-6983-48e4-a9f6-6f3f6b0c4f9c service nova] Releasing lock "refresh_cache-65f3b2f3-6d25-44f0-95cb-f9a7e75f8e82" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 617.761803] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3c202b09-db4f-4022-90de-5f6a974f6155 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 617.770961] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-668269c7-ad24-45e5-9618-8a3ff7a5ad97 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 617.807983] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5d2c8175-079a-4df5-b98b-a5660ae03e76 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 617.819643] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6b4882e2-bc20-4f59-b81a-eee55a789e39 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 617.833770] env[61806]: DEBUG nova.compute.provider_tree [None req-11b5d1a0-518c-42d9-940e-4bf652a6d358 tempest-DeleteServersAdminTestJSON-1553667494 tempest-DeleteServersAdminTestJSON-1553667494-project-member] Inventory has not changed in ProviderTree for provider: a2858be1-fd22-4e08-979e-87ad25293407 {{(pid=61806) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 618.015527] env[61806]: INFO nova.scheduler.client.report [None req-9101cb9e-b191-4241-8bc1-9b126210ca50 tempest-ListImageFiltersTestJSON-680116525 tempest-ListImageFiltersTestJSON-680116525-project-member] Deleted allocations for instance 7e730579-8d0c-4226-86bc-292db16c49ee [ 618.022825] env[61806]: DEBUG oslo_concurrency.lockutils [None req-962d5603-e71b-4a26-895d-6dce1320b70e tempest-ListImageFiltersTestJSON-680116525 tempest-ListImageFiltersTestJSON-680116525-project-member] Releasing lock "refresh_cache-53d3e2f2-f63d-45c7-a5d1-1fa9a9f0ac0f" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 618.022929] env[61806]: DEBUG nova.compute.manager [None req-962d5603-e71b-4a26-895d-6dce1320b70e tempest-ListImageFiltersTestJSON-680116525 tempest-ListImageFiltersTestJSON-680116525-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61806) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 618.023042] env[61806]: DEBUG nova.compute.manager [None req-962d5603-e71b-4a26-895d-6dce1320b70e tempest-ListImageFiltersTestJSON-680116525 tempest-ListImageFiltersTestJSON-680116525-project-member] [instance: 53d3e2f2-f63d-45c7-a5d1-1fa9a9f0ac0f] Deallocating network for instance {{(pid=61806) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 618.023220] env[61806]: DEBUG nova.network.neutron [None req-962d5603-e71b-4a26-895d-6dce1320b70e tempest-ListImageFiltersTestJSON-680116525 tempest-ListImageFiltersTestJSON-680116525-project-member] [instance: 53d3e2f2-f63d-45c7-a5d1-1fa9a9f0ac0f] deallocate_for_instance() {{(pid=61806) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 618.048113] env[61806]: DEBUG nova.network.neutron [None req-962d5603-e71b-4a26-895d-6dce1320b70e tempest-ListImageFiltersTestJSON-680116525 tempest-ListImageFiltersTestJSON-680116525-project-member] [instance: 53d3e2f2-f63d-45c7-a5d1-1fa9a9f0ac0f] Instance cache missing network info. {{(pid=61806) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 618.339399] env[61806]: DEBUG nova.scheduler.client.report [None req-11b5d1a0-518c-42d9-940e-4bf652a6d358 tempest-DeleteServersAdminTestJSON-1553667494 tempest-DeleteServersAdminTestJSON-1553667494-project-member] Inventory has not changed for provider a2858be1-fd22-4e08-979e-87ad25293407 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 139, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61806) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 618.531754] env[61806]: DEBUG oslo_concurrency.lockutils [None req-9101cb9e-b191-4241-8bc1-9b126210ca50 tempest-ListImageFiltersTestJSON-680116525 tempest-ListImageFiltersTestJSON-680116525-project-member] Lock "7e730579-8d0c-4226-86bc-292db16c49ee" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 78.407s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 618.551032] env[61806]: DEBUG nova.network.neutron [None req-962d5603-e71b-4a26-895d-6dce1320b70e tempest-ListImageFiltersTestJSON-680116525 tempest-ListImageFiltersTestJSON-680116525-project-member] [instance: 53d3e2f2-f63d-45c7-a5d1-1fa9a9f0ac0f] Updating instance_info_cache with network_info: [] {{(pid=61806) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 618.845774] env[61806]: DEBUG oslo_concurrency.lockutils [None req-11b5d1a0-518c-42d9-940e-4bf652a6d358 tempest-DeleteServersAdminTestJSON-1553667494 tempest-DeleteServersAdminTestJSON-1553667494-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.019s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 618.846411] env[61806]: ERROR nova.compute.manager [None req-11b5d1a0-518c-42d9-940e-4bf652a6d358 tempest-DeleteServersAdminTestJSON-1553667494 tempest-DeleteServersAdminTestJSON-1553667494-project-member] [instance: ddc801a0-0782-4c6d-bd72-95d0afd01339] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port f65985e6-72fe-4194-a8d3-fc371e6fd74f, please check neutron logs for more information. [ 618.846411] env[61806]: ERROR nova.compute.manager [instance: ddc801a0-0782-4c6d-bd72-95d0afd01339] Traceback (most recent call last): [ 618.846411] env[61806]: ERROR nova.compute.manager [instance: ddc801a0-0782-4c6d-bd72-95d0afd01339] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 618.846411] env[61806]: ERROR nova.compute.manager [instance: ddc801a0-0782-4c6d-bd72-95d0afd01339] self.driver.spawn(context, instance, image_meta, [ 618.846411] env[61806]: ERROR nova.compute.manager [instance: ddc801a0-0782-4c6d-bd72-95d0afd01339] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 618.846411] env[61806]: ERROR nova.compute.manager [instance: ddc801a0-0782-4c6d-bd72-95d0afd01339] self._vmops.spawn(context, instance, image_meta, injected_files, [ 618.846411] env[61806]: ERROR nova.compute.manager [instance: ddc801a0-0782-4c6d-bd72-95d0afd01339] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 618.846411] env[61806]: ERROR nova.compute.manager [instance: ddc801a0-0782-4c6d-bd72-95d0afd01339] vm_ref = self.build_virtual_machine(instance, [ 618.846411] env[61806]: ERROR nova.compute.manager [instance: ddc801a0-0782-4c6d-bd72-95d0afd01339] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 618.846411] env[61806]: ERROR nova.compute.manager [instance: ddc801a0-0782-4c6d-bd72-95d0afd01339] vif_infos = vmwarevif.get_vif_info(self._session, [ 618.846411] env[61806]: ERROR nova.compute.manager [instance: ddc801a0-0782-4c6d-bd72-95d0afd01339] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 618.846693] env[61806]: ERROR nova.compute.manager [instance: ddc801a0-0782-4c6d-bd72-95d0afd01339] for vif in network_info: [ 618.846693] env[61806]: ERROR nova.compute.manager [instance: ddc801a0-0782-4c6d-bd72-95d0afd01339] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 618.846693] env[61806]: ERROR nova.compute.manager [instance: ddc801a0-0782-4c6d-bd72-95d0afd01339] return self._sync_wrapper(fn, *args, **kwargs) [ 618.846693] env[61806]: ERROR nova.compute.manager [instance: ddc801a0-0782-4c6d-bd72-95d0afd01339] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 618.846693] env[61806]: ERROR nova.compute.manager [instance: ddc801a0-0782-4c6d-bd72-95d0afd01339] self.wait() [ 618.846693] env[61806]: ERROR nova.compute.manager [instance: ddc801a0-0782-4c6d-bd72-95d0afd01339] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 618.846693] env[61806]: ERROR nova.compute.manager [instance: ddc801a0-0782-4c6d-bd72-95d0afd01339] self[:] = self._gt.wait() [ 618.846693] env[61806]: ERROR nova.compute.manager [instance: ddc801a0-0782-4c6d-bd72-95d0afd01339] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 618.846693] env[61806]: ERROR nova.compute.manager [instance: ddc801a0-0782-4c6d-bd72-95d0afd01339] return self._exit_event.wait() [ 618.846693] env[61806]: ERROR nova.compute.manager [instance: ddc801a0-0782-4c6d-bd72-95d0afd01339] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 618.846693] env[61806]: ERROR nova.compute.manager [instance: ddc801a0-0782-4c6d-bd72-95d0afd01339] result = hub.switch() [ 618.846693] env[61806]: ERROR nova.compute.manager [instance: ddc801a0-0782-4c6d-bd72-95d0afd01339] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 618.846693] env[61806]: ERROR nova.compute.manager [instance: ddc801a0-0782-4c6d-bd72-95d0afd01339] return self.greenlet.switch() [ 618.846977] env[61806]: ERROR nova.compute.manager [instance: ddc801a0-0782-4c6d-bd72-95d0afd01339] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 618.846977] env[61806]: ERROR nova.compute.manager [instance: ddc801a0-0782-4c6d-bd72-95d0afd01339] result = function(*args, **kwargs) [ 618.846977] env[61806]: ERROR nova.compute.manager [instance: ddc801a0-0782-4c6d-bd72-95d0afd01339] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 618.846977] env[61806]: ERROR nova.compute.manager [instance: ddc801a0-0782-4c6d-bd72-95d0afd01339] return func(*args, **kwargs) [ 618.846977] env[61806]: ERROR nova.compute.manager [instance: ddc801a0-0782-4c6d-bd72-95d0afd01339] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 618.846977] env[61806]: ERROR nova.compute.manager [instance: ddc801a0-0782-4c6d-bd72-95d0afd01339] raise e [ 618.846977] env[61806]: ERROR nova.compute.manager [instance: ddc801a0-0782-4c6d-bd72-95d0afd01339] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 618.846977] env[61806]: ERROR nova.compute.manager [instance: ddc801a0-0782-4c6d-bd72-95d0afd01339] nwinfo = self.network_api.allocate_for_instance( [ 618.846977] env[61806]: ERROR nova.compute.manager [instance: ddc801a0-0782-4c6d-bd72-95d0afd01339] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 618.846977] env[61806]: ERROR nova.compute.manager [instance: ddc801a0-0782-4c6d-bd72-95d0afd01339] created_port_ids = self._update_ports_for_instance( [ 618.846977] env[61806]: ERROR nova.compute.manager [instance: ddc801a0-0782-4c6d-bd72-95d0afd01339] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 618.846977] env[61806]: ERROR nova.compute.manager [instance: ddc801a0-0782-4c6d-bd72-95d0afd01339] with excutils.save_and_reraise_exception(): [ 618.846977] env[61806]: ERROR nova.compute.manager [instance: ddc801a0-0782-4c6d-bd72-95d0afd01339] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 618.847272] env[61806]: ERROR nova.compute.manager [instance: ddc801a0-0782-4c6d-bd72-95d0afd01339] self.force_reraise() [ 618.847272] env[61806]: ERROR nova.compute.manager [instance: ddc801a0-0782-4c6d-bd72-95d0afd01339] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 618.847272] env[61806]: ERROR nova.compute.manager [instance: ddc801a0-0782-4c6d-bd72-95d0afd01339] raise self.value [ 618.847272] env[61806]: ERROR nova.compute.manager [instance: ddc801a0-0782-4c6d-bd72-95d0afd01339] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 618.847272] env[61806]: ERROR nova.compute.manager [instance: ddc801a0-0782-4c6d-bd72-95d0afd01339] updated_port = self._update_port( [ 618.847272] env[61806]: ERROR nova.compute.manager [instance: ddc801a0-0782-4c6d-bd72-95d0afd01339] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 618.847272] env[61806]: ERROR nova.compute.manager [instance: ddc801a0-0782-4c6d-bd72-95d0afd01339] _ensure_no_port_binding_failure(port) [ 618.847272] env[61806]: ERROR nova.compute.manager [instance: ddc801a0-0782-4c6d-bd72-95d0afd01339] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 618.847272] env[61806]: ERROR nova.compute.manager [instance: ddc801a0-0782-4c6d-bd72-95d0afd01339] raise exception.PortBindingFailed(port_id=port['id']) [ 618.847272] env[61806]: ERROR nova.compute.manager [instance: ddc801a0-0782-4c6d-bd72-95d0afd01339] nova.exception.PortBindingFailed: Binding failed for port f65985e6-72fe-4194-a8d3-fc371e6fd74f, please check neutron logs for more information. [ 618.847272] env[61806]: ERROR nova.compute.manager [instance: ddc801a0-0782-4c6d-bd72-95d0afd01339] [ 618.847517] env[61806]: DEBUG nova.compute.utils [None req-11b5d1a0-518c-42d9-940e-4bf652a6d358 tempest-DeleteServersAdminTestJSON-1553667494 tempest-DeleteServersAdminTestJSON-1553667494-project-member] [instance: ddc801a0-0782-4c6d-bd72-95d0afd01339] Binding failed for port f65985e6-72fe-4194-a8d3-fc371e6fd74f, please check neutron logs for more information. {{(pid=61806) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 618.850283] env[61806]: DEBUG nova.compute.manager [None req-11b5d1a0-518c-42d9-940e-4bf652a6d358 tempest-DeleteServersAdminTestJSON-1553667494 tempest-DeleteServersAdminTestJSON-1553667494-project-member] [instance: ddc801a0-0782-4c6d-bd72-95d0afd01339] Build of instance ddc801a0-0782-4c6d-bd72-95d0afd01339 was re-scheduled: Binding failed for port f65985e6-72fe-4194-a8d3-fc371e6fd74f, please check neutron logs for more information. {{(pid=61806) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 618.850283] env[61806]: DEBUG nova.compute.manager [None req-11b5d1a0-518c-42d9-940e-4bf652a6d358 tempest-DeleteServersAdminTestJSON-1553667494 tempest-DeleteServersAdminTestJSON-1553667494-project-member] [instance: ddc801a0-0782-4c6d-bd72-95d0afd01339] Unplugging VIFs for instance {{(pid=61806) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 618.850283] env[61806]: DEBUG oslo_concurrency.lockutils [None req-11b5d1a0-518c-42d9-940e-4bf652a6d358 tempest-DeleteServersAdminTestJSON-1553667494 tempest-DeleteServersAdminTestJSON-1553667494-project-member] Acquiring lock "refresh_cache-ddc801a0-0782-4c6d-bd72-95d0afd01339" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 618.850283] env[61806]: DEBUG oslo_concurrency.lockutils [None req-11b5d1a0-518c-42d9-940e-4bf652a6d358 tempest-DeleteServersAdminTestJSON-1553667494 tempest-DeleteServersAdminTestJSON-1553667494-project-member] Acquired lock "refresh_cache-ddc801a0-0782-4c6d-bd72-95d0afd01339" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 618.850470] env[61806]: DEBUG nova.network.neutron [None req-11b5d1a0-518c-42d9-940e-4bf652a6d358 tempest-DeleteServersAdminTestJSON-1553667494 tempest-DeleteServersAdminTestJSON-1553667494-project-member] [instance: ddc801a0-0782-4c6d-bd72-95d0afd01339] Building network info cache for instance {{(pid=61806) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 618.851646] env[61806]: DEBUG oslo_concurrency.lockutils [None req-a6b8da3a-7829-486a-884d-0f28be02e38c tempest-ServersTestManualDisk-151630353 tempest-ServersTestManualDisk-151630353-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 14.689s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 618.853772] env[61806]: INFO nova.compute.claims [None req-a6b8da3a-7829-486a-884d-0f28be02e38c tempest-ServersTestManualDisk-151630353 tempest-ServersTestManualDisk-151630353-project-member] [instance: 38a79453-939d-4649-9419-a2799636180d] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 619.037209] env[61806]: DEBUG nova.compute.manager [None req-e8db3b70-c93a-419a-89f8-b994313357e1 tempest-ServerShowV254Test-608938657 tempest-ServerShowV254Test-608938657-project-member] [instance: 8016bafc-b287-4df1-bd4b-a00210177a86] Starting instance... {{(pid=61806) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 619.057028] env[61806]: INFO nova.compute.manager [None req-962d5603-e71b-4a26-895d-6dce1320b70e tempest-ListImageFiltersTestJSON-680116525 tempest-ListImageFiltersTestJSON-680116525-project-member] [instance: 53d3e2f2-f63d-45c7-a5d1-1fa9a9f0ac0f] Took 1.03 seconds to deallocate network for instance. [ 619.379355] env[61806]: DEBUG nova.network.neutron [None req-11b5d1a0-518c-42d9-940e-4bf652a6d358 tempest-DeleteServersAdminTestJSON-1553667494 tempest-DeleteServersAdminTestJSON-1553667494-project-member] [instance: ddc801a0-0782-4c6d-bd72-95d0afd01339] Instance cache missing network info. {{(pid=61806) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 619.433037] env[61806]: DEBUG nova.network.neutron [None req-11b5d1a0-518c-42d9-940e-4bf652a6d358 tempest-DeleteServersAdminTestJSON-1553667494 tempest-DeleteServersAdminTestJSON-1553667494-project-member] [instance: ddc801a0-0782-4c6d-bd72-95d0afd01339] Updating instance_info_cache with network_info: [] {{(pid=61806) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 619.574523] env[61806]: DEBUG oslo_concurrency.lockutils [None req-e8db3b70-c93a-419a-89f8-b994313357e1 tempest-ServerShowV254Test-608938657 tempest-ServerShowV254Test-608938657-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 619.936106] env[61806]: DEBUG oslo_concurrency.lockutils [None req-11b5d1a0-518c-42d9-940e-4bf652a6d358 tempest-DeleteServersAdminTestJSON-1553667494 tempest-DeleteServersAdminTestJSON-1553667494-project-member] Releasing lock "refresh_cache-ddc801a0-0782-4c6d-bd72-95d0afd01339" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 619.936106] env[61806]: DEBUG nova.compute.manager [None req-11b5d1a0-518c-42d9-940e-4bf652a6d358 tempest-DeleteServersAdminTestJSON-1553667494 tempest-DeleteServersAdminTestJSON-1553667494-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61806) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 619.936106] env[61806]: DEBUG nova.compute.manager [None req-11b5d1a0-518c-42d9-940e-4bf652a6d358 tempest-DeleteServersAdminTestJSON-1553667494 tempest-DeleteServersAdminTestJSON-1553667494-project-member] [instance: ddc801a0-0782-4c6d-bd72-95d0afd01339] Deallocating network for instance {{(pid=61806) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 619.936106] env[61806]: DEBUG nova.network.neutron [None req-11b5d1a0-518c-42d9-940e-4bf652a6d358 tempest-DeleteServersAdminTestJSON-1553667494 tempest-DeleteServersAdminTestJSON-1553667494-project-member] [instance: ddc801a0-0782-4c6d-bd72-95d0afd01339] deallocate_for_instance() {{(pid=61806) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 619.954167] env[61806]: DEBUG nova.network.neutron [None req-11b5d1a0-518c-42d9-940e-4bf652a6d358 tempest-DeleteServersAdminTestJSON-1553667494 tempest-DeleteServersAdminTestJSON-1553667494-project-member] [instance: ddc801a0-0782-4c6d-bd72-95d0afd01339] Instance cache missing network info. {{(pid=61806) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 620.084418] env[61806]: INFO nova.scheduler.client.report [None req-962d5603-e71b-4a26-895d-6dce1320b70e tempest-ListImageFiltersTestJSON-680116525 tempest-ListImageFiltersTestJSON-680116525-project-member] Deleted allocations for instance 53d3e2f2-f63d-45c7-a5d1-1fa9a9f0ac0f [ 620.260748] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9c0e7b9d-099f-434d-ba7d-d8a603b297be {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 620.268904] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ea100c65-d498-4462-9b55-1c7abe87969f {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 620.302790] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-34f37e31-0416-46d3-b95b-085a22e36ded {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 620.310981] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dfa74794-5afa-451f-860e-a4c3b5b08ce2 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 620.328477] env[61806]: DEBUG nova.compute.provider_tree [None req-a6b8da3a-7829-486a-884d-0f28be02e38c tempest-ServersTestManualDisk-151630353 tempest-ServersTestManualDisk-151630353-project-member] Inventory has not changed in ProviderTree for provider: a2858be1-fd22-4e08-979e-87ad25293407 {{(pid=61806) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 620.456589] env[61806]: DEBUG nova.network.neutron [None req-11b5d1a0-518c-42d9-940e-4bf652a6d358 tempest-DeleteServersAdminTestJSON-1553667494 tempest-DeleteServersAdminTestJSON-1553667494-project-member] [instance: ddc801a0-0782-4c6d-bd72-95d0afd01339] Updating instance_info_cache with network_info: [] {{(pid=61806) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 620.592530] env[61806]: DEBUG oslo_concurrency.lockutils [None req-962d5603-e71b-4a26-895d-6dce1320b70e tempest-ListImageFiltersTestJSON-680116525 tempest-ListImageFiltersTestJSON-680116525-project-member] Lock "53d3e2f2-f63d-45c7-a5d1-1fa9a9f0ac0f" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 78.598s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 620.832356] env[61806]: DEBUG nova.scheduler.client.report [None req-a6b8da3a-7829-486a-884d-0f28be02e38c tempest-ServersTestManualDisk-151630353 tempest-ServersTestManualDisk-151630353-project-member] Inventory has not changed for provider a2858be1-fd22-4e08-979e-87ad25293407 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 139, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61806) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 620.962120] env[61806]: INFO nova.compute.manager [None req-11b5d1a0-518c-42d9-940e-4bf652a6d358 tempest-DeleteServersAdminTestJSON-1553667494 tempest-DeleteServersAdminTestJSON-1553667494-project-member] [instance: ddc801a0-0782-4c6d-bd72-95d0afd01339] Took 1.03 seconds to deallocate network for instance. [ 621.094654] env[61806]: DEBUG nova.compute.manager [None req-88cd9ecf-9bda-478c-b5e6-5a21edd9271f tempest-MigrationsAdminTest-470098049 tempest-MigrationsAdminTest-470098049-project-member] [instance: e66c921c-320e-4347-96c9-27aa80ba7f41] Starting instance... {{(pid=61806) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 621.337873] env[61806]: DEBUG oslo_concurrency.lockutils [None req-a6b8da3a-7829-486a-884d-0f28be02e38c tempest-ServersTestManualDisk-151630353 tempest-ServersTestManualDisk-151630353-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.486s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 621.338788] env[61806]: DEBUG nova.compute.manager [None req-a6b8da3a-7829-486a-884d-0f28be02e38c tempest-ServersTestManualDisk-151630353 tempest-ServersTestManualDisk-151630353-project-member] [instance: 38a79453-939d-4649-9419-a2799636180d] Start building networks asynchronously for instance. {{(pid=61806) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 621.341755] env[61806]: DEBUG oslo_concurrency.lockutils [None req-98d962a0-2632-4a9a-8d12-70ce4acc382b tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 15.336s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 621.343288] env[61806]: INFO nova.compute.claims [None req-98d962a0-2632-4a9a-8d12-70ce4acc382b tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] [instance: b8249dc5-5561-454a-aa01-477f9f1fc0e1] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 621.622082] env[61806]: DEBUG oslo_concurrency.lockutils [None req-88cd9ecf-9bda-478c-b5e6-5a21edd9271f tempest-MigrationsAdminTest-470098049 tempest-MigrationsAdminTest-470098049-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 621.848434] env[61806]: DEBUG nova.compute.utils [None req-a6b8da3a-7829-486a-884d-0f28be02e38c tempest-ServersTestManualDisk-151630353 tempest-ServersTestManualDisk-151630353-project-member] Using /dev/sd instead of None {{(pid=61806) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 621.852344] env[61806]: DEBUG nova.compute.manager [None req-a6b8da3a-7829-486a-884d-0f28be02e38c tempest-ServersTestManualDisk-151630353 tempest-ServersTestManualDisk-151630353-project-member] [instance: 38a79453-939d-4649-9419-a2799636180d] Allocating IP information in the background. {{(pid=61806) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 621.852447] env[61806]: DEBUG nova.network.neutron [None req-a6b8da3a-7829-486a-884d-0f28be02e38c tempest-ServersTestManualDisk-151630353 tempest-ServersTestManualDisk-151630353-project-member] [instance: 38a79453-939d-4649-9419-a2799636180d] allocate_for_instance() {{(pid=61806) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 621.892123] env[61806]: DEBUG nova.policy [None req-a6b8da3a-7829-486a-884d-0f28be02e38c tempest-ServersTestManualDisk-151630353 tempest-ServersTestManualDisk-151630353-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'e9e547b235174e96bef199b9f7687218', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '87612ab1f8ef4b4e806133c1d407e783', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61806) authorize /opt/stack/nova/nova/policy.py:201}} [ 621.994917] env[61806]: INFO nova.scheduler.client.report [None req-11b5d1a0-518c-42d9-940e-4bf652a6d358 tempest-DeleteServersAdminTestJSON-1553667494 tempest-DeleteServersAdminTestJSON-1553667494-project-member] Deleted allocations for instance ddc801a0-0782-4c6d-bd72-95d0afd01339 [ 622.239798] env[61806]: DEBUG nova.network.neutron [None req-a6b8da3a-7829-486a-884d-0f28be02e38c tempest-ServersTestManualDisk-151630353 tempest-ServersTestManualDisk-151630353-project-member] [instance: 38a79453-939d-4649-9419-a2799636180d] Successfully created port: 6016a127-edf4-40d6-aba6-0a14614fe4ff {{(pid=61806) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 622.356813] env[61806]: DEBUG nova.compute.manager [None req-a6b8da3a-7829-486a-884d-0f28be02e38c tempest-ServersTestManualDisk-151630353 tempest-ServersTestManualDisk-151630353-project-member] [instance: 38a79453-939d-4649-9419-a2799636180d] Start building block device mappings for instance. {{(pid=61806) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 622.502715] env[61806]: DEBUG oslo_concurrency.lockutils [None req-11b5d1a0-518c-42d9-940e-4bf652a6d358 tempest-DeleteServersAdminTestJSON-1553667494 tempest-DeleteServersAdminTestJSON-1553667494-project-member] Lock "ddc801a0-0782-4c6d-bd72-95d0afd01339" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 79.665s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 622.688020] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-32bf2ebb-05db-47f1-97a2-c2c4f046e046 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 622.694334] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b8ccf5d2-eda0-4603-a8bc-1c94544522cb {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 622.726189] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9376f97e-95c9-4800-8eff-14f35e94c914 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 622.734221] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2e790d02-0d18-4ca7-b11a-c67a2273fdb4 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 622.748867] env[61806]: DEBUG nova.compute.provider_tree [None req-98d962a0-2632-4a9a-8d12-70ce4acc382b tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] Inventory has not changed in ProviderTree for provider: a2858be1-fd22-4e08-979e-87ad25293407 {{(pid=61806) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 622.868189] env[61806]: DEBUG oslo_concurrency.lockutils [None req-1382c347-5a79-4289-8b32-002f296b7ae3 tempest-TenantUsagesTestJSON-255073140 tempest-TenantUsagesTestJSON-255073140-project-member] Acquiring lock "949c1050-5003-4519-b24c-8904de21a676" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 622.868577] env[61806]: DEBUG oslo_concurrency.lockutils [None req-1382c347-5a79-4289-8b32-002f296b7ae3 tempest-TenantUsagesTestJSON-255073140 tempest-TenantUsagesTestJSON-255073140-project-member] Lock "949c1050-5003-4519-b24c-8904de21a676" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 623.006721] env[61806]: DEBUG nova.compute.manager [None req-44eba412-ca3f-4386-8c36-f42e9734628d tempest-ServersTestBootFromVolume-1815739798 tempest-ServersTestBootFromVolume-1815739798-project-member] [instance: 56d26d08-3c6d-45e8-b238-6dbca6024561] Starting instance... {{(pid=61806) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 623.256626] env[61806]: DEBUG nova.scheduler.client.report [None req-98d962a0-2632-4a9a-8d12-70ce4acc382b tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] Inventory has not changed for provider a2858be1-fd22-4e08-979e-87ad25293407 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 139, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61806) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 623.261842] env[61806]: DEBUG nova.compute.manager [req-896214cf-a27e-4f7b-82d7-2e5987330353 req-43621815-a3b8-4bdb-aa10-683fcdfc5494 service nova] [instance: 38a79453-939d-4649-9419-a2799636180d] Received event network-changed-6016a127-edf4-40d6-aba6-0a14614fe4ff {{(pid=61806) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 623.262098] env[61806]: DEBUG nova.compute.manager [req-896214cf-a27e-4f7b-82d7-2e5987330353 req-43621815-a3b8-4bdb-aa10-683fcdfc5494 service nova] [instance: 38a79453-939d-4649-9419-a2799636180d] Refreshing instance network info cache due to event network-changed-6016a127-edf4-40d6-aba6-0a14614fe4ff. {{(pid=61806) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 623.262249] env[61806]: DEBUG oslo_concurrency.lockutils [req-896214cf-a27e-4f7b-82d7-2e5987330353 req-43621815-a3b8-4bdb-aa10-683fcdfc5494 service nova] Acquiring lock "refresh_cache-38a79453-939d-4649-9419-a2799636180d" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 623.262462] env[61806]: DEBUG oslo_concurrency.lockutils [req-896214cf-a27e-4f7b-82d7-2e5987330353 req-43621815-a3b8-4bdb-aa10-683fcdfc5494 service nova] Acquired lock "refresh_cache-38a79453-939d-4649-9419-a2799636180d" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 623.262592] env[61806]: DEBUG nova.network.neutron [req-896214cf-a27e-4f7b-82d7-2e5987330353 req-43621815-a3b8-4bdb-aa10-683fcdfc5494 service nova] [instance: 38a79453-939d-4649-9419-a2799636180d] Refreshing network info cache for port 6016a127-edf4-40d6-aba6-0a14614fe4ff {{(pid=61806) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 623.366133] env[61806]: DEBUG nova.compute.manager [None req-a6b8da3a-7829-486a-884d-0f28be02e38c tempest-ServersTestManualDisk-151630353 tempest-ServersTestManualDisk-151630353-project-member] [instance: 38a79453-939d-4649-9419-a2799636180d] Start spawning the instance on the hypervisor. {{(pid=61806) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 623.404280] env[61806]: DEBUG nova.virt.hardware [None req-a6b8da3a-7829-486a-884d-0f28be02e38c tempest-ServersTestManualDisk-151630353 tempest-ServersTestManualDisk-151630353-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-15T18:05:10Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-15T18:04:52Z,direct_url=,disk_format='vmdk',id=ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='84ba83fce2c349a6988173c5d6fc3b07',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-15T18:04:53Z,virtual_size=,visibility=), allow threads: False {{(pid=61806) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 623.404539] env[61806]: DEBUG nova.virt.hardware [None req-a6b8da3a-7829-486a-884d-0f28be02e38c tempest-ServersTestManualDisk-151630353 tempest-ServersTestManualDisk-151630353-project-member] Flavor limits 0:0:0 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 623.404695] env[61806]: DEBUG nova.virt.hardware [None req-a6b8da3a-7829-486a-884d-0f28be02e38c tempest-ServersTestManualDisk-151630353 tempest-ServersTestManualDisk-151630353-project-member] Image limits 0:0:0 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 623.404875] env[61806]: DEBUG nova.virt.hardware [None req-a6b8da3a-7829-486a-884d-0f28be02e38c tempest-ServersTestManualDisk-151630353 tempest-ServersTestManualDisk-151630353-project-member] Flavor pref 0:0:0 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 623.405255] env[61806]: DEBUG nova.virt.hardware [None req-a6b8da3a-7829-486a-884d-0f28be02e38c tempest-ServersTestManualDisk-151630353 tempest-ServersTestManualDisk-151630353-project-member] Image pref 0:0:0 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 623.405493] env[61806]: DEBUG nova.virt.hardware [None req-a6b8da3a-7829-486a-884d-0f28be02e38c tempest-ServersTestManualDisk-151630353 tempest-ServersTestManualDisk-151630353-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 623.405712] env[61806]: DEBUG nova.virt.hardware [None req-a6b8da3a-7829-486a-884d-0f28be02e38c tempest-ServersTestManualDisk-151630353 tempest-ServersTestManualDisk-151630353-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61806) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 623.406064] env[61806]: DEBUG nova.virt.hardware [None req-a6b8da3a-7829-486a-884d-0f28be02e38c tempest-ServersTestManualDisk-151630353 tempest-ServersTestManualDisk-151630353-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61806) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 623.406299] env[61806]: DEBUG nova.virt.hardware [None req-a6b8da3a-7829-486a-884d-0f28be02e38c tempest-ServersTestManualDisk-151630353 tempest-ServersTestManualDisk-151630353-project-member] Got 1 possible topologies {{(pid=61806) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 623.406471] env[61806]: DEBUG nova.virt.hardware [None req-a6b8da3a-7829-486a-884d-0f28be02e38c tempest-ServersTestManualDisk-151630353 tempest-ServersTestManualDisk-151630353-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61806) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 623.406644] env[61806]: DEBUG nova.virt.hardware [None req-a6b8da3a-7829-486a-884d-0f28be02e38c tempest-ServersTestManualDisk-151630353 tempest-ServersTestManualDisk-151630353-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61806) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 623.408107] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9e7ff19f-82c4-47ac-984d-820067aa4357 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 623.418320] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0e130c29-0da9-4039-911d-9e980f77322e {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 623.530609] env[61806]: DEBUG oslo_concurrency.lockutils [None req-44eba412-ca3f-4386-8c36-f42e9734628d tempest-ServersTestBootFromVolume-1815739798 tempest-ServersTestBootFromVolume-1815739798-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 623.549033] env[61806]: ERROR nova.compute.manager [None req-a6b8da3a-7829-486a-884d-0f28be02e38c tempest-ServersTestManualDisk-151630353 tempest-ServersTestManualDisk-151630353-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 6016a127-edf4-40d6-aba6-0a14614fe4ff, please check neutron logs for more information. [ 623.549033] env[61806]: ERROR nova.compute.manager Traceback (most recent call last): [ 623.549033] env[61806]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 623.549033] env[61806]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 623.549033] env[61806]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 623.549033] env[61806]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 623.549033] env[61806]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 623.549033] env[61806]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 623.549033] env[61806]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 623.549033] env[61806]: ERROR nova.compute.manager self.force_reraise() [ 623.549033] env[61806]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 623.549033] env[61806]: ERROR nova.compute.manager raise self.value [ 623.549033] env[61806]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 623.549033] env[61806]: ERROR nova.compute.manager updated_port = self._update_port( [ 623.549033] env[61806]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 623.549033] env[61806]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 623.549495] env[61806]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 623.549495] env[61806]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 623.549495] env[61806]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 6016a127-edf4-40d6-aba6-0a14614fe4ff, please check neutron logs for more information. [ 623.549495] env[61806]: ERROR nova.compute.manager [ 623.549495] env[61806]: Traceback (most recent call last): [ 623.549495] env[61806]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 623.549495] env[61806]: listener.cb(fileno) [ 623.549495] env[61806]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 623.549495] env[61806]: result = function(*args, **kwargs) [ 623.549495] env[61806]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 623.549495] env[61806]: return func(*args, **kwargs) [ 623.549495] env[61806]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 623.549495] env[61806]: raise e [ 623.549495] env[61806]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 623.549495] env[61806]: nwinfo = self.network_api.allocate_for_instance( [ 623.549495] env[61806]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 623.549495] env[61806]: created_port_ids = self._update_ports_for_instance( [ 623.549495] env[61806]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 623.549495] env[61806]: with excutils.save_and_reraise_exception(): [ 623.549495] env[61806]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 623.549495] env[61806]: self.force_reraise() [ 623.549495] env[61806]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 623.549495] env[61806]: raise self.value [ 623.549495] env[61806]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 623.549495] env[61806]: updated_port = self._update_port( [ 623.549495] env[61806]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 623.549495] env[61806]: _ensure_no_port_binding_failure(port) [ 623.549495] env[61806]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 623.549495] env[61806]: raise exception.PortBindingFailed(port_id=port['id']) [ 623.550219] env[61806]: nova.exception.PortBindingFailed: Binding failed for port 6016a127-edf4-40d6-aba6-0a14614fe4ff, please check neutron logs for more information. [ 623.550219] env[61806]: Removing descriptor: 16 [ 623.550219] env[61806]: ERROR nova.compute.manager [None req-a6b8da3a-7829-486a-884d-0f28be02e38c tempest-ServersTestManualDisk-151630353 tempest-ServersTestManualDisk-151630353-project-member] [instance: 38a79453-939d-4649-9419-a2799636180d] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 6016a127-edf4-40d6-aba6-0a14614fe4ff, please check neutron logs for more information. [ 623.550219] env[61806]: ERROR nova.compute.manager [instance: 38a79453-939d-4649-9419-a2799636180d] Traceback (most recent call last): [ 623.550219] env[61806]: ERROR nova.compute.manager [instance: 38a79453-939d-4649-9419-a2799636180d] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 623.550219] env[61806]: ERROR nova.compute.manager [instance: 38a79453-939d-4649-9419-a2799636180d] yield resources [ 623.550219] env[61806]: ERROR nova.compute.manager [instance: 38a79453-939d-4649-9419-a2799636180d] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 623.550219] env[61806]: ERROR nova.compute.manager [instance: 38a79453-939d-4649-9419-a2799636180d] self.driver.spawn(context, instance, image_meta, [ 623.550219] env[61806]: ERROR nova.compute.manager [instance: 38a79453-939d-4649-9419-a2799636180d] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 623.550219] env[61806]: ERROR nova.compute.manager [instance: 38a79453-939d-4649-9419-a2799636180d] self._vmops.spawn(context, instance, image_meta, injected_files, [ 623.550219] env[61806]: ERROR nova.compute.manager [instance: 38a79453-939d-4649-9419-a2799636180d] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 623.550219] env[61806]: ERROR nova.compute.manager [instance: 38a79453-939d-4649-9419-a2799636180d] vm_ref = self.build_virtual_machine(instance, [ 623.550571] env[61806]: ERROR nova.compute.manager [instance: 38a79453-939d-4649-9419-a2799636180d] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 623.550571] env[61806]: ERROR nova.compute.manager [instance: 38a79453-939d-4649-9419-a2799636180d] vif_infos = vmwarevif.get_vif_info(self._session, [ 623.550571] env[61806]: ERROR nova.compute.manager [instance: 38a79453-939d-4649-9419-a2799636180d] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 623.550571] env[61806]: ERROR nova.compute.manager [instance: 38a79453-939d-4649-9419-a2799636180d] for vif in network_info: [ 623.550571] env[61806]: ERROR nova.compute.manager [instance: 38a79453-939d-4649-9419-a2799636180d] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 623.550571] env[61806]: ERROR nova.compute.manager [instance: 38a79453-939d-4649-9419-a2799636180d] return self._sync_wrapper(fn, *args, **kwargs) [ 623.550571] env[61806]: ERROR nova.compute.manager [instance: 38a79453-939d-4649-9419-a2799636180d] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 623.550571] env[61806]: ERROR nova.compute.manager [instance: 38a79453-939d-4649-9419-a2799636180d] self.wait() [ 623.550571] env[61806]: ERROR nova.compute.manager [instance: 38a79453-939d-4649-9419-a2799636180d] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 623.550571] env[61806]: ERROR nova.compute.manager [instance: 38a79453-939d-4649-9419-a2799636180d] self[:] = self._gt.wait() [ 623.550571] env[61806]: ERROR nova.compute.manager [instance: 38a79453-939d-4649-9419-a2799636180d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 623.550571] env[61806]: ERROR nova.compute.manager [instance: 38a79453-939d-4649-9419-a2799636180d] return self._exit_event.wait() [ 623.550571] env[61806]: ERROR nova.compute.manager [instance: 38a79453-939d-4649-9419-a2799636180d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 623.550922] env[61806]: ERROR nova.compute.manager [instance: 38a79453-939d-4649-9419-a2799636180d] result = hub.switch() [ 623.550922] env[61806]: ERROR nova.compute.manager [instance: 38a79453-939d-4649-9419-a2799636180d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 623.550922] env[61806]: ERROR nova.compute.manager [instance: 38a79453-939d-4649-9419-a2799636180d] return self.greenlet.switch() [ 623.550922] env[61806]: ERROR nova.compute.manager [instance: 38a79453-939d-4649-9419-a2799636180d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 623.550922] env[61806]: ERROR nova.compute.manager [instance: 38a79453-939d-4649-9419-a2799636180d] result = function(*args, **kwargs) [ 623.550922] env[61806]: ERROR nova.compute.manager [instance: 38a79453-939d-4649-9419-a2799636180d] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 623.550922] env[61806]: ERROR nova.compute.manager [instance: 38a79453-939d-4649-9419-a2799636180d] return func(*args, **kwargs) [ 623.550922] env[61806]: ERROR nova.compute.manager [instance: 38a79453-939d-4649-9419-a2799636180d] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 623.550922] env[61806]: ERROR nova.compute.manager [instance: 38a79453-939d-4649-9419-a2799636180d] raise e [ 623.550922] env[61806]: ERROR nova.compute.manager [instance: 38a79453-939d-4649-9419-a2799636180d] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 623.550922] env[61806]: ERROR nova.compute.manager [instance: 38a79453-939d-4649-9419-a2799636180d] nwinfo = self.network_api.allocate_for_instance( [ 623.550922] env[61806]: ERROR nova.compute.manager [instance: 38a79453-939d-4649-9419-a2799636180d] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 623.550922] env[61806]: ERROR nova.compute.manager [instance: 38a79453-939d-4649-9419-a2799636180d] created_port_ids = self._update_ports_for_instance( [ 623.551238] env[61806]: ERROR nova.compute.manager [instance: 38a79453-939d-4649-9419-a2799636180d] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 623.551238] env[61806]: ERROR nova.compute.manager [instance: 38a79453-939d-4649-9419-a2799636180d] with excutils.save_and_reraise_exception(): [ 623.551238] env[61806]: ERROR nova.compute.manager [instance: 38a79453-939d-4649-9419-a2799636180d] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 623.551238] env[61806]: ERROR nova.compute.manager [instance: 38a79453-939d-4649-9419-a2799636180d] self.force_reraise() [ 623.551238] env[61806]: ERROR nova.compute.manager [instance: 38a79453-939d-4649-9419-a2799636180d] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 623.551238] env[61806]: ERROR nova.compute.manager [instance: 38a79453-939d-4649-9419-a2799636180d] raise self.value [ 623.551238] env[61806]: ERROR nova.compute.manager [instance: 38a79453-939d-4649-9419-a2799636180d] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 623.551238] env[61806]: ERROR nova.compute.manager [instance: 38a79453-939d-4649-9419-a2799636180d] updated_port = self._update_port( [ 623.551238] env[61806]: ERROR nova.compute.manager [instance: 38a79453-939d-4649-9419-a2799636180d] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 623.551238] env[61806]: ERROR nova.compute.manager [instance: 38a79453-939d-4649-9419-a2799636180d] _ensure_no_port_binding_failure(port) [ 623.551238] env[61806]: ERROR nova.compute.manager [instance: 38a79453-939d-4649-9419-a2799636180d] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 623.551238] env[61806]: ERROR nova.compute.manager [instance: 38a79453-939d-4649-9419-a2799636180d] raise exception.PortBindingFailed(port_id=port['id']) [ 623.551505] env[61806]: ERROR nova.compute.manager [instance: 38a79453-939d-4649-9419-a2799636180d] nova.exception.PortBindingFailed: Binding failed for port 6016a127-edf4-40d6-aba6-0a14614fe4ff, please check neutron logs for more information. [ 623.551505] env[61806]: ERROR nova.compute.manager [instance: 38a79453-939d-4649-9419-a2799636180d] [ 623.551505] env[61806]: INFO nova.compute.manager [None req-a6b8da3a-7829-486a-884d-0f28be02e38c tempest-ServersTestManualDisk-151630353 tempest-ServersTestManualDisk-151630353-project-member] [instance: 38a79453-939d-4649-9419-a2799636180d] Terminating instance [ 623.553780] env[61806]: DEBUG oslo_concurrency.lockutils [None req-a6b8da3a-7829-486a-884d-0f28be02e38c tempest-ServersTestManualDisk-151630353 tempest-ServersTestManualDisk-151630353-project-member] Acquiring lock "refresh_cache-38a79453-939d-4649-9419-a2799636180d" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 623.766949] env[61806]: DEBUG oslo_concurrency.lockutils [None req-98d962a0-2632-4a9a-8d12-70ce4acc382b tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.425s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 623.767754] env[61806]: DEBUG nova.compute.manager [None req-98d962a0-2632-4a9a-8d12-70ce4acc382b tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] [instance: b8249dc5-5561-454a-aa01-477f9f1fc0e1] Start building networks asynchronously for instance. {{(pid=61806) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 623.773414] env[61806]: DEBUG oslo_concurrency.lockutils [None req-ebff8349-d5ea-4a95-aee1-e4c9e8b6d3a4 tempest-ServerDiagnosticsTest-195484958 tempest-ServerDiagnosticsTest-195484958-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 14.645s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 623.796203] env[61806]: DEBUG nova.network.neutron [req-896214cf-a27e-4f7b-82d7-2e5987330353 req-43621815-a3b8-4bdb-aa10-683fcdfc5494 service nova] [instance: 38a79453-939d-4649-9419-a2799636180d] Instance cache missing network info. {{(pid=61806) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 623.928211] env[61806]: DEBUG nova.network.neutron [req-896214cf-a27e-4f7b-82d7-2e5987330353 req-43621815-a3b8-4bdb-aa10-683fcdfc5494 service nova] [instance: 38a79453-939d-4649-9419-a2799636180d] Updating instance_info_cache with network_info: [] {{(pid=61806) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 624.278747] env[61806]: DEBUG nova.compute.utils [None req-98d962a0-2632-4a9a-8d12-70ce4acc382b tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] Using /dev/sd instead of None {{(pid=61806) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 624.283257] env[61806]: DEBUG nova.compute.manager [None req-98d962a0-2632-4a9a-8d12-70ce4acc382b tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] [instance: b8249dc5-5561-454a-aa01-477f9f1fc0e1] Allocating IP information in the background. {{(pid=61806) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 624.283438] env[61806]: DEBUG nova.network.neutron [None req-98d962a0-2632-4a9a-8d12-70ce4acc382b tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] [instance: b8249dc5-5561-454a-aa01-477f9f1fc0e1] allocate_for_instance() {{(pid=61806) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 624.346417] env[61806]: DEBUG nova.policy [None req-98d962a0-2632-4a9a-8d12-70ce4acc382b tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'b1eac9072ccb4b3bbd23f240e941a76e', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '6e391e1267014fafbb8f5a3211299819', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61806) authorize /opt/stack/nova/nova/policy.py:201}} [ 624.372662] env[61806]: DEBUG oslo_service.periodic_task [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=61806) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 624.375543] env[61806]: DEBUG oslo_service.periodic_task [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Running periodic task ComputeManager._sync_scheduler_instance_info {{(pid=61806) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 624.430490] env[61806]: DEBUG oslo_concurrency.lockutils [req-896214cf-a27e-4f7b-82d7-2e5987330353 req-43621815-a3b8-4bdb-aa10-683fcdfc5494 service nova] Releasing lock "refresh_cache-38a79453-939d-4649-9419-a2799636180d" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 624.430879] env[61806]: DEBUG oslo_concurrency.lockutils [None req-a6b8da3a-7829-486a-884d-0f28be02e38c tempest-ServersTestManualDisk-151630353 tempest-ServersTestManualDisk-151630353-project-member] Acquired lock "refresh_cache-38a79453-939d-4649-9419-a2799636180d" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 624.431302] env[61806]: DEBUG nova.network.neutron [None req-a6b8da3a-7829-486a-884d-0f28be02e38c tempest-ServersTestManualDisk-151630353 tempest-ServersTestManualDisk-151630353-project-member] [instance: 38a79453-939d-4649-9419-a2799636180d] Building network info cache for instance {{(pid=61806) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 624.665309] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1263e192-f449-45de-b44d-08828e901a98 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 624.672691] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-47dddc8c-f63b-4490-a1a9-16281511b31b {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 624.705860] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-db6df623-4ec7-4ec5-ab3e-9d1c75bcef78 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 624.713637] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-214dddcb-600d-439e-89f6-018fd6ca7bad {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 624.727560] env[61806]: DEBUG nova.compute.provider_tree [None req-ebff8349-d5ea-4a95-aee1-e4c9e8b6d3a4 tempest-ServerDiagnosticsTest-195484958 tempest-ServerDiagnosticsTest-195484958-project-member] Inventory has not changed in ProviderTree for provider: a2858be1-fd22-4e08-979e-87ad25293407 {{(pid=61806) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 624.787031] env[61806]: DEBUG nova.compute.manager [None req-98d962a0-2632-4a9a-8d12-70ce4acc382b tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] [instance: b8249dc5-5561-454a-aa01-477f9f1fc0e1] Start building block device mappings for instance. {{(pid=61806) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 624.888227] env[61806]: DEBUG oslo_service.periodic_task [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=61806) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 624.888227] env[61806]: DEBUG nova.compute.manager [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Starting heal instance info cache {{(pid=61806) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9926}} [ 624.888227] env[61806]: DEBUG nova.compute.manager [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Rebuilding the list of instances to heal {{(pid=61806) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9930}} [ 624.918225] env[61806]: DEBUG nova.network.neutron [None req-98d962a0-2632-4a9a-8d12-70ce4acc382b tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] [instance: b8249dc5-5561-454a-aa01-477f9f1fc0e1] Successfully created port: 340e11ef-b16f-45b1-8a9b-cf7c087aa898 {{(pid=61806) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 624.975377] env[61806]: DEBUG nova.network.neutron [None req-a6b8da3a-7829-486a-884d-0f28be02e38c tempest-ServersTestManualDisk-151630353 tempest-ServersTestManualDisk-151630353-project-member] [instance: 38a79453-939d-4649-9419-a2799636180d] Instance cache missing network info. {{(pid=61806) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 625.225153] env[61806]: DEBUG nova.network.neutron [None req-a6b8da3a-7829-486a-884d-0f28be02e38c tempest-ServersTestManualDisk-151630353 tempest-ServersTestManualDisk-151630353-project-member] [instance: 38a79453-939d-4649-9419-a2799636180d] Updating instance_info_cache with network_info: [] {{(pid=61806) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 625.231972] env[61806]: DEBUG nova.scheduler.client.report [None req-ebff8349-d5ea-4a95-aee1-e4c9e8b6d3a4 tempest-ServerDiagnosticsTest-195484958 tempest-ServerDiagnosticsTest-195484958-project-member] Inventory has not changed for provider a2858be1-fd22-4e08-979e-87ad25293407 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 139, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61806) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 625.369117] env[61806]: DEBUG nova.compute.manager [req-1877df4b-6a32-4acf-9208-6f84e2060c43 req-7a835150-bcbf-4c2a-ad40-329e1e51f5e4 service nova] [instance: 38a79453-939d-4649-9419-a2799636180d] Received event network-vif-deleted-6016a127-edf4-40d6-aba6-0a14614fe4ff {{(pid=61806) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 625.388128] env[61806]: DEBUG nova.compute.manager [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] [instance: c2d1c9f6-f458-4843-982c-afdb2a90edae] Skipping network cache update for instance because it is Building. {{(pid=61806) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9939}} [ 625.388295] env[61806]: DEBUG nova.compute.manager [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] [instance: 65f3b2f3-6d25-44f0-95cb-f9a7e75f8e82] Skipping network cache update for instance because it is Building. {{(pid=61806) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9939}} [ 625.388480] env[61806]: DEBUG nova.compute.manager [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] [instance: 38a79453-939d-4649-9419-a2799636180d] Skipping network cache update for instance because it is Building. {{(pid=61806) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9939}} [ 625.388547] env[61806]: DEBUG nova.compute.manager [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] [instance: b8249dc5-5561-454a-aa01-477f9f1fc0e1] Skipping network cache update for instance because it is Building. {{(pid=61806) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9939}} [ 625.388667] env[61806]: DEBUG nova.compute.manager [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Didn't find any instances for network info cache update. {{(pid=61806) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10012}} [ 625.391022] env[61806]: DEBUG oslo_service.periodic_task [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=61806) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 625.391022] env[61806]: DEBUG oslo_service.periodic_task [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=61806) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 625.391022] env[61806]: DEBUG oslo_service.periodic_task [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=61806) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 625.391022] env[61806]: DEBUG oslo_service.periodic_task [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=61806) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 625.391022] env[61806]: DEBUG oslo_service.periodic_task [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=61806) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 625.391447] env[61806]: DEBUG oslo_service.periodic_task [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=61806) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 625.391447] env[61806]: DEBUG nova.compute.manager [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=61806) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10545}} [ 625.391534] env[61806]: DEBUG oslo_service.periodic_task [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Running periodic task ComputeManager.update_available_resource {{(pid=61806) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 625.728179] env[61806]: DEBUG oslo_concurrency.lockutils [None req-a6b8da3a-7829-486a-884d-0f28be02e38c tempest-ServersTestManualDisk-151630353 tempest-ServersTestManualDisk-151630353-project-member] Releasing lock "refresh_cache-38a79453-939d-4649-9419-a2799636180d" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 625.728562] env[61806]: DEBUG nova.compute.manager [None req-a6b8da3a-7829-486a-884d-0f28be02e38c tempest-ServersTestManualDisk-151630353 tempest-ServersTestManualDisk-151630353-project-member] [instance: 38a79453-939d-4649-9419-a2799636180d] Start destroying the instance on the hypervisor. {{(pid=61806) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 625.728756] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-a6b8da3a-7829-486a-884d-0f28be02e38c tempest-ServersTestManualDisk-151630353 tempest-ServersTestManualDisk-151630353-project-member] [instance: 38a79453-939d-4649-9419-a2799636180d] Destroying instance {{(pid=61806) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 625.729072] env[61806]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-073961c7-65c5-4808-97f6-980f661e7b14 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 625.740568] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-08b5baee-47fe-4d86-8555-9520a484d3f4 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 625.756527] env[61806]: DEBUG oslo_concurrency.lockutils [None req-ebff8349-d5ea-4a95-aee1-e4c9e8b6d3a4 tempest-ServerDiagnosticsTest-195484958 tempest-ServerDiagnosticsTest-195484958-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.982s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 625.756527] env[61806]: ERROR nova.compute.manager [None req-ebff8349-d5ea-4a95-aee1-e4c9e8b6d3a4 tempest-ServerDiagnosticsTest-195484958 tempest-ServerDiagnosticsTest-195484958-project-member] [instance: 79fc9ea2-c427-47ce-acdb-1002ee5d78c9] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port f3ed84a6-47ee-4b8a-875c-72f24cdfd5cb, please check neutron logs for more information. [ 625.756527] env[61806]: ERROR nova.compute.manager [instance: 79fc9ea2-c427-47ce-acdb-1002ee5d78c9] Traceback (most recent call last): [ 625.756527] env[61806]: ERROR nova.compute.manager [instance: 79fc9ea2-c427-47ce-acdb-1002ee5d78c9] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 625.756527] env[61806]: ERROR nova.compute.manager [instance: 79fc9ea2-c427-47ce-acdb-1002ee5d78c9] self.driver.spawn(context, instance, image_meta, [ 625.756527] env[61806]: ERROR nova.compute.manager [instance: 79fc9ea2-c427-47ce-acdb-1002ee5d78c9] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 625.756527] env[61806]: ERROR nova.compute.manager [instance: 79fc9ea2-c427-47ce-acdb-1002ee5d78c9] self._vmops.spawn(context, instance, image_meta, injected_files, [ 625.756527] env[61806]: ERROR nova.compute.manager [instance: 79fc9ea2-c427-47ce-acdb-1002ee5d78c9] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 625.756527] env[61806]: ERROR nova.compute.manager [instance: 79fc9ea2-c427-47ce-acdb-1002ee5d78c9] vm_ref = self.build_virtual_machine(instance, [ 625.756872] env[61806]: ERROR nova.compute.manager [instance: 79fc9ea2-c427-47ce-acdb-1002ee5d78c9] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 625.756872] env[61806]: ERROR nova.compute.manager [instance: 79fc9ea2-c427-47ce-acdb-1002ee5d78c9] vif_infos = vmwarevif.get_vif_info(self._session, [ 625.756872] env[61806]: ERROR nova.compute.manager [instance: 79fc9ea2-c427-47ce-acdb-1002ee5d78c9] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 625.756872] env[61806]: ERROR nova.compute.manager [instance: 79fc9ea2-c427-47ce-acdb-1002ee5d78c9] for vif in network_info: [ 625.756872] env[61806]: ERROR nova.compute.manager [instance: 79fc9ea2-c427-47ce-acdb-1002ee5d78c9] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 625.756872] env[61806]: ERROR nova.compute.manager [instance: 79fc9ea2-c427-47ce-acdb-1002ee5d78c9] return self._sync_wrapper(fn, *args, **kwargs) [ 625.756872] env[61806]: ERROR nova.compute.manager [instance: 79fc9ea2-c427-47ce-acdb-1002ee5d78c9] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 625.756872] env[61806]: ERROR nova.compute.manager [instance: 79fc9ea2-c427-47ce-acdb-1002ee5d78c9] self.wait() [ 625.756872] env[61806]: ERROR nova.compute.manager [instance: 79fc9ea2-c427-47ce-acdb-1002ee5d78c9] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 625.756872] env[61806]: ERROR nova.compute.manager [instance: 79fc9ea2-c427-47ce-acdb-1002ee5d78c9] self[:] = self._gt.wait() [ 625.756872] env[61806]: ERROR nova.compute.manager [instance: 79fc9ea2-c427-47ce-acdb-1002ee5d78c9] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 625.756872] env[61806]: ERROR nova.compute.manager [instance: 79fc9ea2-c427-47ce-acdb-1002ee5d78c9] return self._exit_event.wait() [ 625.756872] env[61806]: ERROR nova.compute.manager [instance: 79fc9ea2-c427-47ce-acdb-1002ee5d78c9] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 625.757236] env[61806]: ERROR nova.compute.manager [instance: 79fc9ea2-c427-47ce-acdb-1002ee5d78c9] result = hub.switch() [ 625.757236] env[61806]: ERROR nova.compute.manager [instance: 79fc9ea2-c427-47ce-acdb-1002ee5d78c9] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 625.757236] env[61806]: ERROR nova.compute.manager [instance: 79fc9ea2-c427-47ce-acdb-1002ee5d78c9] return self.greenlet.switch() [ 625.757236] env[61806]: ERROR nova.compute.manager [instance: 79fc9ea2-c427-47ce-acdb-1002ee5d78c9] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 625.757236] env[61806]: ERROR nova.compute.manager [instance: 79fc9ea2-c427-47ce-acdb-1002ee5d78c9] result = function(*args, **kwargs) [ 625.757236] env[61806]: ERROR nova.compute.manager [instance: 79fc9ea2-c427-47ce-acdb-1002ee5d78c9] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 625.757236] env[61806]: ERROR nova.compute.manager [instance: 79fc9ea2-c427-47ce-acdb-1002ee5d78c9] return func(*args, **kwargs) [ 625.757236] env[61806]: ERROR nova.compute.manager [instance: 79fc9ea2-c427-47ce-acdb-1002ee5d78c9] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 625.757236] env[61806]: ERROR nova.compute.manager [instance: 79fc9ea2-c427-47ce-acdb-1002ee5d78c9] raise e [ 625.757236] env[61806]: ERROR nova.compute.manager [instance: 79fc9ea2-c427-47ce-acdb-1002ee5d78c9] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 625.757236] env[61806]: ERROR nova.compute.manager [instance: 79fc9ea2-c427-47ce-acdb-1002ee5d78c9] nwinfo = self.network_api.allocate_for_instance( [ 625.757236] env[61806]: ERROR nova.compute.manager [instance: 79fc9ea2-c427-47ce-acdb-1002ee5d78c9] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 625.757236] env[61806]: ERROR nova.compute.manager [instance: 79fc9ea2-c427-47ce-acdb-1002ee5d78c9] created_port_ids = self._update_ports_for_instance( [ 625.757586] env[61806]: ERROR nova.compute.manager [instance: 79fc9ea2-c427-47ce-acdb-1002ee5d78c9] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 625.757586] env[61806]: ERROR nova.compute.manager [instance: 79fc9ea2-c427-47ce-acdb-1002ee5d78c9] with excutils.save_and_reraise_exception(): [ 625.757586] env[61806]: ERROR nova.compute.manager [instance: 79fc9ea2-c427-47ce-acdb-1002ee5d78c9] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 625.757586] env[61806]: ERROR nova.compute.manager [instance: 79fc9ea2-c427-47ce-acdb-1002ee5d78c9] self.force_reraise() [ 625.757586] env[61806]: ERROR nova.compute.manager [instance: 79fc9ea2-c427-47ce-acdb-1002ee5d78c9] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 625.757586] env[61806]: ERROR nova.compute.manager [instance: 79fc9ea2-c427-47ce-acdb-1002ee5d78c9] raise self.value [ 625.757586] env[61806]: ERROR nova.compute.manager [instance: 79fc9ea2-c427-47ce-acdb-1002ee5d78c9] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 625.757586] env[61806]: ERROR nova.compute.manager [instance: 79fc9ea2-c427-47ce-acdb-1002ee5d78c9] updated_port = self._update_port( [ 625.757586] env[61806]: ERROR nova.compute.manager [instance: 79fc9ea2-c427-47ce-acdb-1002ee5d78c9] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 625.757586] env[61806]: ERROR nova.compute.manager [instance: 79fc9ea2-c427-47ce-acdb-1002ee5d78c9] _ensure_no_port_binding_failure(port) [ 625.757586] env[61806]: ERROR nova.compute.manager [instance: 79fc9ea2-c427-47ce-acdb-1002ee5d78c9] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 625.757586] env[61806]: ERROR nova.compute.manager [instance: 79fc9ea2-c427-47ce-acdb-1002ee5d78c9] raise exception.PortBindingFailed(port_id=port['id']) [ 625.757912] env[61806]: ERROR nova.compute.manager [instance: 79fc9ea2-c427-47ce-acdb-1002ee5d78c9] nova.exception.PortBindingFailed: Binding failed for port f3ed84a6-47ee-4b8a-875c-72f24cdfd5cb, please check neutron logs for more information. [ 625.757912] env[61806]: ERROR nova.compute.manager [instance: 79fc9ea2-c427-47ce-acdb-1002ee5d78c9] [ 625.757912] env[61806]: DEBUG nova.compute.utils [None req-ebff8349-d5ea-4a95-aee1-e4c9e8b6d3a4 tempest-ServerDiagnosticsTest-195484958 tempest-ServerDiagnosticsTest-195484958-project-member] [instance: 79fc9ea2-c427-47ce-acdb-1002ee5d78c9] Binding failed for port f3ed84a6-47ee-4b8a-875c-72f24cdfd5cb, please check neutron logs for more information. {{(pid=61806) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 625.758667] env[61806]: DEBUG oslo_concurrency.lockutils [None req-24088d08-d3d7-4f62-ad63-91029bd9b22d tempest-ServersTestMultiNic-1480543082 tempest-ServersTestMultiNic-1480543082-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 13.791s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 625.763796] env[61806]: DEBUG nova.compute.manager [None req-ebff8349-d5ea-4a95-aee1-e4c9e8b6d3a4 tempest-ServerDiagnosticsTest-195484958 tempest-ServerDiagnosticsTest-195484958-project-member] [instance: 79fc9ea2-c427-47ce-acdb-1002ee5d78c9] Build of instance 79fc9ea2-c427-47ce-acdb-1002ee5d78c9 was re-scheduled: Binding failed for port f3ed84a6-47ee-4b8a-875c-72f24cdfd5cb, please check neutron logs for more information. {{(pid=61806) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 625.764702] env[61806]: DEBUG nova.compute.manager [None req-ebff8349-d5ea-4a95-aee1-e4c9e8b6d3a4 tempest-ServerDiagnosticsTest-195484958 tempest-ServerDiagnosticsTest-195484958-project-member] [instance: 79fc9ea2-c427-47ce-acdb-1002ee5d78c9] Unplugging VIFs for instance {{(pid=61806) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 625.765381] env[61806]: DEBUG oslo_concurrency.lockutils [None req-ebff8349-d5ea-4a95-aee1-e4c9e8b6d3a4 tempest-ServerDiagnosticsTest-195484958 tempest-ServerDiagnosticsTest-195484958-project-member] Acquiring lock "refresh_cache-79fc9ea2-c427-47ce-acdb-1002ee5d78c9" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 625.765562] env[61806]: DEBUG oslo_concurrency.lockutils [None req-ebff8349-d5ea-4a95-aee1-e4c9e8b6d3a4 tempest-ServerDiagnosticsTest-195484958 tempest-ServerDiagnosticsTest-195484958-project-member] Acquired lock "refresh_cache-79fc9ea2-c427-47ce-acdb-1002ee5d78c9" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 625.765731] env[61806]: DEBUG nova.network.neutron [None req-ebff8349-d5ea-4a95-aee1-e4c9e8b6d3a4 tempest-ServerDiagnosticsTest-195484958 tempest-ServerDiagnosticsTest-195484958-project-member] [instance: 79fc9ea2-c427-47ce-acdb-1002ee5d78c9] Building network info cache for instance {{(pid=61806) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 625.777293] env[61806]: WARNING nova.virt.vmwareapi.vmops [None req-a6b8da3a-7829-486a-884d-0f28be02e38c tempest-ServersTestManualDisk-151630353 tempest-ServersTestManualDisk-151630353-project-member] [instance: 38a79453-939d-4649-9419-a2799636180d] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 38a79453-939d-4649-9419-a2799636180d could not be found. [ 625.777293] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-a6b8da3a-7829-486a-884d-0f28be02e38c tempest-ServersTestManualDisk-151630353 tempest-ServersTestManualDisk-151630353-project-member] [instance: 38a79453-939d-4649-9419-a2799636180d] Instance destroyed {{(pid=61806) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 625.777293] env[61806]: INFO nova.compute.manager [None req-a6b8da3a-7829-486a-884d-0f28be02e38c tempest-ServersTestManualDisk-151630353 tempest-ServersTestManualDisk-151630353-project-member] [instance: 38a79453-939d-4649-9419-a2799636180d] Took 0.05 seconds to destroy the instance on the hypervisor. [ 625.777472] env[61806]: DEBUG oslo.service.loopingcall [None req-a6b8da3a-7829-486a-884d-0f28be02e38c tempest-ServersTestManualDisk-151630353 tempest-ServersTestManualDisk-151630353-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61806) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 625.778143] env[61806]: DEBUG nova.compute.manager [-] [instance: 38a79453-939d-4649-9419-a2799636180d] Deallocating network for instance {{(pid=61806) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 625.778143] env[61806]: DEBUG nova.network.neutron [-] [instance: 38a79453-939d-4649-9419-a2799636180d] deallocate_for_instance() {{(pid=61806) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 625.803527] env[61806]: DEBUG nova.network.neutron [-] [instance: 38a79453-939d-4649-9419-a2799636180d] Instance cache missing network info. {{(pid=61806) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 626.441284] env[61806]: DEBUG nova.compute.manager [None req-98d962a0-2632-4a9a-8d12-70ce4acc382b tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] [instance: b8249dc5-5561-454a-aa01-477f9f1fc0e1] Start spawning the instance on the hypervisor. {{(pid=61806) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 626.443784] env[61806]: DEBUG oslo_concurrency.lockutils [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 626.447619] env[61806]: DEBUG nova.network.neutron [-] [instance: 38a79453-939d-4649-9419-a2799636180d] Updating instance_info_cache with network_info: [] {{(pid=61806) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 626.489658] env[61806]: DEBUG nova.virt.hardware [None req-98d962a0-2632-4a9a-8d12-70ce4acc382b tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-15T18:05:10Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-15T18:04:52Z,direct_url=,disk_format='vmdk',id=ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='84ba83fce2c349a6988173c5d6fc3b07',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-15T18:04:53Z,virtual_size=,visibility=), allow threads: False {{(pid=61806) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 626.490071] env[61806]: DEBUG nova.virt.hardware [None req-98d962a0-2632-4a9a-8d12-70ce4acc382b tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] Flavor limits 0:0:0 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 626.490177] env[61806]: DEBUG nova.virt.hardware [None req-98d962a0-2632-4a9a-8d12-70ce4acc382b tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] Image limits 0:0:0 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 626.490367] env[61806]: DEBUG nova.virt.hardware [None req-98d962a0-2632-4a9a-8d12-70ce4acc382b tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] Flavor pref 0:0:0 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 626.490514] env[61806]: DEBUG nova.virt.hardware [None req-98d962a0-2632-4a9a-8d12-70ce4acc382b tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] Image pref 0:0:0 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 626.490659] env[61806]: DEBUG nova.virt.hardware [None req-98d962a0-2632-4a9a-8d12-70ce4acc382b tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 626.490862] env[61806]: DEBUG nova.virt.hardware [None req-98d962a0-2632-4a9a-8d12-70ce4acc382b tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61806) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 626.491030] env[61806]: DEBUG nova.virt.hardware [None req-98d962a0-2632-4a9a-8d12-70ce4acc382b tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61806) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 626.491198] env[61806]: DEBUG nova.virt.hardware [None req-98d962a0-2632-4a9a-8d12-70ce4acc382b tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] Got 1 possible topologies {{(pid=61806) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 626.491359] env[61806]: DEBUG nova.virt.hardware [None req-98d962a0-2632-4a9a-8d12-70ce4acc382b tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61806) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 626.491524] env[61806]: DEBUG nova.virt.hardware [None req-98d962a0-2632-4a9a-8d12-70ce4acc382b tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61806) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 626.492332] env[61806]: DEBUG nova.network.neutron [None req-ebff8349-d5ea-4a95-aee1-e4c9e8b6d3a4 tempest-ServerDiagnosticsTest-195484958 tempest-ServerDiagnosticsTest-195484958-project-member] [instance: 79fc9ea2-c427-47ce-acdb-1002ee5d78c9] Instance cache missing network info. {{(pid=61806) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 626.494779] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e6fecf5b-2858-41cb-a622-1ee9929347c1 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 626.503040] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3d778285-624a-4e94-8056-914f9dad1718 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 626.635595] env[61806]: DEBUG nova.network.neutron [None req-ebff8349-d5ea-4a95-aee1-e4c9e8b6d3a4 tempest-ServerDiagnosticsTest-195484958 tempest-ServerDiagnosticsTest-195484958-project-member] [instance: 79fc9ea2-c427-47ce-acdb-1002ee5d78c9] Updating instance_info_cache with network_info: [] {{(pid=61806) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 626.718298] env[61806]: ERROR nova.compute.manager [None req-98d962a0-2632-4a9a-8d12-70ce4acc382b tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 340e11ef-b16f-45b1-8a9b-cf7c087aa898, please check neutron logs for more information. [ 626.718298] env[61806]: ERROR nova.compute.manager Traceback (most recent call last): [ 626.718298] env[61806]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 626.718298] env[61806]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 626.718298] env[61806]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 626.718298] env[61806]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 626.718298] env[61806]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 626.718298] env[61806]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 626.718298] env[61806]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 626.718298] env[61806]: ERROR nova.compute.manager self.force_reraise() [ 626.718298] env[61806]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 626.718298] env[61806]: ERROR nova.compute.manager raise self.value [ 626.718298] env[61806]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 626.718298] env[61806]: ERROR nova.compute.manager updated_port = self._update_port( [ 626.718298] env[61806]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 626.718298] env[61806]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 626.718788] env[61806]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 626.718788] env[61806]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 626.718788] env[61806]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 340e11ef-b16f-45b1-8a9b-cf7c087aa898, please check neutron logs for more information. [ 626.718788] env[61806]: ERROR nova.compute.manager [ 626.718788] env[61806]: Traceback (most recent call last): [ 626.718788] env[61806]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 626.718788] env[61806]: listener.cb(fileno) [ 626.718788] env[61806]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 626.718788] env[61806]: result = function(*args, **kwargs) [ 626.718788] env[61806]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 626.718788] env[61806]: return func(*args, **kwargs) [ 626.718788] env[61806]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 626.718788] env[61806]: raise e [ 626.718788] env[61806]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 626.718788] env[61806]: nwinfo = self.network_api.allocate_for_instance( [ 626.718788] env[61806]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 626.718788] env[61806]: created_port_ids = self._update_ports_for_instance( [ 626.718788] env[61806]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 626.718788] env[61806]: with excutils.save_and_reraise_exception(): [ 626.718788] env[61806]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 626.718788] env[61806]: self.force_reraise() [ 626.718788] env[61806]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 626.718788] env[61806]: raise self.value [ 626.718788] env[61806]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 626.718788] env[61806]: updated_port = self._update_port( [ 626.718788] env[61806]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 626.718788] env[61806]: _ensure_no_port_binding_failure(port) [ 626.718788] env[61806]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 626.718788] env[61806]: raise exception.PortBindingFailed(port_id=port['id']) [ 626.719638] env[61806]: nova.exception.PortBindingFailed: Binding failed for port 340e11ef-b16f-45b1-8a9b-cf7c087aa898, please check neutron logs for more information. [ 626.719638] env[61806]: Removing descriptor: 16 [ 626.719638] env[61806]: ERROR nova.compute.manager [None req-98d962a0-2632-4a9a-8d12-70ce4acc382b tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] [instance: b8249dc5-5561-454a-aa01-477f9f1fc0e1] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 340e11ef-b16f-45b1-8a9b-cf7c087aa898, please check neutron logs for more information. [ 626.719638] env[61806]: ERROR nova.compute.manager [instance: b8249dc5-5561-454a-aa01-477f9f1fc0e1] Traceback (most recent call last): [ 626.719638] env[61806]: ERROR nova.compute.manager [instance: b8249dc5-5561-454a-aa01-477f9f1fc0e1] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 626.719638] env[61806]: ERROR nova.compute.manager [instance: b8249dc5-5561-454a-aa01-477f9f1fc0e1] yield resources [ 626.719638] env[61806]: ERROR nova.compute.manager [instance: b8249dc5-5561-454a-aa01-477f9f1fc0e1] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 626.719638] env[61806]: ERROR nova.compute.manager [instance: b8249dc5-5561-454a-aa01-477f9f1fc0e1] self.driver.spawn(context, instance, image_meta, [ 626.719638] env[61806]: ERROR nova.compute.manager [instance: b8249dc5-5561-454a-aa01-477f9f1fc0e1] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 626.719638] env[61806]: ERROR nova.compute.manager [instance: b8249dc5-5561-454a-aa01-477f9f1fc0e1] self._vmops.spawn(context, instance, image_meta, injected_files, [ 626.719638] env[61806]: ERROR nova.compute.manager [instance: b8249dc5-5561-454a-aa01-477f9f1fc0e1] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 626.719638] env[61806]: ERROR nova.compute.manager [instance: b8249dc5-5561-454a-aa01-477f9f1fc0e1] vm_ref = self.build_virtual_machine(instance, [ 626.720017] env[61806]: ERROR nova.compute.manager [instance: b8249dc5-5561-454a-aa01-477f9f1fc0e1] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 626.720017] env[61806]: ERROR nova.compute.manager [instance: b8249dc5-5561-454a-aa01-477f9f1fc0e1] vif_infos = vmwarevif.get_vif_info(self._session, [ 626.720017] env[61806]: ERROR nova.compute.manager [instance: b8249dc5-5561-454a-aa01-477f9f1fc0e1] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 626.720017] env[61806]: ERROR nova.compute.manager [instance: b8249dc5-5561-454a-aa01-477f9f1fc0e1] for vif in network_info: [ 626.720017] env[61806]: ERROR nova.compute.manager [instance: b8249dc5-5561-454a-aa01-477f9f1fc0e1] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 626.720017] env[61806]: ERROR nova.compute.manager [instance: b8249dc5-5561-454a-aa01-477f9f1fc0e1] return self._sync_wrapper(fn, *args, **kwargs) [ 626.720017] env[61806]: ERROR nova.compute.manager [instance: b8249dc5-5561-454a-aa01-477f9f1fc0e1] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 626.720017] env[61806]: ERROR nova.compute.manager [instance: b8249dc5-5561-454a-aa01-477f9f1fc0e1] self.wait() [ 626.720017] env[61806]: ERROR nova.compute.manager [instance: b8249dc5-5561-454a-aa01-477f9f1fc0e1] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 626.720017] env[61806]: ERROR nova.compute.manager [instance: b8249dc5-5561-454a-aa01-477f9f1fc0e1] self[:] = self._gt.wait() [ 626.720017] env[61806]: ERROR nova.compute.manager [instance: b8249dc5-5561-454a-aa01-477f9f1fc0e1] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 626.720017] env[61806]: ERROR nova.compute.manager [instance: b8249dc5-5561-454a-aa01-477f9f1fc0e1] return self._exit_event.wait() [ 626.720017] env[61806]: ERROR nova.compute.manager [instance: b8249dc5-5561-454a-aa01-477f9f1fc0e1] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 626.720489] env[61806]: ERROR nova.compute.manager [instance: b8249dc5-5561-454a-aa01-477f9f1fc0e1] result = hub.switch() [ 626.720489] env[61806]: ERROR nova.compute.manager [instance: b8249dc5-5561-454a-aa01-477f9f1fc0e1] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 626.720489] env[61806]: ERROR nova.compute.manager [instance: b8249dc5-5561-454a-aa01-477f9f1fc0e1] return self.greenlet.switch() [ 626.720489] env[61806]: ERROR nova.compute.manager [instance: b8249dc5-5561-454a-aa01-477f9f1fc0e1] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 626.720489] env[61806]: ERROR nova.compute.manager [instance: b8249dc5-5561-454a-aa01-477f9f1fc0e1] result = function(*args, **kwargs) [ 626.720489] env[61806]: ERROR nova.compute.manager [instance: b8249dc5-5561-454a-aa01-477f9f1fc0e1] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 626.720489] env[61806]: ERROR nova.compute.manager [instance: b8249dc5-5561-454a-aa01-477f9f1fc0e1] return func(*args, **kwargs) [ 626.720489] env[61806]: ERROR nova.compute.manager [instance: b8249dc5-5561-454a-aa01-477f9f1fc0e1] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 626.720489] env[61806]: ERROR nova.compute.manager [instance: b8249dc5-5561-454a-aa01-477f9f1fc0e1] raise e [ 626.720489] env[61806]: ERROR nova.compute.manager [instance: b8249dc5-5561-454a-aa01-477f9f1fc0e1] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 626.720489] env[61806]: ERROR nova.compute.manager [instance: b8249dc5-5561-454a-aa01-477f9f1fc0e1] nwinfo = self.network_api.allocate_for_instance( [ 626.720489] env[61806]: ERROR nova.compute.manager [instance: b8249dc5-5561-454a-aa01-477f9f1fc0e1] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 626.720489] env[61806]: ERROR nova.compute.manager [instance: b8249dc5-5561-454a-aa01-477f9f1fc0e1] created_port_ids = self._update_ports_for_instance( [ 626.720859] env[61806]: ERROR nova.compute.manager [instance: b8249dc5-5561-454a-aa01-477f9f1fc0e1] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 626.720859] env[61806]: ERROR nova.compute.manager [instance: b8249dc5-5561-454a-aa01-477f9f1fc0e1] with excutils.save_and_reraise_exception(): [ 626.720859] env[61806]: ERROR nova.compute.manager [instance: b8249dc5-5561-454a-aa01-477f9f1fc0e1] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 626.720859] env[61806]: ERROR nova.compute.manager [instance: b8249dc5-5561-454a-aa01-477f9f1fc0e1] self.force_reraise() [ 626.720859] env[61806]: ERROR nova.compute.manager [instance: b8249dc5-5561-454a-aa01-477f9f1fc0e1] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 626.720859] env[61806]: ERROR nova.compute.manager [instance: b8249dc5-5561-454a-aa01-477f9f1fc0e1] raise self.value [ 626.720859] env[61806]: ERROR nova.compute.manager [instance: b8249dc5-5561-454a-aa01-477f9f1fc0e1] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 626.720859] env[61806]: ERROR nova.compute.manager [instance: b8249dc5-5561-454a-aa01-477f9f1fc0e1] updated_port = self._update_port( [ 626.720859] env[61806]: ERROR nova.compute.manager [instance: b8249dc5-5561-454a-aa01-477f9f1fc0e1] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 626.720859] env[61806]: ERROR nova.compute.manager [instance: b8249dc5-5561-454a-aa01-477f9f1fc0e1] _ensure_no_port_binding_failure(port) [ 626.720859] env[61806]: ERROR nova.compute.manager [instance: b8249dc5-5561-454a-aa01-477f9f1fc0e1] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 626.720859] env[61806]: ERROR nova.compute.manager [instance: b8249dc5-5561-454a-aa01-477f9f1fc0e1] raise exception.PortBindingFailed(port_id=port['id']) [ 626.721215] env[61806]: ERROR nova.compute.manager [instance: b8249dc5-5561-454a-aa01-477f9f1fc0e1] nova.exception.PortBindingFailed: Binding failed for port 340e11ef-b16f-45b1-8a9b-cf7c087aa898, please check neutron logs for more information. [ 626.721215] env[61806]: ERROR nova.compute.manager [instance: b8249dc5-5561-454a-aa01-477f9f1fc0e1] [ 626.721215] env[61806]: INFO nova.compute.manager [None req-98d962a0-2632-4a9a-8d12-70ce4acc382b tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] [instance: b8249dc5-5561-454a-aa01-477f9f1fc0e1] Terminating instance [ 626.722112] env[61806]: DEBUG oslo_concurrency.lockutils [None req-98d962a0-2632-4a9a-8d12-70ce4acc382b tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] Acquiring lock "refresh_cache-b8249dc5-5561-454a-aa01-477f9f1fc0e1" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 626.722279] env[61806]: DEBUG oslo_concurrency.lockutils [None req-98d962a0-2632-4a9a-8d12-70ce4acc382b tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] Acquired lock "refresh_cache-b8249dc5-5561-454a-aa01-477f9f1fc0e1" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 626.722511] env[61806]: DEBUG nova.network.neutron [None req-98d962a0-2632-4a9a-8d12-70ce4acc382b tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] [instance: b8249dc5-5561-454a-aa01-477f9f1fc0e1] Building network info cache for instance {{(pid=61806) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 626.849316] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cfeb9577-5f1a-485a-adc3-d907873766d5 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 626.860935] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-68964035-2194-4d57-be5d-cddb74caa91b {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 626.901039] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2832607c-7d07-4b7b-882d-8a9ebe7f0e11 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 626.908465] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-981fd07c-dccb-4967-b0a0-e68936bb6a31 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 626.924023] env[61806]: DEBUG nova.compute.provider_tree [None req-24088d08-d3d7-4f62-ad63-91029bd9b22d tempest-ServersTestMultiNic-1480543082 tempest-ServersTestMultiNic-1480543082-project-member] Inventory has not changed in ProviderTree for provider: a2858be1-fd22-4e08-979e-87ad25293407 {{(pid=61806) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 626.953271] env[61806]: INFO nova.compute.manager [-] [instance: 38a79453-939d-4649-9419-a2799636180d] Took 1.18 seconds to deallocate network for instance. [ 626.955804] env[61806]: DEBUG nova.compute.claims [None req-a6b8da3a-7829-486a-884d-0f28be02e38c tempest-ServersTestManualDisk-151630353 tempest-ServersTestManualDisk-151630353-project-member] [instance: 38a79453-939d-4649-9419-a2799636180d] Aborting claim: {{(pid=61806) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 626.956411] env[61806]: DEBUG oslo_concurrency.lockutils [None req-a6b8da3a-7829-486a-884d-0f28be02e38c tempest-ServersTestManualDisk-151630353 tempest-ServersTestManualDisk-151630353-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 627.140942] env[61806]: DEBUG oslo_concurrency.lockutils [None req-ebff8349-d5ea-4a95-aee1-e4c9e8b6d3a4 tempest-ServerDiagnosticsTest-195484958 tempest-ServerDiagnosticsTest-195484958-project-member] Releasing lock "refresh_cache-79fc9ea2-c427-47ce-acdb-1002ee5d78c9" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 627.141677] env[61806]: DEBUG nova.compute.manager [None req-ebff8349-d5ea-4a95-aee1-e4c9e8b6d3a4 tempest-ServerDiagnosticsTest-195484958 tempest-ServerDiagnosticsTest-195484958-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61806) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 627.141945] env[61806]: DEBUG nova.compute.manager [None req-ebff8349-d5ea-4a95-aee1-e4c9e8b6d3a4 tempest-ServerDiagnosticsTest-195484958 tempest-ServerDiagnosticsTest-195484958-project-member] [instance: 79fc9ea2-c427-47ce-acdb-1002ee5d78c9] Deallocating network for instance {{(pid=61806) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 627.142196] env[61806]: DEBUG nova.network.neutron [None req-ebff8349-d5ea-4a95-aee1-e4c9e8b6d3a4 tempest-ServerDiagnosticsTest-195484958 tempest-ServerDiagnosticsTest-195484958-project-member] [instance: 79fc9ea2-c427-47ce-acdb-1002ee5d78c9] deallocate_for_instance() {{(pid=61806) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 627.165255] env[61806]: DEBUG nova.network.neutron [None req-ebff8349-d5ea-4a95-aee1-e4c9e8b6d3a4 tempest-ServerDiagnosticsTest-195484958 tempest-ServerDiagnosticsTest-195484958-project-member] [instance: 79fc9ea2-c427-47ce-acdb-1002ee5d78c9] Instance cache missing network info. {{(pid=61806) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 627.251835] env[61806]: DEBUG nova.network.neutron [None req-98d962a0-2632-4a9a-8d12-70ce4acc382b tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] [instance: b8249dc5-5561-454a-aa01-477f9f1fc0e1] Instance cache missing network info. {{(pid=61806) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 627.428328] env[61806]: DEBUG nova.scheduler.client.report [None req-24088d08-d3d7-4f62-ad63-91029bd9b22d tempest-ServersTestMultiNic-1480543082 tempest-ServersTestMultiNic-1480543082-project-member] Inventory has not changed for provider a2858be1-fd22-4e08-979e-87ad25293407 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 139, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61806) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 627.585094] env[61806]: DEBUG nova.network.neutron [None req-98d962a0-2632-4a9a-8d12-70ce4acc382b tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] [instance: b8249dc5-5561-454a-aa01-477f9f1fc0e1] Updating instance_info_cache with network_info: [] {{(pid=61806) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 627.669167] env[61806]: DEBUG nova.network.neutron [None req-ebff8349-d5ea-4a95-aee1-e4c9e8b6d3a4 tempest-ServerDiagnosticsTest-195484958 tempest-ServerDiagnosticsTest-195484958-project-member] [instance: 79fc9ea2-c427-47ce-acdb-1002ee5d78c9] Updating instance_info_cache with network_info: [] {{(pid=61806) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 627.694618] env[61806]: DEBUG nova.compute.manager [req-ae4f70c7-c9c6-488b-aae5-6ddb5af72c0a req-74d0fd9b-72fc-4ae5-9061-103d30591758 service nova] [instance: b8249dc5-5561-454a-aa01-477f9f1fc0e1] Received event network-changed-340e11ef-b16f-45b1-8a9b-cf7c087aa898 {{(pid=61806) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 627.694813] env[61806]: DEBUG nova.compute.manager [req-ae4f70c7-c9c6-488b-aae5-6ddb5af72c0a req-74d0fd9b-72fc-4ae5-9061-103d30591758 service nova] [instance: b8249dc5-5561-454a-aa01-477f9f1fc0e1] Refreshing instance network info cache due to event network-changed-340e11ef-b16f-45b1-8a9b-cf7c087aa898. {{(pid=61806) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 627.695028] env[61806]: DEBUG oslo_concurrency.lockutils [req-ae4f70c7-c9c6-488b-aae5-6ddb5af72c0a req-74d0fd9b-72fc-4ae5-9061-103d30591758 service nova] Acquiring lock "refresh_cache-b8249dc5-5561-454a-aa01-477f9f1fc0e1" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 627.935013] env[61806]: DEBUG oslo_concurrency.lockutils [None req-24088d08-d3d7-4f62-ad63-91029bd9b22d tempest-ServersTestMultiNic-1480543082 tempest-ServersTestMultiNic-1480543082-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.177s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 627.935677] env[61806]: ERROR nova.compute.manager [None req-24088d08-d3d7-4f62-ad63-91029bd9b22d tempest-ServersTestMultiNic-1480543082 tempest-ServersTestMultiNic-1480543082-project-member] [instance: c2d1c9f6-f458-4843-982c-afdb2a90edae] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 3e148a81-f55b-4f13-adfc-ca47776ae7c5, please check neutron logs for more information. [ 627.935677] env[61806]: ERROR nova.compute.manager [instance: c2d1c9f6-f458-4843-982c-afdb2a90edae] Traceback (most recent call last): [ 627.935677] env[61806]: ERROR nova.compute.manager [instance: c2d1c9f6-f458-4843-982c-afdb2a90edae] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 627.935677] env[61806]: ERROR nova.compute.manager [instance: c2d1c9f6-f458-4843-982c-afdb2a90edae] self.driver.spawn(context, instance, image_meta, [ 627.935677] env[61806]: ERROR nova.compute.manager [instance: c2d1c9f6-f458-4843-982c-afdb2a90edae] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 627.935677] env[61806]: ERROR nova.compute.manager [instance: c2d1c9f6-f458-4843-982c-afdb2a90edae] self._vmops.spawn(context, instance, image_meta, injected_files, [ 627.935677] env[61806]: ERROR nova.compute.manager [instance: c2d1c9f6-f458-4843-982c-afdb2a90edae] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 627.935677] env[61806]: ERROR nova.compute.manager [instance: c2d1c9f6-f458-4843-982c-afdb2a90edae] vm_ref = self.build_virtual_machine(instance, [ 627.935677] env[61806]: ERROR nova.compute.manager [instance: c2d1c9f6-f458-4843-982c-afdb2a90edae] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 627.935677] env[61806]: ERROR nova.compute.manager [instance: c2d1c9f6-f458-4843-982c-afdb2a90edae] vif_infos = vmwarevif.get_vif_info(self._session, [ 627.935677] env[61806]: ERROR nova.compute.manager [instance: c2d1c9f6-f458-4843-982c-afdb2a90edae] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 627.935982] env[61806]: ERROR nova.compute.manager [instance: c2d1c9f6-f458-4843-982c-afdb2a90edae] for vif in network_info: [ 627.935982] env[61806]: ERROR nova.compute.manager [instance: c2d1c9f6-f458-4843-982c-afdb2a90edae] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 627.935982] env[61806]: ERROR nova.compute.manager [instance: c2d1c9f6-f458-4843-982c-afdb2a90edae] return self._sync_wrapper(fn, *args, **kwargs) [ 627.935982] env[61806]: ERROR nova.compute.manager [instance: c2d1c9f6-f458-4843-982c-afdb2a90edae] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 627.935982] env[61806]: ERROR nova.compute.manager [instance: c2d1c9f6-f458-4843-982c-afdb2a90edae] self.wait() [ 627.935982] env[61806]: ERROR nova.compute.manager [instance: c2d1c9f6-f458-4843-982c-afdb2a90edae] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 627.935982] env[61806]: ERROR nova.compute.manager [instance: c2d1c9f6-f458-4843-982c-afdb2a90edae] self[:] = self._gt.wait() [ 627.935982] env[61806]: ERROR nova.compute.manager [instance: c2d1c9f6-f458-4843-982c-afdb2a90edae] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 627.935982] env[61806]: ERROR nova.compute.manager [instance: c2d1c9f6-f458-4843-982c-afdb2a90edae] return self._exit_event.wait() [ 627.935982] env[61806]: ERROR nova.compute.manager [instance: c2d1c9f6-f458-4843-982c-afdb2a90edae] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 627.935982] env[61806]: ERROR nova.compute.manager [instance: c2d1c9f6-f458-4843-982c-afdb2a90edae] result = hub.switch() [ 627.935982] env[61806]: ERROR nova.compute.manager [instance: c2d1c9f6-f458-4843-982c-afdb2a90edae] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 627.935982] env[61806]: ERROR nova.compute.manager [instance: c2d1c9f6-f458-4843-982c-afdb2a90edae] return self.greenlet.switch() [ 627.936293] env[61806]: ERROR nova.compute.manager [instance: c2d1c9f6-f458-4843-982c-afdb2a90edae] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 627.936293] env[61806]: ERROR nova.compute.manager [instance: c2d1c9f6-f458-4843-982c-afdb2a90edae] result = function(*args, **kwargs) [ 627.936293] env[61806]: ERROR nova.compute.manager [instance: c2d1c9f6-f458-4843-982c-afdb2a90edae] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 627.936293] env[61806]: ERROR nova.compute.manager [instance: c2d1c9f6-f458-4843-982c-afdb2a90edae] return func(*args, **kwargs) [ 627.936293] env[61806]: ERROR nova.compute.manager [instance: c2d1c9f6-f458-4843-982c-afdb2a90edae] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 627.936293] env[61806]: ERROR nova.compute.manager [instance: c2d1c9f6-f458-4843-982c-afdb2a90edae] raise e [ 627.936293] env[61806]: ERROR nova.compute.manager [instance: c2d1c9f6-f458-4843-982c-afdb2a90edae] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 627.936293] env[61806]: ERROR nova.compute.manager [instance: c2d1c9f6-f458-4843-982c-afdb2a90edae] nwinfo = self.network_api.allocate_for_instance( [ 627.936293] env[61806]: ERROR nova.compute.manager [instance: c2d1c9f6-f458-4843-982c-afdb2a90edae] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 627.936293] env[61806]: ERROR nova.compute.manager [instance: c2d1c9f6-f458-4843-982c-afdb2a90edae] created_port_ids = self._update_ports_for_instance( [ 627.936293] env[61806]: ERROR nova.compute.manager [instance: c2d1c9f6-f458-4843-982c-afdb2a90edae] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 627.936293] env[61806]: ERROR nova.compute.manager [instance: c2d1c9f6-f458-4843-982c-afdb2a90edae] with excutils.save_and_reraise_exception(): [ 627.936293] env[61806]: ERROR nova.compute.manager [instance: c2d1c9f6-f458-4843-982c-afdb2a90edae] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 627.936588] env[61806]: ERROR nova.compute.manager [instance: c2d1c9f6-f458-4843-982c-afdb2a90edae] self.force_reraise() [ 627.936588] env[61806]: ERROR nova.compute.manager [instance: c2d1c9f6-f458-4843-982c-afdb2a90edae] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 627.936588] env[61806]: ERROR nova.compute.manager [instance: c2d1c9f6-f458-4843-982c-afdb2a90edae] raise self.value [ 627.936588] env[61806]: ERROR nova.compute.manager [instance: c2d1c9f6-f458-4843-982c-afdb2a90edae] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 627.936588] env[61806]: ERROR nova.compute.manager [instance: c2d1c9f6-f458-4843-982c-afdb2a90edae] updated_port = self._update_port( [ 627.936588] env[61806]: ERROR nova.compute.manager [instance: c2d1c9f6-f458-4843-982c-afdb2a90edae] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 627.936588] env[61806]: ERROR nova.compute.manager [instance: c2d1c9f6-f458-4843-982c-afdb2a90edae] _ensure_no_port_binding_failure(port) [ 627.936588] env[61806]: ERROR nova.compute.manager [instance: c2d1c9f6-f458-4843-982c-afdb2a90edae] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 627.936588] env[61806]: ERROR nova.compute.manager [instance: c2d1c9f6-f458-4843-982c-afdb2a90edae] raise exception.PortBindingFailed(port_id=port['id']) [ 627.936588] env[61806]: ERROR nova.compute.manager [instance: c2d1c9f6-f458-4843-982c-afdb2a90edae] nova.exception.PortBindingFailed: Binding failed for port 3e148a81-f55b-4f13-adfc-ca47776ae7c5, please check neutron logs for more information. [ 627.936588] env[61806]: ERROR nova.compute.manager [instance: c2d1c9f6-f458-4843-982c-afdb2a90edae] [ 627.936941] env[61806]: DEBUG nova.compute.utils [None req-24088d08-d3d7-4f62-ad63-91029bd9b22d tempest-ServersTestMultiNic-1480543082 tempest-ServersTestMultiNic-1480543082-project-member] [instance: c2d1c9f6-f458-4843-982c-afdb2a90edae] Binding failed for port 3e148a81-f55b-4f13-adfc-ca47776ae7c5, please check neutron logs for more information. {{(pid=61806) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 627.939156] env[61806]: DEBUG oslo_concurrency.lockutils [None req-3e50272a-5153-4e11-beee-d774a917f2df tempest-ServerPasswordTestJSON-1741413915 tempest-ServerPasswordTestJSON-1741413915-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 14.870s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 627.940368] env[61806]: INFO nova.compute.claims [None req-3e50272a-5153-4e11-beee-d774a917f2df tempest-ServerPasswordTestJSON-1741413915 tempest-ServerPasswordTestJSON-1741413915-project-member] [instance: 0d757385-ff1c-49c9-825f-f9d542ae074f] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 627.943247] env[61806]: DEBUG nova.compute.manager [None req-24088d08-d3d7-4f62-ad63-91029bd9b22d tempest-ServersTestMultiNic-1480543082 tempest-ServersTestMultiNic-1480543082-project-member] [instance: c2d1c9f6-f458-4843-982c-afdb2a90edae] Build of instance c2d1c9f6-f458-4843-982c-afdb2a90edae was re-scheduled: Binding failed for port 3e148a81-f55b-4f13-adfc-ca47776ae7c5, please check neutron logs for more information. {{(pid=61806) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 627.943705] env[61806]: DEBUG nova.compute.manager [None req-24088d08-d3d7-4f62-ad63-91029bd9b22d tempest-ServersTestMultiNic-1480543082 tempest-ServersTestMultiNic-1480543082-project-member] [instance: c2d1c9f6-f458-4843-982c-afdb2a90edae] Unplugging VIFs for instance {{(pid=61806) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 627.943924] env[61806]: DEBUG oslo_concurrency.lockutils [None req-24088d08-d3d7-4f62-ad63-91029bd9b22d tempest-ServersTestMultiNic-1480543082 tempest-ServersTestMultiNic-1480543082-project-member] Acquiring lock "refresh_cache-c2d1c9f6-f458-4843-982c-afdb2a90edae" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 627.944078] env[61806]: DEBUG oslo_concurrency.lockutils [None req-24088d08-d3d7-4f62-ad63-91029bd9b22d tempest-ServersTestMultiNic-1480543082 tempest-ServersTestMultiNic-1480543082-project-member] Acquired lock "refresh_cache-c2d1c9f6-f458-4843-982c-afdb2a90edae" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 627.944248] env[61806]: DEBUG nova.network.neutron [None req-24088d08-d3d7-4f62-ad63-91029bd9b22d tempest-ServersTestMultiNic-1480543082 tempest-ServersTestMultiNic-1480543082-project-member] [instance: c2d1c9f6-f458-4843-982c-afdb2a90edae] Building network info cache for instance {{(pid=61806) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 628.030168] env[61806]: DEBUG oslo_concurrency.lockutils [None req-3bf756e0-c56e-4902-9c2b-b224a1d09a7c tempest-ListServersNegativeTestJSON-272187588 tempest-ListServersNegativeTestJSON-272187588-project-member] Acquiring lock "f2446d04-15b8-4b3d-8932-c37df67b0f8f" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 628.030742] env[61806]: DEBUG oslo_concurrency.lockutils [None req-3bf756e0-c56e-4902-9c2b-b224a1d09a7c tempest-ListServersNegativeTestJSON-272187588 tempest-ListServersNegativeTestJSON-272187588-project-member] Lock "f2446d04-15b8-4b3d-8932-c37df67b0f8f" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 628.069206] env[61806]: DEBUG oslo_concurrency.lockutils [None req-3bf756e0-c56e-4902-9c2b-b224a1d09a7c tempest-ListServersNegativeTestJSON-272187588 tempest-ListServersNegativeTestJSON-272187588-project-member] Acquiring lock "ea8d39a1-066d-4eb4-888d-776b0e45f684" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 628.069487] env[61806]: DEBUG oslo_concurrency.lockutils [None req-3bf756e0-c56e-4902-9c2b-b224a1d09a7c tempest-ListServersNegativeTestJSON-272187588 tempest-ListServersNegativeTestJSON-272187588-project-member] Lock "ea8d39a1-066d-4eb4-888d-776b0e45f684" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 628.090524] env[61806]: DEBUG oslo_concurrency.lockutils [None req-98d962a0-2632-4a9a-8d12-70ce4acc382b tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] Releasing lock "refresh_cache-b8249dc5-5561-454a-aa01-477f9f1fc0e1" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 628.090627] env[61806]: DEBUG nova.compute.manager [None req-98d962a0-2632-4a9a-8d12-70ce4acc382b tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] [instance: b8249dc5-5561-454a-aa01-477f9f1fc0e1] Start destroying the instance on the hypervisor. {{(pid=61806) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 628.090825] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-98d962a0-2632-4a9a-8d12-70ce4acc382b tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] [instance: b8249dc5-5561-454a-aa01-477f9f1fc0e1] Destroying instance {{(pid=61806) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 628.091226] env[61806]: DEBUG oslo_concurrency.lockutils [req-ae4f70c7-c9c6-488b-aae5-6ddb5af72c0a req-74d0fd9b-72fc-4ae5-9061-103d30591758 service nova] Acquired lock "refresh_cache-b8249dc5-5561-454a-aa01-477f9f1fc0e1" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 628.092727] env[61806]: DEBUG nova.network.neutron [req-ae4f70c7-c9c6-488b-aae5-6ddb5af72c0a req-74d0fd9b-72fc-4ae5-9061-103d30591758 service nova] [instance: b8249dc5-5561-454a-aa01-477f9f1fc0e1] Refreshing network info cache for port 340e11ef-b16f-45b1-8a9b-cf7c087aa898 {{(pid=61806) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 628.093575] env[61806]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-b722535b-1999-4687-aa97-ad20f13a8c45 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 628.104344] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1556e639-76a9-4cf0-a0a8-d4e98cf55a76 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 628.129495] env[61806]: DEBUG oslo_concurrency.lockutils [None req-3bf756e0-c56e-4902-9c2b-b224a1d09a7c tempest-ListServersNegativeTestJSON-272187588 tempest-ListServersNegativeTestJSON-272187588-project-member] Acquiring lock "198dc1a4-050d-47b2-8bc4-30af5534a789" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 628.129730] env[61806]: DEBUG oslo_concurrency.lockutils [None req-3bf756e0-c56e-4902-9c2b-b224a1d09a7c tempest-ListServersNegativeTestJSON-272187588 tempest-ListServersNegativeTestJSON-272187588-project-member] Lock "198dc1a4-050d-47b2-8bc4-30af5534a789" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 628.142203] env[61806]: WARNING nova.virt.vmwareapi.vmops [None req-98d962a0-2632-4a9a-8d12-70ce4acc382b tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] [instance: b8249dc5-5561-454a-aa01-477f9f1fc0e1] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance b8249dc5-5561-454a-aa01-477f9f1fc0e1 could not be found. [ 628.142439] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-98d962a0-2632-4a9a-8d12-70ce4acc382b tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] [instance: b8249dc5-5561-454a-aa01-477f9f1fc0e1] Instance destroyed {{(pid=61806) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 628.142620] env[61806]: INFO nova.compute.manager [None req-98d962a0-2632-4a9a-8d12-70ce4acc382b tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] [instance: b8249dc5-5561-454a-aa01-477f9f1fc0e1] Took 0.05 seconds to destroy the instance on the hypervisor. [ 628.142876] env[61806]: DEBUG oslo.service.loopingcall [None req-98d962a0-2632-4a9a-8d12-70ce4acc382b tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61806) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 628.143301] env[61806]: DEBUG nova.compute.manager [-] [instance: b8249dc5-5561-454a-aa01-477f9f1fc0e1] Deallocating network for instance {{(pid=61806) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 628.143399] env[61806]: DEBUG nova.network.neutron [-] [instance: b8249dc5-5561-454a-aa01-477f9f1fc0e1] deallocate_for_instance() {{(pid=61806) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 628.161542] env[61806]: DEBUG nova.network.neutron [-] [instance: b8249dc5-5561-454a-aa01-477f9f1fc0e1] Instance cache missing network info. {{(pid=61806) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 628.174235] env[61806]: INFO nova.compute.manager [None req-ebff8349-d5ea-4a95-aee1-e4c9e8b6d3a4 tempest-ServerDiagnosticsTest-195484958 tempest-ServerDiagnosticsTest-195484958-project-member] [instance: 79fc9ea2-c427-47ce-acdb-1002ee5d78c9] Took 1.03 seconds to deallocate network for instance. [ 628.465429] env[61806]: DEBUG nova.network.neutron [None req-24088d08-d3d7-4f62-ad63-91029bd9b22d tempest-ServersTestMultiNic-1480543082 tempest-ServersTestMultiNic-1480543082-project-member] [instance: c2d1c9f6-f458-4843-982c-afdb2a90edae] Instance cache missing network info. {{(pid=61806) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 628.560707] env[61806]: DEBUG nova.network.neutron [None req-24088d08-d3d7-4f62-ad63-91029bd9b22d tempest-ServersTestMultiNic-1480543082 tempest-ServersTestMultiNic-1480543082-project-member] [instance: c2d1c9f6-f458-4843-982c-afdb2a90edae] Updating instance_info_cache with network_info: [] {{(pid=61806) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 628.623076] env[61806]: DEBUG nova.network.neutron [req-ae4f70c7-c9c6-488b-aae5-6ddb5af72c0a req-74d0fd9b-72fc-4ae5-9061-103d30591758 service nova] [instance: b8249dc5-5561-454a-aa01-477f9f1fc0e1] Instance cache missing network info. {{(pid=61806) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 628.664887] env[61806]: DEBUG nova.network.neutron [-] [instance: b8249dc5-5561-454a-aa01-477f9f1fc0e1] Updating instance_info_cache with network_info: [] {{(pid=61806) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 628.719493] env[61806]: DEBUG nova.network.neutron [req-ae4f70c7-c9c6-488b-aae5-6ddb5af72c0a req-74d0fd9b-72fc-4ae5-9061-103d30591758 service nova] [instance: b8249dc5-5561-454a-aa01-477f9f1fc0e1] Updating instance_info_cache with network_info: [] {{(pid=61806) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 629.063409] env[61806]: DEBUG oslo_concurrency.lockutils [None req-24088d08-d3d7-4f62-ad63-91029bd9b22d tempest-ServersTestMultiNic-1480543082 tempest-ServersTestMultiNic-1480543082-project-member] Releasing lock "refresh_cache-c2d1c9f6-f458-4843-982c-afdb2a90edae" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 629.063667] env[61806]: DEBUG nova.compute.manager [None req-24088d08-d3d7-4f62-ad63-91029bd9b22d tempest-ServersTestMultiNic-1480543082 tempest-ServersTestMultiNic-1480543082-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61806) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 629.063827] env[61806]: DEBUG nova.compute.manager [None req-24088d08-d3d7-4f62-ad63-91029bd9b22d tempest-ServersTestMultiNic-1480543082 tempest-ServersTestMultiNic-1480543082-project-member] [instance: c2d1c9f6-f458-4843-982c-afdb2a90edae] Deallocating network for instance {{(pid=61806) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 629.063989] env[61806]: DEBUG nova.network.neutron [None req-24088d08-d3d7-4f62-ad63-91029bd9b22d tempest-ServersTestMultiNic-1480543082 tempest-ServersTestMultiNic-1480543082-project-member] [instance: c2d1c9f6-f458-4843-982c-afdb2a90edae] deallocate_for_instance() {{(pid=61806) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 629.100285] env[61806]: DEBUG nova.network.neutron [None req-24088d08-d3d7-4f62-ad63-91029bd9b22d tempest-ServersTestMultiNic-1480543082 tempest-ServersTestMultiNic-1480543082-project-member] [instance: c2d1c9f6-f458-4843-982c-afdb2a90edae] Instance cache missing network info. {{(pid=61806) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 629.158314] env[61806]: DEBUG oslo_concurrency.lockutils [None req-f0377b51-dab9-4f68-8592-ca0f31b4ba51 tempest-FloatingIPsAssociationNegativeTestJSON-30494360 tempest-FloatingIPsAssociationNegativeTestJSON-30494360-project-member] Acquiring lock "37c07b8d-e2a0-4aa6-b7ba-194feed837e5" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 629.158676] env[61806]: DEBUG oslo_concurrency.lockutils [None req-f0377b51-dab9-4f68-8592-ca0f31b4ba51 tempest-FloatingIPsAssociationNegativeTestJSON-30494360 tempest-FloatingIPsAssociationNegativeTestJSON-30494360-project-member] Lock "37c07b8d-e2a0-4aa6-b7ba-194feed837e5" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 629.168359] env[61806]: INFO nova.compute.manager [-] [instance: b8249dc5-5561-454a-aa01-477f9f1fc0e1] Took 1.02 seconds to deallocate network for instance. [ 629.168935] env[61806]: DEBUG nova.compute.claims [None req-98d962a0-2632-4a9a-8d12-70ce4acc382b tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] [instance: b8249dc5-5561-454a-aa01-477f9f1fc0e1] Aborting claim: {{(pid=61806) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 629.170057] env[61806]: DEBUG oslo_concurrency.lockutils [None req-98d962a0-2632-4a9a-8d12-70ce4acc382b tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 629.203267] env[61806]: DEBUG oslo_concurrency.lockutils [None req-da1d2f3c-f6ea-47f5-8c64-65c86bf2d676 tempest-ServerRescueTestJSON-811948373 tempest-ServerRescueTestJSON-811948373-project-member] Acquiring lock "c61df5d3-7f43-48d4-a2c0-a8972372c0ba" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 629.205266] env[61806]: DEBUG oslo_concurrency.lockutils [None req-da1d2f3c-f6ea-47f5-8c64-65c86bf2d676 tempest-ServerRescueTestJSON-811948373 tempest-ServerRescueTestJSON-811948373-project-member] Lock "c61df5d3-7f43-48d4-a2c0-a8972372c0ba" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 629.212544] env[61806]: INFO nova.scheduler.client.report [None req-ebff8349-d5ea-4a95-aee1-e4c9e8b6d3a4 tempest-ServerDiagnosticsTest-195484958 tempest-ServerDiagnosticsTest-195484958-project-member] Deleted allocations for instance 79fc9ea2-c427-47ce-acdb-1002ee5d78c9 [ 629.224171] env[61806]: DEBUG oslo_concurrency.lockutils [req-ae4f70c7-c9c6-488b-aae5-6ddb5af72c0a req-74d0fd9b-72fc-4ae5-9061-103d30591758 service nova] Releasing lock "refresh_cache-b8249dc5-5561-454a-aa01-477f9f1fc0e1" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 629.225120] env[61806]: DEBUG nova.compute.manager [req-ae4f70c7-c9c6-488b-aae5-6ddb5af72c0a req-74d0fd9b-72fc-4ae5-9061-103d30591758 service nova] [instance: b8249dc5-5561-454a-aa01-477f9f1fc0e1] Received event network-vif-deleted-340e11ef-b16f-45b1-8a9b-cf7c087aa898 {{(pid=61806) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 629.495933] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-66d62043-7446-4f05-b5b8-4828cd163ecb {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 629.503994] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-45d9449e-40a1-4e31-87df-321a4424db08 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 629.534701] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d574ca9d-4fb5-45c4-9d16-8653624fa45e {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 629.542444] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-75ee7189-9dc5-4bf4-941f-ec824a5d653b {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 629.557221] env[61806]: DEBUG nova.compute.provider_tree [None req-3e50272a-5153-4e11-beee-d774a917f2df tempest-ServerPasswordTestJSON-1741413915 tempest-ServerPasswordTestJSON-1741413915-project-member] Inventory has not changed in ProviderTree for provider: a2858be1-fd22-4e08-979e-87ad25293407 {{(pid=61806) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 629.607736] env[61806]: DEBUG nova.network.neutron [None req-24088d08-d3d7-4f62-ad63-91029bd9b22d tempest-ServersTestMultiNic-1480543082 tempest-ServersTestMultiNic-1480543082-project-member] [instance: c2d1c9f6-f458-4843-982c-afdb2a90edae] Updating instance_info_cache with network_info: [] {{(pid=61806) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 629.725738] env[61806]: DEBUG oslo_concurrency.lockutils [None req-ebff8349-d5ea-4a95-aee1-e4c9e8b6d3a4 tempest-ServerDiagnosticsTest-195484958 tempest-ServerDiagnosticsTest-195484958-project-member] Lock "79fc9ea2-c427-47ce-acdb-1002ee5d78c9" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 82.798s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 630.060071] env[61806]: DEBUG nova.scheduler.client.report [None req-3e50272a-5153-4e11-beee-d774a917f2df tempest-ServerPasswordTestJSON-1741413915 tempest-ServerPasswordTestJSON-1741413915-project-member] Inventory has not changed for provider a2858be1-fd22-4e08-979e-87ad25293407 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 139, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61806) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 630.111951] env[61806]: INFO nova.compute.manager [None req-24088d08-d3d7-4f62-ad63-91029bd9b22d tempest-ServersTestMultiNic-1480543082 tempest-ServersTestMultiNic-1480543082-project-member] [instance: c2d1c9f6-f458-4843-982c-afdb2a90edae] Took 1.05 seconds to deallocate network for instance. [ 630.231747] env[61806]: DEBUG nova.compute.manager [None req-4803fdfc-9054-49cf-8525-e438672eeedd tempest-ServersAdmin275Test-1945525728 tempest-ServersAdmin275Test-1945525728-project-member] [instance: 9db80c3c-e2c3-4034-a79a-d7be924c4ff6] Starting instance... {{(pid=61806) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 630.568562] env[61806]: DEBUG oslo_concurrency.lockutils [None req-3e50272a-5153-4e11-beee-d774a917f2df tempest-ServerPasswordTestJSON-1741413915 tempest-ServerPasswordTestJSON-1741413915-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.631s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 630.569091] env[61806]: DEBUG nova.compute.manager [None req-3e50272a-5153-4e11-beee-d774a917f2df tempest-ServerPasswordTestJSON-1741413915 tempest-ServerPasswordTestJSON-1741413915-project-member] [instance: 0d757385-ff1c-49c9-825f-f9d542ae074f] Start building networks asynchronously for instance. {{(pid=61806) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 630.571727] env[61806]: DEBUG oslo_concurrency.lockutils [None req-f8b2f38e-90f5-4b03-89bc-ca1926adb40c tempest-FloatingIPsAssociationTestJSON-35718468 tempest-FloatingIPsAssociationTestJSON-35718468-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 15.463s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 630.573096] env[61806]: INFO nova.compute.claims [None req-f8b2f38e-90f5-4b03-89bc-ca1926adb40c tempest-FloatingIPsAssociationTestJSON-35718468 tempest-FloatingIPsAssociationTestJSON-35718468-project-member] [instance: 88cb0961-ad34-4ffc-8fb2-6034f08ee740] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 630.761874] env[61806]: DEBUG oslo_concurrency.lockutils [None req-4803fdfc-9054-49cf-8525-e438672eeedd tempest-ServersAdmin275Test-1945525728 tempest-ServersAdmin275Test-1945525728-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 631.077774] env[61806]: DEBUG nova.compute.utils [None req-3e50272a-5153-4e11-beee-d774a917f2df tempest-ServerPasswordTestJSON-1741413915 tempest-ServerPasswordTestJSON-1741413915-project-member] Using /dev/sd instead of None {{(pid=61806) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 631.081417] env[61806]: DEBUG nova.compute.manager [None req-3e50272a-5153-4e11-beee-d774a917f2df tempest-ServerPasswordTestJSON-1741413915 tempest-ServerPasswordTestJSON-1741413915-project-member] [instance: 0d757385-ff1c-49c9-825f-f9d542ae074f] Allocating IP information in the background. {{(pid=61806) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 631.081677] env[61806]: DEBUG nova.network.neutron [None req-3e50272a-5153-4e11-beee-d774a917f2df tempest-ServerPasswordTestJSON-1741413915 tempest-ServerPasswordTestJSON-1741413915-project-member] [instance: 0d757385-ff1c-49c9-825f-f9d542ae074f] allocate_for_instance() {{(pid=61806) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 631.126129] env[61806]: DEBUG nova.policy [None req-3e50272a-5153-4e11-beee-d774a917f2df tempest-ServerPasswordTestJSON-1741413915 tempest-ServerPasswordTestJSON-1741413915-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '37fed767ddda4e2682b17e3b162da01d', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '2acd17eca07249d4aace0f3f6243c98e', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61806) authorize /opt/stack/nova/nova/policy.py:201}} [ 631.149754] env[61806]: INFO nova.scheduler.client.report [None req-24088d08-d3d7-4f62-ad63-91029bd9b22d tempest-ServersTestMultiNic-1480543082 tempest-ServersTestMultiNic-1480543082-project-member] Deleted allocations for instance c2d1c9f6-f458-4843-982c-afdb2a90edae [ 631.474661] env[61806]: DEBUG nova.network.neutron [None req-3e50272a-5153-4e11-beee-d774a917f2df tempest-ServerPasswordTestJSON-1741413915 tempest-ServerPasswordTestJSON-1741413915-project-member] [instance: 0d757385-ff1c-49c9-825f-f9d542ae074f] Successfully created port: fc109b8a-4096-4169-bf3b-f19ce33059d3 {{(pid=61806) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 631.585226] env[61806]: DEBUG nova.compute.manager [None req-3e50272a-5153-4e11-beee-d774a917f2df tempest-ServerPasswordTestJSON-1741413915 tempest-ServerPasswordTestJSON-1741413915-project-member] [instance: 0d757385-ff1c-49c9-825f-f9d542ae074f] Start building block device mappings for instance. {{(pid=61806) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 631.663652] env[61806]: DEBUG oslo_concurrency.lockutils [None req-24088d08-d3d7-4f62-ad63-91029bd9b22d tempest-ServersTestMultiNic-1480543082 tempest-ServersTestMultiNic-1480543082-project-member] Lock "c2d1c9f6-f458-4843-982c-afdb2a90edae" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 82.560s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 632.039439] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eb67b1a9-63e5-4136-bdcf-ae05c4f9c398 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 632.047393] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2ec3a982-b559-4d06-96c6-b5ed1a9a3ae3 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 632.076971] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1cc595a9-6458-456c-b27d-e3cd4006758e {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 632.084535] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5cb4fc95-2b09-4448-9a91-ca9703daca39 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 632.101829] env[61806]: DEBUG nova.compute.provider_tree [None req-f8b2f38e-90f5-4b03-89bc-ca1926adb40c tempest-FloatingIPsAssociationTestJSON-35718468 tempest-FloatingIPsAssociationTestJSON-35718468-project-member] Inventory has not changed in ProviderTree for provider: a2858be1-fd22-4e08-979e-87ad25293407 {{(pid=61806) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 632.168856] env[61806]: DEBUG nova.compute.manager [None req-a93ef3a4-6e83-4dfe-9292-bac21f14d678 tempest-ServerExternalEventsTest-1847900285 tempest-ServerExternalEventsTest-1847900285-project-member] [instance: d5434dd2-206f-4a1e-b97b-bd6f691b6ddf] Starting instance... {{(pid=61806) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 632.604329] env[61806]: DEBUG nova.compute.manager [None req-3e50272a-5153-4e11-beee-d774a917f2df tempest-ServerPasswordTestJSON-1741413915 tempest-ServerPasswordTestJSON-1741413915-project-member] [instance: 0d757385-ff1c-49c9-825f-f9d542ae074f] Start spawning the instance on the hypervisor. {{(pid=61806) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 632.606816] env[61806]: DEBUG nova.scheduler.client.report [None req-f8b2f38e-90f5-4b03-89bc-ca1926adb40c tempest-FloatingIPsAssociationTestJSON-35718468 tempest-FloatingIPsAssociationTestJSON-35718468-project-member] Inventory has not changed for provider a2858be1-fd22-4e08-979e-87ad25293407 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 139, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61806) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 632.622406] env[61806]: DEBUG nova.compute.manager [req-a509107e-9dae-4385-a90b-2541d22a9a22 req-2bb902d7-13b8-418a-a658-44772410bca6 service nova] [instance: 0d757385-ff1c-49c9-825f-f9d542ae074f] Received event network-changed-fc109b8a-4096-4169-bf3b-f19ce33059d3 {{(pid=61806) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 632.622623] env[61806]: DEBUG nova.compute.manager [req-a509107e-9dae-4385-a90b-2541d22a9a22 req-2bb902d7-13b8-418a-a658-44772410bca6 service nova] [instance: 0d757385-ff1c-49c9-825f-f9d542ae074f] Refreshing instance network info cache due to event network-changed-fc109b8a-4096-4169-bf3b-f19ce33059d3. {{(pid=61806) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 632.622838] env[61806]: DEBUG oslo_concurrency.lockutils [req-a509107e-9dae-4385-a90b-2541d22a9a22 req-2bb902d7-13b8-418a-a658-44772410bca6 service nova] Acquiring lock "refresh_cache-0d757385-ff1c-49c9-825f-f9d542ae074f" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 632.623253] env[61806]: DEBUG oslo_concurrency.lockutils [req-a509107e-9dae-4385-a90b-2541d22a9a22 req-2bb902d7-13b8-418a-a658-44772410bca6 service nova] Acquired lock "refresh_cache-0d757385-ff1c-49c9-825f-f9d542ae074f" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 632.623253] env[61806]: DEBUG nova.network.neutron [req-a509107e-9dae-4385-a90b-2541d22a9a22 req-2bb902d7-13b8-418a-a658-44772410bca6 service nova] [instance: 0d757385-ff1c-49c9-825f-f9d542ae074f] Refreshing network info cache for port fc109b8a-4096-4169-bf3b-f19ce33059d3 {{(pid=61806) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 632.637015] env[61806]: DEBUG nova.virt.hardware [None req-3e50272a-5153-4e11-beee-d774a917f2df tempest-ServerPasswordTestJSON-1741413915 tempest-ServerPasswordTestJSON-1741413915-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-15T18:05:10Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-15T18:04:52Z,direct_url=,disk_format='vmdk',id=ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='84ba83fce2c349a6988173c5d6fc3b07',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-15T18:04:53Z,virtual_size=,visibility=), allow threads: False {{(pid=61806) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 632.637282] env[61806]: DEBUG nova.virt.hardware [None req-3e50272a-5153-4e11-beee-d774a917f2df tempest-ServerPasswordTestJSON-1741413915 tempest-ServerPasswordTestJSON-1741413915-project-member] Flavor limits 0:0:0 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 632.637441] env[61806]: DEBUG nova.virt.hardware [None req-3e50272a-5153-4e11-beee-d774a917f2df tempest-ServerPasswordTestJSON-1741413915 tempest-ServerPasswordTestJSON-1741413915-project-member] Image limits 0:0:0 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 632.637622] env[61806]: DEBUG nova.virt.hardware [None req-3e50272a-5153-4e11-beee-d774a917f2df tempest-ServerPasswordTestJSON-1741413915 tempest-ServerPasswordTestJSON-1741413915-project-member] Flavor pref 0:0:0 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 632.637770] env[61806]: DEBUG nova.virt.hardware [None req-3e50272a-5153-4e11-beee-d774a917f2df tempest-ServerPasswordTestJSON-1741413915 tempest-ServerPasswordTestJSON-1741413915-project-member] Image pref 0:0:0 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 632.637916] env[61806]: DEBUG nova.virt.hardware [None req-3e50272a-5153-4e11-beee-d774a917f2df tempest-ServerPasswordTestJSON-1741413915 tempest-ServerPasswordTestJSON-1741413915-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 632.638468] env[61806]: DEBUG nova.virt.hardware [None req-3e50272a-5153-4e11-beee-d774a917f2df tempest-ServerPasswordTestJSON-1741413915 tempest-ServerPasswordTestJSON-1741413915-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61806) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 632.638544] env[61806]: DEBUG nova.virt.hardware [None req-3e50272a-5153-4e11-beee-d774a917f2df tempest-ServerPasswordTestJSON-1741413915 tempest-ServerPasswordTestJSON-1741413915-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61806) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 632.638697] env[61806]: DEBUG nova.virt.hardware [None req-3e50272a-5153-4e11-beee-d774a917f2df tempest-ServerPasswordTestJSON-1741413915 tempest-ServerPasswordTestJSON-1741413915-project-member] Got 1 possible topologies {{(pid=61806) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 632.638865] env[61806]: DEBUG nova.virt.hardware [None req-3e50272a-5153-4e11-beee-d774a917f2df tempest-ServerPasswordTestJSON-1741413915 tempest-ServerPasswordTestJSON-1741413915-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61806) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 632.639211] env[61806]: DEBUG nova.virt.hardware [None req-3e50272a-5153-4e11-beee-d774a917f2df tempest-ServerPasswordTestJSON-1741413915 tempest-ServerPasswordTestJSON-1741413915-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61806) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 632.640276] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7b4b5491-1618-41f0-b98b-df54fb44944d {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 632.649953] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-82c56899-2183-4df6-b77b-aa09d94fef48 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 632.694983] env[61806]: DEBUG oslo_concurrency.lockutils [None req-a93ef3a4-6e83-4dfe-9292-bac21f14d678 tempest-ServerExternalEventsTest-1847900285 tempest-ServerExternalEventsTest-1847900285-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 632.787908] env[61806]: ERROR nova.compute.manager [None req-3e50272a-5153-4e11-beee-d774a917f2df tempest-ServerPasswordTestJSON-1741413915 tempest-ServerPasswordTestJSON-1741413915-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port fc109b8a-4096-4169-bf3b-f19ce33059d3, please check neutron logs for more information. [ 632.787908] env[61806]: ERROR nova.compute.manager Traceback (most recent call last): [ 632.787908] env[61806]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 632.787908] env[61806]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 632.787908] env[61806]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 632.787908] env[61806]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 632.787908] env[61806]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 632.787908] env[61806]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 632.787908] env[61806]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 632.787908] env[61806]: ERROR nova.compute.manager self.force_reraise() [ 632.787908] env[61806]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 632.787908] env[61806]: ERROR nova.compute.manager raise self.value [ 632.787908] env[61806]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 632.787908] env[61806]: ERROR nova.compute.manager updated_port = self._update_port( [ 632.787908] env[61806]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 632.787908] env[61806]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 632.788386] env[61806]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 632.788386] env[61806]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 632.788386] env[61806]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port fc109b8a-4096-4169-bf3b-f19ce33059d3, please check neutron logs for more information. [ 632.788386] env[61806]: ERROR nova.compute.manager [ 632.788386] env[61806]: Traceback (most recent call last): [ 632.788386] env[61806]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 632.788386] env[61806]: listener.cb(fileno) [ 632.788386] env[61806]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 632.788386] env[61806]: result = function(*args, **kwargs) [ 632.788386] env[61806]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 632.788386] env[61806]: return func(*args, **kwargs) [ 632.788386] env[61806]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 632.788386] env[61806]: raise e [ 632.788386] env[61806]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 632.788386] env[61806]: nwinfo = self.network_api.allocate_for_instance( [ 632.788386] env[61806]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 632.788386] env[61806]: created_port_ids = self._update_ports_for_instance( [ 632.788386] env[61806]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 632.788386] env[61806]: with excutils.save_and_reraise_exception(): [ 632.788386] env[61806]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 632.788386] env[61806]: self.force_reraise() [ 632.788386] env[61806]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 632.788386] env[61806]: raise self.value [ 632.788386] env[61806]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 632.788386] env[61806]: updated_port = self._update_port( [ 632.788386] env[61806]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 632.788386] env[61806]: _ensure_no_port_binding_failure(port) [ 632.788386] env[61806]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 632.788386] env[61806]: raise exception.PortBindingFailed(port_id=port['id']) [ 632.789315] env[61806]: nova.exception.PortBindingFailed: Binding failed for port fc109b8a-4096-4169-bf3b-f19ce33059d3, please check neutron logs for more information. [ 632.789315] env[61806]: Removing descriptor: 16 [ 632.789481] env[61806]: ERROR nova.compute.manager [None req-3e50272a-5153-4e11-beee-d774a917f2df tempest-ServerPasswordTestJSON-1741413915 tempest-ServerPasswordTestJSON-1741413915-project-member] [instance: 0d757385-ff1c-49c9-825f-f9d542ae074f] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port fc109b8a-4096-4169-bf3b-f19ce33059d3, please check neutron logs for more information. [ 632.789481] env[61806]: ERROR nova.compute.manager [instance: 0d757385-ff1c-49c9-825f-f9d542ae074f] Traceback (most recent call last): [ 632.789481] env[61806]: ERROR nova.compute.manager [instance: 0d757385-ff1c-49c9-825f-f9d542ae074f] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 632.789481] env[61806]: ERROR nova.compute.manager [instance: 0d757385-ff1c-49c9-825f-f9d542ae074f] yield resources [ 632.789481] env[61806]: ERROR nova.compute.manager [instance: 0d757385-ff1c-49c9-825f-f9d542ae074f] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 632.789481] env[61806]: ERROR nova.compute.manager [instance: 0d757385-ff1c-49c9-825f-f9d542ae074f] self.driver.spawn(context, instance, image_meta, [ 632.789481] env[61806]: ERROR nova.compute.manager [instance: 0d757385-ff1c-49c9-825f-f9d542ae074f] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 632.789481] env[61806]: ERROR nova.compute.manager [instance: 0d757385-ff1c-49c9-825f-f9d542ae074f] self._vmops.spawn(context, instance, image_meta, injected_files, [ 632.789481] env[61806]: ERROR nova.compute.manager [instance: 0d757385-ff1c-49c9-825f-f9d542ae074f] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 632.789481] env[61806]: ERROR nova.compute.manager [instance: 0d757385-ff1c-49c9-825f-f9d542ae074f] vm_ref = self.build_virtual_machine(instance, [ 632.789481] env[61806]: ERROR nova.compute.manager [instance: 0d757385-ff1c-49c9-825f-f9d542ae074f] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 632.789764] env[61806]: ERROR nova.compute.manager [instance: 0d757385-ff1c-49c9-825f-f9d542ae074f] vif_infos = vmwarevif.get_vif_info(self._session, [ 632.789764] env[61806]: ERROR nova.compute.manager [instance: 0d757385-ff1c-49c9-825f-f9d542ae074f] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 632.789764] env[61806]: ERROR nova.compute.manager [instance: 0d757385-ff1c-49c9-825f-f9d542ae074f] for vif in network_info: [ 632.789764] env[61806]: ERROR nova.compute.manager [instance: 0d757385-ff1c-49c9-825f-f9d542ae074f] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 632.789764] env[61806]: ERROR nova.compute.manager [instance: 0d757385-ff1c-49c9-825f-f9d542ae074f] return self._sync_wrapper(fn, *args, **kwargs) [ 632.789764] env[61806]: ERROR nova.compute.manager [instance: 0d757385-ff1c-49c9-825f-f9d542ae074f] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 632.789764] env[61806]: ERROR nova.compute.manager [instance: 0d757385-ff1c-49c9-825f-f9d542ae074f] self.wait() [ 632.789764] env[61806]: ERROR nova.compute.manager [instance: 0d757385-ff1c-49c9-825f-f9d542ae074f] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 632.789764] env[61806]: ERROR nova.compute.manager [instance: 0d757385-ff1c-49c9-825f-f9d542ae074f] self[:] = self._gt.wait() [ 632.789764] env[61806]: ERROR nova.compute.manager [instance: 0d757385-ff1c-49c9-825f-f9d542ae074f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 632.789764] env[61806]: ERROR nova.compute.manager [instance: 0d757385-ff1c-49c9-825f-f9d542ae074f] return self._exit_event.wait() [ 632.789764] env[61806]: ERROR nova.compute.manager [instance: 0d757385-ff1c-49c9-825f-f9d542ae074f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 632.789764] env[61806]: ERROR nova.compute.manager [instance: 0d757385-ff1c-49c9-825f-f9d542ae074f] result = hub.switch() [ 632.790095] env[61806]: ERROR nova.compute.manager [instance: 0d757385-ff1c-49c9-825f-f9d542ae074f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 632.790095] env[61806]: ERROR nova.compute.manager [instance: 0d757385-ff1c-49c9-825f-f9d542ae074f] return self.greenlet.switch() [ 632.790095] env[61806]: ERROR nova.compute.manager [instance: 0d757385-ff1c-49c9-825f-f9d542ae074f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 632.790095] env[61806]: ERROR nova.compute.manager [instance: 0d757385-ff1c-49c9-825f-f9d542ae074f] result = function(*args, **kwargs) [ 632.790095] env[61806]: ERROR nova.compute.manager [instance: 0d757385-ff1c-49c9-825f-f9d542ae074f] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 632.790095] env[61806]: ERROR nova.compute.manager [instance: 0d757385-ff1c-49c9-825f-f9d542ae074f] return func(*args, **kwargs) [ 632.790095] env[61806]: ERROR nova.compute.manager [instance: 0d757385-ff1c-49c9-825f-f9d542ae074f] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 632.790095] env[61806]: ERROR nova.compute.manager [instance: 0d757385-ff1c-49c9-825f-f9d542ae074f] raise e [ 632.790095] env[61806]: ERROR nova.compute.manager [instance: 0d757385-ff1c-49c9-825f-f9d542ae074f] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 632.790095] env[61806]: ERROR nova.compute.manager [instance: 0d757385-ff1c-49c9-825f-f9d542ae074f] nwinfo = self.network_api.allocate_for_instance( [ 632.790095] env[61806]: ERROR nova.compute.manager [instance: 0d757385-ff1c-49c9-825f-f9d542ae074f] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 632.790095] env[61806]: ERROR nova.compute.manager [instance: 0d757385-ff1c-49c9-825f-f9d542ae074f] created_port_ids = self._update_ports_for_instance( [ 632.790095] env[61806]: ERROR nova.compute.manager [instance: 0d757385-ff1c-49c9-825f-f9d542ae074f] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 632.790485] env[61806]: ERROR nova.compute.manager [instance: 0d757385-ff1c-49c9-825f-f9d542ae074f] with excutils.save_and_reraise_exception(): [ 632.790485] env[61806]: ERROR nova.compute.manager [instance: 0d757385-ff1c-49c9-825f-f9d542ae074f] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 632.790485] env[61806]: ERROR nova.compute.manager [instance: 0d757385-ff1c-49c9-825f-f9d542ae074f] self.force_reraise() [ 632.790485] env[61806]: ERROR nova.compute.manager [instance: 0d757385-ff1c-49c9-825f-f9d542ae074f] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 632.790485] env[61806]: ERROR nova.compute.manager [instance: 0d757385-ff1c-49c9-825f-f9d542ae074f] raise self.value [ 632.790485] env[61806]: ERROR nova.compute.manager [instance: 0d757385-ff1c-49c9-825f-f9d542ae074f] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 632.790485] env[61806]: ERROR nova.compute.manager [instance: 0d757385-ff1c-49c9-825f-f9d542ae074f] updated_port = self._update_port( [ 632.790485] env[61806]: ERROR nova.compute.manager [instance: 0d757385-ff1c-49c9-825f-f9d542ae074f] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 632.790485] env[61806]: ERROR nova.compute.manager [instance: 0d757385-ff1c-49c9-825f-f9d542ae074f] _ensure_no_port_binding_failure(port) [ 632.790485] env[61806]: ERROR nova.compute.manager [instance: 0d757385-ff1c-49c9-825f-f9d542ae074f] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 632.790485] env[61806]: ERROR nova.compute.manager [instance: 0d757385-ff1c-49c9-825f-f9d542ae074f] raise exception.PortBindingFailed(port_id=port['id']) [ 632.790485] env[61806]: ERROR nova.compute.manager [instance: 0d757385-ff1c-49c9-825f-f9d542ae074f] nova.exception.PortBindingFailed: Binding failed for port fc109b8a-4096-4169-bf3b-f19ce33059d3, please check neutron logs for more information. [ 632.790485] env[61806]: ERROR nova.compute.manager [instance: 0d757385-ff1c-49c9-825f-f9d542ae074f] [ 632.791084] env[61806]: INFO nova.compute.manager [None req-3e50272a-5153-4e11-beee-d774a917f2df tempest-ServerPasswordTestJSON-1741413915 tempest-ServerPasswordTestJSON-1741413915-project-member] [instance: 0d757385-ff1c-49c9-825f-f9d542ae074f] Terminating instance [ 632.792958] env[61806]: DEBUG oslo_concurrency.lockutils [None req-3e50272a-5153-4e11-beee-d774a917f2df tempest-ServerPasswordTestJSON-1741413915 tempest-ServerPasswordTestJSON-1741413915-project-member] Acquiring lock "refresh_cache-0d757385-ff1c-49c9-825f-f9d542ae074f" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 633.115017] env[61806]: DEBUG oslo_concurrency.lockutils [None req-f8b2f38e-90f5-4b03-89bc-ca1926adb40c tempest-FloatingIPsAssociationTestJSON-35718468 tempest-FloatingIPsAssociationTestJSON-35718468-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.541s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 633.115017] env[61806]: DEBUG nova.compute.manager [None req-f8b2f38e-90f5-4b03-89bc-ca1926adb40c tempest-FloatingIPsAssociationTestJSON-35718468 tempest-FloatingIPsAssociationTestJSON-35718468-project-member] [instance: 88cb0961-ad34-4ffc-8fb2-6034f08ee740] Start building networks asynchronously for instance. {{(pid=61806) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 633.123616] env[61806]: DEBUG oslo_concurrency.lockutils [None req-e34d2bcf-266f-4faa-8f31-6bd93f6ab176 tempest-ImagesTestJSON-618852556 tempest-ImagesTestJSON-618852556-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 15.672s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 633.169586] env[61806]: DEBUG nova.network.neutron [req-a509107e-9dae-4385-a90b-2541d22a9a22 req-2bb902d7-13b8-418a-a658-44772410bca6 service nova] [instance: 0d757385-ff1c-49c9-825f-f9d542ae074f] Instance cache missing network info. {{(pid=61806) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 633.386812] env[61806]: DEBUG nova.network.neutron [req-a509107e-9dae-4385-a90b-2541d22a9a22 req-2bb902d7-13b8-418a-a658-44772410bca6 service nova] [instance: 0d757385-ff1c-49c9-825f-f9d542ae074f] Updating instance_info_cache with network_info: [] {{(pid=61806) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 633.625358] env[61806]: DEBUG nova.compute.utils [None req-f8b2f38e-90f5-4b03-89bc-ca1926adb40c tempest-FloatingIPsAssociationTestJSON-35718468 tempest-FloatingIPsAssociationTestJSON-35718468-project-member] Using /dev/sd instead of None {{(pid=61806) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 633.627098] env[61806]: DEBUG nova.compute.manager [None req-f8b2f38e-90f5-4b03-89bc-ca1926adb40c tempest-FloatingIPsAssociationTestJSON-35718468 tempest-FloatingIPsAssociationTestJSON-35718468-project-member] [instance: 88cb0961-ad34-4ffc-8fb2-6034f08ee740] Allocating IP information in the background. {{(pid=61806) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 633.627273] env[61806]: DEBUG nova.network.neutron [None req-f8b2f38e-90f5-4b03-89bc-ca1926adb40c tempest-FloatingIPsAssociationTestJSON-35718468 tempest-FloatingIPsAssociationTestJSON-35718468-project-member] [instance: 88cb0961-ad34-4ffc-8fb2-6034f08ee740] allocate_for_instance() {{(pid=61806) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 633.698438] env[61806]: DEBUG nova.policy [None req-f8b2f38e-90f5-4b03-89bc-ca1926adb40c tempest-FloatingIPsAssociationTestJSON-35718468 tempest-FloatingIPsAssociationTestJSON-35718468-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '1e85c4dbfae9470e948dadbe9dda5a78', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'd1d4c5b594ae4ccc93fafde392b39aea', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61806) authorize /opt/stack/nova/nova/policy.py:201}} [ 633.890313] env[61806]: DEBUG oslo_concurrency.lockutils [req-a509107e-9dae-4385-a90b-2541d22a9a22 req-2bb902d7-13b8-418a-a658-44772410bca6 service nova] Releasing lock "refresh_cache-0d757385-ff1c-49c9-825f-f9d542ae074f" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 633.890742] env[61806]: DEBUG oslo_concurrency.lockutils [None req-3e50272a-5153-4e11-beee-d774a917f2df tempest-ServerPasswordTestJSON-1741413915 tempest-ServerPasswordTestJSON-1741413915-project-member] Acquired lock "refresh_cache-0d757385-ff1c-49c9-825f-f9d542ae074f" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 633.890923] env[61806]: DEBUG nova.network.neutron [None req-3e50272a-5153-4e11-beee-d774a917f2df tempest-ServerPasswordTestJSON-1741413915 tempest-ServerPasswordTestJSON-1741413915-project-member] [instance: 0d757385-ff1c-49c9-825f-f9d542ae074f] Building network info cache for instance {{(pid=61806) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 634.066421] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-10e046ea-0305-4663-a55c-af2d136fa7ca {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 634.074314] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-71dc490b-8c85-42f4-8527-8f96acdcc1c6 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 634.104608] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-26ff37e9-5b60-4540-915d-9cbe0a51d189 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 634.111993] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2649c530-e9f5-4b31-899b-abf6d1806bd2 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 634.124870] env[61806]: DEBUG nova.compute.provider_tree [None req-e34d2bcf-266f-4faa-8f31-6bd93f6ab176 tempest-ImagesTestJSON-618852556 tempest-ImagesTestJSON-618852556-project-member] Inventory has not changed in ProviderTree for provider: a2858be1-fd22-4e08-979e-87ad25293407 {{(pid=61806) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 634.131910] env[61806]: DEBUG nova.compute.manager [None req-f8b2f38e-90f5-4b03-89bc-ca1926adb40c tempest-FloatingIPsAssociationTestJSON-35718468 tempest-FloatingIPsAssociationTestJSON-35718468-project-member] [instance: 88cb0961-ad34-4ffc-8fb2-6034f08ee740] Start building block device mappings for instance. {{(pid=61806) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 634.421762] env[61806]: DEBUG nova.network.neutron [None req-3e50272a-5153-4e11-beee-d774a917f2df tempest-ServerPasswordTestJSON-1741413915 tempest-ServerPasswordTestJSON-1741413915-project-member] [instance: 0d757385-ff1c-49c9-825f-f9d542ae074f] Instance cache missing network info. {{(pid=61806) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 634.466376] env[61806]: DEBUG nova.network.neutron [None req-f8b2f38e-90f5-4b03-89bc-ca1926adb40c tempest-FloatingIPsAssociationTestJSON-35718468 tempest-FloatingIPsAssociationTestJSON-35718468-project-member] [instance: 88cb0961-ad34-4ffc-8fb2-6034f08ee740] Successfully created port: 3eb55d85-873d-4f86-bf99-5f748157a7af {{(pid=61806) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 634.521209] env[61806]: DEBUG nova.network.neutron [None req-3e50272a-5153-4e11-beee-d774a917f2df tempest-ServerPasswordTestJSON-1741413915 tempest-ServerPasswordTestJSON-1741413915-project-member] [instance: 0d757385-ff1c-49c9-825f-f9d542ae074f] Updating instance_info_cache with network_info: [] {{(pid=61806) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 634.628317] env[61806]: DEBUG nova.scheduler.client.report [None req-e34d2bcf-266f-4faa-8f31-6bd93f6ab176 tempest-ImagesTestJSON-618852556 tempest-ImagesTestJSON-618852556-project-member] Inventory has not changed for provider a2858be1-fd22-4e08-979e-87ad25293407 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 139, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61806) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 634.906369] env[61806]: DEBUG nova.compute.manager [req-ac65d8a9-5fa9-4955-b792-dcb1b7adbed2 req-39a55996-10ba-42b1-8d09-c5322fe0dff0 service nova] [instance: 0d757385-ff1c-49c9-825f-f9d542ae074f] Received event network-vif-deleted-fc109b8a-4096-4169-bf3b-f19ce33059d3 {{(pid=61806) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 635.027384] env[61806]: DEBUG oslo_concurrency.lockutils [None req-3e50272a-5153-4e11-beee-d774a917f2df tempest-ServerPasswordTestJSON-1741413915 tempest-ServerPasswordTestJSON-1741413915-project-member] Releasing lock "refresh_cache-0d757385-ff1c-49c9-825f-f9d542ae074f" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 635.028188] env[61806]: DEBUG nova.compute.manager [None req-3e50272a-5153-4e11-beee-d774a917f2df tempest-ServerPasswordTestJSON-1741413915 tempest-ServerPasswordTestJSON-1741413915-project-member] [instance: 0d757385-ff1c-49c9-825f-f9d542ae074f] Start destroying the instance on the hypervisor. {{(pid=61806) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 635.028654] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-3e50272a-5153-4e11-beee-d774a917f2df tempest-ServerPasswordTestJSON-1741413915 tempest-ServerPasswordTestJSON-1741413915-project-member] [instance: 0d757385-ff1c-49c9-825f-f9d542ae074f] Destroying instance {{(pid=61806) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 635.029621] env[61806]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-4516c398-2b71-4770-88ef-96b816ac15ec {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 635.039951] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-974a1d8a-3889-450e-bca2-b9b384433d16 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 635.064733] env[61806]: WARNING nova.virt.vmwareapi.vmops [None req-3e50272a-5153-4e11-beee-d774a917f2df tempest-ServerPasswordTestJSON-1741413915 tempest-ServerPasswordTestJSON-1741413915-project-member] [instance: 0d757385-ff1c-49c9-825f-f9d542ae074f] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 0d757385-ff1c-49c9-825f-f9d542ae074f could not be found. [ 635.065076] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-3e50272a-5153-4e11-beee-d774a917f2df tempest-ServerPasswordTestJSON-1741413915 tempest-ServerPasswordTestJSON-1741413915-project-member] [instance: 0d757385-ff1c-49c9-825f-f9d542ae074f] Instance destroyed {{(pid=61806) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 635.065355] env[61806]: INFO nova.compute.manager [None req-3e50272a-5153-4e11-beee-d774a917f2df tempest-ServerPasswordTestJSON-1741413915 tempest-ServerPasswordTestJSON-1741413915-project-member] [instance: 0d757385-ff1c-49c9-825f-f9d542ae074f] Took 0.04 seconds to destroy the instance on the hypervisor. [ 635.065704] env[61806]: DEBUG oslo.service.loopingcall [None req-3e50272a-5153-4e11-beee-d774a917f2df tempest-ServerPasswordTestJSON-1741413915 tempest-ServerPasswordTestJSON-1741413915-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61806) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 635.066081] env[61806]: DEBUG nova.compute.manager [-] [instance: 0d757385-ff1c-49c9-825f-f9d542ae074f] Deallocating network for instance {{(pid=61806) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 635.066268] env[61806]: DEBUG nova.network.neutron [-] [instance: 0d757385-ff1c-49c9-825f-f9d542ae074f] deallocate_for_instance() {{(pid=61806) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 635.092552] env[61806]: DEBUG nova.network.neutron [-] [instance: 0d757385-ff1c-49c9-825f-f9d542ae074f] Instance cache missing network info. {{(pid=61806) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 635.135226] env[61806]: DEBUG oslo_concurrency.lockutils [None req-e34d2bcf-266f-4faa-8f31-6bd93f6ab176 tempest-ImagesTestJSON-618852556 tempest-ImagesTestJSON-618852556-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.012s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 635.135894] env[61806]: ERROR nova.compute.manager [None req-e34d2bcf-266f-4faa-8f31-6bd93f6ab176 tempest-ImagesTestJSON-618852556 tempest-ImagesTestJSON-618852556-project-member] [instance: 65f3b2f3-6d25-44f0-95cb-f9a7e75f8e82] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port ddedeb2e-1da8-4774-925f-5f0e09c0cae8, please check neutron logs for more information. [ 635.135894] env[61806]: ERROR nova.compute.manager [instance: 65f3b2f3-6d25-44f0-95cb-f9a7e75f8e82] Traceback (most recent call last): [ 635.135894] env[61806]: ERROR nova.compute.manager [instance: 65f3b2f3-6d25-44f0-95cb-f9a7e75f8e82] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 635.135894] env[61806]: ERROR nova.compute.manager [instance: 65f3b2f3-6d25-44f0-95cb-f9a7e75f8e82] self.driver.spawn(context, instance, image_meta, [ 635.135894] env[61806]: ERROR nova.compute.manager [instance: 65f3b2f3-6d25-44f0-95cb-f9a7e75f8e82] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 635.135894] env[61806]: ERROR nova.compute.manager [instance: 65f3b2f3-6d25-44f0-95cb-f9a7e75f8e82] self._vmops.spawn(context, instance, image_meta, injected_files, [ 635.135894] env[61806]: ERROR nova.compute.manager [instance: 65f3b2f3-6d25-44f0-95cb-f9a7e75f8e82] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 635.135894] env[61806]: ERROR nova.compute.manager [instance: 65f3b2f3-6d25-44f0-95cb-f9a7e75f8e82] vm_ref = self.build_virtual_machine(instance, [ 635.135894] env[61806]: ERROR nova.compute.manager [instance: 65f3b2f3-6d25-44f0-95cb-f9a7e75f8e82] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 635.135894] env[61806]: ERROR nova.compute.manager [instance: 65f3b2f3-6d25-44f0-95cb-f9a7e75f8e82] vif_infos = vmwarevif.get_vif_info(self._session, [ 635.135894] env[61806]: ERROR nova.compute.manager [instance: 65f3b2f3-6d25-44f0-95cb-f9a7e75f8e82] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 635.136227] env[61806]: ERROR nova.compute.manager [instance: 65f3b2f3-6d25-44f0-95cb-f9a7e75f8e82] for vif in network_info: [ 635.136227] env[61806]: ERROR nova.compute.manager [instance: 65f3b2f3-6d25-44f0-95cb-f9a7e75f8e82] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 635.136227] env[61806]: ERROR nova.compute.manager [instance: 65f3b2f3-6d25-44f0-95cb-f9a7e75f8e82] return self._sync_wrapper(fn, *args, **kwargs) [ 635.136227] env[61806]: ERROR nova.compute.manager [instance: 65f3b2f3-6d25-44f0-95cb-f9a7e75f8e82] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 635.136227] env[61806]: ERROR nova.compute.manager [instance: 65f3b2f3-6d25-44f0-95cb-f9a7e75f8e82] self.wait() [ 635.136227] env[61806]: ERROR nova.compute.manager [instance: 65f3b2f3-6d25-44f0-95cb-f9a7e75f8e82] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 635.136227] env[61806]: ERROR nova.compute.manager [instance: 65f3b2f3-6d25-44f0-95cb-f9a7e75f8e82] self[:] = self._gt.wait() [ 635.136227] env[61806]: ERROR nova.compute.manager [instance: 65f3b2f3-6d25-44f0-95cb-f9a7e75f8e82] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 635.136227] env[61806]: ERROR nova.compute.manager [instance: 65f3b2f3-6d25-44f0-95cb-f9a7e75f8e82] return self._exit_event.wait() [ 635.136227] env[61806]: ERROR nova.compute.manager [instance: 65f3b2f3-6d25-44f0-95cb-f9a7e75f8e82] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 635.136227] env[61806]: ERROR nova.compute.manager [instance: 65f3b2f3-6d25-44f0-95cb-f9a7e75f8e82] result = hub.switch() [ 635.136227] env[61806]: ERROR nova.compute.manager [instance: 65f3b2f3-6d25-44f0-95cb-f9a7e75f8e82] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 635.136227] env[61806]: ERROR nova.compute.manager [instance: 65f3b2f3-6d25-44f0-95cb-f9a7e75f8e82] return self.greenlet.switch() [ 635.136526] env[61806]: ERROR nova.compute.manager [instance: 65f3b2f3-6d25-44f0-95cb-f9a7e75f8e82] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 635.136526] env[61806]: ERROR nova.compute.manager [instance: 65f3b2f3-6d25-44f0-95cb-f9a7e75f8e82] result = function(*args, **kwargs) [ 635.136526] env[61806]: ERROR nova.compute.manager [instance: 65f3b2f3-6d25-44f0-95cb-f9a7e75f8e82] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 635.136526] env[61806]: ERROR nova.compute.manager [instance: 65f3b2f3-6d25-44f0-95cb-f9a7e75f8e82] return func(*args, **kwargs) [ 635.136526] env[61806]: ERROR nova.compute.manager [instance: 65f3b2f3-6d25-44f0-95cb-f9a7e75f8e82] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 635.136526] env[61806]: ERROR nova.compute.manager [instance: 65f3b2f3-6d25-44f0-95cb-f9a7e75f8e82] raise e [ 635.136526] env[61806]: ERROR nova.compute.manager [instance: 65f3b2f3-6d25-44f0-95cb-f9a7e75f8e82] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 635.136526] env[61806]: ERROR nova.compute.manager [instance: 65f3b2f3-6d25-44f0-95cb-f9a7e75f8e82] nwinfo = self.network_api.allocate_for_instance( [ 635.136526] env[61806]: ERROR nova.compute.manager [instance: 65f3b2f3-6d25-44f0-95cb-f9a7e75f8e82] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 635.136526] env[61806]: ERROR nova.compute.manager [instance: 65f3b2f3-6d25-44f0-95cb-f9a7e75f8e82] created_port_ids = self._update_ports_for_instance( [ 635.136526] env[61806]: ERROR nova.compute.manager [instance: 65f3b2f3-6d25-44f0-95cb-f9a7e75f8e82] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 635.136526] env[61806]: ERROR nova.compute.manager [instance: 65f3b2f3-6d25-44f0-95cb-f9a7e75f8e82] with excutils.save_and_reraise_exception(): [ 635.136526] env[61806]: ERROR nova.compute.manager [instance: 65f3b2f3-6d25-44f0-95cb-f9a7e75f8e82] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 635.136828] env[61806]: ERROR nova.compute.manager [instance: 65f3b2f3-6d25-44f0-95cb-f9a7e75f8e82] self.force_reraise() [ 635.136828] env[61806]: ERROR nova.compute.manager [instance: 65f3b2f3-6d25-44f0-95cb-f9a7e75f8e82] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 635.136828] env[61806]: ERROR nova.compute.manager [instance: 65f3b2f3-6d25-44f0-95cb-f9a7e75f8e82] raise self.value [ 635.136828] env[61806]: ERROR nova.compute.manager [instance: 65f3b2f3-6d25-44f0-95cb-f9a7e75f8e82] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 635.136828] env[61806]: ERROR nova.compute.manager [instance: 65f3b2f3-6d25-44f0-95cb-f9a7e75f8e82] updated_port = self._update_port( [ 635.136828] env[61806]: ERROR nova.compute.manager [instance: 65f3b2f3-6d25-44f0-95cb-f9a7e75f8e82] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 635.136828] env[61806]: ERROR nova.compute.manager [instance: 65f3b2f3-6d25-44f0-95cb-f9a7e75f8e82] _ensure_no_port_binding_failure(port) [ 635.136828] env[61806]: ERROR nova.compute.manager [instance: 65f3b2f3-6d25-44f0-95cb-f9a7e75f8e82] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 635.136828] env[61806]: ERROR nova.compute.manager [instance: 65f3b2f3-6d25-44f0-95cb-f9a7e75f8e82] raise exception.PortBindingFailed(port_id=port['id']) [ 635.136828] env[61806]: ERROR nova.compute.manager [instance: 65f3b2f3-6d25-44f0-95cb-f9a7e75f8e82] nova.exception.PortBindingFailed: Binding failed for port ddedeb2e-1da8-4774-925f-5f0e09c0cae8, please check neutron logs for more information. [ 635.136828] env[61806]: ERROR nova.compute.manager [instance: 65f3b2f3-6d25-44f0-95cb-f9a7e75f8e82] [ 635.137176] env[61806]: DEBUG nova.compute.utils [None req-e34d2bcf-266f-4faa-8f31-6bd93f6ab176 tempest-ImagesTestJSON-618852556 tempest-ImagesTestJSON-618852556-project-member] [instance: 65f3b2f3-6d25-44f0-95cb-f9a7e75f8e82] Binding failed for port ddedeb2e-1da8-4774-925f-5f0e09c0cae8, please check neutron logs for more information. {{(pid=61806) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 635.138436] env[61806]: DEBUG nova.compute.manager [None req-e34d2bcf-266f-4faa-8f31-6bd93f6ab176 tempest-ImagesTestJSON-618852556 tempest-ImagesTestJSON-618852556-project-member] [instance: 65f3b2f3-6d25-44f0-95cb-f9a7e75f8e82] Build of instance 65f3b2f3-6d25-44f0-95cb-f9a7e75f8e82 was re-scheduled: Binding failed for port ddedeb2e-1da8-4774-925f-5f0e09c0cae8, please check neutron logs for more information. {{(pid=61806) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 635.138877] env[61806]: DEBUG nova.compute.manager [None req-e34d2bcf-266f-4faa-8f31-6bd93f6ab176 tempest-ImagesTestJSON-618852556 tempest-ImagesTestJSON-618852556-project-member] [instance: 65f3b2f3-6d25-44f0-95cb-f9a7e75f8e82] Unplugging VIFs for instance {{(pid=61806) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 635.139152] env[61806]: DEBUG oslo_concurrency.lockutils [None req-e34d2bcf-266f-4faa-8f31-6bd93f6ab176 tempest-ImagesTestJSON-618852556 tempest-ImagesTestJSON-618852556-project-member] Acquiring lock "refresh_cache-65f3b2f3-6d25-44f0-95cb-f9a7e75f8e82" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 635.139331] env[61806]: DEBUG oslo_concurrency.lockutils [None req-e34d2bcf-266f-4faa-8f31-6bd93f6ab176 tempest-ImagesTestJSON-618852556 tempest-ImagesTestJSON-618852556-project-member] Acquired lock "refresh_cache-65f3b2f3-6d25-44f0-95cb-f9a7e75f8e82" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 635.139517] env[61806]: DEBUG nova.network.neutron [None req-e34d2bcf-266f-4faa-8f31-6bd93f6ab176 tempest-ImagesTestJSON-618852556 tempest-ImagesTestJSON-618852556-project-member] [instance: 65f3b2f3-6d25-44f0-95cb-f9a7e75f8e82] Building network info cache for instance {{(pid=61806) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 635.140706] env[61806]: DEBUG oslo_concurrency.lockutils [None req-e8db3b70-c93a-419a-89f8-b994313357e1 tempest-ServerShowV254Test-608938657 tempest-ServerShowV254Test-608938657-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 15.566s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 635.145014] env[61806]: INFO nova.compute.claims [None req-e8db3b70-c93a-419a-89f8-b994313357e1 tempest-ServerShowV254Test-608938657 tempest-ServerShowV254Test-608938657-project-member] [instance: 8016bafc-b287-4df1-bd4b-a00210177a86] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 635.149740] env[61806]: DEBUG nova.compute.manager [None req-f8b2f38e-90f5-4b03-89bc-ca1926adb40c tempest-FloatingIPsAssociationTestJSON-35718468 tempest-FloatingIPsAssociationTestJSON-35718468-project-member] [instance: 88cb0961-ad34-4ffc-8fb2-6034f08ee740] Start spawning the instance on the hypervisor. {{(pid=61806) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 635.182109] env[61806]: DEBUG nova.virt.hardware [None req-f8b2f38e-90f5-4b03-89bc-ca1926adb40c tempest-FloatingIPsAssociationTestJSON-35718468 tempest-FloatingIPsAssociationTestJSON-35718468-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-15T18:05:10Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-15T18:04:52Z,direct_url=,disk_format='vmdk',id=ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='84ba83fce2c349a6988173c5d6fc3b07',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-15T18:04:53Z,virtual_size=,visibility=), allow threads: False {{(pid=61806) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 635.182109] env[61806]: DEBUG nova.virt.hardware [None req-f8b2f38e-90f5-4b03-89bc-ca1926adb40c tempest-FloatingIPsAssociationTestJSON-35718468 tempest-FloatingIPsAssociationTestJSON-35718468-project-member] Flavor limits 0:0:0 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 635.182109] env[61806]: DEBUG nova.virt.hardware [None req-f8b2f38e-90f5-4b03-89bc-ca1926adb40c tempest-FloatingIPsAssociationTestJSON-35718468 tempest-FloatingIPsAssociationTestJSON-35718468-project-member] Image limits 0:0:0 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 635.182322] env[61806]: DEBUG nova.virt.hardware [None req-f8b2f38e-90f5-4b03-89bc-ca1926adb40c tempest-FloatingIPsAssociationTestJSON-35718468 tempest-FloatingIPsAssociationTestJSON-35718468-project-member] Flavor pref 0:0:0 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 635.182322] env[61806]: DEBUG nova.virt.hardware [None req-f8b2f38e-90f5-4b03-89bc-ca1926adb40c tempest-FloatingIPsAssociationTestJSON-35718468 tempest-FloatingIPsAssociationTestJSON-35718468-project-member] Image pref 0:0:0 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 635.182439] env[61806]: DEBUG nova.virt.hardware [None req-f8b2f38e-90f5-4b03-89bc-ca1926adb40c tempest-FloatingIPsAssociationTestJSON-35718468 tempest-FloatingIPsAssociationTestJSON-35718468-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 635.182727] env[61806]: DEBUG nova.virt.hardware [None req-f8b2f38e-90f5-4b03-89bc-ca1926adb40c tempest-FloatingIPsAssociationTestJSON-35718468 tempest-FloatingIPsAssociationTestJSON-35718468-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61806) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 635.182986] env[61806]: DEBUG nova.virt.hardware [None req-f8b2f38e-90f5-4b03-89bc-ca1926adb40c tempest-FloatingIPsAssociationTestJSON-35718468 tempest-FloatingIPsAssociationTestJSON-35718468-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61806) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 635.183416] env[61806]: DEBUG nova.virt.hardware [None req-f8b2f38e-90f5-4b03-89bc-ca1926adb40c tempest-FloatingIPsAssociationTestJSON-35718468 tempest-FloatingIPsAssociationTestJSON-35718468-project-member] Got 1 possible topologies {{(pid=61806) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 635.183653] env[61806]: DEBUG nova.virt.hardware [None req-f8b2f38e-90f5-4b03-89bc-ca1926adb40c tempest-FloatingIPsAssociationTestJSON-35718468 tempest-FloatingIPsAssociationTestJSON-35718468-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61806) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 635.183900] env[61806]: DEBUG nova.virt.hardware [None req-f8b2f38e-90f5-4b03-89bc-ca1926adb40c tempest-FloatingIPsAssociationTestJSON-35718468 tempest-FloatingIPsAssociationTestJSON-35718468-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61806) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 635.185692] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b9f5d127-bbdb-416b-974a-ea121bbc6978 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 635.195734] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-df3192cc-c778-4eba-be31-df822cce8dc6 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 635.337399] env[61806]: DEBUG oslo_concurrency.lockutils [None req-b5a6ece2-b279-4bbd-a3d1-16582c4d8894 tempest-ServersTestMultiNic-1480543082 tempest-ServersTestMultiNic-1480543082-project-member] Acquiring lock "24c83131-1451-429f-84fe-dcd070898d65" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 635.337622] env[61806]: DEBUG oslo_concurrency.lockutils [None req-b5a6ece2-b279-4bbd-a3d1-16582c4d8894 tempest-ServersTestMultiNic-1480543082 tempest-ServersTestMultiNic-1480543082-project-member] Lock "24c83131-1451-429f-84fe-dcd070898d65" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 635.597770] env[61806]: DEBUG nova.network.neutron [-] [instance: 0d757385-ff1c-49c9-825f-f9d542ae074f] Updating instance_info_cache with network_info: [] {{(pid=61806) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 635.676649] env[61806]: ERROR nova.compute.manager [None req-f8b2f38e-90f5-4b03-89bc-ca1926adb40c tempest-FloatingIPsAssociationTestJSON-35718468 tempest-FloatingIPsAssociationTestJSON-35718468-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 3eb55d85-873d-4f86-bf99-5f748157a7af, please check neutron logs for more information. [ 635.676649] env[61806]: ERROR nova.compute.manager Traceback (most recent call last): [ 635.676649] env[61806]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 635.676649] env[61806]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 635.676649] env[61806]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 635.676649] env[61806]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 635.676649] env[61806]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 635.676649] env[61806]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 635.676649] env[61806]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 635.676649] env[61806]: ERROR nova.compute.manager self.force_reraise() [ 635.676649] env[61806]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 635.676649] env[61806]: ERROR nova.compute.manager raise self.value [ 635.676649] env[61806]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 635.676649] env[61806]: ERROR nova.compute.manager updated_port = self._update_port( [ 635.676649] env[61806]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 635.676649] env[61806]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 635.677074] env[61806]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 635.677074] env[61806]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 635.677074] env[61806]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 3eb55d85-873d-4f86-bf99-5f748157a7af, please check neutron logs for more information. [ 635.677074] env[61806]: ERROR nova.compute.manager [ 635.677074] env[61806]: Traceback (most recent call last): [ 635.677074] env[61806]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 635.677074] env[61806]: listener.cb(fileno) [ 635.677074] env[61806]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 635.677074] env[61806]: result = function(*args, **kwargs) [ 635.677074] env[61806]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 635.677074] env[61806]: return func(*args, **kwargs) [ 635.677074] env[61806]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 635.677074] env[61806]: raise e [ 635.677074] env[61806]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 635.677074] env[61806]: nwinfo = self.network_api.allocate_for_instance( [ 635.677074] env[61806]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 635.677074] env[61806]: created_port_ids = self._update_ports_for_instance( [ 635.677074] env[61806]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 635.677074] env[61806]: with excutils.save_and_reraise_exception(): [ 635.677074] env[61806]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 635.677074] env[61806]: self.force_reraise() [ 635.677074] env[61806]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 635.677074] env[61806]: raise self.value [ 635.677074] env[61806]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 635.677074] env[61806]: updated_port = self._update_port( [ 635.677074] env[61806]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 635.677074] env[61806]: _ensure_no_port_binding_failure(port) [ 635.677074] env[61806]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 635.677074] env[61806]: raise exception.PortBindingFailed(port_id=port['id']) [ 635.677723] env[61806]: nova.exception.PortBindingFailed: Binding failed for port 3eb55d85-873d-4f86-bf99-5f748157a7af, please check neutron logs for more information. [ 635.677723] env[61806]: Removing descriptor: 16 [ 635.677723] env[61806]: ERROR nova.compute.manager [None req-f8b2f38e-90f5-4b03-89bc-ca1926adb40c tempest-FloatingIPsAssociationTestJSON-35718468 tempest-FloatingIPsAssociationTestJSON-35718468-project-member] [instance: 88cb0961-ad34-4ffc-8fb2-6034f08ee740] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 3eb55d85-873d-4f86-bf99-5f748157a7af, please check neutron logs for more information. [ 635.677723] env[61806]: ERROR nova.compute.manager [instance: 88cb0961-ad34-4ffc-8fb2-6034f08ee740] Traceback (most recent call last): [ 635.677723] env[61806]: ERROR nova.compute.manager [instance: 88cb0961-ad34-4ffc-8fb2-6034f08ee740] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 635.677723] env[61806]: ERROR nova.compute.manager [instance: 88cb0961-ad34-4ffc-8fb2-6034f08ee740] yield resources [ 635.677723] env[61806]: ERROR nova.compute.manager [instance: 88cb0961-ad34-4ffc-8fb2-6034f08ee740] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 635.677723] env[61806]: ERROR nova.compute.manager [instance: 88cb0961-ad34-4ffc-8fb2-6034f08ee740] self.driver.spawn(context, instance, image_meta, [ 635.677723] env[61806]: ERROR nova.compute.manager [instance: 88cb0961-ad34-4ffc-8fb2-6034f08ee740] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 635.677723] env[61806]: ERROR nova.compute.manager [instance: 88cb0961-ad34-4ffc-8fb2-6034f08ee740] self._vmops.spawn(context, instance, image_meta, injected_files, [ 635.677723] env[61806]: ERROR nova.compute.manager [instance: 88cb0961-ad34-4ffc-8fb2-6034f08ee740] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 635.677723] env[61806]: ERROR nova.compute.manager [instance: 88cb0961-ad34-4ffc-8fb2-6034f08ee740] vm_ref = self.build_virtual_machine(instance, [ 635.678031] env[61806]: ERROR nova.compute.manager [instance: 88cb0961-ad34-4ffc-8fb2-6034f08ee740] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 635.678031] env[61806]: ERROR nova.compute.manager [instance: 88cb0961-ad34-4ffc-8fb2-6034f08ee740] vif_infos = vmwarevif.get_vif_info(self._session, [ 635.678031] env[61806]: ERROR nova.compute.manager [instance: 88cb0961-ad34-4ffc-8fb2-6034f08ee740] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 635.678031] env[61806]: ERROR nova.compute.manager [instance: 88cb0961-ad34-4ffc-8fb2-6034f08ee740] for vif in network_info: [ 635.678031] env[61806]: ERROR nova.compute.manager [instance: 88cb0961-ad34-4ffc-8fb2-6034f08ee740] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 635.678031] env[61806]: ERROR nova.compute.manager [instance: 88cb0961-ad34-4ffc-8fb2-6034f08ee740] return self._sync_wrapper(fn, *args, **kwargs) [ 635.678031] env[61806]: ERROR nova.compute.manager [instance: 88cb0961-ad34-4ffc-8fb2-6034f08ee740] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 635.678031] env[61806]: ERROR nova.compute.manager [instance: 88cb0961-ad34-4ffc-8fb2-6034f08ee740] self.wait() [ 635.678031] env[61806]: ERROR nova.compute.manager [instance: 88cb0961-ad34-4ffc-8fb2-6034f08ee740] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 635.678031] env[61806]: ERROR nova.compute.manager [instance: 88cb0961-ad34-4ffc-8fb2-6034f08ee740] self[:] = self._gt.wait() [ 635.678031] env[61806]: ERROR nova.compute.manager [instance: 88cb0961-ad34-4ffc-8fb2-6034f08ee740] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 635.678031] env[61806]: ERROR nova.compute.manager [instance: 88cb0961-ad34-4ffc-8fb2-6034f08ee740] return self._exit_event.wait() [ 635.678031] env[61806]: ERROR nova.compute.manager [instance: 88cb0961-ad34-4ffc-8fb2-6034f08ee740] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 635.678365] env[61806]: ERROR nova.compute.manager [instance: 88cb0961-ad34-4ffc-8fb2-6034f08ee740] result = hub.switch() [ 635.678365] env[61806]: ERROR nova.compute.manager [instance: 88cb0961-ad34-4ffc-8fb2-6034f08ee740] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 635.678365] env[61806]: ERROR nova.compute.manager [instance: 88cb0961-ad34-4ffc-8fb2-6034f08ee740] return self.greenlet.switch() [ 635.678365] env[61806]: ERROR nova.compute.manager [instance: 88cb0961-ad34-4ffc-8fb2-6034f08ee740] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 635.678365] env[61806]: ERROR nova.compute.manager [instance: 88cb0961-ad34-4ffc-8fb2-6034f08ee740] result = function(*args, **kwargs) [ 635.678365] env[61806]: ERROR nova.compute.manager [instance: 88cb0961-ad34-4ffc-8fb2-6034f08ee740] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 635.678365] env[61806]: ERROR nova.compute.manager [instance: 88cb0961-ad34-4ffc-8fb2-6034f08ee740] return func(*args, **kwargs) [ 635.678365] env[61806]: ERROR nova.compute.manager [instance: 88cb0961-ad34-4ffc-8fb2-6034f08ee740] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 635.678365] env[61806]: ERROR nova.compute.manager [instance: 88cb0961-ad34-4ffc-8fb2-6034f08ee740] raise e [ 635.678365] env[61806]: ERROR nova.compute.manager [instance: 88cb0961-ad34-4ffc-8fb2-6034f08ee740] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 635.678365] env[61806]: ERROR nova.compute.manager [instance: 88cb0961-ad34-4ffc-8fb2-6034f08ee740] nwinfo = self.network_api.allocate_for_instance( [ 635.678365] env[61806]: ERROR nova.compute.manager [instance: 88cb0961-ad34-4ffc-8fb2-6034f08ee740] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 635.678365] env[61806]: ERROR nova.compute.manager [instance: 88cb0961-ad34-4ffc-8fb2-6034f08ee740] created_port_ids = self._update_ports_for_instance( [ 635.678822] env[61806]: ERROR nova.compute.manager [instance: 88cb0961-ad34-4ffc-8fb2-6034f08ee740] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 635.678822] env[61806]: ERROR nova.compute.manager [instance: 88cb0961-ad34-4ffc-8fb2-6034f08ee740] with excutils.save_and_reraise_exception(): [ 635.678822] env[61806]: ERROR nova.compute.manager [instance: 88cb0961-ad34-4ffc-8fb2-6034f08ee740] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 635.678822] env[61806]: ERROR nova.compute.manager [instance: 88cb0961-ad34-4ffc-8fb2-6034f08ee740] self.force_reraise() [ 635.678822] env[61806]: ERROR nova.compute.manager [instance: 88cb0961-ad34-4ffc-8fb2-6034f08ee740] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 635.678822] env[61806]: ERROR nova.compute.manager [instance: 88cb0961-ad34-4ffc-8fb2-6034f08ee740] raise self.value [ 635.678822] env[61806]: ERROR nova.compute.manager [instance: 88cb0961-ad34-4ffc-8fb2-6034f08ee740] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 635.678822] env[61806]: ERROR nova.compute.manager [instance: 88cb0961-ad34-4ffc-8fb2-6034f08ee740] updated_port = self._update_port( [ 635.678822] env[61806]: ERROR nova.compute.manager [instance: 88cb0961-ad34-4ffc-8fb2-6034f08ee740] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 635.678822] env[61806]: ERROR nova.compute.manager [instance: 88cb0961-ad34-4ffc-8fb2-6034f08ee740] _ensure_no_port_binding_failure(port) [ 635.678822] env[61806]: ERROR nova.compute.manager [instance: 88cb0961-ad34-4ffc-8fb2-6034f08ee740] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 635.678822] env[61806]: ERROR nova.compute.manager [instance: 88cb0961-ad34-4ffc-8fb2-6034f08ee740] raise exception.PortBindingFailed(port_id=port['id']) [ 635.679134] env[61806]: ERROR nova.compute.manager [instance: 88cb0961-ad34-4ffc-8fb2-6034f08ee740] nova.exception.PortBindingFailed: Binding failed for port 3eb55d85-873d-4f86-bf99-5f748157a7af, please check neutron logs for more information. [ 635.679134] env[61806]: ERROR nova.compute.manager [instance: 88cb0961-ad34-4ffc-8fb2-6034f08ee740] [ 635.679134] env[61806]: INFO nova.compute.manager [None req-f8b2f38e-90f5-4b03-89bc-ca1926adb40c tempest-FloatingIPsAssociationTestJSON-35718468 tempest-FloatingIPsAssociationTestJSON-35718468-project-member] [instance: 88cb0961-ad34-4ffc-8fb2-6034f08ee740] Terminating instance [ 635.683242] env[61806]: DEBUG oslo_concurrency.lockutils [None req-f8b2f38e-90f5-4b03-89bc-ca1926adb40c tempest-FloatingIPsAssociationTestJSON-35718468 tempest-FloatingIPsAssociationTestJSON-35718468-project-member] Acquiring lock "refresh_cache-88cb0961-ad34-4ffc-8fb2-6034f08ee740" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 635.683660] env[61806]: DEBUG oslo_concurrency.lockutils [None req-f8b2f38e-90f5-4b03-89bc-ca1926adb40c tempest-FloatingIPsAssociationTestJSON-35718468 tempest-FloatingIPsAssociationTestJSON-35718468-project-member] Acquired lock "refresh_cache-88cb0961-ad34-4ffc-8fb2-6034f08ee740" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 635.683660] env[61806]: DEBUG nova.network.neutron [None req-f8b2f38e-90f5-4b03-89bc-ca1926adb40c tempest-FloatingIPsAssociationTestJSON-35718468 tempest-FloatingIPsAssociationTestJSON-35718468-project-member] [instance: 88cb0961-ad34-4ffc-8fb2-6034f08ee740] Building network info cache for instance {{(pid=61806) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 635.685738] env[61806]: DEBUG nova.network.neutron [None req-e34d2bcf-266f-4faa-8f31-6bd93f6ab176 tempest-ImagesTestJSON-618852556 tempest-ImagesTestJSON-618852556-project-member] [instance: 65f3b2f3-6d25-44f0-95cb-f9a7e75f8e82] Instance cache missing network info. {{(pid=61806) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 635.824035] env[61806]: DEBUG nova.network.neutron [None req-e34d2bcf-266f-4faa-8f31-6bd93f6ab176 tempest-ImagesTestJSON-618852556 tempest-ImagesTestJSON-618852556-project-member] [instance: 65f3b2f3-6d25-44f0-95cb-f9a7e75f8e82] Updating instance_info_cache with network_info: [] {{(pid=61806) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 636.102011] env[61806]: INFO nova.compute.manager [-] [instance: 0d757385-ff1c-49c9-825f-f9d542ae074f] Took 1.03 seconds to deallocate network for instance. [ 636.110441] env[61806]: DEBUG nova.compute.claims [None req-3e50272a-5153-4e11-beee-d774a917f2df tempest-ServerPasswordTestJSON-1741413915 tempest-ServerPasswordTestJSON-1741413915-project-member] [instance: 0d757385-ff1c-49c9-825f-f9d542ae074f] Aborting claim: {{(pid=61806) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 636.110645] env[61806]: DEBUG oslo_concurrency.lockutils [None req-3e50272a-5153-4e11-beee-d774a917f2df tempest-ServerPasswordTestJSON-1741413915 tempest-ServerPasswordTestJSON-1741413915-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 636.325941] env[61806]: DEBUG oslo_concurrency.lockutils [None req-e34d2bcf-266f-4faa-8f31-6bd93f6ab176 tempest-ImagesTestJSON-618852556 tempest-ImagesTestJSON-618852556-project-member] Releasing lock "refresh_cache-65f3b2f3-6d25-44f0-95cb-f9a7e75f8e82" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 636.326209] env[61806]: DEBUG nova.compute.manager [None req-e34d2bcf-266f-4faa-8f31-6bd93f6ab176 tempest-ImagesTestJSON-618852556 tempest-ImagesTestJSON-618852556-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61806) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 636.326450] env[61806]: DEBUG nova.compute.manager [None req-e34d2bcf-266f-4faa-8f31-6bd93f6ab176 tempest-ImagesTestJSON-618852556 tempest-ImagesTestJSON-618852556-project-member] [instance: 65f3b2f3-6d25-44f0-95cb-f9a7e75f8e82] Deallocating network for instance {{(pid=61806) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 636.326644] env[61806]: DEBUG nova.network.neutron [None req-e34d2bcf-266f-4faa-8f31-6bd93f6ab176 tempest-ImagesTestJSON-618852556 tempest-ImagesTestJSON-618852556-project-member] [instance: 65f3b2f3-6d25-44f0-95cb-f9a7e75f8e82] deallocate_for_instance() {{(pid=61806) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 636.341904] env[61806]: DEBUG nova.network.neutron [None req-e34d2bcf-266f-4faa-8f31-6bd93f6ab176 tempest-ImagesTestJSON-618852556 tempest-ImagesTestJSON-618852556-project-member] [instance: 65f3b2f3-6d25-44f0-95cb-f9a7e75f8e82] Instance cache missing network info. {{(pid=61806) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 636.368871] env[61806]: DEBUG nova.network.neutron [None req-f8b2f38e-90f5-4b03-89bc-ca1926adb40c tempest-FloatingIPsAssociationTestJSON-35718468 tempest-FloatingIPsAssociationTestJSON-35718468-project-member] [instance: 88cb0961-ad34-4ffc-8fb2-6034f08ee740] Instance cache missing network info. {{(pid=61806) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 636.464704] env[61806]: DEBUG nova.network.neutron [None req-f8b2f38e-90f5-4b03-89bc-ca1926adb40c tempest-FloatingIPsAssociationTestJSON-35718468 tempest-FloatingIPsAssociationTestJSON-35718468-project-member] [instance: 88cb0961-ad34-4ffc-8fb2-6034f08ee740] Updating instance_info_cache with network_info: [] {{(pid=61806) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 636.573919] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5b18280f-634b-4ad3-8f4b-26cde7b417b4 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 636.580353] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-824a6a2b-6a3f-45ef-8fe4-9c95683c3642 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 636.612309] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2ebb7b4e-cbe7-4604-b7d0-b6767c14e865 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 636.619967] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9d2b0870-9a0c-4747-a1ba-ae0114b678e9 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 636.634587] env[61806]: DEBUG nova.compute.provider_tree [None req-e8db3b70-c93a-419a-89f8-b994313357e1 tempest-ServerShowV254Test-608938657 tempest-ServerShowV254Test-608938657-project-member] Inventory has not changed in ProviderTree for provider: a2858be1-fd22-4e08-979e-87ad25293407 {{(pid=61806) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 636.846236] env[61806]: DEBUG nova.network.neutron [None req-e34d2bcf-266f-4faa-8f31-6bd93f6ab176 tempest-ImagesTestJSON-618852556 tempest-ImagesTestJSON-618852556-project-member] [instance: 65f3b2f3-6d25-44f0-95cb-f9a7e75f8e82] Updating instance_info_cache with network_info: [] {{(pid=61806) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 636.930028] env[61806]: DEBUG nova.compute.manager [req-e863ba3f-710b-4a93-ba78-49277a4ffc43 req-6d0d91b5-4057-48d9-807e-012cc4560b52 service nova] [instance: 88cb0961-ad34-4ffc-8fb2-6034f08ee740] Received event network-changed-3eb55d85-873d-4f86-bf99-5f748157a7af {{(pid=61806) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 636.930028] env[61806]: DEBUG nova.compute.manager [req-e863ba3f-710b-4a93-ba78-49277a4ffc43 req-6d0d91b5-4057-48d9-807e-012cc4560b52 service nova] [instance: 88cb0961-ad34-4ffc-8fb2-6034f08ee740] Refreshing instance network info cache due to event network-changed-3eb55d85-873d-4f86-bf99-5f748157a7af. {{(pid=61806) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 636.930415] env[61806]: DEBUG oslo_concurrency.lockutils [req-e863ba3f-710b-4a93-ba78-49277a4ffc43 req-6d0d91b5-4057-48d9-807e-012cc4560b52 service nova] Acquiring lock "refresh_cache-88cb0961-ad34-4ffc-8fb2-6034f08ee740" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 636.969722] env[61806]: DEBUG oslo_concurrency.lockutils [None req-f8b2f38e-90f5-4b03-89bc-ca1926adb40c tempest-FloatingIPsAssociationTestJSON-35718468 tempest-FloatingIPsAssociationTestJSON-35718468-project-member] Releasing lock "refresh_cache-88cb0961-ad34-4ffc-8fb2-6034f08ee740" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 636.970156] env[61806]: DEBUG nova.compute.manager [None req-f8b2f38e-90f5-4b03-89bc-ca1926adb40c tempest-FloatingIPsAssociationTestJSON-35718468 tempest-FloatingIPsAssociationTestJSON-35718468-project-member] [instance: 88cb0961-ad34-4ffc-8fb2-6034f08ee740] Start destroying the instance on the hypervisor. {{(pid=61806) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 636.970420] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-f8b2f38e-90f5-4b03-89bc-ca1926adb40c tempest-FloatingIPsAssociationTestJSON-35718468 tempest-FloatingIPsAssociationTestJSON-35718468-project-member] [instance: 88cb0961-ad34-4ffc-8fb2-6034f08ee740] Destroying instance {{(pid=61806) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 636.970734] env[61806]: DEBUG oslo_concurrency.lockutils [req-e863ba3f-710b-4a93-ba78-49277a4ffc43 req-6d0d91b5-4057-48d9-807e-012cc4560b52 service nova] Acquired lock "refresh_cache-88cb0961-ad34-4ffc-8fb2-6034f08ee740" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 636.970906] env[61806]: DEBUG nova.network.neutron [req-e863ba3f-710b-4a93-ba78-49277a4ffc43 req-6d0d91b5-4057-48d9-807e-012cc4560b52 service nova] [instance: 88cb0961-ad34-4ffc-8fb2-6034f08ee740] Refreshing network info cache for port 3eb55d85-873d-4f86-bf99-5f748157a7af {{(pid=61806) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 636.971955] env[61806]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-fab8b80f-aac2-4f88-a8c6-68efd04909a4 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 636.980483] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0e1f0832-cf21-4fd6-a8e8-3dcc7321fd40 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 637.002145] env[61806]: WARNING nova.virt.vmwareapi.vmops [None req-f8b2f38e-90f5-4b03-89bc-ca1926adb40c tempest-FloatingIPsAssociationTestJSON-35718468 tempest-FloatingIPsAssociationTestJSON-35718468-project-member] [instance: 88cb0961-ad34-4ffc-8fb2-6034f08ee740] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 88cb0961-ad34-4ffc-8fb2-6034f08ee740 could not be found. [ 637.002464] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-f8b2f38e-90f5-4b03-89bc-ca1926adb40c tempest-FloatingIPsAssociationTestJSON-35718468 tempest-FloatingIPsAssociationTestJSON-35718468-project-member] [instance: 88cb0961-ad34-4ffc-8fb2-6034f08ee740] Instance destroyed {{(pid=61806) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 637.002566] env[61806]: INFO nova.compute.manager [None req-f8b2f38e-90f5-4b03-89bc-ca1926adb40c tempest-FloatingIPsAssociationTestJSON-35718468 tempest-FloatingIPsAssociationTestJSON-35718468-project-member] [instance: 88cb0961-ad34-4ffc-8fb2-6034f08ee740] Took 0.03 seconds to destroy the instance on the hypervisor. [ 637.002807] env[61806]: DEBUG oslo.service.loopingcall [None req-f8b2f38e-90f5-4b03-89bc-ca1926adb40c tempest-FloatingIPsAssociationTestJSON-35718468 tempest-FloatingIPsAssociationTestJSON-35718468-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61806) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 637.003091] env[61806]: DEBUG nova.compute.manager [-] [instance: 88cb0961-ad34-4ffc-8fb2-6034f08ee740] Deallocating network for instance {{(pid=61806) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 637.003191] env[61806]: DEBUG nova.network.neutron [-] [instance: 88cb0961-ad34-4ffc-8fb2-6034f08ee740] deallocate_for_instance() {{(pid=61806) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 637.018409] env[61806]: DEBUG nova.network.neutron [-] [instance: 88cb0961-ad34-4ffc-8fb2-6034f08ee740] Instance cache missing network info. {{(pid=61806) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 637.080811] env[61806]: DEBUG oslo_concurrency.lockutils [None req-35a8f95d-1ddd-492f-a93c-adb382770732 tempest-ServersV294TestFqdnHostnames-1709116086 tempest-ServersV294TestFqdnHostnames-1709116086-project-member] Acquiring lock "c25ddf7c-d0cc-4b73-96da-1dcd6012f072" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 637.081551] env[61806]: DEBUG oslo_concurrency.lockutils [None req-35a8f95d-1ddd-492f-a93c-adb382770732 tempest-ServersV294TestFqdnHostnames-1709116086 tempest-ServersV294TestFqdnHostnames-1709116086-project-member] Lock "c25ddf7c-d0cc-4b73-96da-1dcd6012f072" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 637.137957] env[61806]: DEBUG nova.scheduler.client.report [None req-e8db3b70-c93a-419a-89f8-b994313357e1 tempest-ServerShowV254Test-608938657 tempest-ServerShowV254Test-608938657-project-member] Inventory has not changed for provider a2858be1-fd22-4e08-979e-87ad25293407 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 139, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61806) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 637.349115] env[61806]: INFO nova.compute.manager [None req-e34d2bcf-266f-4faa-8f31-6bd93f6ab176 tempest-ImagesTestJSON-618852556 tempest-ImagesTestJSON-618852556-project-member] [instance: 65f3b2f3-6d25-44f0-95cb-f9a7e75f8e82] Took 1.02 seconds to deallocate network for instance. [ 637.489336] env[61806]: DEBUG nova.network.neutron [req-e863ba3f-710b-4a93-ba78-49277a4ffc43 req-6d0d91b5-4057-48d9-807e-012cc4560b52 service nova] [instance: 88cb0961-ad34-4ffc-8fb2-6034f08ee740] Instance cache missing network info. {{(pid=61806) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 637.520906] env[61806]: DEBUG nova.network.neutron [-] [instance: 88cb0961-ad34-4ffc-8fb2-6034f08ee740] Updating instance_info_cache with network_info: [] {{(pid=61806) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 637.587941] env[61806]: DEBUG nova.network.neutron [req-e863ba3f-710b-4a93-ba78-49277a4ffc43 req-6d0d91b5-4057-48d9-807e-012cc4560b52 service nova] [instance: 88cb0961-ad34-4ffc-8fb2-6034f08ee740] Updating instance_info_cache with network_info: [] {{(pid=61806) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 637.642865] env[61806]: DEBUG oslo_concurrency.lockutils [None req-e8db3b70-c93a-419a-89f8-b994313357e1 tempest-ServerShowV254Test-608938657 tempest-ServerShowV254Test-608938657-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.502s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 637.643410] env[61806]: DEBUG nova.compute.manager [None req-e8db3b70-c93a-419a-89f8-b994313357e1 tempest-ServerShowV254Test-608938657 tempest-ServerShowV254Test-608938657-project-member] [instance: 8016bafc-b287-4df1-bd4b-a00210177a86] Start building networks asynchronously for instance. {{(pid=61806) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 637.645899] env[61806]: DEBUG oslo_concurrency.lockutils [None req-88cd9ecf-9bda-478c-b5e6-5a21edd9271f tempest-MigrationsAdminTest-470098049 tempest-MigrationsAdminTest-470098049-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 16.024s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 637.647304] env[61806]: INFO nova.compute.claims [None req-88cd9ecf-9bda-478c-b5e6-5a21edd9271f tempest-MigrationsAdminTest-470098049 tempest-MigrationsAdminTest-470098049-project-member] [instance: e66c921c-320e-4347-96c9-27aa80ba7f41] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 638.023436] env[61806]: INFO nova.compute.manager [-] [instance: 88cb0961-ad34-4ffc-8fb2-6034f08ee740] Took 1.02 seconds to deallocate network for instance. [ 638.025878] env[61806]: DEBUG nova.compute.claims [None req-f8b2f38e-90f5-4b03-89bc-ca1926adb40c tempest-FloatingIPsAssociationTestJSON-35718468 tempest-FloatingIPsAssociationTestJSON-35718468-project-member] [instance: 88cb0961-ad34-4ffc-8fb2-6034f08ee740] Aborting claim: {{(pid=61806) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 638.026075] env[61806]: DEBUG oslo_concurrency.lockutils [None req-f8b2f38e-90f5-4b03-89bc-ca1926adb40c tempest-FloatingIPsAssociationTestJSON-35718468 tempest-FloatingIPsAssociationTestJSON-35718468-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 638.094528] env[61806]: DEBUG oslo_concurrency.lockutils [req-e863ba3f-710b-4a93-ba78-49277a4ffc43 req-6d0d91b5-4057-48d9-807e-012cc4560b52 service nova] Releasing lock "refresh_cache-88cb0961-ad34-4ffc-8fb2-6034f08ee740" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 638.094528] env[61806]: DEBUG nova.compute.manager [req-e863ba3f-710b-4a93-ba78-49277a4ffc43 req-6d0d91b5-4057-48d9-807e-012cc4560b52 service nova] [instance: 88cb0961-ad34-4ffc-8fb2-6034f08ee740] Received event network-vif-deleted-3eb55d85-873d-4f86-bf99-5f748157a7af {{(pid=61806) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 638.152162] env[61806]: DEBUG nova.compute.utils [None req-e8db3b70-c93a-419a-89f8-b994313357e1 tempest-ServerShowV254Test-608938657 tempest-ServerShowV254Test-608938657-project-member] Using /dev/sd instead of None {{(pid=61806) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 638.156123] env[61806]: DEBUG nova.compute.manager [None req-e8db3b70-c93a-419a-89f8-b994313357e1 tempest-ServerShowV254Test-608938657 tempest-ServerShowV254Test-608938657-project-member] [instance: 8016bafc-b287-4df1-bd4b-a00210177a86] Not allocating networking since 'none' was specified. {{(pid=61806) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1977}} [ 638.380160] env[61806]: INFO nova.scheduler.client.report [None req-e34d2bcf-266f-4faa-8f31-6bd93f6ab176 tempest-ImagesTestJSON-618852556 tempest-ImagesTestJSON-618852556-project-member] Deleted allocations for instance 65f3b2f3-6d25-44f0-95cb-f9a7e75f8e82 [ 638.657391] env[61806]: DEBUG nova.compute.manager [None req-e8db3b70-c93a-419a-89f8-b994313357e1 tempest-ServerShowV254Test-608938657 tempest-ServerShowV254Test-608938657-project-member] [instance: 8016bafc-b287-4df1-bd4b-a00210177a86] Start building block device mappings for instance. {{(pid=61806) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 638.890766] env[61806]: DEBUG oslo_concurrency.lockutils [None req-e34d2bcf-266f-4faa-8f31-6bd93f6ab176 tempest-ImagesTestJSON-618852556 tempest-ImagesTestJSON-618852556-project-member] Lock "65f3b2f3-6d25-44f0-95cb-f9a7e75f8e82" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 87.405s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 639.017086] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f7296679-6114-451e-aa1b-9378e5141af1 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 639.024779] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a6121967-8413-4e10-a468-76d04d8bd9fd {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 639.055881] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7b2ee30e-fdf5-4611-8eaa-6baccdcc8e72 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 639.064941] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6a6dc561-deb6-45b0-abcc-d0186d1437a9 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 639.078683] env[61806]: DEBUG nova.compute.provider_tree [None req-88cd9ecf-9bda-478c-b5e6-5a21edd9271f tempest-MigrationsAdminTest-470098049 tempest-MigrationsAdminTest-470098049-project-member] Inventory has not changed in ProviderTree for provider: a2858be1-fd22-4e08-979e-87ad25293407 {{(pid=61806) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 639.395591] env[61806]: DEBUG nova.compute.manager [None req-b4447f02-a3c7-4b6e-89ba-6bc34ec03205 tempest-InstanceActionsTestJSON-949444324 tempest-InstanceActionsTestJSON-949444324-project-member] [instance: 2681766c-0468-4170-8c51-03389e2bd1ab] Starting instance... {{(pid=61806) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 639.583142] env[61806]: DEBUG nova.scheduler.client.report [None req-88cd9ecf-9bda-478c-b5e6-5a21edd9271f tempest-MigrationsAdminTest-470098049 tempest-MigrationsAdminTest-470098049-project-member] Inventory has not changed for provider a2858be1-fd22-4e08-979e-87ad25293407 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 139, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61806) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 639.669514] env[61806]: DEBUG nova.compute.manager [None req-e8db3b70-c93a-419a-89f8-b994313357e1 tempest-ServerShowV254Test-608938657 tempest-ServerShowV254Test-608938657-project-member] [instance: 8016bafc-b287-4df1-bd4b-a00210177a86] Start spawning the instance on the hypervisor. {{(pid=61806) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 639.691434] env[61806]: DEBUG nova.virt.hardware [None req-e8db3b70-c93a-419a-89f8-b994313357e1 tempest-ServerShowV254Test-608938657 tempest-ServerShowV254Test-608938657-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-15T18:05:10Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-15T18:04:52Z,direct_url=,disk_format='vmdk',id=ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='84ba83fce2c349a6988173c5d6fc3b07',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-15T18:04:53Z,virtual_size=,visibility=), allow threads: False {{(pid=61806) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 639.691700] env[61806]: DEBUG nova.virt.hardware [None req-e8db3b70-c93a-419a-89f8-b994313357e1 tempest-ServerShowV254Test-608938657 tempest-ServerShowV254Test-608938657-project-member] Flavor limits 0:0:0 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 639.691921] env[61806]: DEBUG nova.virt.hardware [None req-e8db3b70-c93a-419a-89f8-b994313357e1 tempest-ServerShowV254Test-608938657 tempest-ServerShowV254Test-608938657-project-member] Image limits 0:0:0 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 639.692039] env[61806]: DEBUG nova.virt.hardware [None req-e8db3b70-c93a-419a-89f8-b994313357e1 tempest-ServerShowV254Test-608938657 tempest-ServerShowV254Test-608938657-project-member] Flavor pref 0:0:0 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 639.692185] env[61806]: DEBUG nova.virt.hardware [None req-e8db3b70-c93a-419a-89f8-b994313357e1 tempest-ServerShowV254Test-608938657 tempest-ServerShowV254Test-608938657-project-member] Image pref 0:0:0 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 639.692328] env[61806]: DEBUG nova.virt.hardware [None req-e8db3b70-c93a-419a-89f8-b994313357e1 tempest-ServerShowV254Test-608938657 tempest-ServerShowV254Test-608938657-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 639.692541] env[61806]: DEBUG nova.virt.hardware [None req-e8db3b70-c93a-419a-89f8-b994313357e1 tempest-ServerShowV254Test-608938657 tempest-ServerShowV254Test-608938657-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61806) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 639.692685] env[61806]: DEBUG nova.virt.hardware [None req-e8db3b70-c93a-419a-89f8-b994313357e1 tempest-ServerShowV254Test-608938657 tempest-ServerShowV254Test-608938657-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61806) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 639.692849] env[61806]: DEBUG nova.virt.hardware [None req-e8db3b70-c93a-419a-89f8-b994313357e1 tempest-ServerShowV254Test-608938657 tempest-ServerShowV254Test-608938657-project-member] Got 1 possible topologies {{(pid=61806) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 639.693071] env[61806]: DEBUG nova.virt.hardware [None req-e8db3b70-c93a-419a-89f8-b994313357e1 tempest-ServerShowV254Test-608938657 tempest-ServerShowV254Test-608938657-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61806) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 639.693267] env[61806]: DEBUG nova.virt.hardware [None req-e8db3b70-c93a-419a-89f8-b994313357e1 tempest-ServerShowV254Test-608938657 tempest-ServerShowV254Test-608938657-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61806) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 639.694187] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1f78e27f-2a56-4548-9db9-23fb003465a2 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 639.702031] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-797ab62e-e504-4a70-b86b-fb01eaac22d1 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 639.715999] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-e8db3b70-c93a-419a-89f8-b994313357e1 tempest-ServerShowV254Test-608938657 tempest-ServerShowV254Test-608938657-project-member] [instance: 8016bafc-b287-4df1-bd4b-a00210177a86] Instance VIF info [] {{(pid=61806) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 639.729354] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [None req-e8db3b70-c93a-419a-89f8-b994313357e1 tempest-ServerShowV254Test-608938657 tempest-ServerShowV254Test-608938657-project-member] Creating folder: OpenStack. Parent ref: group-v4. {{(pid=61806) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 639.729604] env[61806]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-861ad192-742f-494e-be3b-65493fe477e6 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 639.742935] env[61806]: WARNING suds.client [-] Web service reported a SOAP processing fault using an unexpected HTTP status code 200. Reporting as an internal server error. [ 639.743120] env[61806]: DEBUG oslo_vmware.api [-] Fault list: [DuplicateName] {{(pid=61806) _invoke_api /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:337}} [ 639.743447] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [None req-e8db3b70-c93a-419a-89f8-b994313357e1 tempest-ServerShowV254Test-608938657 tempest-ServerShowV254Test-608938657-project-member] Folder already exists: OpenStack. Parent ref: group-v4. {{(pid=61806) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 639.743637] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [None req-e8db3b70-c93a-419a-89f8-b994313357e1 tempest-ServerShowV254Test-608938657 tempest-ServerShowV254Test-608938657-project-member] Creating folder: Project (b4cc4067092e4135b02bada3a5642857). Parent ref: group-v277609. {{(pid=61806) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 639.743863] env[61806]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-79bd3a58-38c3-4d6c-914c-863160cf6a3f {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 639.752386] env[61806]: INFO nova.virt.vmwareapi.vm_util [None req-e8db3b70-c93a-419a-89f8-b994313357e1 tempest-ServerShowV254Test-608938657 tempest-ServerShowV254Test-608938657-project-member] Created folder: Project (b4cc4067092e4135b02bada3a5642857) in parent group-v277609. [ 639.752587] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [None req-e8db3b70-c93a-419a-89f8-b994313357e1 tempest-ServerShowV254Test-608938657 tempest-ServerShowV254Test-608938657-project-member] Creating folder: Instances. Parent ref: group-v277614. {{(pid=61806) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 639.752802] env[61806]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-5a71fa0e-f334-4772-a85b-aa8b195da1e1 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 639.760299] env[61806]: INFO nova.virt.vmwareapi.vm_util [None req-e8db3b70-c93a-419a-89f8-b994313357e1 tempest-ServerShowV254Test-608938657 tempest-ServerShowV254Test-608938657-project-member] Created folder: Instances in parent group-v277614. [ 639.760523] env[61806]: DEBUG oslo.service.loopingcall [None req-e8db3b70-c93a-419a-89f8-b994313357e1 tempest-ServerShowV254Test-608938657 tempest-ServerShowV254Test-608938657-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61806) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 639.760699] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 8016bafc-b287-4df1-bd4b-a00210177a86] Creating VM on the ESX host {{(pid=61806) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 639.760878] env[61806]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-e6cc2afb-04d0-4f58-a4c2-2e3cb3360f13 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 639.777305] env[61806]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 639.777305] env[61806]: value = "task-1294452" [ 639.777305] env[61806]: _type = "Task" [ 639.777305] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 639.785719] env[61806]: DEBUG oslo_vmware.api [-] Task: {'id': task-1294452, 'name': CreateVM_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 639.919946] env[61806]: DEBUG oslo_concurrency.lockutils [None req-b4447f02-a3c7-4b6e-89ba-6bc34ec03205 tempest-InstanceActionsTestJSON-949444324 tempest-InstanceActionsTestJSON-949444324-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 640.087825] env[61806]: DEBUG oslo_concurrency.lockutils [None req-88cd9ecf-9bda-478c-b5e6-5a21edd9271f tempest-MigrationsAdminTest-470098049 tempest-MigrationsAdminTest-470098049-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.442s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 640.088443] env[61806]: DEBUG nova.compute.manager [None req-88cd9ecf-9bda-478c-b5e6-5a21edd9271f tempest-MigrationsAdminTest-470098049 tempest-MigrationsAdminTest-470098049-project-member] [instance: e66c921c-320e-4347-96c9-27aa80ba7f41] Start building networks asynchronously for instance. {{(pid=61806) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 640.091628] env[61806]: DEBUG oslo_concurrency.lockutils [None req-44eba412-ca3f-4386-8c36-f42e9734628d tempest-ServersTestBootFromVolume-1815739798 tempest-ServersTestBootFromVolume-1815739798-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 16.561s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 640.093887] env[61806]: INFO nova.compute.claims [None req-44eba412-ca3f-4386-8c36-f42e9734628d tempest-ServersTestBootFromVolume-1815739798 tempest-ServersTestBootFromVolume-1815739798-project-member] [instance: 56d26d08-3c6d-45e8-b238-6dbca6024561] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 640.288571] env[61806]: DEBUG oslo_vmware.api [-] Task: {'id': task-1294452, 'name': CreateVM_Task, 'duration_secs': 0.272305} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 640.288740] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 8016bafc-b287-4df1-bd4b-a00210177a86] Created VM on the ESX host {{(pid=61806) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 640.289730] env[61806]: DEBUG oslo_vmware.service [None req-e8db3b70-c93a-419a-89f8-b994313357e1 tempest-ServerShowV254Test-608938657 tempest-ServerShowV254Test-608938657-project-member] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d77ce900-cfd9-4f70-84a5-3c44f31b15e4 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 640.295367] env[61806]: DEBUG oslo_concurrency.lockutils [None req-e8db3b70-c93a-419a-89f8-b994313357e1 tempest-ServerShowV254Test-608938657 tempest-ServerShowV254Test-608938657-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 640.295531] env[61806]: DEBUG oslo_concurrency.lockutils [None req-e8db3b70-c93a-419a-89f8-b994313357e1 tempest-ServerShowV254Test-608938657 tempest-ServerShowV254Test-608938657-project-member] Acquired lock "[datastore1] devstack-image-cache_base/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 640.296168] env[61806]: DEBUG oslo_concurrency.lockutils [None req-e8db3b70-c93a-419a-89f8-b994313357e1 tempest-ServerShowV254Test-608938657 tempest-ServerShowV254Test-608938657-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 640.296408] env[61806]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-aed9d683-da0c-42df-809e-bdb4d9376ea9 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 640.300580] env[61806]: DEBUG oslo_vmware.api [None req-e8db3b70-c93a-419a-89f8-b994313357e1 tempest-ServerShowV254Test-608938657 tempest-ServerShowV254Test-608938657-project-member] Waiting for the task: (returnval){ [ 640.300580] env[61806]: value = "session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]52aa4771-855d-39b1-db61-d6d99c2e666a" [ 640.300580] env[61806]: _type = "Task" [ 640.300580] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 640.307831] env[61806]: DEBUG oslo_vmware.api [None req-e8db3b70-c93a-419a-89f8-b994313357e1 tempest-ServerShowV254Test-608938657 tempest-ServerShowV254Test-608938657-project-member] Task: {'id': session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]52aa4771-855d-39b1-db61-d6d99c2e666a, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 640.420433] env[61806]: DEBUG oslo_concurrency.lockutils [None req-f91d1db1-2f8a-42e8-9cfa-f4915f7b77dc tempest-ImagesTestJSON-618852556 tempest-ImagesTestJSON-618852556-project-member] Acquiring lock "e0ef0a35-82a5-495b-9d5c-5805e8306390" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 640.420740] env[61806]: DEBUG oslo_concurrency.lockutils [None req-f91d1db1-2f8a-42e8-9cfa-f4915f7b77dc tempest-ImagesTestJSON-618852556 tempest-ImagesTestJSON-618852556-project-member] Lock "e0ef0a35-82a5-495b-9d5c-5805e8306390" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 640.598008] env[61806]: DEBUG nova.compute.utils [None req-88cd9ecf-9bda-478c-b5e6-5a21edd9271f tempest-MigrationsAdminTest-470098049 tempest-MigrationsAdminTest-470098049-project-member] Using /dev/sd instead of None {{(pid=61806) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 640.599464] env[61806]: DEBUG nova.compute.manager [None req-88cd9ecf-9bda-478c-b5e6-5a21edd9271f tempest-MigrationsAdminTest-470098049 tempest-MigrationsAdminTest-470098049-project-member] [instance: e66c921c-320e-4347-96c9-27aa80ba7f41] Allocating IP information in the background. {{(pid=61806) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 640.599650] env[61806]: DEBUG nova.network.neutron [None req-88cd9ecf-9bda-478c-b5e6-5a21edd9271f tempest-MigrationsAdminTest-470098049 tempest-MigrationsAdminTest-470098049-project-member] [instance: e66c921c-320e-4347-96c9-27aa80ba7f41] allocate_for_instance() {{(pid=61806) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 640.663252] env[61806]: DEBUG nova.policy [None req-88cd9ecf-9bda-478c-b5e6-5a21edd9271f tempest-MigrationsAdminTest-470098049 tempest-MigrationsAdminTest-470098049-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'b248b20100be4ecbaeaa9e7ccd1bcfc3', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'ca6e40a2aeee479388f3a098945dba6e', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61806) authorize /opt/stack/nova/nova/policy.py:201}} [ 640.816195] env[61806]: DEBUG oslo_concurrency.lockutils [None req-e8db3b70-c93a-419a-89f8-b994313357e1 tempest-ServerShowV254Test-608938657 tempest-ServerShowV254Test-608938657-project-member] Releasing lock "[datastore1] devstack-image-cache_base/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 640.816876] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-e8db3b70-c93a-419a-89f8-b994313357e1 tempest-ServerShowV254Test-608938657 tempest-ServerShowV254Test-608938657-project-member] [instance: 8016bafc-b287-4df1-bd4b-a00210177a86] Processing image ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa {{(pid=61806) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 640.816876] env[61806]: DEBUG oslo_concurrency.lockutils [None req-e8db3b70-c93a-419a-89f8-b994313357e1 tempest-ServerShowV254Test-608938657 tempest-ServerShowV254Test-608938657-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa.vmdk" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 640.818447] env[61806]: DEBUG oslo_concurrency.lockutils [None req-e8db3b70-c93a-419a-89f8-b994313357e1 tempest-ServerShowV254Test-608938657 tempest-ServerShowV254Test-608938657-project-member] Acquired lock "[datastore1] devstack-image-cache_base/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa.vmdk" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 640.818447] env[61806]: DEBUG nova.virt.vmwareapi.ds_util [None req-e8db3b70-c93a-419a-89f8-b994313357e1 tempest-ServerShowV254Test-608938657 tempest-ServerShowV254Test-608938657-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61806) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 640.818447] env[61806]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-b4a2079e-c172-4145-94ce-f8d3c1cdfe95 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 640.834965] env[61806]: DEBUG nova.virt.vmwareapi.ds_util [None req-e8db3b70-c93a-419a-89f8-b994313357e1 tempest-ServerShowV254Test-608938657 tempest-ServerShowV254Test-608938657-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61806) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 640.835256] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-e8db3b70-c93a-419a-89f8-b994313357e1 tempest-ServerShowV254Test-608938657 tempest-ServerShowV254Test-608938657-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61806) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 640.836084] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a53a32fe-88f5-4aef-bd10-1b64cde28fd6 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 640.842568] env[61806]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-95bc6167-f4be-4cd7-ad95-d7825e4f4296 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 640.847411] env[61806]: DEBUG oslo_vmware.api [None req-e8db3b70-c93a-419a-89f8-b994313357e1 tempest-ServerShowV254Test-608938657 tempest-ServerShowV254Test-608938657-project-member] Waiting for the task: (returnval){ [ 640.847411] env[61806]: value = "session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]52bb90fe-b0ac-5f99-0884-02b55621dc17" [ 640.847411] env[61806]: _type = "Task" [ 640.847411] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 640.855065] env[61806]: DEBUG oslo_vmware.api [None req-e8db3b70-c93a-419a-89f8-b994313357e1 tempest-ServerShowV254Test-608938657 tempest-ServerShowV254Test-608938657-project-member] Task: {'id': session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]52bb90fe-b0ac-5f99-0884-02b55621dc17, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 640.962119] env[61806]: DEBUG nova.network.neutron [None req-88cd9ecf-9bda-478c-b5e6-5a21edd9271f tempest-MigrationsAdminTest-470098049 tempest-MigrationsAdminTest-470098049-project-member] [instance: e66c921c-320e-4347-96c9-27aa80ba7f41] Successfully created port: e5cb5794-df67-4846-a386-fcf7657ff23f {{(pid=61806) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 641.104225] env[61806]: DEBUG nova.compute.manager [None req-88cd9ecf-9bda-478c-b5e6-5a21edd9271f tempest-MigrationsAdminTest-470098049 tempest-MigrationsAdminTest-470098049-project-member] [instance: e66c921c-320e-4347-96c9-27aa80ba7f41] Start building block device mappings for instance. {{(pid=61806) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 641.358271] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-e8db3b70-c93a-419a-89f8-b994313357e1 tempest-ServerShowV254Test-608938657 tempest-ServerShowV254Test-608938657-project-member] [instance: 8016bafc-b287-4df1-bd4b-a00210177a86] Preparing fetch location {{(pid=61806) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:633}} [ 641.358535] env[61806]: DEBUG nova.virt.vmwareapi.ds_util [None req-e8db3b70-c93a-419a-89f8-b994313357e1 tempest-ServerShowV254Test-608938657 tempest-ServerShowV254Test-608938657-project-member] Creating directory with path [datastore1] vmware_temp/1ba31a81-3b83-488e-89ed-462bb868cd0d/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa {{(pid=61806) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 641.358806] env[61806]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-2e654d8a-c77d-48a9-b5d6-040d9befd827 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 641.383838] env[61806]: DEBUG nova.virt.vmwareapi.ds_util [None req-e8db3b70-c93a-419a-89f8-b994313357e1 tempest-ServerShowV254Test-608938657 tempest-ServerShowV254Test-608938657-project-member] Created directory with path [datastore1] vmware_temp/1ba31a81-3b83-488e-89ed-462bb868cd0d/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa {{(pid=61806) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 641.384049] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-e8db3b70-c93a-419a-89f8-b994313357e1 tempest-ServerShowV254Test-608938657 tempest-ServerShowV254Test-608938657-project-member] [instance: 8016bafc-b287-4df1-bd4b-a00210177a86] Fetch image to [datastore1] vmware_temp/1ba31a81-3b83-488e-89ed-462bb868cd0d/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa/tmp-sparse.vmdk {{(pid=61806) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:635}} [ 641.384238] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-e8db3b70-c93a-419a-89f8-b994313357e1 tempest-ServerShowV254Test-608938657 tempest-ServerShowV254Test-608938657-project-member] [instance: 8016bafc-b287-4df1-bd4b-a00210177a86] Downloading image file data ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa to [datastore1] vmware_temp/1ba31a81-3b83-488e-89ed-462bb868cd0d/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa/tmp-sparse.vmdk on the data store datastore1 {{(pid=61806) _fetch_image_as_file /opt/stack/nova/nova/virt/vmwareapi/vmops.py:399}} [ 641.384997] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-108ce022-b83d-42c2-a45f-3077b27748b1 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 641.394372] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-85c9cd23-0583-4567-84d5-f6e482eeab20 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 641.405907] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-854b5f6a-9ee4-415a-936d-18a82388b650 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 641.440398] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ec6ad2b3-e9a7-4c9a-9f5a-349ccc3a16b7 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 641.445698] env[61806]: DEBUG oslo_vmware.service [-] Invoking SessionManager.AcquireGenericServiceTicket with opID=oslo.vmware-2702738d-e8a9-4f34-94f3-830c8ae4ce74 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 641.477238] env[61806]: DEBUG nova.virt.vmwareapi.images [None req-e8db3b70-c93a-419a-89f8-b994313357e1 tempest-ServerShowV254Test-608938657 tempest-ServerShowV254Test-608938657-project-member] [instance: 8016bafc-b287-4df1-bd4b-a00210177a86] Downloading image file data ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa to the data store datastore1 {{(pid=61806) fetch_image /opt/stack/nova/nova/virt/vmwareapi/images.py:245}} [ 641.532243] env[61806]: DEBUG oslo_vmware.rw_handles [None req-e8db3b70-c93a-419a-89f8-b994313357e1 tempest-ServerShowV254Test-608938657 tempest-ServerShowV254Test-608938657-project-member] Creating HTTP connection to write to file with size = 21318656 and URL = https://esx7c1n2.openstack.eu-de-1.cloud.sap:443/folder/vmware_temp/1ba31a81-3b83-488e-89ed-462bb868cd0d/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa/tmp-sparse.vmdk?dcPath=ha-datacenter&dsName=datastore1. {{(pid=61806) _create_write_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:122}} [ 641.590333] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b086480e-8fee-4cda-8e0d-534ba8e4a686 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 641.599619] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b3052ce1-fcfc-494b-93e8-f8473282d4ec {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 641.639034] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e72765ea-66f8-42a3-9d59-abd406b24abf {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 641.647639] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b99af9a1-3920-4f2c-b310-073ac1b1a33c {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 641.661972] env[61806]: DEBUG nova.compute.provider_tree [None req-44eba412-ca3f-4386-8c36-f42e9734628d tempest-ServersTestBootFromVolume-1815739798 tempest-ServersTestBootFromVolume-1815739798-project-member] Inventory has not changed in ProviderTree for provider: a2858be1-fd22-4e08-979e-87ad25293407 {{(pid=61806) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 641.797602] env[61806]: DEBUG nova.compute.manager [req-2fb38071-5f74-47aa-a64f-fc8a9a02f1f8 req-d44d75af-6207-473e-bf11-1eb8be428a31 service nova] [instance: e66c921c-320e-4347-96c9-27aa80ba7f41] Received event network-changed-e5cb5794-df67-4846-a386-fcf7657ff23f {{(pid=61806) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 641.797831] env[61806]: DEBUG nova.compute.manager [req-2fb38071-5f74-47aa-a64f-fc8a9a02f1f8 req-d44d75af-6207-473e-bf11-1eb8be428a31 service nova] [instance: e66c921c-320e-4347-96c9-27aa80ba7f41] Refreshing instance network info cache due to event network-changed-e5cb5794-df67-4846-a386-fcf7657ff23f. {{(pid=61806) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 641.798143] env[61806]: DEBUG oslo_concurrency.lockutils [req-2fb38071-5f74-47aa-a64f-fc8a9a02f1f8 req-d44d75af-6207-473e-bf11-1eb8be428a31 service nova] Acquiring lock "refresh_cache-e66c921c-320e-4347-96c9-27aa80ba7f41" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 641.798314] env[61806]: DEBUG oslo_concurrency.lockutils [req-2fb38071-5f74-47aa-a64f-fc8a9a02f1f8 req-d44d75af-6207-473e-bf11-1eb8be428a31 service nova] Acquired lock "refresh_cache-e66c921c-320e-4347-96c9-27aa80ba7f41" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 641.798507] env[61806]: DEBUG nova.network.neutron [req-2fb38071-5f74-47aa-a64f-fc8a9a02f1f8 req-d44d75af-6207-473e-bf11-1eb8be428a31 service nova] [instance: e66c921c-320e-4347-96c9-27aa80ba7f41] Refreshing network info cache for port e5cb5794-df67-4846-a386-fcf7657ff23f {{(pid=61806) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 641.989191] env[61806]: ERROR nova.compute.manager [None req-88cd9ecf-9bda-478c-b5e6-5a21edd9271f tempest-MigrationsAdminTest-470098049 tempest-MigrationsAdminTest-470098049-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port e5cb5794-df67-4846-a386-fcf7657ff23f, please check neutron logs for more information. [ 641.989191] env[61806]: ERROR nova.compute.manager Traceback (most recent call last): [ 641.989191] env[61806]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 641.989191] env[61806]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 641.989191] env[61806]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 641.989191] env[61806]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 641.989191] env[61806]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 641.989191] env[61806]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 641.989191] env[61806]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 641.989191] env[61806]: ERROR nova.compute.manager self.force_reraise() [ 641.989191] env[61806]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 641.989191] env[61806]: ERROR nova.compute.manager raise self.value [ 641.989191] env[61806]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 641.989191] env[61806]: ERROR nova.compute.manager updated_port = self._update_port( [ 641.989191] env[61806]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 641.989191] env[61806]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 641.989626] env[61806]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 641.989626] env[61806]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 641.989626] env[61806]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port e5cb5794-df67-4846-a386-fcf7657ff23f, please check neutron logs for more information. [ 641.989626] env[61806]: ERROR nova.compute.manager [ 641.989626] env[61806]: Traceback (most recent call last): [ 641.989626] env[61806]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 641.989626] env[61806]: listener.cb(fileno) [ 641.989626] env[61806]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 641.989626] env[61806]: result = function(*args, **kwargs) [ 641.989626] env[61806]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 641.989626] env[61806]: return func(*args, **kwargs) [ 641.989626] env[61806]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 641.989626] env[61806]: raise e [ 641.989626] env[61806]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 641.989626] env[61806]: nwinfo = self.network_api.allocate_for_instance( [ 641.989626] env[61806]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 641.989626] env[61806]: created_port_ids = self._update_ports_for_instance( [ 641.989626] env[61806]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 641.989626] env[61806]: with excutils.save_and_reraise_exception(): [ 641.989626] env[61806]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 641.989626] env[61806]: self.force_reraise() [ 641.989626] env[61806]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 641.989626] env[61806]: raise self.value [ 641.989626] env[61806]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 641.989626] env[61806]: updated_port = self._update_port( [ 641.989626] env[61806]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 641.989626] env[61806]: _ensure_no_port_binding_failure(port) [ 641.989626] env[61806]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 641.989626] env[61806]: raise exception.PortBindingFailed(port_id=port['id']) [ 641.990486] env[61806]: nova.exception.PortBindingFailed: Binding failed for port e5cb5794-df67-4846-a386-fcf7657ff23f, please check neutron logs for more information. [ 641.990486] env[61806]: Removing descriptor: 14 [ 642.145815] env[61806]: DEBUG nova.compute.manager [None req-88cd9ecf-9bda-478c-b5e6-5a21edd9271f tempest-MigrationsAdminTest-470098049 tempest-MigrationsAdminTest-470098049-project-member] [instance: e66c921c-320e-4347-96c9-27aa80ba7f41] Start spawning the instance on the hypervisor. {{(pid=61806) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 642.167517] env[61806]: DEBUG nova.scheduler.client.report [None req-44eba412-ca3f-4386-8c36-f42e9734628d tempest-ServersTestBootFromVolume-1815739798 tempest-ServersTestBootFromVolume-1815739798-project-member] Inventory has not changed for provider a2858be1-fd22-4e08-979e-87ad25293407 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 139, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61806) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 642.173036] env[61806]: DEBUG nova.virt.hardware [None req-88cd9ecf-9bda-478c-b5e6-5a21edd9271f tempest-MigrationsAdminTest-470098049 tempest-MigrationsAdminTest-470098049-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-15T18:07:39Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='79c98b1c-559f-4be1-ada7-a89b19a9e09e',id=37,is_public=True,memory_mb=192,name='tempest-test_resize_flavor_-29819944',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-15T18:04:52Z,direct_url=,disk_format='vmdk',id=ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='84ba83fce2c349a6988173c5d6fc3b07',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-15T18:04:53Z,virtual_size=,visibility=), allow threads: False {{(pid=61806) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 642.173278] env[61806]: DEBUG nova.virt.hardware [None req-88cd9ecf-9bda-478c-b5e6-5a21edd9271f tempest-MigrationsAdminTest-470098049 tempest-MigrationsAdminTest-470098049-project-member] Flavor limits 0:0:0 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 642.173462] env[61806]: DEBUG nova.virt.hardware [None req-88cd9ecf-9bda-478c-b5e6-5a21edd9271f tempest-MigrationsAdminTest-470098049 tempest-MigrationsAdminTest-470098049-project-member] Image limits 0:0:0 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 642.173644] env[61806]: DEBUG nova.virt.hardware [None req-88cd9ecf-9bda-478c-b5e6-5a21edd9271f tempest-MigrationsAdminTest-470098049 tempest-MigrationsAdminTest-470098049-project-member] Flavor pref 0:0:0 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 642.173807] env[61806]: DEBUG nova.virt.hardware [None req-88cd9ecf-9bda-478c-b5e6-5a21edd9271f tempest-MigrationsAdminTest-470098049 tempest-MigrationsAdminTest-470098049-project-member] Image pref 0:0:0 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 642.174876] env[61806]: DEBUG nova.virt.hardware [None req-88cd9ecf-9bda-478c-b5e6-5a21edd9271f tempest-MigrationsAdminTest-470098049 tempest-MigrationsAdminTest-470098049-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 642.174876] env[61806]: DEBUG nova.virt.hardware [None req-88cd9ecf-9bda-478c-b5e6-5a21edd9271f tempest-MigrationsAdminTest-470098049 tempest-MigrationsAdminTest-470098049-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61806) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 642.174876] env[61806]: DEBUG nova.virt.hardware [None req-88cd9ecf-9bda-478c-b5e6-5a21edd9271f tempest-MigrationsAdminTest-470098049 tempest-MigrationsAdminTest-470098049-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61806) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 642.174876] env[61806]: DEBUG nova.virt.hardware [None req-88cd9ecf-9bda-478c-b5e6-5a21edd9271f tempest-MigrationsAdminTest-470098049 tempest-MigrationsAdminTest-470098049-project-member] Got 1 possible topologies {{(pid=61806) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 642.174876] env[61806]: DEBUG nova.virt.hardware [None req-88cd9ecf-9bda-478c-b5e6-5a21edd9271f tempest-MigrationsAdminTest-470098049 tempest-MigrationsAdminTest-470098049-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61806) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 642.175109] env[61806]: DEBUG nova.virt.hardware [None req-88cd9ecf-9bda-478c-b5e6-5a21edd9271f tempest-MigrationsAdminTest-470098049 tempest-MigrationsAdminTest-470098049-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61806) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 642.176112] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f1500425-bdb0-42d5-935f-b80be81ba9de {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 642.184940] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cfc5a273-1e7e-43fc-8fc1-e98f936aa6ef {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 642.205816] env[61806]: ERROR nova.compute.manager [None req-88cd9ecf-9bda-478c-b5e6-5a21edd9271f tempest-MigrationsAdminTest-470098049 tempest-MigrationsAdminTest-470098049-project-member] [instance: e66c921c-320e-4347-96c9-27aa80ba7f41] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port e5cb5794-df67-4846-a386-fcf7657ff23f, please check neutron logs for more information. [ 642.205816] env[61806]: ERROR nova.compute.manager [instance: e66c921c-320e-4347-96c9-27aa80ba7f41] Traceback (most recent call last): [ 642.205816] env[61806]: ERROR nova.compute.manager [instance: e66c921c-320e-4347-96c9-27aa80ba7f41] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 642.205816] env[61806]: ERROR nova.compute.manager [instance: e66c921c-320e-4347-96c9-27aa80ba7f41] yield resources [ 642.205816] env[61806]: ERROR nova.compute.manager [instance: e66c921c-320e-4347-96c9-27aa80ba7f41] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 642.205816] env[61806]: ERROR nova.compute.manager [instance: e66c921c-320e-4347-96c9-27aa80ba7f41] self.driver.spawn(context, instance, image_meta, [ 642.205816] env[61806]: ERROR nova.compute.manager [instance: e66c921c-320e-4347-96c9-27aa80ba7f41] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 642.205816] env[61806]: ERROR nova.compute.manager [instance: e66c921c-320e-4347-96c9-27aa80ba7f41] self._vmops.spawn(context, instance, image_meta, injected_files, [ 642.205816] env[61806]: ERROR nova.compute.manager [instance: e66c921c-320e-4347-96c9-27aa80ba7f41] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 642.205816] env[61806]: ERROR nova.compute.manager [instance: e66c921c-320e-4347-96c9-27aa80ba7f41] vm_ref = self.build_virtual_machine(instance, [ 642.205816] env[61806]: ERROR nova.compute.manager [instance: e66c921c-320e-4347-96c9-27aa80ba7f41] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 642.206228] env[61806]: ERROR nova.compute.manager [instance: e66c921c-320e-4347-96c9-27aa80ba7f41] vif_infos = vmwarevif.get_vif_info(self._session, [ 642.206228] env[61806]: ERROR nova.compute.manager [instance: e66c921c-320e-4347-96c9-27aa80ba7f41] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 642.206228] env[61806]: ERROR nova.compute.manager [instance: e66c921c-320e-4347-96c9-27aa80ba7f41] for vif in network_info: [ 642.206228] env[61806]: ERROR nova.compute.manager [instance: e66c921c-320e-4347-96c9-27aa80ba7f41] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 642.206228] env[61806]: ERROR nova.compute.manager [instance: e66c921c-320e-4347-96c9-27aa80ba7f41] return self._sync_wrapper(fn, *args, **kwargs) [ 642.206228] env[61806]: ERROR nova.compute.manager [instance: e66c921c-320e-4347-96c9-27aa80ba7f41] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 642.206228] env[61806]: ERROR nova.compute.manager [instance: e66c921c-320e-4347-96c9-27aa80ba7f41] self.wait() [ 642.206228] env[61806]: ERROR nova.compute.manager [instance: e66c921c-320e-4347-96c9-27aa80ba7f41] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 642.206228] env[61806]: ERROR nova.compute.manager [instance: e66c921c-320e-4347-96c9-27aa80ba7f41] self[:] = self._gt.wait() [ 642.206228] env[61806]: ERROR nova.compute.manager [instance: e66c921c-320e-4347-96c9-27aa80ba7f41] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 642.206228] env[61806]: ERROR nova.compute.manager [instance: e66c921c-320e-4347-96c9-27aa80ba7f41] return self._exit_event.wait() [ 642.206228] env[61806]: ERROR nova.compute.manager [instance: e66c921c-320e-4347-96c9-27aa80ba7f41] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 642.206228] env[61806]: ERROR nova.compute.manager [instance: e66c921c-320e-4347-96c9-27aa80ba7f41] current.throw(*self._exc) [ 642.206628] env[61806]: ERROR nova.compute.manager [instance: e66c921c-320e-4347-96c9-27aa80ba7f41] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 642.206628] env[61806]: ERROR nova.compute.manager [instance: e66c921c-320e-4347-96c9-27aa80ba7f41] result = function(*args, **kwargs) [ 642.206628] env[61806]: ERROR nova.compute.manager [instance: e66c921c-320e-4347-96c9-27aa80ba7f41] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 642.206628] env[61806]: ERROR nova.compute.manager [instance: e66c921c-320e-4347-96c9-27aa80ba7f41] return func(*args, **kwargs) [ 642.206628] env[61806]: ERROR nova.compute.manager [instance: e66c921c-320e-4347-96c9-27aa80ba7f41] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 642.206628] env[61806]: ERROR nova.compute.manager [instance: e66c921c-320e-4347-96c9-27aa80ba7f41] raise e [ 642.206628] env[61806]: ERROR nova.compute.manager [instance: e66c921c-320e-4347-96c9-27aa80ba7f41] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 642.206628] env[61806]: ERROR nova.compute.manager [instance: e66c921c-320e-4347-96c9-27aa80ba7f41] nwinfo = self.network_api.allocate_for_instance( [ 642.206628] env[61806]: ERROR nova.compute.manager [instance: e66c921c-320e-4347-96c9-27aa80ba7f41] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 642.206628] env[61806]: ERROR nova.compute.manager [instance: e66c921c-320e-4347-96c9-27aa80ba7f41] created_port_ids = self._update_ports_for_instance( [ 642.206628] env[61806]: ERROR nova.compute.manager [instance: e66c921c-320e-4347-96c9-27aa80ba7f41] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 642.206628] env[61806]: ERROR nova.compute.manager [instance: e66c921c-320e-4347-96c9-27aa80ba7f41] with excutils.save_and_reraise_exception(): [ 642.206628] env[61806]: ERROR nova.compute.manager [instance: e66c921c-320e-4347-96c9-27aa80ba7f41] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 642.206988] env[61806]: ERROR nova.compute.manager [instance: e66c921c-320e-4347-96c9-27aa80ba7f41] self.force_reraise() [ 642.206988] env[61806]: ERROR nova.compute.manager [instance: e66c921c-320e-4347-96c9-27aa80ba7f41] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 642.206988] env[61806]: ERROR nova.compute.manager [instance: e66c921c-320e-4347-96c9-27aa80ba7f41] raise self.value [ 642.206988] env[61806]: ERROR nova.compute.manager [instance: e66c921c-320e-4347-96c9-27aa80ba7f41] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 642.206988] env[61806]: ERROR nova.compute.manager [instance: e66c921c-320e-4347-96c9-27aa80ba7f41] updated_port = self._update_port( [ 642.206988] env[61806]: ERROR nova.compute.manager [instance: e66c921c-320e-4347-96c9-27aa80ba7f41] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 642.206988] env[61806]: ERROR nova.compute.manager [instance: e66c921c-320e-4347-96c9-27aa80ba7f41] _ensure_no_port_binding_failure(port) [ 642.206988] env[61806]: ERROR nova.compute.manager [instance: e66c921c-320e-4347-96c9-27aa80ba7f41] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 642.206988] env[61806]: ERROR nova.compute.manager [instance: e66c921c-320e-4347-96c9-27aa80ba7f41] raise exception.PortBindingFailed(port_id=port['id']) [ 642.206988] env[61806]: ERROR nova.compute.manager [instance: e66c921c-320e-4347-96c9-27aa80ba7f41] nova.exception.PortBindingFailed: Binding failed for port e5cb5794-df67-4846-a386-fcf7657ff23f, please check neutron logs for more information. [ 642.206988] env[61806]: ERROR nova.compute.manager [instance: e66c921c-320e-4347-96c9-27aa80ba7f41] [ 642.206988] env[61806]: INFO nova.compute.manager [None req-88cd9ecf-9bda-478c-b5e6-5a21edd9271f tempest-MigrationsAdminTest-470098049 tempest-MigrationsAdminTest-470098049-project-member] [instance: e66c921c-320e-4347-96c9-27aa80ba7f41] Terminating instance [ 642.207353] env[61806]: DEBUG oslo_concurrency.lockutils [None req-88cd9ecf-9bda-478c-b5e6-5a21edd9271f tempest-MigrationsAdminTest-470098049 tempest-MigrationsAdminTest-470098049-project-member] Acquiring lock "refresh_cache-e66c921c-320e-4347-96c9-27aa80ba7f41" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 642.208161] env[61806]: DEBUG oslo_vmware.rw_handles [None req-e8db3b70-c93a-419a-89f8-b994313357e1 tempest-ServerShowV254Test-608938657 tempest-ServerShowV254Test-608938657-project-member] Completed reading data from the image iterator. {{(pid=61806) read /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:765}} [ 642.208407] env[61806]: DEBUG oslo_vmware.rw_handles [None req-e8db3b70-c93a-419a-89f8-b994313357e1 tempest-ServerShowV254Test-608938657 tempest-ServerShowV254Test-608938657-project-member] Closing write handle for https://esx7c1n2.openstack.eu-de-1.cloud.sap:443/folder/vmware_temp/1ba31a81-3b83-488e-89ed-462bb868cd0d/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa/tmp-sparse.vmdk?dcPath=ha-datacenter&dsName=datastore1. {{(pid=61806) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:281}} [ 642.317531] env[61806]: DEBUG nova.network.neutron [req-2fb38071-5f74-47aa-a64f-fc8a9a02f1f8 req-d44d75af-6207-473e-bf11-1eb8be428a31 service nova] [instance: e66c921c-320e-4347-96c9-27aa80ba7f41] Instance cache missing network info. {{(pid=61806) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 642.346499] env[61806]: DEBUG nova.virt.vmwareapi.images [None req-e8db3b70-c93a-419a-89f8-b994313357e1 tempest-ServerShowV254Test-608938657 tempest-ServerShowV254Test-608938657-project-member] [instance: 8016bafc-b287-4df1-bd4b-a00210177a86] Downloaded image file data ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa to vmware_temp/1ba31a81-3b83-488e-89ed-462bb868cd0d/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa/tmp-sparse.vmdk on the data store datastore1 {{(pid=61806) fetch_image /opt/stack/nova/nova/virt/vmwareapi/images.py:258}} [ 642.348419] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-e8db3b70-c93a-419a-89f8-b994313357e1 tempest-ServerShowV254Test-608938657 tempest-ServerShowV254Test-608938657-project-member] [instance: 8016bafc-b287-4df1-bd4b-a00210177a86] Caching image {{(pid=61806) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:638}} [ 642.348675] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [None req-e8db3b70-c93a-419a-89f8-b994313357e1 tempest-ServerShowV254Test-608938657 tempest-ServerShowV254Test-608938657-project-member] Copying Virtual Disk [datastore1] vmware_temp/1ba31a81-3b83-488e-89ed-462bb868cd0d/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa/tmp-sparse.vmdk to [datastore1] vmware_temp/1ba31a81-3b83-488e-89ed-462bb868cd0d/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa.vmdk {{(pid=61806) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 642.348951] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-285f6c7b-ef6b-4313-868a-1d40de695466 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 642.356691] env[61806]: DEBUG oslo_vmware.api [None req-e8db3b70-c93a-419a-89f8-b994313357e1 tempest-ServerShowV254Test-608938657 tempest-ServerShowV254Test-608938657-project-member] Waiting for the task: (returnval){ [ 642.356691] env[61806]: value = "task-1294453" [ 642.356691] env[61806]: _type = "Task" [ 642.356691] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 642.365857] env[61806]: DEBUG oslo_vmware.api [None req-e8db3b70-c93a-419a-89f8-b994313357e1 tempest-ServerShowV254Test-608938657 tempest-ServerShowV254Test-608938657-project-member] Task: {'id': task-1294453, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 642.384347] env[61806]: DEBUG nova.network.neutron [req-2fb38071-5f74-47aa-a64f-fc8a9a02f1f8 req-d44d75af-6207-473e-bf11-1eb8be428a31 service nova] [instance: e66c921c-320e-4347-96c9-27aa80ba7f41] Updating instance_info_cache with network_info: [] {{(pid=61806) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 642.680608] env[61806]: DEBUG oslo_concurrency.lockutils [None req-44eba412-ca3f-4386-8c36-f42e9734628d tempest-ServersTestBootFromVolume-1815739798 tempest-ServersTestBootFromVolume-1815739798-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.589s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 642.681092] env[61806]: DEBUG nova.compute.manager [None req-44eba412-ca3f-4386-8c36-f42e9734628d tempest-ServersTestBootFromVolume-1815739798 tempest-ServersTestBootFromVolume-1815739798-project-member] [instance: 56d26d08-3c6d-45e8-b238-6dbca6024561] Start building networks asynchronously for instance. {{(pid=61806) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 642.684282] env[61806]: DEBUG oslo_concurrency.lockutils [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 16.240s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 642.684423] env[61806]: DEBUG oslo_concurrency.lockutils [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 642.684577] env[61806]: DEBUG nova.compute.resource_tracker [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=61806) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 642.684863] env[61806]: DEBUG oslo_concurrency.lockutils [None req-a6b8da3a-7829-486a-884d-0f28be02e38c tempest-ServersTestManualDisk-151630353 tempest-ServersTestManualDisk-151630353-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 15.729s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 642.688732] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2e5ce964-a35b-4521-9454-5f15c45c5bd1 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 642.696761] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6e7a8fef-491e-43a2-84a6-cd9f1c904546 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 642.711374] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-298694eb-1462-4429-9ead-2b2bc88bb173 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 642.717708] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0febf021-d72a-4c8a-8d25-6c9eef5f0c6b {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 642.747277] env[61806]: DEBUG nova.compute.resource_tracker [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=181521MB free_disk=139GB free_vcpus=48 pci_devices=None {{(pid=61806) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 642.747438] env[61806]: DEBUG oslo_concurrency.lockutils [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 642.867283] env[61806]: DEBUG oslo_vmware.api [None req-e8db3b70-c93a-419a-89f8-b994313357e1 tempest-ServerShowV254Test-608938657 tempest-ServerShowV254Test-608938657-project-member] Task: {'id': task-1294453, 'name': CopyVirtualDisk_Task} progress is 100%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 642.886498] env[61806]: DEBUG oslo_concurrency.lockutils [req-2fb38071-5f74-47aa-a64f-fc8a9a02f1f8 req-d44d75af-6207-473e-bf11-1eb8be428a31 service nova] Releasing lock "refresh_cache-e66c921c-320e-4347-96c9-27aa80ba7f41" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 642.886927] env[61806]: DEBUG oslo_concurrency.lockutils [None req-88cd9ecf-9bda-478c-b5e6-5a21edd9271f tempest-MigrationsAdminTest-470098049 tempest-MigrationsAdminTest-470098049-project-member] Acquired lock "refresh_cache-e66c921c-320e-4347-96c9-27aa80ba7f41" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 642.887132] env[61806]: DEBUG nova.network.neutron [None req-88cd9ecf-9bda-478c-b5e6-5a21edd9271f tempest-MigrationsAdminTest-470098049 tempest-MigrationsAdminTest-470098049-project-member] [instance: e66c921c-320e-4347-96c9-27aa80ba7f41] Building network info cache for instance {{(pid=61806) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 643.190303] env[61806]: DEBUG nova.compute.utils [None req-44eba412-ca3f-4386-8c36-f42e9734628d tempest-ServersTestBootFromVolume-1815739798 tempest-ServersTestBootFromVolume-1815739798-project-member] Using /dev/sd instead of None {{(pid=61806) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 643.192355] env[61806]: DEBUG nova.compute.manager [None req-44eba412-ca3f-4386-8c36-f42e9734628d tempest-ServersTestBootFromVolume-1815739798 tempest-ServersTestBootFromVolume-1815739798-project-member] [instance: 56d26d08-3c6d-45e8-b238-6dbca6024561] Allocating IP information in the background. {{(pid=61806) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 643.192554] env[61806]: DEBUG nova.network.neutron [None req-44eba412-ca3f-4386-8c36-f42e9734628d tempest-ServersTestBootFromVolume-1815739798 tempest-ServersTestBootFromVolume-1815739798-project-member] [instance: 56d26d08-3c6d-45e8-b238-6dbca6024561] allocate_for_instance() {{(pid=61806) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 643.233080] env[61806]: DEBUG nova.policy [None req-44eba412-ca3f-4386-8c36-f42e9734628d tempest-ServersTestBootFromVolume-1815739798 tempest-ServersTestBootFromVolume-1815739798-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'de484b686b994982b209128b2f798690', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'c479e0ae3b2845faa8000b9be6a35334', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61806) authorize /opt/stack/nova/nova/policy.py:201}} [ 643.373052] env[61806]: DEBUG oslo_vmware.api [None req-e8db3b70-c93a-419a-89f8-b994313357e1 tempest-ServerShowV254Test-608938657 tempest-ServerShowV254Test-608938657-project-member] Task: {'id': task-1294453, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.688302} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 643.373052] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [None req-e8db3b70-c93a-419a-89f8-b994313357e1 tempest-ServerShowV254Test-608938657 tempest-ServerShowV254Test-608938657-project-member] Copied Virtual Disk [datastore1] vmware_temp/1ba31a81-3b83-488e-89ed-462bb868cd0d/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa/tmp-sparse.vmdk to [datastore1] vmware_temp/1ba31a81-3b83-488e-89ed-462bb868cd0d/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa.vmdk {{(pid=61806) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 643.373052] env[61806]: DEBUG nova.virt.vmwareapi.ds_util [None req-e8db3b70-c93a-419a-89f8-b994313357e1 tempest-ServerShowV254Test-608938657 tempest-ServerShowV254Test-608938657-project-member] Deleting the datastore file [datastore1] vmware_temp/1ba31a81-3b83-488e-89ed-462bb868cd0d/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa/tmp-sparse.vmdk {{(pid=61806) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 643.373323] env[61806]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-6a774bc0-dd13-4e6e-85bc-58df28a5789e {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 643.382111] env[61806]: DEBUG oslo_vmware.api [None req-e8db3b70-c93a-419a-89f8-b994313357e1 tempest-ServerShowV254Test-608938657 tempest-ServerShowV254Test-608938657-project-member] Waiting for the task: (returnval){ [ 643.382111] env[61806]: value = "task-1294454" [ 643.382111] env[61806]: _type = "Task" [ 643.382111] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 643.392968] env[61806]: DEBUG oslo_vmware.api [None req-e8db3b70-c93a-419a-89f8-b994313357e1 tempest-ServerShowV254Test-608938657 tempest-ServerShowV254Test-608938657-project-member] Task: {'id': task-1294454, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 643.409747] env[61806]: DEBUG nova.network.neutron [None req-88cd9ecf-9bda-478c-b5e6-5a21edd9271f tempest-MigrationsAdminTest-470098049 tempest-MigrationsAdminTest-470098049-project-member] [instance: e66c921c-320e-4347-96c9-27aa80ba7f41] Instance cache missing network info. {{(pid=61806) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 643.473149] env[61806]: DEBUG nova.network.neutron [None req-88cd9ecf-9bda-478c-b5e6-5a21edd9271f tempest-MigrationsAdminTest-470098049 tempest-MigrationsAdminTest-470098049-project-member] [instance: e66c921c-320e-4347-96c9-27aa80ba7f41] Updating instance_info_cache with network_info: [] {{(pid=61806) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 643.578698] env[61806]: DEBUG nova.network.neutron [None req-44eba412-ca3f-4386-8c36-f42e9734628d tempest-ServersTestBootFromVolume-1815739798 tempest-ServersTestBootFromVolume-1815739798-project-member] [instance: 56d26d08-3c6d-45e8-b238-6dbca6024561] Successfully created port: 1570039f-d77a-49a9-a526-42d234745ef6 {{(pid=61806) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 643.584107] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dd95e4b9-7728-41d0-9192-f4c9bb38b2aa {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 643.591503] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9d0a1913-a78c-4400-a746-756c2bd7fc55 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 643.621846] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-32b2c862-7462-4d3e-bc8c-b01fdfe27e8b {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 643.628590] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a516906b-fb26-4eaa-8d5d-c387172e24da {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 643.642041] env[61806]: DEBUG nova.compute.provider_tree [None req-a6b8da3a-7829-486a-884d-0f28be02e38c tempest-ServersTestManualDisk-151630353 tempest-ServersTestManualDisk-151630353-project-member] Inventory has not changed in ProviderTree for provider: a2858be1-fd22-4e08-979e-87ad25293407 {{(pid=61806) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 643.697948] env[61806]: DEBUG nova.compute.manager [None req-44eba412-ca3f-4386-8c36-f42e9734628d tempest-ServersTestBootFromVolume-1815739798 tempest-ServersTestBootFromVolume-1815739798-project-member] [instance: 56d26d08-3c6d-45e8-b238-6dbca6024561] Start building block device mappings for instance. {{(pid=61806) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 643.825902] env[61806]: DEBUG nova.compute.manager [req-7f63316e-8ab5-4745-af70-85f22d420786 req-d25f4a9f-d3e7-4c47-a1fd-92cc402c3c57 service nova] [instance: e66c921c-320e-4347-96c9-27aa80ba7f41] Received event network-vif-deleted-e5cb5794-df67-4846-a386-fcf7657ff23f {{(pid=61806) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 643.893692] env[61806]: DEBUG oslo_vmware.api [None req-e8db3b70-c93a-419a-89f8-b994313357e1 tempest-ServerShowV254Test-608938657 tempest-ServerShowV254Test-608938657-project-member] Task: {'id': task-1294454, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.087093} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 643.893953] env[61806]: DEBUG nova.virt.vmwareapi.ds_util [None req-e8db3b70-c93a-419a-89f8-b994313357e1 tempest-ServerShowV254Test-608938657 tempest-ServerShowV254Test-608938657-project-member] Deleted the datastore file {{(pid=61806) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 643.894179] env[61806]: DEBUG nova.virt.vmwareapi.ds_util [None req-e8db3b70-c93a-419a-89f8-b994313357e1 tempest-ServerShowV254Test-608938657 tempest-ServerShowV254Test-608938657-project-member] Moving file from [datastore1] vmware_temp/1ba31a81-3b83-488e-89ed-462bb868cd0d/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa to [datastore1] devstack-image-cache_base/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa. {{(pid=61806) file_move /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:334}} [ 643.894434] env[61806]: DEBUG oslo_vmware.service [-] Invoking FileManager.MoveDatastoreFile_Task with opID=oslo.vmware-b1cf2ee7-43c6-4370-b368-1ed11d06ff39 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 643.903467] env[61806]: DEBUG oslo_vmware.api [None req-e8db3b70-c93a-419a-89f8-b994313357e1 tempest-ServerShowV254Test-608938657 tempest-ServerShowV254Test-608938657-project-member] Waiting for the task: (returnval){ [ 643.903467] env[61806]: value = "task-1294455" [ 643.903467] env[61806]: _type = "Task" [ 643.903467] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 643.913151] env[61806]: DEBUG oslo_vmware.api [None req-e8db3b70-c93a-419a-89f8-b994313357e1 tempest-ServerShowV254Test-608938657 tempest-ServerShowV254Test-608938657-project-member] Task: {'id': task-1294455, 'name': MoveDatastoreFile_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 643.975493] env[61806]: DEBUG oslo_concurrency.lockutils [None req-88cd9ecf-9bda-478c-b5e6-5a21edd9271f tempest-MigrationsAdminTest-470098049 tempest-MigrationsAdminTest-470098049-project-member] Releasing lock "refresh_cache-e66c921c-320e-4347-96c9-27aa80ba7f41" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 643.976153] env[61806]: DEBUG nova.compute.manager [None req-88cd9ecf-9bda-478c-b5e6-5a21edd9271f tempest-MigrationsAdminTest-470098049 tempest-MigrationsAdminTest-470098049-project-member] [instance: e66c921c-320e-4347-96c9-27aa80ba7f41] Start destroying the instance on the hypervisor. {{(pid=61806) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 643.976153] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-88cd9ecf-9bda-478c-b5e6-5a21edd9271f tempest-MigrationsAdminTest-470098049 tempest-MigrationsAdminTest-470098049-project-member] [instance: e66c921c-320e-4347-96c9-27aa80ba7f41] Destroying instance {{(pid=61806) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 643.977907] env[61806]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-0ec67756-c592-4bbf-9e7c-ff7f9469d148 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 643.985068] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8c778cc1-c965-49ca-99c2-1d7fb1f6855b {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 644.006066] env[61806]: WARNING nova.virt.vmwareapi.vmops [None req-88cd9ecf-9bda-478c-b5e6-5a21edd9271f tempest-MigrationsAdminTest-470098049 tempest-MigrationsAdminTest-470098049-project-member] [instance: e66c921c-320e-4347-96c9-27aa80ba7f41] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance e66c921c-320e-4347-96c9-27aa80ba7f41 could not be found. [ 644.006281] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-88cd9ecf-9bda-478c-b5e6-5a21edd9271f tempest-MigrationsAdminTest-470098049 tempest-MigrationsAdminTest-470098049-project-member] [instance: e66c921c-320e-4347-96c9-27aa80ba7f41] Instance destroyed {{(pid=61806) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 644.006500] env[61806]: INFO nova.compute.manager [None req-88cd9ecf-9bda-478c-b5e6-5a21edd9271f tempest-MigrationsAdminTest-470098049 tempest-MigrationsAdminTest-470098049-project-member] [instance: e66c921c-320e-4347-96c9-27aa80ba7f41] Took 0.03 seconds to destroy the instance on the hypervisor. [ 644.006754] env[61806]: DEBUG oslo.service.loopingcall [None req-88cd9ecf-9bda-478c-b5e6-5a21edd9271f tempest-MigrationsAdminTest-470098049 tempest-MigrationsAdminTest-470098049-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61806) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 644.006962] env[61806]: DEBUG nova.compute.manager [-] [instance: e66c921c-320e-4347-96c9-27aa80ba7f41] Deallocating network for instance {{(pid=61806) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 644.007068] env[61806]: DEBUG nova.network.neutron [-] [instance: e66c921c-320e-4347-96c9-27aa80ba7f41] deallocate_for_instance() {{(pid=61806) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 644.021878] env[61806]: DEBUG nova.network.neutron [-] [instance: e66c921c-320e-4347-96c9-27aa80ba7f41] Instance cache missing network info. {{(pid=61806) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 644.144997] env[61806]: DEBUG nova.scheduler.client.report [None req-a6b8da3a-7829-486a-884d-0f28be02e38c tempest-ServersTestManualDisk-151630353 tempest-ServersTestManualDisk-151630353-project-member] Inventory has not changed for provider a2858be1-fd22-4e08-979e-87ad25293407 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 139, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61806) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 644.203356] env[61806]: INFO nova.virt.block_device [None req-44eba412-ca3f-4386-8c36-f42e9734628d tempest-ServersTestBootFromVolume-1815739798 tempest-ServersTestBootFromVolume-1815739798-project-member] [instance: 56d26d08-3c6d-45e8-b238-6dbca6024561] Booting with volume 53e62409-4921-4079-adbb-d45cb5a680ff at /dev/sda [ 644.256746] env[61806]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-60b226bf-29e8-400c-9049-c168f78e2f93 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 644.265870] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8498c3e7-5d11-41f6-9c85-3cabe9f3aee1 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 644.288535] env[61806]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-73a7f285-68aa-4ee4-a537-526d9628e11a {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 644.297384] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2307361c-99a1-453f-9e6c-8a3a9ccf996f {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 644.320718] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-843790a5-de2a-4355-96df-e25ba03fef97 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 644.327799] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4332d034-4eff-4496-ba9c-d530803ed00b {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 644.341843] env[61806]: DEBUG nova.virt.block_device [None req-44eba412-ca3f-4386-8c36-f42e9734628d tempest-ServersTestBootFromVolume-1815739798 tempest-ServersTestBootFromVolume-1815739798-project-member] [instance: 56d26d08-3c6d-45e8-b238-6dbca6024561] Updating existing volume attachment record: 57b1a40c-a686-4b02-9b1e-d58f504a583e {{(pid=61806) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 644.413826] env[61806]: DEBUG oslo_vmware.api [None req-e8db3b70-c93a-419a-89f8-b994313357e1 tempest-ServerShowV254Test-608938657 tempest-ServerShowV254Test-608938657-project-member] Task: {'id': task-1294455, 'name': MoveDatastoreFile_Task, 'duration_secs': 0.026976} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 644.414098] env[61806]: DEBUG nova.virt.vmwareapi.ds_util [None req-e8db3b70-c93a-419a-89f8-b994313357e1 tempest-ServerShowV254Test-608938657 tempest-ServerShowV254Test-608938657-project-member] File moved {{(pid=61806) file_move /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:346}} [ 644.415324] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-e8db3b70-c93a-419a-89f8-b994313357e1 tempest-ServerShowV254Test-608938657 tempest-ServerShowV254Test-608938657-project-member] [instance: 8016bafc-b287-4df1-bd4b-a00210177a86] Cleaning up location [datastore1] vmware_temp/1ba31a81-3b83-488e-89ed-462bb868cd0d {{(pid=61806) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:640}} [ 644.415324] env[61806]: DEBUG nova.virt.vmwareapi.ds_util [None req-e8db3b70-c93a-419a-89f8-b994313357e1 tempest-ServerShowV254Test-608938657 tempest-ServerShowV254Test-608938657-project-member] Deleting the datastore file [datastore1] vmware_temp/1ba31a81-3b83-488e-89ed-462bb868cd0d {{(pid=61806) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 644.415324] env[61806]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-df489cc4-223f-41f8-b1c0-a87d6f6ae6d0 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 644.421091] env[61806]: DEBUG oslo_vmware.api [None req-e8db3b70-c93a-419a-89f8-b994313357e1 tempest-ServerShowV254Test-608938657 tempest-ServerShowV254Test-608938657-project-member] Waiting for the task: (returnval){ [ 644.421091] env[61806]: value = "task-1294456" [ 644.421091] env[61806]: _type = "Task" [ 644.421091] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 644.432330] env[61806]: DEBUG oslo_vmware.api [None req-e8db3b70-c93a-419a-89f8-b994313357e1 tempest-ServerShowV254Test-608938657 tempest-ServerShowV254Test-608938657-project-member] Task: {'id': task-1294456, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 644.524242] env[61806]: DEBUG nova.network.neutron [-] [instance: e66c921c-320e-4347-96c9-27aa80ba7f41] Updating instance_info_cache with network_info: [] {{(pid=61806) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 644.569637] env[61806]: ERROR nova.compute.manager [None req-44eba412-ca3f-4386-8c36-f42e9734628d tempest-ServersTestBootFromVolume-1815739798 tempest-ServersTestBootFromVolume-1815739798-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 1570039f-d77a-49a9-a526-42d234745ef6, please check neutron logs for more information. [ 644.569637] env[61806]: ERROR nova.compute.manager Traceback (most recent call last): [ 644.569637] env[61806]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 644.569637] env[61806]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 644.569637] env[61806]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 644.569637] env[61806]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 644.569637] env[61806]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 644.569637] env[61806]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 644.569637] env[61806]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 644.569637] env[61806]: ERROR nova.compute.manager self.force_reraise() [ 644.569637] env[61806]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 644.569637] env[61806]: ERROR nova.compute.manager raise self.value [ 644.569637] env[61806]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 644.569637] env[61806]: ERROR nova.compute.manager updated_port = self._update_port( [ 644.569637] env[61806]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 644.569637] env[61806]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 644.570392] env[61806]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 644.570392] env[61806]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 644.570392] env[61806]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 1570039f-d77a-49a9-a526-42d234745ef6, please check neutron logs for more information. [ 644.570392] env[61806]: ERROR nova.compute.manager [ 644.570392] env[61806]: Traceback (most recent call last): [ 644.570392] env[61806]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 644.570392] env[61806]: listener.cb(fileno) [ 644.570392] env[61806]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 644.570392] env[61806]: result = function(*args, **kwargs) [ 644.570392] env[61806]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 644.570392] env[61806]: return func(*args, **kwargs) [ 644.570392] env[61806]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 644.570392] env[61806]: raise e [ 644.570392] env[61806]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 644.570392] env[61806]: nwinfo = self.network_api.allocate_for_instance( [ 644.570392] env[61806]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 644.570392] env[61806]: created_port_ids = self._update_ports_for_instance( [ 644.570392] env[61806]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 644.570392] env[61806]: with excutils.save_and_reraise_exception(): [ 644.570392] env[61806]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 644.570392] env[61806]: self.force_reraise() [ 644.570392] env[61806]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 644.570392] env[61806]: raise self.value [ 644.570392] env[61806]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 644.570392] env[61806]: updated_port = self._update_port( [ 644.570392] env[61806]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 644.570392] env[61806]: _ensure_no_port_binding_failure(port) [ 644.570392] env[61806]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 644.570392] env[61806]: raise exception.PortBindingFailed(port_id=port['id']) [ 644.571580] env[61806]: nova.exception.PortBindingFailed: Binding failed for port 1570039f-d77a-49a9-a526-42d234745ef6, please check neutron logs for more information. [ 644.571580] env[61806]: Removing descriptor: 14 [ 644.652313] env[61806]: DEBUG oslo_concurrency.lockutils [None req-a6b8da3a-7829-486a-884d-0f28be02e38c tempest-ServersTestManualDisk-151630353 tempest-ServersTestManualDisk-151630353-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.967s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 644.652977] env[61806]: ERROR nova.compute.manager [None req-a6b8da3a-7829-486a-884d-0f28be02e38c tempest-ServersTestManualDisk-151630353 tempest-ServersTestManualDisk-151630353-project-member] [instance: 38a79453-939d-4649-9419-a2799636180d] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 6016a127-edf4-40d6-aba6-0a14614fe4ff, please check neutron logs for more information. [ 644.652977] env[61806]: ERROR nova.compute.manager [instance: 38a79453-939d-4649-9419-a2799636180d] Traceback (most recent call last): [ 644.652977] env[61806]: ERROR nova.compute.manager [instance: 38a79453-939d-4649-9419-a2799636180d] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 644.652977] env[61806]: ERROR nova.compute.manager [instance: 38a79453-939d-4649-9419-a2799636180d] self.driver.spawn(context, instance, image_meta, [ 644.652977] env[61806]: ERROR nova.compute.manager [instance: 38a79453-939d-4649-9419-a2799636180d] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 644.652977] env[61806]: ERROR nova.compute.manager [instance: 38a79453-939d-4649-9419-a2799636180d] self._vmops.spawn(context, instance, image_meta, injected_files, [ 644.652977] env[61806]: ERROR nova.compute.manager [instance: 38a79453-939d-4649-9419-a2799636180d] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 644.652977] env[61806]: ERROR nova.compute.manager [instance: 38a79453-939d-4649-9419-a2799636180d] vm_ref = self.build_virtual_machine(instance, [ 644.652977] env[61806]: ERROR nova.compute.manager [instance: 38a79453-939d-4649-9419-a2799636180d] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 644.652977] env[61806]: ERROR nova.compute.manager [instance: 38a79453-939d-4649-9419-a2799636180d] vif_infos = vmwarevif.get_vif_info(self._session, [ 644.652977] env[61806]: ERROR nova.compute.manager [instance: 38a79453-939d-4649-9419-a2799636180d] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 644.653498] env[61806]: ERROR nova.compute.manager [instance: 38a79453-939d-4649-9419-a2799636180d] for vif in network_info: [ 644.653498] env[61806]: ERROR nova.compute.manager [instance: 38a79453-939d-4649-9419-a2799636180d] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 644.653498] env[61806]: ERROR nova.compute.manager [instance: 38a79453-939d-4649-9419-a2799636180d] return self._sync_wrapper(fn, *args, **kwargs) [ 644.653498] env[61806]: ERROR nova.compute.manager [instance: 38a79453-939d-4649-9419-a2799636180d] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 644.653498] env[61806]: ERROR nova.compute.manager [instance: 38a79453-939d-4649-9419-a2799636180d] self.wait() [ 644.653498] env[61806]: ERROR nova.compute.manager [instance: 38a79453-939d-4649-9419-a2799636180d] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 644.653498] env[61806]: ERROR nova.compute.manager [instance: 38a79453-939d-4649-9419-a2799636180d] self[:] = self._gt.wait() [ 644.653498] env[61806]: ERROR nova.compute.manager [instance: 38a79453-939d-4649-9419-a2799636180d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 644.653498] env[61806]: ERROR nova.compute.manager [instance: 38a79453-939d-4649-9419-a2799636180d] return self._exit_event.wait() [ 644.653498] env[61806]: ERROR nova.compute.manager [instance: 38a79453-939d-4649-9419-a2799636180d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 644.653498] env[61806]: ERROR nova.compute.manager [instance: 38a79453-939d-4649-9419-a2799636180d] result = hub.switch() [ 644.653498] env[61806]: ERROR nova.compute.manager [instance: 38a79453-939d-4649-9419-a2799636180d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 644.653498] env[61806]: ERROR nova.compute.manager [instance: 38a79453-939d-4649-9419-a2799636180d] return self.greenlet.switch() [ 644.654046] env[61806]: ERROR nova.compute.manager [instance: 38a79453-939d-4649-9419-a2799636180d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 644.654046] env[61806]: ERROR nova.compute.manager [instance: 38a79453-939d-4649-9419-a2799636180d] result = function(*args, **kwargs) [ 644.654046] env[61806]: ERROR nova.compute.manager [instance: 38a79453-939d-4649-9419-a2799636180d] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 644.654046] env[61806]: ERROR nova.compute.manager [instance: 38a79453-939d-4649-9419-a2799636180d] return func(*args, **kwargs) [ 644.654046] env[61806]: ERROR nova.compute.manager [instance: 38a79453-939d-4649-9419-a2799636180d] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 644.654046] env[61806]: ERROR nova.compute.manager [instance: 38a79453-939d-4649-9419-a2799636180d] raise e [ 644.654046] env[61806]: ERROR nova.compute.manager [instance: 38a79453-939d-4649-9419-a2799636180d] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 644.654046] env[61806]: ERROR nova.compute.manager [instance: 38a79453-939d-4649-9419-a2799636180d] nwinfo = self.network_api.allocate_for_instance( [ 644.654046] env[61806]: ERROR nova.compute.manager [instance: 38a79453-939d-4649-9419-a2799636180d] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 644.654046] env[61806]: ERROR nova.compute.manager [instance: 38a79453-939d-4649-9419-a2799636180d] created_port_ids = self._update_ports_for_instance( [ 644.654046] env[61806]: ERROR nova.compute.manager [instance: 38a79453-939d-4649-9419-a2799636180d] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 644.654046] env[61806]: ERROR nova.compute.manager [instance: 38a79453-939d-4649-9419-a2799636180d] with excutils.save_and_reraise_exception(): [ 644.654046] env[61806]: ERROR nova.compute.manager [instance: 38a79453-939d-4649-9419-a2799636180d] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 644.654559] env[61806]: ERROR nova.compute.manager [instance: 38a79453-939d-4649-9419-a2799636180d] self.force_reraise() [ 644.654559] env[61806]: ERROR nova.compute.manager [instance: 38a79453-939d-4649-9419-a2799636180d] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 644.654559] env[61806]: ERROR nova.compute.manager [instance: 38a79453-939d-4649-9419-a2799636180d] raise self.value [ 644.654559] env[61806]: ERROR nova.compute.manager [instance: 38a79453-939d-4649-9419-a2799636180d] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 644.654559] env[61806]: ERROR nova.compute.manager [instance: 38a79453-939d-4649-9419-a2799636180d] updated_port = self._update_port( [ 644.654559] env[61806]: ERROR nova.compute.manager [instance: 38a79453-939d-4649-9419-a2799636180d] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 644.654559] env[61806]: ERROR nova.compute.manager [instance: 38a79453-939d-4649-9419-a2799636180d] _ensure_no_port_binding_failure(port) [ 644.654559] env[61806]: ERROR nova.compute.manager [instance: 38a79453-939d-4649-9419-a2799636180d] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 644.654559] env[61806]: ERROR nova.compute.manager [instance: 38a79453-939d-4649-9419-a2799636180d] raise exception.PortBindingFailed(port_id=port['id']) [ 644.654559] env[61806]: ERROR nova.compute.manager [instance: 38a79453-939d-4649-9419-a2799636180d] nova.exception.PortBindingFailed: Binding failed for port 6016a127-edf4-40d6-aba6-0a14614fe4ff, please check neutron logs for more information. [ 644.654559] env[61806]: ERROR nova.compute.manager [instance: 38a79453-939d-4649-9419-a2799636180d] [ 644.655011] env[61806]: DEBUG nova.compute.utils [None req-a6b8da3a-7829-486a-884d-0f28be02e38c tempest-ServersTestManualDisk-151630353 tempest-ServersTestManualDisk-151630353-project-member] [instance: 38a79453-939d-4649-9419-a2799636180d] Binding failed for port 6016a127-edf4-40d6-aba6-0a14614fe4ff, please check neutron logs for more information. {{(pid=61806) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 644.655063] env[61806]: DEBUG oslo_concurrency.lockutils [None req-98d962a0-2632-4a9a-8d12-70ce4acc382b tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 15.485s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 644.657950] env[61806]: DEBUG nova.compute.manager [None req-a6b8da3a-7829-486a-884d-0f28be02e38c tempest-ServersTestManualDisk-151630353 tempest-ServersTestManualDisk-151630353-project-member] [instance: 38a79453-939d-4649-9419-a2799636180d] Build of instance 38a79453-939d-4649-9419-a2799636180d was re-scheduled: Binding failed for port 6016a127-edf4-40d6-aba6-0a14614fe4ff, please check neutron logs for more information. {{(pid=61806) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 644.658442] env[61806]: DEBUG nova.compute.manager [None req-a6b8da3a-7829-486a-884d-0f28be02e38c tempest-ServersTestManualDisk-151630353 tempest-ServersTestManualDisk-151630353-project-member] [instance: 38a79453-939d-4649-9419-a2799636180d] Unplugging VIFs for instance {{(pid=61806) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 644.658667] env[61806]: DEBUG oslo_concurrency.lockutils [None req-a6b8da3a-7829-486a-884d-0f28be02e38c tempest-ServersTestManualDisk-151630353 tempest-ServersTestManualDisk-151630353-project-member] Acquiring lock "refresh_cache-38a79453-939d-4649-9419-a2799636180d" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 644.658813] env[61806]: DEBUG oslo_concurrency.lockutils [None req-a6b8da3a-7829-486a-884d-0f28be02e38c tempest-ServersTestManualDisk-151630353 tempest-ServersTestManualDisk-151630353-project-member] Acquired lock "refresh_cache-38a79453-939d-4649-9419-a2799636180d" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 644.658971] env[61806]: DEBUG nova.network.neutron [None req-a6b8da3a-7829-486a-884d-0f28be02e38c tempest-ServersTestManualDisk-151630353 tempest-ServersTestManualDisk-151630353-project-member] [instance: 38a79453-939d-4649-9419-a2799636180d] Building network info cache for instance {{(pid=61806) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 644.931722] env[61806]: DEBUG oslo_vmware.api [None req-e8db3b70-c93a-419a-89f8-b994313357e1 tempest-ServerShowV254Test-608938657 tempest-ServerShowV254Test-608938657-project-member] Task: {'id': task-1294456, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.028291} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 644.932020] env[61806]: DEBUG nova.virt.vmwareapi.ds_util [None req-e8db3b70-c93a-419a-89f8-b994313357e1 tempest-ServerShowV254Test-608938657 tempest-ServerShowV254Test-608938657-project-member] Deleted the datastore file {{(pid=61806) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 644.933543] env[61806]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-85ab7d8e-b9db-4464-93b0-10e8fc223ba8 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 644.939406] env[61806]: DEBUG oslo_vmware.api [None req-e8db3b70-c93a-419a-89f8-b994313357e1 tempest-ServerShowV254Test-608938657 tempest-ServerShowV254Test-608938657-project-member] Waiting for the task: (returnval){ [ 644.939406] env[61806]: value = "session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]5223adf4-578d-cf68-4e2e-65930722babe" [ 644.939406] env[61806]: _type = "Task" [ 644.939406] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 644.947989] env[61806]: DEBUG oslo_vmware.api [None req-e8db3b70-c93a-419a-89f8-b994313357e1 tempest-ServerShowV254Test-608938657 tempest-ServerShowV254Test-608938657-project-member] Task: {'id': session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]5223adf4-578d-cf68-4e2e-65930722babe, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 645.026460] env[61806]: INFO nova.compute.manager [-] [instance: e66c921c-320e-4347-96c9-27aa80ba7f41] Took 1.02 seconds to deallocate network for instance. [ 645.028997] env[61806]: DEBUG nova.compute.claims [None req-88cd9ecf-9bda-478c-b5e6-5a21edd9271f tempest-MigrationsAdminTest-470098049 tempest-MigrationsAdminTest-470098049-project-member] [instance: e66c921c-320e-4347-96c9-27aa80ba7f41] Aborting claim: {{(pid=61806) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 645.029192] env[61806]: DEBUG oslo_concurrency.lockutils [None req-88cd9ecf-9bda-478c-b5e6-5a21edd9271f tempest-MigrationsAdminTest-470098049 tempest-MigrationsAdminTest-470098049-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 645.182325] env[61806]: DEBUG nova.network.neutron [None req-a6b8da3a-7829-486a-884d-0f28be02e38c tempest-ServersTestManualDisk-151630353 tempest-ServersTestManualDisk-151630353-project-member] [instance: 38a79453-939d-4649-9419-a2799636180d] Instance cache missing network info. {{(pid=61806) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 645.275738] env[61806]: DEBUG nova.network.neutron [None req-a6b8da3a-7829-486a-884d-0f28be02e38c tempest-ServersTestManualDisk-151630353 tempest-ServersTestManualDisk-151630353-project-member] [instance: 38a79453-939d-4649-9419-a2799636180d] Updating instance_info_cache with network_info: [] {{(pid=61806) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 645.453803] env[61806]: DEBUG oslo_vmware.api [None req-e8db3b70-c93a-419a-89f8-b994313357e1 tempest-ServerShowV254Test-608938657 tempest-ServerShowV254Test-608938657-project-member] Task: {'id': session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]5223adf4-578d-cf68-4e2e-65930722babe, 'name': SearchDatastore_Task, 'duration_secs': 0.012914} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 645.454222] env[61806]: DEBUG oslo_concurrency.lockutils [None req-e8db3b70-c93a-419a-89f8-b994313357e1 tempest-ServerShowV254Test-608938657 tempest-ServerShowV254Test-608938657-project-member] Releasing lock "[datastore1] devstack-image-cache_base/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa.vmdk" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 645.454602] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [None req-e8db3b70-c93a-419a-89f8-b994313357e1 tempest-ServerShowV254Test-608938657 tempest-ServerShowV254Test-608938657-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa.vmdk to [datastore1] 8016bafc-b287-4df1-bd4b-a00210177a86/8016bafc-b287-4df1-bd4b-a00210177a86.vmdk {{(pid=61806) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 645.455029] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-55f075a7-320c-4229-82b4-eb447c5f29cb {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 645.466117] env[61806]: DEBUG oslo_vmware.api [None req-e8db3b70-c93a-419a-89f8-b994313357e1 tempest-ServerShowV254Test-608938657 tempest-ServerShowV254Test-608938657-project-member] Waiting for the task: (returnval){ [ 645.466117] env[61806]: value = "task-1294457" [ 645.466117] env[61806]: _type = "Task" [ 645.466117] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 645.476032] env[61806]: DEBUG oslo_vmware.api [None req-e8db3b70-c93a-419a-89f8-b994313357e1 tempest-ServerShowV254Test-608938657 tempest-ServerShowV254Test-608938657-project-member] Task: {'id': task-1294457, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 645.659162] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-947c7e85-aaf7-4c37-a668-95be6d1cd503 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 645.667341] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-10162fc9-8acb-402c-b0e7-87dbd55ff5d3 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 645.699926] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b5f067b9-2909-4a4b-a7b4-2b6a31263fc2 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 645.709055] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d05ac228-b0ce-4378-883c-64f403fa0e7f {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 645.723737] env[61806]: DEBUG nova.compute.provider_tree [None req-98d962a0-2632-4a9a-8d12-70ce4acc382b tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] Inventory has not changed in ProviderTree for provider: a2858be1-fd22-4e08-979e-87ad25293407 {{(pid=61806) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 645.779868] env[61806]: DEBUG oslo_concurrency.lockutils [None req-a6b8da3a-7829-486a-884d-0f28be02e38c tempest-ServersTestManualDisk-151630353 tempest-ServersTestManualDisk-151630353-project-member] Releasing lock "refresh_cache-38a79453-939d-4649-9419-a2799636180d" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 645.780261] env[61806]: DEBUG nova.compute.manager [None req-a6b8da3a-7829-486a-884d-0f28be02e38c tempest-ServersTestManualDisk-151630353 tempest-ServersTestManualDisk-151630353-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61806) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 645.780693] env[61806]: DEBUG nova.compute.manager [None req-a6b8da3a-7829-486a-884d-0f28be02e38c tempest-ServersTestManualDisk-151630353 tempest-ServersTestManualDisk-151630353-project-member] [instance: 38a79453-939d-4649-9419-a2799636180d] Deallocating network for instance {{(pid=61806) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 645.781033] env[61806]: DEBUG nova.network.neutron [None req-a6b8da3a-7829-486a-884d-0f28be02e38c tempest-ServersTestManualDisk-151630353 tempest-ServersTestManualDisk-151630353-project-member] [instance: 38a79453-939d-4649-9419-a2799636180d] deallocate_for_instance() {{(pid=61806) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 645.799485] env[61806]: DEBUG nova.network.neutron [None req-a6b8da3a-7829-486a-884d-0f28be02e38c tempest-ServersTestManualDisk-151630353 tempest-ServersTestManualDisk-151630353-project-member] [instance: 38a79453-939d-4649-9419-a2799636180d] Instance cache missing network info. {{(pid=61806) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 645.863236] env[61806]: DEBUG nova.compute.manager [req-0d4f4869-d8a5-4a98-9582-fe758d1be0a6 req-fc8a0659-8afd-401e-8a13-8b894e8dc341 service nova] [instance: 56d26d08-3c6d-45e8-b238-6dbca6024561] Received event network-changed-1570039f-d77a-49a9-a526-42d234745ef6 {{(pid=61806) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 645.863515] env[61806]: DEBUG nova.compute.manager [req-0d4f4869-d8a5-4a98-9582-fe758d1be0a6 req-fc8a0659-8afd-401e-8a13-8b894e8dc341 service nova] [instance: 56d26d08-3c6d-45e8-b238-6dbca6024561] Refreshing instance network info cache due to event network-changed-1570039f-d77a-49a9-a526-42d234745ef6. {{(pid=61806) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 645.863809] env[61806]: DEBUG oslo_concurrency.lockutils [req-0d4f4869-d8a5-4a98-9582-fe758d1be0a6 req-fc8a0659-8afd-401e-8a13-8b894e8dc341 service nova] Acquiring lock "refresh_cache-56d26d08-3c6d-45e8-b238-6dbca6024561" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 645.863997] env[61806]: DEBUG oslo_concurrency.lockutils [req-0d4f4869-d8a5-4a98-9582-fe758d1be0a6 req-fc8a0659-8afd-401e-8a13-8b894e8dc341 service nova] Acquired lock "refresh_cache-56d26d08-3c6d-45e8-b238-6dbca6024561" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 645.864214] env[61806]: DEBUG nova.network.neutron [req-0d4f4869-d8a5-4a98-9582-fe758d1be0a6 req-fc8a0659-8afd-401e-8a13-8b894e8dc341 service nova] [instance: 56d26d08-3c6d-45e8-b238-6dbca6024561] Refreshing network info cache for port 1570039f-d77a-49a9-a526-42d234745ef6 {{(pid=61806) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 645.976573] env[61806]: DEBUG oslo_vmware.api [None req-e8db3b70-c93a-419a-89f8-b994313357e1 tempest-ServerShowV254Test-608938657 tempest-ServerShowV254Test-608938657-project-member] Task: {'id': task-1294457, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 646.227935] env[61806]: DEBUG nova.scheduler.client.report [None req-98d962a0-2632-4a9a-8d12-70ce4acc382b tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] Inventory has not changed for provider a2858be1-fd22-4e08-979e-87ad25293407 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 139, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61806) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 646.303372] env[61806]: DEBUG nova.network.neutron [None req-a6b8da3a-7829-486a-884d-0f28be02e38c tempest-ServersTestManualDisk-151630353 tempest-ServersTestManualDisk-151630353-project-member] [instance: 38a79453-939d-4649-9419-a2799636180d] Updating instance_info_cache with network_info: [] {{(pid=61806) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 646.382105] env[61806]: DEBUG nova.network.neutron [req-0d4f4869-d8a5-4a98-9582-fe758d1be0a6 req-fc8a0659-8afd-401e-8a13-8b894e8dc341 service nova] [instance: 56d26d08-3c6d-45e8-b238-6dbca6024561] Instance cache missing network info. {{(pid=61806) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 646.444447] env[61806]: DEBUG nova.compute.manager [None req-44eba412-ca3f-4386-8c36-f42e9734628d tempest-ServersTestBootFromVolume-1815739798 tempest-ServersTestBootFromVolume-1815739798-project-member] [instance: 56d26d08-3c6d-45e8-b238-6dbca6024561] Start spawning the instance on the hypervisor. {{(pid=61806) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 646.444944] env[61806]: DEBUG nova.virt.hardware [None req-44eba412-ca3f-4386-8c36-f42e9734628d tempest-ServersTestBootFromVolume-1815739798 tempest-ServersTestBootFromVolume-1815739798-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-15T18:05:10Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format=,created_at=,direct_url=,disk_format=,id=,min_disk=0,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=1073741824,status='active',tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=61806) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 646.445173] env[61806]: DEBUG nova.virt.hardware [None req-44eba412-ca3f-4386-8c36-f42e9734628d tempest-ServersTestBootFromVolume-1815739798 tempest-ServersTestBootFromVolume-1815739798-project-member] Flavor limits 0:0:0 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 646.445385] env[61806]: DEBUG nova.virt.hardware [None req-44eba412-ca3f-4386-8c36-f42e9734628d tempest-ServersTestBootFromVolume-1815739798 tempest-ServersTestBootFromVolume-1815739798-project-member] Image limits 0:0:0 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 646.445528] env[61806]: DEBUG nova.virt.hardware [None req-44eba412-ca3f-4386-8c36-f42e9734628d tempest-ServersTestBootFromVolume-1815739798 tempest-ServersTestBootFromVolume-1815739798-project-member] Flavor pref 0:0:0 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 646.445676] env[61806]: DEBUG nova.virt.hardware [None req-44eba412-ca3f-4386-8c36-f42e9734628d tempest-ServersTestBootFromVolume-1815739798 tempest-ServersTestBootFromVolume-1815739798-project-member] Image pref 0:0:0 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 646.445824] env[61806]: DEBUG nova.virt.hardware [None req-44eba412-ca3f-4386-8c36-f42e9734628d tempest-ServersTestBootFromVolume-1815739798 tempest-ServersTestBootFromVolume-1815739798-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 646.446032] env[61806]: DEBUG nova.virt.hardware [None req-44eba412-ca3f-4386-8c36-f42e9734628d tempest-ServersTestBootFromVolume-1815739798 tempest-ServersTestBootFromVolume-1815739798-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61806) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 646.446196] env[61806]: DEBUG nova.virt.hardware [None req-44eba412-ca3f-4386-8c36-f42e9734628d tempest-ServersTestBootFromVolume-1815739798 tempest-ServersTestBootFromVolume-1815739798-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61806) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 646.446361] env[61806]: DEBUG nova.virt.hardware [None req-44eba412-ca3f-4386-8c36-f42e9734628d tempest-ServersTestBootFromVolume-1815739798 tempest-ServersTestBootFromVolume-1815739798-project-member] Got 1 possible topologies {{(pid=61806) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 646.446521] env[61806]: DEBUG nova.virt.hardware [None req-44eba412-ca3f-4386-8c36-f42e9734628d tempest-ServersTestBootFromVolume-1815739798 tempest-ServersTestBootFromVolume-1815739798-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61806) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 646.446688] env[61806]: DEBUG nova.virt.hardware [None req-44eba412-ca3f-4386-8c36-f42e9734628d tempest-ServersTestBootFromVolume-1815739798 tempest-ServersTestBootFromVolume-1815739798-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61806) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 646.447846] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a1033b77-b11a-4380-8c4c-5dfa0ef6c5b9 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 646.457050] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2aa0d475-e1c3-4eb9-9e7d-b532d6dbd698 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 646.461103] env[61806]: DEBUG nova.network.neutron [req-0d4f4869-d8a5-4a98-9582-fe758d1be0a6 req-fc8a0659-8afd-401e-8a13-8b894e8dc341 service nova] [instance: 56d26d08-3c6d-45e8-b238-6dbca6024561] Updating instance_info_cache with network_info: [] {{(pid=61806) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 646.472228] env[61806]: ERROR nova.compute.manager [None req-44eba412-ca3f-4386-8c36-f42e9734628d tempest-ServersTestBootFromVolume-1815739798 tempest-ServersTestBootFromVolume-1815739798-project-member] [instance: 56d26d08-3c6d-45e8-b238-6dbca6024561] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 1570039f-d77a-49a9-a526-42d234745ef6, please check neutron logs for more information. [ 646.472228] env[61806]: ERROR nova.compute.manager [instance: 56d26d08-3c6d-45e8-b238-6dbca6024561] Traceback (most recent call last): [ 646.472228] env[61806]: ERROR nova.compute.manager [instance: 56d26d08-3c6d-45e8-b238-6dbca6024561] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 646.472228] env[61806]: ERROR nova.compute.manager [instance: 56d26d08-3c6d-45e8-b238-6dbca6024561] yield resources [ 646.472228] env[61806]: ERROR nova.compute.manager [instance: 56d26d08-3c6d-45e8-b238-6dbca6024561] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 646.472228] env[61806]: ERROR nova.compute.manager [instance: 56d26d08-3c6d-45e8-b238-6dbca6024561] self.driver.spawn(context, instance, image_meta, [ 646.472228] env[61806]: ERROR nova.compute.manager [instance: 56d26d08-3c6d-45e8-b238-6dbca6024561] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 646.472228] env[61806]: ERROR nova.compute.manager [instance: 56d26d08-3c6d-45e8-b238-6dbca6024561] self._vmops.spawn(context, instance, image_meta, injected_files, [ 646.472228] env[61806]: ERROR nova.compute.manager [instance: 56d26d08-3c6d-45e8-b238-6dbca6024561] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 646.472228] env[61806]: ERROR nova.compute.manager [instance: 56d26d08-3c6d-45e8-b238-6dbca6024561] vm_ref = self.build_virtual_machine(instance, [ 646.472228] env[61806]: ERROR nova.compute.manager [instance: 56d26d08-3c6d-45e8-b238-6dbca6024561] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 646.472520] env[61806]: ERROR nova.compute.manager [instance: 56d26d08-3c6d-45e8-b238-6dbca6024561] vif_infos = vmwarevif.get_vif_info(self._session, [ 646.472520] env[61806]: ERROR nova.compute.manager [instance: 56d26d08-3c6d-45e8-b238-6dbca6024561] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 646.472520] env[61806]: ERROR nova.compute.manager [instance: 56d26d08-3c6d-45e8-b238-6dbca6024561] for vif in network_info: [ 646.472520] env[61806]: ERROR nova.compute.manager [instance: 56d26d08-3c6d-45e8-b238-6dbca6024561] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 646.472520] env[61806]: ERROR nova.compute.manager [instance: 56d26d08-3c6d-45e8-b238-6dbca6024561] return self._sync_wrapper(fn, *args, **kwargs) [ 646.472520] env[61806]: ERROR nova.compute.manager [instance: 56d26d08-3c6d-45e8-b238-6dbca6024561] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 646.472520] env[61806]: ERROR nova.compute.manager [instance: 56d26d08-3c6d-45e8-b238-6dbca6024561] self.wait() [ 646.472520] env[61806]: ERROR nova.compute.manager [instance: 56d26d08-3c6d-45e8-b238-6dbca6024561] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 646.472520] env[61806]: ERROR nova.compute.manager [instance: 56d26d08-3c6d-45e8-b238-6dbca6024561] self[:] = self._gt.wait() [ 646.472520] env[61806]: ERROR nova.compute.manager [instance: 56d26d08-3c6d-45e8-b238-6dbca6024561] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 646.472520] env[61806]: ERROR nova.compute.manager [instance: 56d26d08-3c6d-45e8-b238-6dbca6024561] return self._exit_event.wait() [ 646.472520] env[61806]: ERROR nova.compute.manager [instance: 56d26d08-3c6d-45e8-b238-6dbca6024561] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 646.472520] env[61806]: ERROR nova.compute.manager [instance: 56d26d08-3c6d-45e8-b238-6dbca6024561] current.throw(*self._exc) [ 646.472826] env[61806]: ERROR nova.compute.manager [instance: 56d26d08-3c6d-45e8-b238-6dbca6024561] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 646.472826] env[61806]: ERROR nova.compute.manager [instance: 56d26d08-3c6d-45e8-b238-6dbca6024561] result = function(*args, **kwargs) [ 646.472826] env[61806]: ERROR nova.compute.manager [instance: 56d26d08-3c6d-45e8-b238-6dbca6024561] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 646.472826] env[61806]: ERROR nova.compute.manager [instance: 56d26d08-3c6d-45e8-b238-6dbca6024561] return func(*args, **kwargs) [ 646.472826] env[61806]: ERROR nova.compute.manager [instance: 56d26d08-3c6d-45e8-b238-6dbca6024561] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 646.472826] env[61806]: ERROR nova.compute.manager [instance: 56d26d08-3c6d-45e8-b238-6dbca6024561] raise e [ 646.472826] env[61806]: ERROR nova.compute.manager [instance: 56d26d08-3c6d-45e8-b238-6dbca6024561] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 646.472826] env[61806]: ERROR nova.compute.manager [instance: 56d26d08-3c6d-45e8-b238-6dbca6024561] nwinfo = self.network_api.allocate_for_instance( [ 646.472826] env[61806]: ERROR nova.compute.manager [instance: 56d26d08-3c6d-45e8-b238-6dbca6024561] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 646.472826] env[61806]: ERROR nova.compute.manager [instance: 56d26d08-3c6d-45e8-b238-6dbca6024561] created_port_ids = self._update_ports_for_instance( [ 646.472826] env[61806]: ERROR nova.compute.manager [instance: 56d26d08-3c6d-45e8-b238-6dbca6024561] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 646.472826] env[61806]: ERROR nova.compute.manager [instance: 56d26d08-3c6d-45e8-b238-6dbca6024561] with excutils.save_and_reraise_exception(): [ 646.472826] env[61806]: ERROR nova.compute.manager [instance: 56d26d08-3c6d-45e8-b238-6dbca6024561] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 646.473142] env[61806]: ERROR nova.compute.manager [instance: 56d26d08-3c6d-45e8-b238-6dbca6024561] self.force_reraise() [ 646.473142] env[61806]: ERROR nova.compute.manager [instance: 56d26d08-3c6d-45e8-b238-6dbca6024561] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 646.473142] env[61806]: ERROR nova.compute.manager [instance: 56d26d08-3c6d-45e8-b238-6dbca6024561] raise self.value [ 646.473142] env[61806]: ERROR nova.compute.manager [instance: 56d26d08-3c6d-45e8-b238-6dbca6024561] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 646.473142] env[61806]: ERROR nova.compute.manager [instance: 56d26d08-3c6d-45e8-b238-6dbca6024561] updated_port = self._update_port( [ 646.473142] env[61806]: ERROR nova.compute.manager [instance: 56d26d08-3c6d-45e8-b238-6dbca6024561] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 646.473142] env[61806]: ERROR nova.compute.manager [instance: 56d26d08-3c6d-45e8-b238-6dbca6024561] _ensure_no_port_binding_failure(port) [ 646.473142] env[61806]: ERROR nova.compute.manager [instance: 56d26d08-3c6d-45e8-b238-6dbca6024561] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 646.473142] env[61806]: ERROR nova.compute.manager [instance: 56d26d08-3c6d-45e8-b238-6dbca6024561] raise exception.PortBindingFailed(port_id=port['id']) [ 646.473142] env[61806]: ERROR nova.compute.manager [instance: 56d26d08-3c6d-45e8-b238-6dbca6024561] nova.exception.PortBindingFailed: Binding failed for port 1570039f-d77a-49a9-a526-42d234745ef6, please check neutron logs for more information. [ 646.473142] env[61806]: ERROR nova.compute.manager [instance: 56d26d08-3c6d-45e8-b238-6dbca6024561] [ 646.473142] env[61806]: INFO nova.compute.manager [None req-44eba412-ca3f-4386-8c36-f42e9734628d tempest-ServersTestBootFromVolume-1815739798 tempest-ServersTestBootFromVolume-1815739798-project-member] [instance: 56d26d08-3c6d-45e8-b238-6dbca6024561] Terminating instance [ 646.478160] env[61806]: DEBUG oslo_concurrency.lockutils [None req-44eba412-ca3f-4386-8c36-f42e9734628d tempest-ServersTestBootFromVolume-1815739798 tempest-ServersTestBootFromVolume-1815739798-project-member] Acquiring lock "refresh_cache-56d26d08-3c6d-45e8-b238-6dbca6024561" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 646.484386] env[61806]: DEBUG oslo_vmware.api [None req-e8db3b70-c93a-419a-89f8-b994313357e1 tempest-ServerShowV254Test-608938657 tempest-ServerShowV254Test-608938657-project-member] Task: {'id': task-1294457, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.556225} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 646.484635] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [None req-e8db3b70-c93a-419a-89f8-b994313357e1 tempest-ServerShowV254Test-608938657 tempest-ServerShowV254Test-608938657-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa.vmdk to [datastore1] 8016bafc-b287-4df1-bd4b-a00210177a86/8016bafc-b287-4df1-bd4b-a00210177a86.vmdk {{(pid=61806) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 646.484868] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-e8db3b70-c93a-419a-89f8-b994313357e1 tempest-ServerShowV254Test-608938657 tempest-ServerShowV254Test-608938657-project-member] [instance: 8016bafc-b287-4df1-bd4b-a00210177a86] Extending root virtual disk to 1048576 {{(pid=61806) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 646.485120] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-b8d6f91a-4d70-4ecb-95b2-80a94222c5a4 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 646.491485] env[61806]: DEBUG oslo_vmware.api [None req-e8db3b70-c93a-419a-89f8-b994313357e1 tempest-ServerShowV254Test-608938657 tempest-ServerShowV254Test-608938657-project-member] Waiting for the task: (returnval){ [ 646.491485] env[61806]: value = "task-1294458" [ 646.491485] env[61806]: _type = "Task" [ 646.491485] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 646.500091] env[61806]: DEBUG oslo_vmware.api [None req-e8db3b70-c93a-419a-89f8-b994313357e1 tempest-ServerShowV254Test-608938657 tempest-ServerShowV254Test-608938657-project-member] Task: {'id': task-1294458, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 646.733826] env[61806]: DEBUG oslo_concurrency.lockutils [None req-98d962a0-2632-4a9a-8d12-70ce4acc382b tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.079s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 646.734479] env[61806]: ERROR nova.compute.manager [None req-98d962a0-2632-4a9a-8d12-70ce4acc382b tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] [instance: b8249dc5-5561-454a-aa01-477f9f1fc0e1] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 340e11ef-b16f-45b1-8a9b-cf7c087aa898, please check neutron logs for more information. [ 646.734479] env[61806]: ERROR nova.compute.manager [instance: b8249dc5-5561-454a-aa01-477f9f1fc0e1] Traceback (most recent call last): [ 646.734479] env[61806]: ERROR nova.compute.manager [instance: b8249dc5-5561-454a-aa01-477f9f1fc0e1] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 646.734479] env[61806]: ERROR nova.compute.manager [instance: b8249dc5-5561-454a-aa01-477f9f1fc0e1] self.driver.spawn(context, instance, image_meta, [ 646.734479] env[61806]: ERROR nova.compute.manager [instance: b8249dc5-5561-454a-aa01-477f9f1fc0e1] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 646.734479] env[61806]: ERROR nova.compute.manager [instance: b8249dc5-5561-454a-aa01-477f9f1fc0e1] self._vmops.spawn(context, instance, image_meta, injected_files, [ 646.734479] env[61806]: ERROR nova.compute.manager [instance: b8249dc5-5561-454a-aa01-477f9f1fc0e1] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 646.734479] env[61806]: ERROR nova.compute.manager [instance: b8249dc5-5561-454a-aa01-477f9f1fc0e1] vm_ref = self.build_virtual_machine(instance, [ 646.734479] env[61806]: ERROR nova.compute.manager [instance: b8249dc5-5561-454a-aa01-477f9f1fc0e1] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 646.734479] env[61806]: ERROR nova.compute.manager [instance: b8249dc5-5561-454a-aa01-477f9f1fc0e1] vif_infos = vmwarevif.get_vif_info(self._session, [ 646.734479] env[61806]: ERROR nova.compute.manager [instance: b8249dc5-5561-454a-aa01-477f9f1fc0e1] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 646.734794] env[61806]: ERROR nova.compute.manager [instance: b8249dc5-5561-454a-aa01-477f9f1fc0e1] for vif in network_info: [ 646.734794] env[61806]: ERROR nova.compute.manager [instance: b8249dc5-5561-454a-aa01-477f9f1fc0e1] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 646.734794] env[61806]: ERROR nova.compute.manager [instance: b8249dc5-5561-454a-aa01-477f9f1fc0e1] return self._sync_wrapper(fn, *args, **kwargs) [ 646.734794] env[61806]: ERROR nova.compute.manager [instance: b8249dc5-5561-454a-aa01-477f9f1fc0e1] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 646.734794] env[61806]: ERROR nova.compute.manager [instance: b8249dc5-5561-454a-aa01-477f9f1fc0e1] self.wait() [ 646.734794] env[61806]: ERROR nova.compute.manager [instance: b8249dc5-5561-454a-aa01-477f9f1fc0e1] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 646.734794] env[61806]: ERROR nova.compute.manager [instance: b8249dc5-5561-454a-aa01-477f9f1fc0e1] self[:] = self._gt.wait() [ 646.734794] env[61806]: ERROR nova.compute.manager [instance: b8249dc5-5561-454a-aa01-477f9f1fc0e1] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 646.734794] env[61806]: ERROR nova.compute.manager [instance: b8249dc5-5561-454a-aa01-477f9f1fc0e1] return self._exit_event.wait() [ 646.734794] env[61806]: ERROR nova.compute.manager [instance: b8249dc5-5561-454a-aa01-477f9f1fc0e1] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 646.734794] env[61806]: ERROR nova.compute.manager [instance: b8249dc5-5561-454a-aa01-477f9f1fc0e1] result = hub.switch() [ 646.734794] env[61806]: ERROR nova.compute.manager [instance: b8249dc5-5561-454a-aa01-477f9f1fc0e1] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 646.734794] env[61806]: ERROR nova.compute.manager [instance: b8249dc5-5561-454a-aa01-477f9f1fc0e1] return self.greenlet.switch() [ 646.735113] env[61806]: ERROR nova.compute.manager [instance: b8249dc5-5561-454a-aa01-477f9f1fc0e1] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 646.735113] env[61806]: ERROR nova.compute.manager [instance: b8249dc5-5561-454a-aa01-477f9f1fc0e1] result = function(*args, **kwargs) [ 646.735113] env[61806]: ERROR nova.compute.manager [instance: b8249dc5-5561-454a-aa01-477f9f1fc0e1] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 646.735113] env[61806]: ERROR nova.compute.manager [instance: b8249dc5-5561-454a-aa01-477f9f1fc0e1] return func(*args, **kwargs) [ 646.735113] env[61806]: ERROR nova.compute.manager [instance: b8249dc5-5561-454a-aa01-477f9f1fc0e1] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 646.735113] env[61806]: ERROR nova.compute.manager [instance: b8249dc5-5561-454a-aa01-477f9f1fc0e1] raise e [ 646.735113] env[61806]: ERROR nova.compute.manager [instance: b8249dc5-5561-454a-aa01-477f9f1fc0e1] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 646.735113] env[61806]: ERROR nova.compute.manager [instance: b8249dc5-5561-454a-aa01-477f9f1fc0e1] nwinfo = self.network_api.allocate_for_instance( [ 646.735113] env[61806]: ERROR nova.compute.manager [instance: b8249dc5-5561-454a-aa01-477f9f1fc0e1] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 646.735113] env[61806]: ERROR nova.compute.manager [instance: b8249dc5-5561-454a-aa01-477f9f1fc0e1] created_port_ids = self._update_ports_for_instance( [ 646.735113] env[61806]: ERROR nova.compute.manager [instance: b8249dc5-5561-454a-aa01-477f9f1fc0e1] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 646.735113] env[61806]: ERROR nova.compute.manager [instance: b8249dc5-5561-454a-aa01-477f9f1fc0e1] with excutils.save_and_reraise_exception(): [ 646.735113] env[61806]: ERROR nova.compute.manager [instance: b8249dc5-5561-454a-aa01-477f9f1fc0e1] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 646.735424] env[61806]: ERROR nova.compute.manager [instance: b8249dc5-5561-454a-aa01-477f9f1fc0e1] self.force_reraise() [ 646.735424] env[61806]: ERROR nova.compute.manager [instance: b8249dc5-5561-454a-aa01-477f9f1fc0e1] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 646.735424] env[61806]: ERROR nova.compute.manager [instance: b8249dc5-5561-454a-aa01-477f9f1fc0e1] raise self.value [ 646.735424] env[61806]: ERROR nova.compute.manager [instance: b8249dc5-5561-454a-aa01-477f9f1fc0e1] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 646.735424] env[61806]: ERROR nova.compute.manager [instance: b8249dc5-5561-454a-aa01-477f9f1fc0e1] updated_port = self._update_port( [ 646.735424] env[61806]: ERROR nova.compute.manager [instance: b8249dc5-5561-454a-aa01-477f9f1fc0e1] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 646.735424] env[61806]: ERROR nova.compute.manager [instance: b8249dc5-5561-454a-aa01-477f9f1fc0e1] _ensure_no_port_binding_failure(port) [ 646.735424] env[61806]: ERROR nova.compute.manager [instance: b8249dc5-5561-454a-aa01-477f9f1fc0e1] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 646.735424] env[61806]: ERROR nova.compute.manager [instance: b8249dc5-5561-454a-aa01-477f9f1fc0e1] raise exception.PortBindingFailed(port_id=port['id']) [ 646.735424] env[61806]: ERROR nova.compute.manager [instance: b8249dc5-5561-454a-aa01-477f9f1fc0e1] nova.exception.PortBindingFailed: Binding failed for port 340e11ef-b16f-45b1-8a9b-cf7c087aa898, please check neutron logs for more information. [ 646.735424] env[61806]: ERROR nova.compute.manager [instance: b8249dc5-5561-454a-aa01-477f9f1fc0e1] [ 646.735691] env[61806]: DEBUG nova.compute.utils [None req-98d962a0-2632-4a9a-8d12-70ce4acc382b tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] [instance: b8249dc5-5561-454a-aa01-477f9f1fc0e1] Binding failed for port 340e11ef-b16f-45b1-8a9b-cf7c087aa898, please check neutron logs for more information. {{(pid=61806) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 646.736742] env[61806]: DEBUG oslo_concurrency.lockutils [None req-4803fdfc-9054-49cf-8525-e438672eeedd tempest-ServersAdmin275Test-1945525728 tempest-ServersAdmin275Test-1945525728-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 15.976s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 646.738154] env[61806]: INFO nova.compute.claims [None req-4803fdfc-9054-49cf-8525-e438672eeedd tempest-ServersAdmin275Test-1945525728 tempest-ServersAdmin275Test-1945525728-project-member] [instance: 9db80c3c-e2c3-4034-a79a-d7be924c4ff6] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 646.740840] env[61806]: DEBUG nova.compute.manager [None req-98d962a0-2632-4a9a-8d12-70ce4acc382b tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] [instance: b8249dc5-5561-454a-aa01-477f9f1fc0e1] Build of instance b8249dc5-5561-454a-aa01-477f9f1fc0e1 was re-scheduled: Binding failed for port 340e11ef-b16f-45b1-8a9b-cf7c087aa898, please check neutron logs for more information. {{(pid=61806) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 646.742037] env[61806]: DEBUG nova.compute.manager [None req-98d962a0-2632-4a9a-8d12-70ce4acc382b tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] [instance: b8249dc5-5561-454a-aa01-477f9f1fc0e1] Unplugging VIFs for instance {{(pid=61806) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 646.742037] env[61806]: DEBUG oslo_concurrency.lockutils [None req-98d962a0-2632-4a9a-8d12-70ce4acc382b tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] Acquiring lock "refresh_cache-b8249dc5-5561-454a-aa01-477f9f1fc0e1" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 646.742037] env[61806]: DEBUG oslo_concurrency.lockutils [None req-98d962a0-2632-4a9a-8d12-70ce4acc382b tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] Acquired lock "refresh_cache-b8249dc5-5561-454a-aa01-477f9f1fc0e1" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 646.742037] env[61806]: DEBUG nova.network.neutron [None req-98d962a0-2632-4a9a-8d12-70ce4acc382b tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] [instance: b8249dc5-5561-454a-aa01-477f9f1fc0e1] Building network info cache for instance {{(pid=61806) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 646.806515] env[61806]: INFO nova.compute.manager [None req-a6b8da3a-7829-486a-884d-0f28be02e38c tempest-ServersTestManualDisk-151630353 tempest-ServersTestManualDisk-151630353-project-member] [instance: 38a79453-939d-4649-9419-a2799636180d] Took 1.03 seconds to deallocate network for instance. [ 646.964023] env[61806]: DEBUG oslo_concurrency.lockutils [req-0d4f4869-d8a5-4a98-9582-fe758d1be0a6 req-fc8a0659-8afd-401e-8a13-8b894e8dc341 service nova] Releasing lock "refresh_cache-56d26d08-3c6d-45e8-b238-6dbca6024561" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 646.964396] env[61806]: DEBUG nova.compute.manager [req-0d4f4869-d8a5-4a98-9582-fe758d1be0a6 req-fc8a0659-8afd-401e-8a13-8b894e8dc341 service nova] [instance: 56d26d08-3c6d-45e8-b238-6dbca6024561] Received event network-vif-deleted-1570039f-d77a-49a9-a526-42d234745ef6 {{(pid=61806) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 646.964817] env[61806]: DEBUG oslo_concurrency.lockutils [None req-44eba412-ca3f-4386-8c36-f42e9734628d tempest-ServersTestBootFromVolume-1815739798 tempest-ServersTestBootFromVolume-1815739798-project-member] Acquired lock "refresh_cache-56d26d08-3c6d-45e8-b238-6dbca6024561" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 646.965086] env[61806]: DEBUG nova.network.neutron [None req-44eba412-ca3f-4386-8c36-f42e9734628d tempest-ServersTestBootFromVolume-1815739798 tempest-ServersTestBootFromVolume-1815739798-project-member] [instance: 56d26d08-3c6d-45e8-b238-6dbca6024561] Building network info cache for instance {{(pid=61806) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 647.000811] env[61806]: DEBUG oslo_vmware.api [None req-e8db3b70-c93a-419a-89f8-b994313357e1 tempest-ServerShowV254Test-608938657 tempest-ServerShowV254Test-608938657-project-member] Task: {'id': task-1294458, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.080342} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 647.001109] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-e8db3b70-c93a-419a-89f8-b994313357e1 tempest-ServerShowV254Test-608938657 tempest-ServerShowV254Test-608938657-project-member] [instance: 8016bafc-b287-4df1-bd4b-a00210177a86] Extended root virtual disk {{(pid=61806) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 647.001917] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3378feac-7cb9-46b8-b60f-f14acbef9400 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 647.022972] env[61806]: DEBUG nova.virt.vmwareapi.volumeops [None req-e8db3b70-c93a-419a-89f8-b994313357e1 tempest-ServerShowV254Test-608938657 tempest-ServerShowV254Test-608938657-project-member] [instance: 8016bafc-b287-4df1-bd4b-a00210177a86] Reconfiguring VM instance instance-0000001c to attach disk [datastore1] 8016bafc-b287-4df1-bd4b-a00210177a86/8016bafc-b287-4df1-bd4b-a00210177a86.vmdk or device None with type sparse {{(pid=61806) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 647.023333] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-3730fc47-caa1-4142-8c86-0dcd3781bfb2 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 647.042496] env[61806]: DEBUG oslo_vmware.api [None req-e8db3b70-c93a-419a-89f8-b994313357e1 tempest-ServerShowV254Test-608938657 tempest-ServerShowV254Test-608938657-project-member] Waiting for the task: (returnval){ [ 647.042496] env[61806]: value = "task-1294459" [ 647.042496] env[61806]: _type = "Task" [ 647.042496] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 647.050205] env[61806]: DEBUG oslo_vmware.api [None req-e8db3b70-c93a-419a-89f8-b994313357e1 tempest-ServerShowV254Test-608938657 tempest-ServerShowV254Test-608938657-project-member] Task: {'id': task-1294459, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 647.260679] env[61806]: DEBUG nova.network.neutron [None req-98d962a0-2632-4a9a-8d12-70ce4acc382b tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] [instance: b8249dc5-5561-454a-aa01-477f9f1fc0e1] Instance cache missing network info. {{(pid=61806) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 647.353366] env[61806]: DEBUG nova.network.neutron [None req-98d962a0-2632-4a9a-8d12-70ce4acc382b tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] [instance: b8249dc5-5561-454a-aa01-477f9f1fc0e1] Updating instance_info_cache with network_info: [] {{(pid=61806) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 647.482964] env[61806]: DEBUG nova.network.neutron [None req-44eba412-ca3f-4386-8c36-f42e9734628d tempest-ServersTestBootFromVolume-1815739798 tempest-ServersTestBootFromVolume-1815739798-project-member] [instance: 56d26d08-3c6d-45e8-b238-6dbca6024561] Instance cache missing network info. {{(pid=61806) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 647.551886] env[61806]: DEBUG oslo_vmware.api [None req-e8db3b70-c93a-419a-89f8-b994313357e1 tempest-ServerShowV254Test-608938657 tempest-ServerShowV254Test-608938657-project-member] Task: {'id': task-1294459, 'name': ReconfigVM_Task, 'duration_secs': 0.29183} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 647.552177] env[61806]: DEBUG nova.virt.vmwareapi.volumeops [None req-e8db3b70-c93a-419a-89f8-b994313357e1 tempest-ServerShowV254Test-608938657 tempest-ServerShowV254Test-608938657-project-member] [instance: 8016bafc-b287-4df1-bd4b-a00210177a86] Reconfigured VM instance instance-0000001c to attach disk [datastore1] 8016bafc-b287-4df1-bd4b-a00210177a86/8016bafc-b287-4df1-bd4b-a00210177a86.vmdk or device None with type sparse {{(pid=61806) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 647.552852] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-3d602624-db4d-46c2-85f8-08a5baef14d2 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 647.557981] env[61806]: DEBUG nova.network.neutron [None req-44eba412-ca3f-4386-8c36-f42e9734628d tempest-ServersTestBootFromVolume-1815739798 tempest-ServersTestBootFromVolume-1815739798-project-member] [instance: 56d26d08-3c6d-45e8-b238-6dbca6024561] Updating instance_info_cache with network_info: [] {{(pid=61806) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 647.560149] env[61806]: DEBUG oslo_vmware.api [None req-e8db3b70-c93a-419a-89f8-b994313357e1 tempest-ServerShowV254Test-608938657 tempest-ServerShowV254Test-608938657-project-member] Waiting for the task: (returnval){ [ 647.560149] env[61806]: value = "task-1294460" [ 647.560149] env[61806]: _type = "Task" [ 647.560149] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 647.568749] env[61806]: DEBUG oslo_vmware.api [None req-e8db3b70-c93a-419a-89f8-b994313357e1 tempest-ServerShowV254Test-608938657 tempest-ServerShowV254Test-608938657-project-member] Task: {'id': task-1294460, 'name': Rename_Task} progress is 5%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 647.832103] env[61806]: INFO nova.scheduler.client.report [None req-a6b8da3a-7829-486a-884d-0f28be02e38c tempest-ServersTestManualDisk-151630353 tempest-ServersTestManualDisk-151630353-project-member] Deleted allocations for instance 38a79453-939d-4649-9419-a2799636180d [ 647.856859] env[61806]: DEBUG oslo_concurrency.lockutils [None req-98d962a0-2632-4a9a-8d12-70ce4acc382b tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] Releasing lock "refresh_cache-b8249dc5-5561-454a-aa01-477f9f1fc0e1" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 647.856859] env[61806]: DEBUG nova.compute.manager [None req-98d962a0-2632-4a9a-8d12-70ce4acc382b tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61806) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 647.857148] env[61806]: DEBUG nova.compute.manager [None req-98d962a0-2632-4a9a-8d12-70ce4acc382b tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] [instance: b8249dc5-5561-454a-aa01-477f9f1fc0e1] Deallocating network for instance {{(pid=61806) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 647.857148] env[61806]: DEBUG nova.network.neutron [None req-98d962a0-2632-4a9a-8d12-70ce4acc382b tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] [instance: b8249dc5-5561-454a-aa01-477f9f1fc0e1] deallocate_for_instance() {{(pid=61806) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 647.875029] env[61806]: DEBUG nova.network.neutron [None req-98d962a0-2632-4a9a-8d12-70ce4acc382b tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] [instance: b8249dc5-5561-454a-aa01-477f9f1fc0e1] Instance cache missing network info. {{(pid=61806) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 648.061668] env[61806]: DEBUG oslo_concurrency.lockutils [None req-44eba412-ca3f-4386-8c36-f42e9734628d tempest-ServersTestBootFromVolume-1815739798 tempest-ServersTestBootFromVolume-1815739798-project-member] Releasing lock "refresh_cache-56d26d08-3c6d-45e8-b238-6dbca6024561" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 648.063019] env[61806]: DEBUG nova.compute.manager [None req-44eba412-ca3f-4386-8c36-f42e9734628d tempest-ServersTestBootFromVolume-1815739798 tempest-ServersTestBootFromVolume-1815739798-project-member] [instance: 56d26d08-3c6d-45e8-b238-6dbca6024561] Start destroying the instance on the hypervisor. {{(pid=61806) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 648.063019] env[61806]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-db23ccaa-6c20-4d53-9f73-50e3257a10c8 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 648.078257] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-514bb063-7784-495b-8adb-6e407e4ada98 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 648.090401] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-65dc6888-a961-470e-a352-2e90631850d0 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 648.096284] env[61806]: DEBUG oslo_vmware.api [None req-e8db3b70-c93a-419a-89f8-b994313357e1 tempest-ServerShowV254Test-608938657 tempest-ServerShowV254Test-608938657-project-member] Task: {'id': task-1294460, 'name': Rename_Task, 'duration_secs': 0.16631} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 648.097499] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [None req-e8db3b70-c93a-419a-89f8-b994313357e1 tempest-ServerShowV254Test-608938657 tempest-ServerShowV254Test-608938657-project-member] [instance: 8016bafc-b287-4df1-bd4b-a00210177a86] Powering on the VM {{(pid=61806) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 648.097731] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-3ddba613-9013-4b32-97cc-4d6369e6c7a8 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 648.103604] env[61806]: DEBUG oslo_vmware.api [None req-e8db3b70-c93a-419a-89f8-b994313357e1 tempest-ServerShowV254Test-608938657 tempest-ServerShowV254Test-608938657-project-member] Waiting for the task: (returnval){ [ 648.103604] env[61806]: value = "task-1294461" [ 648.103604] env[61806]: _type = "Task" [ 648.103604] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 648.108280] env[61806]: WARNING nova.virt.vmwareapi.driver [None req-44eba412-ca3f-4386-8c36-f42e9734628d tempest-ServersTestBootFromVolume-1815739798 tempest-ServersTestBootFromVolume-1815739798-project-member] [instance: 56d26d08-3c6d-45e8-b238-6dbca6024561] Instance does not exists. Proceeding to delete instance properties on datastore: nova.exception.InstanceNotFound: Instance 56d26d08-3c6d-45e8-b238-6dbca6024561 could not be found. [ 648.108540] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-44eba412-ca3f-4386-8c36-f42e9734628d tempest-ServersTestBootFromVolume-1815739798 tempest-ServersTestBootFromVolume-1815739798-project-member] [instance: 56d26d08-3c6d-45e8-b238-6dbca6024561] Destroying instance {{(pid=61806) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 648.112782] env[61806]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-43714abe-4303-45cb-a95d-c7846c5c1109 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 648.115241] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-00f171bc-4af0-4ea2-bf71-27393b59464c {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 648.122900] env[61806]: DEBUG oslo_vmware.api [None req-e8db3b70-c93a-419a-89f8-b994313357e1 tempest-ServerShowV254Test-608938657 tempest-ServerShowV254Test-608938657-project-member] Task: {'id': task-1294461, 'name': PowerOnVM_Task} progress is 33%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 648.152550] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bde83cc9-2d4d-4cce-aee0-526ac03b58c7 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 648.163585] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3ef4cadb-430d-40cc-912f-67343b5f8dcd {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 648.171302] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2f584778-db69-4cda-968a-78685618862f {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 648.179198] env[61806]: WARNING nova.virt.vmwareapi.vmops [None req-44eba412-ca3f-4386-8c36-f42e9734628d tempest-ServersTestBootFromVolume-1815739798 tempest-ServersTestBootFromVolume-1815739798-project-member] [instance: 56d26d08-3c6d-45e8-b238-6dbca6024561] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 56d26d08-3c6d-45e8-b238-6dbca6024561 could not be found. [ 648.179420] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-44eba412-ca3f-4386-8c36-f42e9734628d tempest-ServersTestBootFromVolume-1815739798 tempest-ServersTestBootFromVolume-1815739798-project-member] [instance: 56d26d08-3c6d-45e8-b238-6dbca6024561] Instance destroyed {{(pid=61806) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 648.179613] env[61806]: INFO nova.compute.manager [None req-44eba412-ca3f-4386-8c36-f42e9734628d tempest-ServersTestBootFromVolume-1815739798 tempest-ServersTestBootFromVolume-1815739798-project-member] [instance: 56d26d08-3c6d-45e8-b238-6dbca6024561] Took 0.12 seconds to destroy the instance on the hypervisor. [ 648.179847] env[61806]: DEBUG oslo.service.loopingcall [None req-44eba412-ca3f-4386-8c36-f42e9734628d tempest-ServersTestBootFromVolume-1815739798 tempest-ServersTestBootFromVolume-1815739798-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61806) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 648.180426] env[61806]: DEBUG nova.compute.manager [-] [instance: 56d26d08-3c6d-45e8-b238-6dbca6024561] Deallocating network for instance {{(pid=61806) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 648.180542] env[61806]: DEBUG nova.network.neutron [-] [instance: 56d26d08-3c6d-45e8-b238-6dbca6024561] deallocate_for_instance() {{(pid=61806) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 648.189773] env[61806]: DEBUG nova.compute.provider_tree [None req-4803fdfc-9054-49cf-8525-e438672eeedd tempest-ServersAdmin275Test-1945525728 tempest-ServersAdmin275Test-1945525728-project-member] Inventory has not changed in ProviderTree for provider: a2858be1-fd22-4e08-979e-87ad25293407 {{(pid=61806) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 648.341643] env[61806]: DEBUG nova.network.neutron [-] [instance: 56d26d08-3c6d-45e8-b238-6dbca6024561] Instance cache missing network info. {{(pid=61806) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 648.343320] env[61806]: DEBUG oslo_concurrency.lockutils [None req-a6b8da3a-7829-486a-884d-0f28be02e38c tempest-ServersTestManualDisk-151630353 tempest-ServersTestManualDisk-151630353-project-member] Lock "38a79453-939d-4649-9419-a2799636180d" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 94.385s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 648.377659] env[61806]: DEBUG nova.network.neutron [None req-98d962a0-2632-4a9a-8d12-70ce4acc382b tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] [instance: b8249dc5-5561-454a-aa01-477f9f1fc0e1] Updating instance_info_cache with network_info: [] {{(pid=61806) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 648.617376] env[61806]: DEBUG oslo_vmware.api [None req-e8db3b70-c93a-419a-89f8-b994313357e1 tempest-ServerShowV254Test-608938657 tempest-ServerShowV254Test-608938657-project-member] Task: {'id': task-1294461, 'name': PowerOnVM_Task, 'duration_secs': 0.42242} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 648.617641] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [None req-e8db3b70-c93a-419a-89f8-b994313357e1 tempest-ServerShowV254Test-608938657 tempest-ServerShowV254Test-608938657-project-member] [instance: 8016bafc-b287-4df1-bd4b-a00210177a86] Powered on the VM {{(pid=61806) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 648.617828] env[61806]: INFO nova.compute.manager [None req-e8db3b70-c93a-419a-89f8-b994313357e1 tempest-ServerShowV254Test-608938657 tempest-ServerShowV254Test-608938657-project-member] [instance: 8016bafc-b287-4df1-bd4b-a00210177a86] Took 8.95 seconds to spawn the instance on the hypervisor. [ 648.618088] env[61806]: DEBUG nova.compute.manager [None req-e8db3b70-c93a-419a-89f8-b994313357e1 tempest-ServerShowV254Test-608938657 tempest-ServerShowV254Test-608938657-project-member] [instance: 8016bafc-b287-4df1-bd4b-a00210177a86] Checking state {{(pid=61806) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 648.618845] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-65ad299c-5c99-475f-a4d4-459da99c59ab {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 648.692411] env[61806]: DEBUG nova.scheduler.client.report [None req-4803fdfc-9054-49cf-8525-e438672eeedd tempest-ServersAdmin275Test-1945525728 tempest-ServersAdmin275Test-1945525728-project-member] Inventory has not changed for provider a2858be1-fd22-4e08-979e-87ad25293407 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 139, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61806) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 648.845092] env[61806]: DEBUG nova.network.neutron [-] [instance: 56d26d08-3c6d-45e8-b238-6dbca6024561] Updating instance_info_cache with network_info: [] {{(pid=61806) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 648.846190] env[61806]: DEBUG nova.compute.manager [None req-7bb45eee-0cbe-4995-9898-530114f6f104 tempest-ImagesOneServerNegativeTestJSON-606465252 tempest-ImagesOneServerNegativeTestJSON-606465252-project-member] [instance: 8d0cdc81-56ed-4a7a-8221-30bb906e16b2] Starting instance... {{(pid=61806) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 648.880655] env[61806]: INFO nova.compute.manager [None req-98d962a0-2632-4a9a-8d12-70ce4acc382b tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] [instance: b8249dc5-5561-454a-aa01-477f9f1fc0e1] Took 1.02 seconds to deallocate network for instance. [ 649.135938] env[61806]: INFO nova.compute.manager [None req-e8db3b70-c93a-419a-89f8-b994313357e1 tempest-ServerShowV254Test-608938657 tempest-ServerShowV254Test-608938657-project-member] [instance: 8016bafc-b287-4df1-bd4b-a00210177a86] Took 29.59 seconds to build instance. [ 649.197353] env[61806]: DEBUG oslo_concurrency.lockutils [None req-4803fdfc-9054-49cf-8525-e438672eeedd tempest-ServersAdmin275Test-1945525728 tempest-ServersAdmin275Test-1945525728-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.460s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 649.197883] env[61806]: DEBUG nova.compute.manager [None req-4803fdfc-9054-49cf-8525-e438672eeedd tempest-ServersAdmin275Test-1945525728 tempest-ServersAdmin275Test-1945525728-project-member] [instance: 9db80c3c-e2c3-4034-a79a-d7be924c4ff6] Start building networks asynchronously for instance. {{(pid=61806) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 649.202766] env[61806]: DEBUG oslo_concurrency.lockutils [None req-a93ef3a4-6e83-4dfe-9292-bac21f14d678 tempest-ServerExternalEventsTest-1847900285 tempest-ServerExternalEventsTest-1847900285-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 16.508s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 649.207024] env[61806]: INFO nova.compute.claims [None req-a93ef3a4-6e83-4dfe-9292-bac21f14d678 tempest-ServerExternalEventsTest-1847900285 tempest-ServerExternalEventsTest-1847900285-project-member] [instance: d5434dd2-206f-4a1e-b97b-bd6f691b6ddf] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 649.353196] env[61806]: INFO nova.compute.manager [-] [instance: 56d26d08-3c6d-45e8-b238-6dbca6024561] Took 1.17 seconds to deallocate network for instance. [ 649.371303] env[61806]: DEBUG oslo_concurrency.lockutils [None req-7bb45eee-0cbe-4995-9898-530114f6f104 tempest-ImagesOneServerNegativeTestJSON-606465252 tempest-ImagesOneServerNegativeTestJSON-606465252-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 649.637861] env[61806]: DEBUG oslo_concurrency.lockutils [None req-e8db3b70-c93a-419a-89f8-b994313357e1 tempest-ServerShowV254Test-608938657 tempest-ServerShowV254Test-608938657-project-member] Lock "8016bafc-b287-4df1-bd4b-a00210177a86" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 91.397s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 649.710848] env[61806]: DEBUG nova.compute.utils [None req-4803fdfc-9054-49cf-8525-e438672eeedd tempest-ServersAdmin275Test-1945525728 tempest-ServersAdmin275Test-1945525728-project-member] Using /dev/sd instead of None {{(pid=61806) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 649.714046] env[61806]: DEBUG nova.compute.manager [None req-4803fdfc-9054-49cf-8525-e438672eeedd tempest-ServersAdmin275Test-1945525728 tempest-ServersAdmin275Test-1945525728-project-member] [instance: 9db80c3c-e2c3-4034-a79a-d7be924c4ff6] Not allocating networking since 'none' was specified. {{(pid=61806) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1977}} [ 649.750508] env[61806]: INFO nova.compute.manager [None req-fe158576-4234-4a6b-b48d-5ca8bb977c40 tempest-ServerShowV254Test-608938657 tempest-ServerShowV254Test-608938657-project-member] [instance: 8016bafc-b287-4df1-bd4b-a00210177a86] Rebuilding instance [ 649.801587] env[61806]: DEBUG nova.compute.manager [None req-fe158576-4234-4a6b-b48d-5ca8bb977c40 tempest-ServerShowV254Test-608938657 tempest-ServerShowV254Test-608938657-project-member] [instance: 8016bafc-b287-4df1-bd4b-a00210177a86] Checking state {{(pid=61806) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 649.802827] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8e0f81ee-61cf-4c2c-92da-0aad1f834733 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 649.915138] env[61806]: INFO nova.scheduler.client.report [None req-98d962a0-2632-4a9a-8d12-70ce4acc382b tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] Deleted allocations for instance b8249dc5-5561-454a-aa01-477f9f1fc0e1 [ 649.930301] env[61806]: INFO nova.compute.manager [None req-44eba412-ca3f-4386-8c36-f42e9734628d tempest-ServersTestBootFromVolume-1815739798 tempest-ServersTestBootFromVolume-1815739798-project-member] [instance: 56d26d08-3c6d-45e8-b238-6dbca6024561] Took 0.58 seconds to detach 1 volumes for instance. [ 649.939469] env[61806]: DEBUG nova.compute.claims [None req-44eba412-ca3f-4386-8c36-f42e9734628d tempest-ServersTestBootFromVolume-1815739798 tempest-ServersTestBootFromVolume-1815739798-project-member] [instance: 56d26d08-3c6d-45e8-b238-6dbca6024561] Aborting claim: {{(pid=61806) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 649.941096] env[61806]: DEBUG oslo_concurrency.lockutils [None req-44eba412-ca3f-4386-8c36-f42e9734628d tempest-ServersTestBootFromVolume-1815739798 tempest-ServersTestBootFromVolume-1815739798-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 650.140899] env[61806]: DEBUG nova.compute.manager [None req-865353ee-538a-40fc-bfba-502aba0b6a20 tempest-VolumesAdminNegativeTest-1387177911 tempest-VolumesAdminNegativeTest-1387177911-project-member] [instance: b7d55e87-95db-4c45-9417-bbece4fbb98d] Starting instance... {{(pid=61806) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 650.215712] env[61806]: DEBUG nova.compute.manager [None req-4803fdfc-9054-49cf-8525-e438672eeedd tempest-ServersAdmin275Test-1945525728 tempest-ServersAdmin275Test-1945525728-project-member] [instance: 9db80c3c-e2c3-4034-a79a-d7be924c4ff6] Start building block device mappings for instance. {{(pid=61806) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 650.316337] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [None req-fe158576-4234-4a6b-b48d-5ca8bb977c40 tempest-ServerShowV254Test-608938657 tempest-ServerShowV254Test-608938657-project-member] [instance: 8016bafc-b287-4df1-bd4b-a00210177a86] Powering off the VM {{(pid=61806) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 650.316724] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-f2093c20-4419-41cc-8171-57a24e051c67 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 650.329193] env[61806]: DEBUG oslo_vmware.api [None req-fe158576-4234-4a6b-b48d-5ca8bb977c40 tempest-ServerShowV254Test-608938657 tempest-ServerShowV254Test-608938657-project-member] Waiting for the task: (returnval){ [ 650.329193] env[61806]: value = "task-1294462" [ 650.329193] env[61806]: _type = "Task" [ 650.329193] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 650.338180] env[61806]: DEBUG oslo_vmware.api [None req-fe158576-4234-4a6b-b48d-5ca8bb977c40 tempest-ServerShowV254Test-608938657 tempest-ServerShowV254Test-608938657-project-member] Task: {'id': task-1294462, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 650.425639] env[61806]: DEBUG oslo_concurrency.lockutils [None req-98d962a0-2632-4a9a-8d12-70ce4acc382b tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] Lock "b8249dc5-5561-454a-aa01-477f9f1fc0e1" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 96.066s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 650.672162] env[61806]: DEBUG oslo_concurrency.lockutils [None req-865353ee-538a-40fc-bfba-502aba0b6a20 tempest-VolumesAdminNegativeTest-1387177911 tempest-VolumesAdminNegativeTest-1387177911-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 650.673867] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2b7ea09d-448d-4e3a-a009-5e8277137262 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 650.684549] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e026101a-9e76-47e9-a687-b27c40895257 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 650.727501] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cbed799f-ec48-46fc-ac84-066a3643b903 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 650.735662] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-46c508d8-5afd-45e1-af29-7e8d3e866870 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 650.756569] env[61806]: DEBUG nova.compute.provider_tree [None req-a93ef3a4-6e83-4dfe-9292-bac21f14d678 tempest-ServerExternalEventsTest-1847900285 tempest-ServerExternalEventsTest-1847900285-project-member] Inventory has not changed in ProviderTree for provider: a2858be1-fd22-4e08-979e-87ad25293407 {{(pid=61806) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 650.841905] env[61806]: DEBUG oslo_vmware.api [None req-fe158576-4234-4a6b-b48d-5ca8bb977c40 tempest-ServerShowV254Test-608938657 tempest-ServerShowV254Test-608938657-project-member] Task: {'id': task-1294462, 'name': PowerOffVM_Task, 'duration_secs': 0.200473} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 650.842132] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [None req-fe158576-4234-4a6b-b48d-5ca8bb977c40 tempest-ServerShowV254Test-608938657 tempest-ServerShowV254Test-608938657-project-member] [instance: 8016bafc-b287-4df1-bd4b-a00210177a86] Powered off the VM {{(pid=61806) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 650.842394] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-fe158576-4234-4a6b-b48d-5ca8bb977c40 tempest-ServerShowV254Test-608938657 tempest-ServerShowV254Test-608938657-project-member] [instance: 8016bafc-b287-4df1-bd4b-a00210177a86] Destroying instance {{(pid=61806) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 650.844358] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a637afba-4ac2-4733-ab23-a7ae573e0a33 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 650.850596] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-fe158576-4234-4a6b-b48d-5ca8bb977c40 tempest-ServerShowV254Test-608938657 tempest-ServerShowV254Test-608938657-project-member] [instance: 8016bafc-b287-4df1-bd4b-a00210177a86] Unregistering the VM {{(pid=61806) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 650.850830] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-97d3fa7a-ddac-42cb-a401-832ad76a0d1c {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 650.876741] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-fe158576-4234-4a6b-b48d-5ca8bb977c40 tempest-ServerShowV254Test-608938657 tempest-ServerShowV254Test-608938657-project-member] [instance: 8016bafc-b287-4df1-bd4b-a00210177a86] Unregistered the VM {{(pid=61806) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 650.876741] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-fe158576-4234-4a6b-b48d-5ca8bb977c40 tempest-ServerShowV254Test-608938657 tempest-ServerShowV254Test-608938657-project-member] [instance: 8016bafc-b287-4df1-bd4b-a00210177a86] Deleting contents of the VM from datastore datastore1 {{(pid=61806) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 650.876741] env[61806]: DEBUG nova.virt.vmwareapi.ds_util [None req-fe158576-4234-4a6b-b48d-5ca8bb977c40 tempest-ServerShowV254Test-608938657 tempest-ServerShowV254Test-608938657-project-member] Deleting the datastore file [datastore1] 8016bafc-b287-4df1-bd4b-a00210177a86 {{(pid=61806) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 650.876741] env[61806]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-7a71d2d3-2e0c-418b-89fb-d3cfe6b55a50 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 650.888106] env[61806]: DEBUG oslo_vmware.api [None req-fe158576-4234-4a6b-b48d-5ca8bb977c40 tempest-ServerShowV254Test-608938657 tempest-ServerShowV254Test-608938657-project-member] Waiting for the task: (returnval){ [ 650.888106] env[61806]: value = "task-1294464" [ 650.888106] env[61806]: _type = "Task" [ 650.888106] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 650.897049] env[61806]: DEBUG oslo_vmware.api [None req-fe158576-4234-4a6b-b48d-5ca8bb977c40 tempest-ServerShowV254Test-608938657 tempest-ServerShowV254Test-608938657-project-member] Task: {'id': task-1294464, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 650.928074] env[61806]: DEBUG nova.compute.manager [None req-81a55da1-45c4-45f8-af56-f8f7d424c8aa tempest-ServersTestJSON-430716070 tempest-ServersTestJSON-430716070-project-member] [instance: eca2e3da-b38c-482e-b9b6-4ad75d83caa1] Starting instance... {{(pid=61806) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 651.234065] env[61806]: DEBUG nova.compute.manager [None req-4803fdfc-9054-49cf-8525-e438672eeedd tempest-ServersAdmin275Test-1945525728 tempest-ServersAdmin275Test-1945525728-project-member] [instance: 9db80c3c-e2c3-4034-a79a-d7be924c4ff6] Start spawning the instance on the hypervisor. {{(pid=61806) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 651.257905] env[61806]: DEBUG nova.virt.hardware [None req-4803fdfc-9054-49cf-8525-e438672eeedd tempest-ServersAdmin275Test-1945525728 tempest-ServersAdmin275Test-1945525728-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-15T18:05:10Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-15T18:04:52Z,direct_url=,disk_format='vmdk',id=ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='84ba83fce2c349a6988173c5d6fc3b07',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-15T18:04:53Z,virtual_size=,visibility=), allow threads: False {{(pid=61806) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 651.258172] env[61806]: DEBUG nova.virt.hardware [None req-4803fdfc-9054-49cf-8525-e438672eeedd tempest-ServersAdmin275Test-1945525728 tempest-ServersAdmin275Test-1945525728-project-member] Flavor limits 0:0:0 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 651.258345] env[61806]: DEBUG nova.virt.hardware [None req-4803fdfc-9054-49cf-8525-e438672eeedd tempest-ServersAdmin275Test-1945525728 tempest-ServersAdmin275Test-1945525728-project-member] Image limits 0:0:0 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 651.258526] env[61806]: DEBUG nova.virt.hardware [None req-4803fdfc-9054-49cf-8525-e438672eeedd tempest-ServersAdmin275Test-1945525728 tempest-ServersAdmin275Test-1945525728-project-member] Flavor pref 0:0:0 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 651.258684] env[61806]: DEBUG nova.virt.hardware [None req-4803fdfc-9054-49cf-8525-e438672eeedd tempest-ServersAdmin275Test-1945525728 tempest-ServersAdmin275Test-1945525728-project-member] Image pref 0:0:0 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 651.258815] env[61806]: DEBUG nova.virt.hardware [None req-4803fdfc-9054-49cf-8525-e438672eeedd tempest-ServersAdmin275Test-1945525728 tempest-ServersAdmin275Test-1945525728-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 651.259109] env[61806]: DEBUG nova.virt.hardware [None req-4803fdfc-9054-49cf-8525-e438672eeedd tempest-ServersAdmin275Test-1945525728 tempest-ServersAdmin275Test-1945525728-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61806) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 651.259327] env[61806]: DEBUG nova.virt.hardware [None req-4803fdfc-9054-49cf-8525-e438672eeedd tempest-ServersAdmin275Test-1945525728 tempest-ServersAdmin275Test-1945525728-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61806) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 651.259541] env[61806]: DEBUG nova.virt.hardware [None req-4803fdfc-9054-49cf-8525-e438672eeedd tempest-ServersAdmin275Test-1945525728 tempest-ServersAdmin275Test-1945525728-project-member] Got 1 possible topologies {{(pid=61806) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 651.260276] env[61806]: DEBUG nova.virt.hardware [None req-4803fdfc-9054-49cf-8525-e438672eeedd tempest-ServersAdmin275Test-1945525728 tempest-ServersAdmin275Test-1945525728-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61806) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 651.260276] env[61806]: DEBUG nova.virt.hardware [None req-4803fdfc-9054-49cf-8525-e438672eeedd tempest-ServersAdmin275Test-1945525728 tempest-ServersAdmin275Test-1945525728-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61806) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 651.260979] env[61806]: DEBUG nova.scheduler.client.report [None req-a93ef3a4-6e83-4dfe-9292-bac21f14d678 tempest-ServerExternalEventsTest-1847900285 tempest-ServerExternalEventsTest-1847900285-project-member] Inventory has not changed for provider a2858be1-fd22-4e08-979e-87ad25293407 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 139, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61806) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 651.264741] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-55eb2875-51d4-404b-a295-5d07d9c8a49f {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 651.276222] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-50a6a369-1544-451c-a355-25d2cf7d7f28 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 651.293126] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-4803fdfc-9054-49cf-8525-e438672eeedd tempest-ServersAdmin275Test-1945525728 tempest-ServersAdmin275Test-1945525728-project-member] [instance: 9db80c3c-e2c3-4034-a79a-d7be924c4ff6] Instance VIF info [] {{(pid=61806) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 651.298586] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [None req-4803fdfc-9054-49cf-8525-e438672eeedd tempest-ServersAdmin275Test-1945525728 tempest-ServersAdmin275Test-1945525728-project-member] Creating folder: Project (5a245ce147a84c9fa0475cc4f35bfb4b). Parent ref: group-v277609. {{(pid=61806) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 651.299141] env[61806]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-899e2ad6-64de-4542-8e56-cb1916de990d {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 651.308698] env[61806]: INFO nova.virt.vmwareapi.vm_util [None req-4803fdfc-9054-49cf-8525-e438672eeedd tempest-ServersAdmin275Test-1945525728 tempest-ServersAdmin275Test-1945525728-project-member] Created folder: Project (5a245ce147a84c9fa0475cc4f35bfb4b) in parent group-v277609. [ 651.308698] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [None req-4803fdfc-9054-49cf-8525-e438672eeedd tempest-ServersAdmin275Test-1945525728 tempest-ServersAdmin275Test-1945525728-project-member] Creating folder: Instances. Parent ref: group-v277617. {{(pid=61806) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 651.308966] env[61806]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-8e59e57a-acfc-48af-88b3-f081a5c67b1e {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 651.317422] env[61806]: INFO nova.virt.vmwareapi.vm_util [None req-4803fdfc-9054-49cf-8525-e438672eeedd tempest-ServersAdmin275Test-1945525728 tempest-ServersAdmin275Test-1945525728-project-member] Created folder: Instances in parent group-v277617. [ 651.317602] env[61806]: DEBUG oslo.service.loopingcall [None req-4803fdfc-9054-49cf-8525-e438672eeedd tempest-ServersAdmin275Test-1945525728 tempest-ServersAdmin275Test-1945525728-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61806) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 651.317782] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 9db80c3c-e2c3-4034-a79a-d7be924c4ff6] Creating VM on the ESX host {{(pid=61806) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 651.317971] env[61806]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-a60bb26f-e75f-4727-a9b6-bd0509d84bf4 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 651.333592] env[61806]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 651.333592] env[61806]: value = "task-1294467" [ 651.333592] env[61806]: _type = "Task" [ 651.333592] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 651.341248] env[61806]: DEBUG oslo_vmware.api [-] Task: {'id': task-1294467, 'name': CreateVM_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 651.397798] env[61806]: DEBUG oslo_vmware.api [None req-fe158576-4234-4a6b-b48d-5ca8bb977c40 tempest-ServerShowV254Test-608938657 tempest-ServerShowV254Test-608938657-project-member] Task: {'id': task-1294464, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.113176} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 651.397798] env[61806]: DEBUG nova.virt.vmwareapi.ds_util [None req-fe158576-4234-4a6b-b48d-5ca8bb977c40 tempest-ServerShowV254Test-608938657 tempest-ServerShowV254Test-608938657-project-member] Deleted the datastore file {{(pid=61806) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 651.397979] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-fe158576-4234-4a6b-b48d-5ca8bb977c40 tempest-ServerShowV254Test-608938657 tempest-ServerShowV254Test-608938657-project-member] [instance: 8016bafc-b287-4df1-bd4b-a00210177a86] Deleted contents of the VM from datastore datastore1 {{(pid=61806) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 651.398039] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-fe158576-4234-4a6b-b48d-5ca8bb977c40 tempest-ServerShowV254Test-608938657 tempest-ServerShowV254Test-608938657-project-member] [instance: 8016bafc-b287-4df1-bd4b-a00210177a86] Instance destroyed {{(pid=61806) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 651.457098] env[61806]: DEBUG oslo_concurrency.lockutils [None req-81a55da1-45c4-45f8-af56-f8f7d424c8aa tempest-ServersTestJSON-430716070 tempest-ServersTestJSON-430716070-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 651.772632] env[61806]: DEBUG oslo_concurrency.lockutils [None req-a93ef3a4-6e83-4dfe-9292-bac21f14d678 tempest-ServerExternalEventsTest-1847900285 tempest-ServerExternalEventsTest-1847900285-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.570s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 651.773175] env[61806]: DEBUG nova.compute.manager [None req-a93ef3a4-6e83-4dfe-9292-bac21f14d678 tempest-ServerExternalEventsTest-1847900285 tempest-ServerExternalEventsTest-1847900285-project-member] [instance: d5434dd2-206f-4a1e-b97b-bd6f691b6ddf] Start building networks asynchronously for instance. {{(pid=61806) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 651.775878] env[61806]: DEBUG oslo_concurrency.lockutils [None req-3e50272a-5153-4e11-beee-d774a917f2df tempest-ServerPasswordTestJSON-1741413915 tempest-ServerPasswordTestJSON-1741413915-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 15.665s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 651.843892] env[61806]: DEBUG oslo_vmware.api [-] Task: {'id': task-1294467, 'name': CreateVM_Task, 'duration_secs': 0.313656} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 651.844179] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 9db80c3c-e2c3-4034-a79a-d7be924c4ff6] Created VM on the ESX host {{(pid=61806) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 651.845073] env[61806]: DEBUG oslo_vmware.service [None req-4803fdfc-9054-49cf-8525-e438672eeedd tempest-ServersAdmin275Test-1945525728 tempest-ServersAdmin275Test-1945525728-project-member] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-841c769e-70bf-448a-8669-466f20385603 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 651.850407] env[61806]: DEBUG oslo_concurrency.lockutils [None req-4803fdfc-9054-49cf-8525-e438672eeedd tempest-ServersAdmin275Test-1945525728 tempest-ServersAdmin275Test-1945525728-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 651.850596] env[61806]: DEBUG oslo_concurrency.lockutils [None req-4803fdfc-9054-49cf-8525-e438672eeedd tempest-ServersAdmin275Test-1945525728 tempest-ServersAdmin275Test-1945525728-project-member] Acquired lock "[datastore2] devstack-image-cache_base/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 651.850955] env[61806]: DEBUG oslo_concurrency.lockutils [None req-4803fdfc-9054-49cf-8525-e438672eeedd tempest-ServersAdmin275Test-1945525728 tempest-ServersAdmin275Test-1945525728-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 651.851206] env[61806]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-1058aebe-c5fa-4fee-be83-8148015cbf0f {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 651.855386] env[61806]: DEBUG oslo_vmware.api [None req-4803fdfc-9054-49cf-8525-e438672eeedd tempest-ServersAdmin275Test-1945525728 tempest-ServersAdmin275Test-1945525728-project-member] Waiting for the task: (returnval){ [ 651.855386] env[61806]: value = "session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]52dbb129-1058-0729-a1ee-a473406cd38a" [ 651.855386] env[61806]: _type = "Task" [ 651.855386] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 651.862179] env[61806]: DEBUG oslo_vmware.api [None req-4803fdfc-9054-49cf-8525-e438672eeedd tempest-ServersAdmin275Test-1945525728 tempest-ServersAdmin275Test-1945525728-project-member] Task: {'id': session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]52dbb129-1058-0729-a1ee-a473406cd38a, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 652.282025] env[61806]: DEBUG nova.compute.utils [None req-a93ef3a4-6e83-4dfe-9292-bac21f14d678 tempest-ServerExternalEventsTest-1847900285 tempest-ServerExternalEventsTest-1847900285-project-member] Using /dev/sd instead of None {{(pid=61806) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 652.285838] env[61806]: DEBUG nova.compute.manager [None req-a93ef3a4-6e83-4dfe-9292-bac21f14d678 tempest-ServerExternalEventsTest-1847900285 tempest-ServerExternalEventsTest-1847900285-project-member] [instance: d5434dd2-206f-4a1e-b97b-bd6f691b6ddf] Allocating IP information in the background. {{(pid=61806) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 652.287127] env[61806]: DEBUG nova.network.neutron [None req-a93ef3a4-6e83-4dfe-9292-bac21f14d678 tempest-ServerExternalEventsTest-1847900285 tempest-ServerExternalEventsTest-1847900285-project-member] [instance: d5434dd2-206f-4a1e-b97b-bd6f691b6ddf] allocate_for_instance() {{(pid=61806) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 652.341956] env[61806]: DEBUG nova.policy [None req-a93ef3a4-6e83-4dfe-9292-bac21f14d678 tempest-ServerExternalEventsTest-1847900285 tempest-ServerExternalEventsTest-1847900285-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '9fa9e3a40cc145a5a68f0c5650d04ee0', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'fe46298a6ea746d1bc157e27bf9a5245', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61806) authorize /opt/stack/nova/nova/policy.py:201}} [ 652.369567] env[61806]: DEBUG oslo_concurrency.lockutils [None req-4803fdfc-9054-49cf-8525-e438672eeedd tempest-ServersAdmin275Test-1945525728 tempest-ServersAdmin275Test-1945525728-project-member] Releasing lock "[datastore2] devstack-image-cache_base/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 652.369823] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-4803fdfc-9054-49cf-8525-e438672eeedd tempest-ServersAdmin275Test-1945525728 tempest-ServersAdmin275Test-1945525728-project-member] [instance: 9db80c3c-e2c3-4034-a79a-d7be924c4ff6] Processing image ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa {{(pid=61806) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 652.370065] env[61806]: DEBUG oslo_concurrency.lockutils [None req-4803fdfc-9054-49cf-8525-e438672eeedd tempest-ServersAdmin275Test-1945525728 tempest-ServersAdmin275Test-1945525728-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa.vmdk" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 652.370219] env[61806]: DEBUG oslo_concurrency.lockutils [None req-4803fdfc-9054-49cf-8525-e438672eeedd tempest-ServersAdmin275Test-1945525728 tempest-ServersAdmin275Test-1945525728-project-member] Acquired lock "[datastore2] devstack-image-cache_base/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa.vmdk" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 652.370424] env[61806]: DEBUG nova.virt.vmwareapi.ds_util [None req-4803fdfc-9054-49cf-8525-e438672eeedd tempest-ServersAdmin275Test-1945525728 tempest-ServersAdmin275Test-1945525728-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61806) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 652.370906] env[61806]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-ea71ee32-186e-41c4-8d9b-82da615fe713 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 652.381386] env[61806]: DEBUG nova.virt.vmwareapi.ds_util [None req-4803fdfc-9054-49cf-8525-e438672eeedd tempest-ServersAdmin275Test-1945525728 tempest-ServersAdmin275Test-1945525728-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61806) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 652.381603] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-4803fdfc-9054-49cf-8525-e438672eeedd tempest-ServersAdmin275Test-1945525728 tempest-ServersAdmin275Test-1945525728-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61806) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 652.382348] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ae0624c2-254e-43f9-a12c-10b4c76910c7 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 652.390581] env[61806]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-89cf501d-80be-463c-ad67-85407e122c9a {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 652.395314] env[61806]: DEBUG oslo_vmware.api [None req-4803fdfc-9054-49cf-8525-e438672eeedd tempest-ServersAdmin275Test-1945525728 tempest-ServersAdmin275Test-1945525728-project-member] Waiting for the task: (returnval){ [ 652.395314] env[61806]: value = "session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]524ddd89-96c3-8a73-5cae-1a79026ed54a" [ 652.395314] env[61806]: _type = "Task" [ 652.395314] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 652.411832] env[61806]: DEBUG oslo_vmware.api [None req-4803fdfc-9054-49cf-8525-e438672eeedd tempest-ServersAdmin275Test-1945525728 tempest-ServersAdmin275Test-1945525728-project-member] Task: {'id': session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]524ddd89-96c3-8a73-5cae-1a79026ed54a, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 652.451536] env[61806]: DEBUG nova.virt.hardware [None req-fe158576-4234-4a6b-b48d-5ca8bb977c40 tempest-ServerShowV254Test-608938657 tempest-ServerShowV254Test-608938657-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-15T18:05:10Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-15T18:04:52Z,direct_url=,disk_format='vmdk',id=ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='84ba83fce2c349a6988173c5d6fc3b07',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-15T18:04:53Z,virtual_size=,visibility=), allow threads: False {{(pid=61806) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 652.451775] env[61806]: DEBUG nova.virt.hardware [None req-fe158576-4234-4a6b-b48d-5ca8bb977c40 tempest-ServerShowV254Test-608938657 tempest-ServerShowV254Test-608938657-project-member] Flavor limits 0:0:0 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 652.451933] env[61806]: DEBUG nova.virt.hardware [None req-fe158576-4234-4a6b-b48d-5ca8bb977c40 tempest-ServerShowV254Test-608938657 tempest-ServerShowV254Test-608938657-project-member] Image limits 0:0:0 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 652.452132] env[61806]: DEBUG nova.virt.hardware [None req-fe158576-4234-4a6b-b48d-5ca8bb977c40 tempest-ServerShowV254Test-608938657 tempest-ServerShowV254Test-608938657-project-member] Flavor pref 0:0:0 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 652.452282] env[61806]: DEBUG nova.virt.hardware [None req-fe158576-4234-4a6b-b48d-5ca8bb977c40 tempest-ServerShowV254Test-608938657 tempest-ServerShowV254Test-608938657-project-member] Image pref 0:0:0 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 652.452431] env[61806]: DEBUG nova.virt.hardware [None req-fe158576-4234-4a6b-b48d-5ca8bb977c40 tempest-ServerShowV254Test-608938657 tempest-ServerShowV254Test-608938657-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 652.452663] env[61806]: DEBUG nova.virt.hardware [None req-fe158576-4234-4a6b-b48d-5ca8bb977c40 tempest-ServerShowV254Test-608938657 tempest-ServerShowV254Test-608938657-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61806) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 652.452825] env[61806]: DEBUG nova.virt.hardware [None req-fe158576-4234-4a6b-b48d-5ca8bb977c40 tempest-ServerShowV254Test-608938657 tempest-ServerShowV254Test-608938657-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61806) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 652.453019] env[61806]: DEBUG nova.virt.hardware [None req-fe158576-4234-4a6b-b48d-5ca8bb977c40 tempest-ServerShowV254Test-608938657 tempest-ServerShowV254Test-608938657-project-member] Got 1 possible topologies {{(pid=61806) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 652.453193] env[61806]: DEBUG nova.virt.hardware [None req-fe158576-4234-4a6b-b48d-5ca8bb977c40 tempest-ServerShowV254Test-608938657 tempest-ServerShowV254Test-608938657-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61806) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 652.453365] env[61806]: DEBUG nova.virt.hardware [None req-fe158576-4234-4a6b-b48d-5ca8bb977c40 tempest-ServerShowV254Test-608938657 tempest-ServerShowV254Test-608938657-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61806) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 652.454881] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aad5b35e-c625-4fc7-904b-7cc9e599e4e9 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 652.462574] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a6944576-b4b6-49b5-8cee-ac2952b90db5 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 652.480178] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-fe158576-4234-4a6b-b48d-5ca8bb977c40 tempest-ServerShowV254Test-608938657 tempest-ServerShowV254Test-608938657-project-member] [instance: 8016bafc-b287-4df1-bd4b-a00210177a86] Instance VIF info [] {{(pid=61806) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 652.486217] env[61806]: DEBUG oslo.service.loopingcall [None req-fe158576-4234-4a6b-b48d-5ca8bb977c40 tempest-ServerShowV254Test-608938657 tempest-ServerShowV254Test-608938657-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61806) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 652.488754] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 8016bafc-b287-4df1-bd4b-a00210177a86] Creating VM on the ESX host {{(pid=61806) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 652.490178] env[61806]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-6520839a-a11b-41cc-8eaa-5e06b77f5aa4 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 652.505940] env[61806]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 652.505940] env[61806]: value = "task-1294468" [ 652.505940] env[61806]: _type = "Task" [ 652.505940] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 652.514345] env[61806]: DEBUG oslo_vmware.api [-] Task: {'id': task-1294468, 'name': CreateVM_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 652.624456] env[61806]: DEBUG oslo_concurrency.lockutils [None req-0b87073d-3de3-486a-876d-8ca5794993c1 tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] Acquiring lock "2a3f81e3-b84f-4370-a598-2f6a607c3d47" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 652.625035] env[61806]: DEBUG oslo_concurrency.lockutils [None req-0b87073d-3de3-486a-876d-8ca5794993c1 tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] Lock "2a3f81e3-b84f-4370-a598-2f6a607c3d47" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 652.712962] env[61806]: DEBUG nova.network.neutron [None req-a93ef3a4-6e83-4dfe-9292-bac21f14d678 tempest-ServerExternalEventsTest-1847900285 tempest-ServerExternalEventsTest-1847900285-project-member] [instance: d5434dd2-206f-4a1e-b97b-bd6f691b6ddf] Successfully created port: 6b28a2b6-fc7b-4db9-84db-18b0d6bbe032 {{(pid=61806) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 652.721527] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6bb3804a-b539-4aa6-bfda-0432fa8ebfe9 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 652.728281] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0704ee49-a2aa-4421-ba7c-6dedfbbe0ce5 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 652.761340] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eb750039-4c8b-4b1f-9af6-aa69e01c4dd2 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 652.769043] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3e42a9ea-3400-4ae7-85f2-2f36f399d297 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 652.784817] env[61806]: DEBUG nova.compute.provider_tree [None req-3e50272a-5153-4e11-beee-d774a917f2df tempest-ServerPasswordTestJSON-1741413915 tempest-ServerPasswordTestJSON-1741413915-project-member] Inventory has not changed in ProviderTree for provider: a2858be1-fd22-4e08-979e-87ad25293407 {{(pid=61806) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 652.786412] env[61806]: DEBUG nova.compute.manager [None req-a93ef3a4-6e83-4dfe-9292-bac21f14d678 tempest-ServerExternalEventsTest-1847900285 tempest-ServerExternalEventsTest-1847900285-project-member] [instance: d5434dd2-206f-4a1e-b97b-bd6f691b6ddf] Start building block device mappings for instance. {{(pid=61806) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 652.905604] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-4803fdfc-9054-49cf-8525-e438672eeedd tempest-ServersAdmin275Test-1945525728 tempest-ServersAdmin275Test-1945525728-project-member] [instance: 9db80c3c-e2c3-4034-a79a-d7be924c4ff6] Preparing fetch location {{(pid=61806) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:633}} [ 652.905868] env[61806]: DEBUG nova.virt.vmwareapi.ds_util [None req-4803fdfc-9054-49cf-8525-e438672eeedd tempest-ServersAdmin275Test-1945525728 tempest-ServersAdmin275Test-1945525728-project-member] Creating directory with path [datastore2] vmware_temp/7890e5b2-17c3-4ad0-b191-91ffb399ba28/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa {{(pid=61806) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 652.906115] env[61806]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-49529a53-091a-4c1b-854b-a2390390fc84 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 652.940023] env[61806]: DEBUG nova.virt.vmwareapi.ds_util [None req-4803fdfc-9054-49cf-8525-e438672eeedd tempest-ServersAdmin275Test-1945525728 tempest-ServersAdmin275Test-1945525728-project-member] Created directory with path [datastore2] vmware_temp/7890e5b2-17c3-4ad0-b191-91ffb399ba28/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa {{(pid=61806) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 652.940023] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-4803fdfc-9054-49cf-8525-e438672eeedd tempest-ServersAdmin275Test-1945525728 tempest-ServersAdmin275Test-1945525728-project-member] [instance: 9db80c3c-e2c3-4034-a79a-d7be924c4ff6] Fetch image to [datastore2] vmware_temp/7890e5b2-17c3-4ad0-b191-91ffb399ba28/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa/tmp-sparse.vmdk {{(pid=61806) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:635}} [ 652.940023] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-4803fdfc-9054-49cf-8525-e438672eeedd tempest-ServersAdmin275Test-1945525728 tempest-ServersAdmin275Test-1945525728-project-member] [instance: 9db80c3c-e2c3-4034-a79a-d7be924c4ff6] Downloading image file data ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa to [datastore2] vmware_temp/7890e5b2-17c3-4ad0-b191-91ffb399ba28/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa/tmp-sparse.vmdk on the data store datastore2 {{(pid=61806) _fetch_image_as_file /opt/stack/nova/nova/virt/vmwareapi/vmops.py:399}} [ 652.940023] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-925f16df-b993-4d97-b6b5-222a0d700aee {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 652.946396] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9eb3ba48-038a-4e90-94af-dcbe72ae823d {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 652.955891] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c3bb0641-cc70-455c-ace0-08e901d059aa {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 652.986794] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2f7b9bba-8661-4510-ab98-b7de981d4d96 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 652.992889] env[61806]: DEBUG oslo_vmware.service [-] Invoking SessionManager.AcquireGenericServiceTicket with opID=oslo.vmware-6ff676f5-28b1-4e56-9805-bf0be5c7b9a3 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 653.015348] env[61806]: DEBUG oslo_vmware.api [-] Task: {'id': task-1294468, 'name': CreateVM_Task, 'duration_secs': 0.310257} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 653.015517] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 8016bafc-b287-4df1-bd4b-a00210177a86] Created VM on the ESX host {{(pid=61806) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 653.015937] env[61806]: DEBUG oslo_concurrency.lockutils [None req-fe158576-4234-4a6b-b48d-5ca8bb977c40 tempest-ServerShowV254Test-608938657 tempest-ServerShowV254Test-608938657-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 653.016142] env[61806]: DEBUG oslo_concurrency.lockutils [None req-fe158576-4234-4a6b-b48d-5ca8bb977c40 tempest-ServerShowV254Test-608938657 tempest-ServerShowV254Test-608938657-project-member] Acquired lock "[datastore2] devstack-image-cache_base/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 653.016465] env[61806]: DEBUG oslo_concurrency.lockutils [None req-fe158576-4234-4a6b-b48d-5ca8bb977c40 tempest-ServerShowV254Test-608938657 tempest-ServerShowV254Test-608938657-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 653.016706] env[61806]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-1f5e0a77-7548-4bbc-8e5f-787a473032aa {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 653.020601] env[61806]: DEBUG nova.virt.vmwareapi.images [None req-4803fdfc-9054-49cf-8525-e438672eeedd tempest-ServersAdmin275Test-1945525728 tempest-ServersAdmin275Test-1945525728-project-member] [instance: 9db80c3c-e2c3-4034-a79a-d7be924c4ff6] Downloading image file data ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa to the data store datastore2 {{(pid=61806) fetch_image /opt/stack/nova/nova/virt/vmwareapi/images.py:245}} [ 653.024041] env[61806]: DEBUG oslo_vmware.api [None req-fe158576-4234-4a6b-b48d-5ca8bb977c40 tempest-ServerShowV254Test-608938657 tempest-ServerShowV254Test-608938657-project-member] Waiting for the task: (returnval){ [ 653.024041] env[61806]: value = "session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]5211313b-fdb2-25cf-f279-022f4c2be706" [ 653.024041] env[61806]: _type = "Task" [ 653.024041] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 653.034887] env[61806]: DEBUG oslo_vmware.api [None req-fe158576-4234-4a6b-b48d-5ca8bb977c40 tempest-ServerShowV254Test-608938657 tempest-ServerShowV254Test-608938657-project-member] Task: {'id': session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]5211313b-fdb2-25cf-f279-022f4c2be706, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 653.084097] env[61806]: DEBUG oslo_vmware.rw_handles [None req-4803fdfc-9054-49cf-8525-e438672eeedd tempest-ServersAdmin275Test-1945525728 tempest-ServersAdmin275Test-1945525728-project-member] Creating HTTP connection to write to file with size = 21318656 and URL = https://esx7c1n3.openstack.eu-de-1.cloud.sap:443/folder/vmware_temp/7890e5b2-17c3-4ad0-b191-91ffb399ba28/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa/tmp-sparse.vmdk?dcPath=ha-datacenter&dsName=datastore2. {{(pid=61806) _create_write_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:122}} [ 653.297752] env[61806]: DEBUG nova.scheduler.client.report [None req-3e50272a-5153-4e11-beee-d774a917f2df tempest-ServerPasswordTestJSON-1741413915 tempest-ServerPasswordTestJSON-1741413915-project-member] Inventory has not changed for provider a2858be1-fd22-4e08-979e-87ad25293407 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 139, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61806) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 653.535562] env[61806]: DEBUG oslo_concurrency.lockutils [None req-fe158576-4234-4a6b-b48d-5ca8bb977c40 tempest-ServerShowV254Test-608938657 tempest-ServerShowV254Test-608938657-project-member] Releasing lock "[datastore2] devstack-image-cache_base/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 653.536268] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-fe158576-4234-4a6b-b48d-5ca8bb977c40 tempest-ServerShowV254Test-608938657 tempest-ServerShowV254Test-608938657-project-member] [instance: 8016bafc-b287-4df1-bd4b-a00210177a86] Processing image ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa {{(pid=61806) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 653.536768] env[61806]: DEBUG oslo_concurrency.lockutils [None req-fe158576-4234-4a6b-b48d-5ca8bb977c40 tempest-ServerShowV254Test-608938657 tempest-ServerShowV254Test-608938657-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa.vmdk" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 653.589326] env[61806]: DEBUG nova.compute.manager [req-ae408ac2-8895-4a7d-a83d-e973c892c14c req-1c1cd9a1-48df-40a2-9c05-8f95fa9eca04 service nova] [instance: d5434dd2-206f-4a1e-b97b-bd6f691b6ddf] Received event network-changed-6b28a2b6-fc7b-4db9-84db-18b0d6bbe032 {{(pid=61806) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 653.589326] env[61806]: DEBUG nova.compute.manager [req-ae408ac2-8895-4a7d-a83d-e973c892c14c req-1c1cd9a1-48df-40a2-9c05-8f95fa9eca04 service nova] [instance: d5434dd2-206f-4a1e-b97b-bd6f691b6ddf] Refreshing instance network info cache due to event network-changed-6b28a2b6-fc7b-4db9-84db-18b0d6bbe032. {{(pid=61806) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 653.589326] env[61806]: DEBUG oslo_concurrency.lockutils [req-ae408ac2-8895-4a7d-a83d-e973c892c14c req-1c1cd9a1-48df-40a2-9c05-8f95fa9eca04 service nova] Acquiring lock "refresh_cache-d5434dd2-206f-4a1e-b97b-bd6f691b6ddf" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 653.589326] env[61806]: DEBUG oslo_concurrency.lockutils [req-ae408ac2-8895-4a7d-a83d-e973c892c14c req-1c1cd9a1-48df-40a2-9c05-8f95fa9eca04 service nova] Acquired lock "refresh_cache-d5434dd2-206f-4a1e-b97b-bd6f691b6ddf" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 653.589326] env[61806]: DEBUG nova.network.neutron [req-ae408ac2-8895-4a7d-a83d-e973c892c14c req-1c1cd9a1-48df-40a2-9c05-8f95fa9eca04 service nova] [instance: d5434dd2-206f-4a1e-b97b-bd6f691b6ddf] Refreshing network info cache for port 6b28a2b6-fc7b-4db9-84db-18b0d6bbe032 {{(pid=61806) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 653.809834] env[61806]: ERROR nova.compute.manager [None req-a93ef3a4-6e83-4dfe-9292-bac21f14d678 tempest-ServerExternalEventsTest-1847900285 tempest-ServerExternalEventsTest-1847900285-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 6b28a2b6-fc7b-4db9-84db-18b0d6bbe032, please check neutron logs for more information. [ 653.809834] env[61806]: ERROR nova.compute.manager Traceback (most recent call last): [ 653.809834] env[61806]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 653.809834] env[61806]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 653.809834] env[61806]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 653.809834] env[61806]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 653.809834] env[61806]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 653.809834] env[61806]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 653.809834] env[61806]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 653.809834] env[61806]: ERROR nova.compute.manager self.force_reraise() [ 653.809834] env[61806]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 653.809834] env[61806]: ERROR nova.compute.manager raise self.value [ 653.809834] env[61806]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 653.809834] env[61806]: ERROR nova.compute.manager updated_port = self._update_port( [ 653.809834] env[61806]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 653.809834] env[61806]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 653.810727] env[61806]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 653.810727] env[61806]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 653.810727] env[61806]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 6b28a2b6-fc7b-4db9-84db-18b0d6bbe032, please check neutron logs for more information. [ 653.810727] env[61806]: ERROR nova.compute.manager [ 653.810727] env[61806]: Traceback (most recent call last): [ 653.810727] env[61806]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 653.810727] env[61806]: listener.cb(fileno) [ 653.810727] env[61806]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 653.810727] env[61806]: result = function(*args, **kwargs) [ 653.810727] env[61806]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 653.810727] env[61806]: return func(*args, **kwargs) [ 653.810727] env[61806]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 653.810727] env[61806]: raise e [ 653.810727] env[61806]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 653.810727] env[61806]: nwinfo = self.network_api.allocate_for_instance( [ 653.810727] env[61806]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 653.810727] env[61806]: created_port_ids = self._update_ports_for_instance( [ 653.810727] env[61806]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 653.810727] env[61806]: with excutils.save_and_reraise_exception(): [ 653.810727] env[61806]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 653.810727] env[61806]: self.force_reraise() [ 653.810727] env[61806]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 653.810727] env[61806]: raise self.value [ 653.810727] env[61806]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 653.810727] env[61806]: updated_port = self._update_port( [ 653.810727] env[61806]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 653.810727] env[61806]: _ensure_no_port_binding_failure(port) [ 653.810727] env[61806]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 653.810727] env[61806]: raise exception.PortBindingFailed(port_id=port['id']) [ 653.812038] env[61806]: nova.exception.PortBindingFailed: Binding failed for port 6b28a2b6-fc7b-4db9-84db-18b0d6bbe032, please check neutron logs for more information. [ 653.812038] env[61806]: Removing descriptor: 14 [ 653.812038] env[61806]: DEBUG oslo_concurrency.lockutils [None req-3e50272a-5153-4e11-beee-d774a917f2df tempest-ServerPasswordTestJSON-1741413915 tempest-ServerPasswordTestJSON-1741413915-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.035s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 653.812038] env[61806]: ERROR nova.compute.manager [None req-3e50272a-5153-4e11-beee-d774a917f2df tempest-ServerPasswordTestJSON-1741413915 tempest-ServerPasswordTestJSON-1741413915-project-member] [instance: 0d757385-ff1c-49c9-825f-f9d542ae074f] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port fc109b8a-4096-4169-bf3b-f19ce33059d3, please check neutron logs for more information. [ 653.812038] env[61806]: ERROR nova.compute.manager [instance: 0d757385-ff1c-49c9-825f-f9d542ae074f] Traceback (most recent call last): [ 653.812038] env[61806]: ERROR nova.compute.manager [instance: 0d757385-ff1c-49c9-825f-f9d542ae074f] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 653.812038] env[61806]: ERROR nova.compute.manager [instance: 0d757385-ff1c-49c9-825f-f9d542ae074f] self.driver.spawn(context, instance, image_meta, [ 653.812038] env[61806]: ERROR nova.compute.manager [instance: 0d757385-ff1c-49c9-825f-f9d542ae074f] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 653.812038] env[61806]: ERROR nova.compute.manager [instance: 0d757385-ff1c-49c9-825f-f9d542ae074f] self._vmops.spawn(context, instance, image_meta, injected_files, [ 653.812038] env[61806]: ERROR nova.compute.manager [instance: 0d757385-ff1c-49c9-825f-f9d542ae074f] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 653.812636] env[61806]: ERROR nova.compute.manager [instance: 0d757385-ff1c-49c9-825f-f9d542ae074f] vm_ref = self.build_virtual_machine(instance, [ 653.812636] env[61806]: ERROR nova.compute.manager [instance: 0d757385-ff1c-49c9-825f-f9d542ae074f] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 653.812636] env[61806]: ERROR nova.compute.manager [instance: 0d757385-ff1c-49c9-825f-f9d542ae074f] vif_infos = vmwarevif.get_vif_info(self._session, [ 653.812636] env[61806]: ERROR nova.compute.manager [instance: 0d757385-ff1c-49c9-825f-f9d542ae074f] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 653.812636] env[61806]: ERROR nova.compute.manager [instance: 0d757385-ff1c-49c9-825f-f9d542ae074f] for vif in network_info: [ 653.812636] env[61806]: ERROR nova.compute.manager [instance: 0d757385-ff1c-49c9-825f-f9d542ae074f] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 653.812636] env[61806]: ERROR nova.compute.manager [instance: 0d757385-ff1c-49c9-825f-f9d542ae074f] return self._sync_wrapper(fn, *args, **kwargs) [ 653.812636] env[61806]: ERROR nova.compute.manager [instance: 0d757385-ff1c-49c9-825f-f9d542ae074f] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 653.812636] env[61806]: ERROR nova.compute.manager [instance: 0d757385-ff1c-49c9-825f-f9d542ae074f] self.wait() [ 653.812636] env[61806]: ERROR nova.compute.manager [instance: 0d757385-ff1c-49c9-825f-f9d542ae074f] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 653.812636] env[61806]: ERROR nova.compute.manager [instance: 0d757385-ff1c-49c9-825f-f9d542ae074f] self[:] = self._gt.wait() [ 653.812636] env[61806]: ERROR nova.compute.manager [instance: 0d757385-ff1c-49c9-825f-f9d542ae074f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 653.812636] env[61806]: ERROR nova.compute.manager [instance: 0d757385-ff1c-49c9-825f-f9d542ae074f] return self._exit_event.wait() [ 653.813291] env[61806]: ERROR nova.compute.manager [instance: 0d757385-ff1c-49c9-825f-f9d542ae074f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 653.813291] env[61806]: ERROR nova.compute.manager [instance: 0d757385-ff1c-49c9-825f-f9d542ae074f] result = hub.switch() [ 653.813291] env[61806]: ERROR nova.compute.manager [instance: 0d757385-ff1c-49c9-825f-f9d542ae074f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 653.813291] env[61806]: ERROR nova.compute.manager [instance: 0d757385-ff1c-49c9-825f-f9d542ae074f] return self.greenlet.switch() [ 653.813291] env[61806]: ERROR nova.compute.manager [instance: 0d757385-ff1c-49c9-825f-f9d542ae074f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 653.813291] env[61806]: ERROR nova.compute.manager [instance: 0d757385-ff1c-49c9-825f-f9d542ae074f] result = function(*args, **kwargs) [ 653.813291] env[61806]: ERROR nova.compute.manager [instance: 0d757385-ff1c-49c9-825f-f9d542ae074f] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 653.813291] env[61806]: ERROR nova.compute.manager [instance: 0d757385-ff1c-49c9-825f-f9d542ae074f] return func(*args, **kwargs) [ 653.813291] env[61806]: ERROR nova.compute.manager [instance: 0d757385-ff1c-49c9-825f-f9d542ae074f] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 653.813291] env[61806]: ERROR nova.compute.manager [instance: 0d757385-ff1c-49c9-825f-f9d542ae074f] raise e [ 653.813291] env[61806]: ERROR nova.compute.manager [instance: 0d757385-ff1c-49c9-825f-f9d542ae074f] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 653.813291] env[61806]: ERROR nova.compute.manager [instance: 0d757385-ff1c-49c9-825f-f9d542ae074f] nwinfo = self.network_api.allocate_for_instance( [ 653.813291] env[61806]: ERROR nova.compute.manager [instance: 0d757385-ff1c-49c9-825f-f9d542ae074f] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 653.814192] env[61806]: ERROR nova.compute.manager [instance: 0d757385-ff1c-49c9-825f-f9d542ae074f] created_port_ids = self._update_ports_for_instance( [ 653.814192] env[61806]: ERROR nova.compute.manager [instance: 0d757385-ff1c-49c9-825f-f9d542ae074f] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 653.814192] env[61806]: ERROR nova.compute.manager [instance: 0d757385-ff1c-49c9-825f-f9d542ae074f] with excutils.save_and_reraise_exception(): [ 653.814192] env[61806]: ERROR nova.compute.manager [instance: 0d757385-ff1c-49c9-825f-f9d542ae074f] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 653.814192] env[61806]: ERROR nova.compute.manager [instance: 0d757385-ff1c-49c9-825f-f9d542ae074f] self.force_reraise() [ 653.814192] env[61806]: ERROR nova.compute.manager [instance: 0d757385-ff1c-49c9-825f-f9d542ae074f] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 653.814192] env[61806]: ERROR nova.compute.manager [instance: 0d757385-ff1c-49c9-825f-f9d542ae074f] raise self.value [ 653.814192] env[61806]: ERROR nova.compute.manager [instance: 0d757385-ff1c-49c9-825f-f9d542ae074f] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 653.814192] env[61806]: ERROR nova.compute.manager [instance: 0d757385-ff1c-49c9-825f-f9d542ae074f] updated_port = self._update_port( [ 653.814192] env[61806]: ERROR nova.compute.manager [instance: 0d757385-ff1c-49c9-825f-f9d542ae074f] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 653.814192] env[61806]: ERROR nova.compute.manager [instance: 0d757385-ff1c-49c9-825f-f9d542ae074f] _ensure_no_port_binding_failure(port) [ 653.814192] env[61806]: ERROR nova.compute.manager [instance: 0d757385-ff1c-49c9-825f-f9d542ae074f] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 653.814192] env[61806]: ERROR nova.compute.manager [instance: 0d757385-ff1c-49c9-825f-f9d542ae074f] raise exception.PortBindingFailed(port_id=port['id']) [ 653.814925] env[61806]: ERROR nova.compute.manager [instance: 0d757385-ff1c-49c9-825f-f9d542ae074f] nova.exception.PortBindingFailed: Binding failed for port fc109b8a-4096-4169-bf3b-f19ce33059d3, please check neutron logs for more information. [ 653.814925] env[61806]: ERROR nova.compute.manager [instance: 0d757385-ff1c-49c9-825f-f9d542ae074f] [ 653.814925] env[61806]: DEBUG nova.compute.utils [None req-3e50272a-5153-4e11-beee-d774a917f2df tempest-ServerPasswordTestJSON-1741413915 tempest-ServerPasswordTestJSON-1741413915-project-member] [instance: 0d757385-ff1c-49c9-825f-f9d542ae074f] Binding failed for port fc109b8a-4096-4169-bf3b-f19ce33059d3, please check neutron logs for more information. {{(pid=61806) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 653.814925] env[61806]: DEBUG nova.compute.manager [None req-a93ef3a4-6e83-4dfe-9292-bac21f14d678 tempest-ServerExternalEventsTest-1847900285 tempest-ServerExternalEventsTest-1847900285-project-member] [instance: d5434dd2-206f-4a1e-b97b-bd6f691b6ddf] Start spawning the instance on the hypervisor. {{(pid=61806) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 653.819919] env[61806]: DEBUG nova.compute.manager [None req-3e50272a-5153-4e11-beee-d774a917f2df tempest-ServerPasswordTestJSON-1741413915 tempest-ServerPasswordTestJSON-1741413915-project-member] [instance: 0d757385-ff1c-49c9-825f-f9d542ae074f] Build of instance 0d757385-ff1c-49c9-825f-f9d542ae074f was re-scheduled: Binding failed for port fc109b8a-4096-4169-bf3b-f19ce33059d3, please check neutron logs for more information. {{(pid=61806) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 653.822842] env[61806]: DEBUG nova.compute.manager [None req-3e50272a-5153-4e11-beee-d774a917f2df tempest-ServerPasswordTestJSON-1741413915 tempest-ServerPasswordTestJSON-1741413915-project-member] [instance: 0d757385-ff1c-49c9-825f-f9d542ae074f] Unplugging VIFs for instance {{(pid=61806) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 653.822842] env[61806]: DEBUG oslo_concurrency.lockutils [None req-3e50272a-5153-4e11-beee-d774a917f2df tempest-ServerPasswordTestJSON-1741413915 tempest-ServerPasswordTestJSON-1741413915-project-member] Acquiring lock "refresh_cache-0d757385-ff1c-49c9-825f-f9d542ae074f" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 653.822842] env[61806]: DEBUG oslo_concurrency.lockutils [None req-3e50272a-5153-4e11-beee-d774a917f2df tempest-ServerPasswordTestJSON-1741413915 tempest-ServerPasswordTestJSON-1741413915-project-member] Acquired lock "refresh_cache-0d757385-ff1c-49c9-825f-f9d542ae074f" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 653.822842] env[61806]: DEBUG nova.network.neutron [None req-3e50272a-5153-4e11-beee-d774a917f2df tempest-ServerPasswordTestJSON-1741413915 tempest-ServerPasswordTestJSON-1741413915-project-member] [instance: 0d757385-ff1c-49c9-825f-f9d542ae074f] Building network info cache for instance {{(pid=61806) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 653.822842] env[61806]: DEBUG oslo_concurrency.lockutils [None req-f8b2f38e-90f5-4b03-89bc-ca1926adb40c tempest-FloatingIPsAssociationTestJSON-35718468 tempest-FloatingIPsAssociationTestJSON-35718468-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 15.796s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 653.829214] env[61806]: DEBUG oslo_vmware.rw_handles [None req-4803fdfc-9054-49cf-8525-e438672eeedd tempest-ServersAdmin275Test-1945525728 tempest-ServersAdmin275Test-1945525728-project-member] Completed reading data from the image iterator. {{(pid=61806) read /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:765}} [ 653.831274] env[61806]: DEBUG oslo_vmware.rw_handles [None req-4803fdfc-9054-49cf-8525-e438672eeedd tempest-ServersAdmin275Test-1945525728 tempest-ServersAdmin275Test-1945525728-project-member] Closing write handle for https://esx7c1n3.openstack.eu-de-1.cloud.sap:443/folder/vmware_temp/7890e5b2-17c3-4ad0-b191-91ffb399ba28/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa/tmp-sparse.vmdk?dcPath=ha-datacenter&dsName=datastore2. {{(pid=61806) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:281}} [ 653.859071] env[61806]: DEBUG nova.virt.hardware [None req-a93ef3a4-6e83-4dfe-9292-bac21f14d678 tempest-ServerExternalEventsTest-1847900285 tempest-ServerExternalEventsTest-1847900285-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-15T18:05:10Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-15T18:04:52Z,direct_url=,disk_format='vmdk',id=ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='84ba83fce2c349a6988173c5d6fc3b07',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-15T18:04:53Z,virtual_size=,visibility=), allow threads: False {{(pid=61806) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 653.859337] env[61806]: DEBUG nova.virt.hardware [None req-a93ef3a4-6e83-4dfe-9292-bac21f14d678 tempest-ServerExternalEventsTest-1847900285 tempest-ServerExternalEventsTest-1847900285-project-member] Flavor limits 0:0:0 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 653.859496] env[61806]: DEBUG nova.virt.hardware [None req-a93ef3a4-6e83-4dfe-9292-bac21f14d678 tempest-ServerExternalEventsTest-1847900285 tempest-ServerExternalEventsTest-1847900285-project-member] Image limits 0:0:0 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 653.859683] env[61806]: DEBUG nova.virt.hardware [None req-a93ef3a4-6e83-4dfe-9292-bac21f14d678 tempest-ServerExternalEventsTest-1847900285 tempest-ServerExternalEventsTest-1847900285-project-member] Flavor pref 0:0:0 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 653.859831] env[61806]: DEBUG nova.virt.hardware [None req-a93ef3a4-6e83-4dfe-9292-bac21f14d678 tempest-ServerExternalEventsTest-1847900285 tempest-ServerExternalEventsTest-1847900285-project-member] Image pref 0:0:0 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 653.859976] env[61806]: DEBUG nova.virt.hardware [None req-a93ef3a4-6e83-4dfe-9292-bac21f14d678 tempest-ServerExternalEventsTest-1847900285 tempest-ServerExternalEventsTest-1847900285-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 653.860201] env[61806]: DEBUG nova.virt.hardware [None req-a93ef3a4-6e83-4dfe-9292-bac21f14d678 tempest-ServerExternalEventsTest-1847900285 tempest-ServerExternalEventsTest-1847900285-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61806) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 653.860358] env[61806]: DEBUG nova.virt.hardware [None req-a93ef3a4-6e83-4dfe-9292-bac21f14d678 tempest-ServerExternalEventsTest-1847900285 tempest-ServerExternalEventsTest-1847900285-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61806) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 653.860557] env[61806]: DEBUG nova.virt.hardware [None req-a93ef3a4-6e83-4dfe-9292-bac21f14d678 tempest-ServerExternalEventsTest-1847900285 tempest-ServerExternalEventsTest-1847900285-project-member] Got 1 possible topologies {{(pid=61806) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 653.860727] env[61806]: DEBUG nova.virt.hardware [None req-a93ef3a4-6e83-4dfe-9292-bac21f14d678 tempest-ServerExternalEventsTest-1847900285 tempest-ServerExternalEventsTest-1847900285-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61806) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 653.860902] env[61806]: DEBUG nova.virt.hardware [None req-a93ef3a4-6e83-4dfe-9292-bac21f14d678 tempest-ServerExternalEventsTest-1847900285 tempest-ServerExternalEventsTest-1847900285-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61806) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 653.861807] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b63b63de-2279-4aaa-a5bd-f284f0eb2c8d {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 653.875422] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e498fa4d-2cdc-43fb-a040-6e49c1219870 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 653.891287] env[61806]: ERROR nova.compute.manager [None req-a93ef3a4-6e83-4dfe-9292-bac21f14d678 tempest-ServerExternalEventsTest-1847900285 tempest-ServerExternalEventsTest-1847900285-project-member] [instance: d5434dd2-206f-4a1e-b97b-bd6f691b6ddf] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 6b28a2b6-fc7b-4db9-84db-18b0d6bbe032, please check neutron logs for more information. [ 653.891287] env[61806]: ERROR nova.compute.manager [instance: d5434dd2-206f-4a1e-b97b-bd6f691b6ddf] Traceback (most recent call last): [ 653.891287] env[61806]: ERROR nova.compute.manager [instance: d5434dd2-206f-4a1e-b97b-bd6f691b6ddf] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 653.891287] env[61806]: ERROR nova.compute.manager [instance: d5434dd2-206f-4a1e-b97b-bd6f691b6ddf] yield resources [ 653.891287] env[61806]: ERROR nova.compute.manager [instance: d5434dd2-206f-4a1e-b97b-bd6f691b6ddf] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 653.891287] env[61806]: ERROR nova.compute.manager [instance: d5434dd2-206f-4a1e-b97b-bd6f691b6ddf] self.driver.spawn(context, instance, image_meta, [ 653.891287] env[61806]: ERROR nova.compute.manager [instance: d5434dd2-206f-4a1e-b97b-bd6f691b6ddf] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 653.891287] env[61806]: ERROR nova.compute.manager [instance: d5434dd2-206f-4a1e-b97b-bd6f691b6ddf] self._vmops.spawn(context, instance, image_meta, injected_files, [ 653.891287] env[61806]: ERROR nova.compute.manager [instance: d5434dd2-206f-4a1e-b97b-bd6f691b6ddf] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 653.891287] env[61806]: ERROR nova.compute.manager [instance: d5434dd2-206f-4a1e-b97b-bd6f691b6ddf] vm_ref = self.build_virtual_machine(instance, [ 653.891287] env[61806]: ERROR nova.compute.manager [instance: d5434dd2-206f-4a1e-b97b-bd6f691b6ddf] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 653.891659] env[61806]: ERROR nova.compute.manager [instance: d5434dd2-206f-4a1e-b97b-bd6f691b6ddf] vif_infos = vmwarevif.get_vif_info(self._session, [ 653.891659] env[61806]: ERROR nova.compute.manager [instance: d5434dd2-206f-4a1e-b97b-bd6f691b6ddf] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 653.891659] env[61806]: ERROR nova.compute.manager [instance: d5434dd2-206f-4a1e-b97b-bd6f691b6ddf] for vif in network_info: [ 653.891659] env[61806]: ERROR nova.compute.manager [instance: d5434dd2-206f-4a1e-b97b-bd6f691b6ddf] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 653.891659] env[61806]: ERROR nova.compute.manager [instance: d5434dd2-206f-4a1e-b97b-bd6f691b6ddf] return self._sync_wrapper(fn, *args, **kwargs) [ 653.891659] env[61806]: ERROR nova.compute.manager [instance: d5434dd2-206f-4a1e-b97b-bd6f691b6ddf] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 653.891659] env[61806]: ERROR nova.compute.manager [instance: d5434dd2-206f-4a1e-b97b-bd6f691b6ddf] self.wait() [ 653.891659] env[61806]: ERROR nova.compute.manager [instance: d5434dd2-206f-4a1e-b97b-bd6f691b6ddf] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 653.891659] env[61806]: ERROR nova.compute.manager [instance: d5434dd2-206f-4a1e-b97b-bd6f691b6ddf] self[:] = self._gt.wait() [ 653.891659] env[61806]: ERROR nova.compute.manager [instance: d5434dd2-206f-4a1e-b97b-bd6f691b6ddf] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 653.891659] env[61806]: ERROR nova.compute.manager [instance: d5434dd2-206f-4a1e-b97b-bd6f691b6ddf] return self._exit_event.wait() [ 653.891659] env[61806]: ERROR nova.compute.manager [instance: d5434dd2-206f-4a1e-b97b-bd6f691b6ddf] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 653.891659] env[61806]: ERROR nova.compute.manager [instance: d5434dd2-206f-4a1e-b97b-bd6f691b6ddf] current.throw(*self._exc) [ 653.891971] env[61806]: ERROR nova.compute.manager [instance: d5434dd2-206f-4a1e-b97b-bd6f691b6ddf] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 653.891971] env[61806]: ERROR nova.compute.manager [instance: d5434dd2-206f-4a1e-b97b-bd6f691b6ddf] result = function(*args, **kwargs) [ 653.891971] env[61806]: ERROR nova.compute.manager [instance: d5434dd2-206f-4a1e-b97b-bd6f691b6ddf] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 653.891971] env[61806]: ERROR nova.compute.manager [instance: d5434dd2-206f-4a1e-b97b-bd6f691b6ddf] return func(*args, **kwargs) [ 653.891971] env[61806]: ERROR nova.compute.manager [instance: d5434dd2-206f-4a1e-b97b-bd6f691b6ddf] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 653.891971] env[61806]: ERROR nova.compute.manager [instance: d5434dd2-206f-4a1e-b97b-bd6f691b6ddf] raise e [ 653.891971] env[61806]: ERROR nova.compute.manager [instance: d5434dd2-206f-4a1e-b97b-bd6f691b6ddf] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 653.891971] env[61806]: ERROR nova.compute.manager [instance: d5434dd2-206f-4a1e-b97b-bd6f691b6ddf] nwinfo = self.network_api.allocate_for_instance( [ 653.891971] env[61806]: ERROR nova.compute.manager [instance: d5434dd2-206f-4a1e-b97b-bd6f691b6ddf] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 653.891971] env[61806]: ERROR nova.compute.manager [instance: d5434dd2-206f-4a1e-b97b-bd6f691b6ddf] created_port_ids = self._update_ports_for_instance( [ 653.891971] env[61806]: ERROR nova.compute.manager [instance: d5434dd2-206f-4a1e-b97b-bd6f691b6ddf] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 653.891971] env[61806]: ERROR nova.compute.manager [instance: d5434dd2-206f-4a1e-b97b-bd6f691b6ddf] with excutils.save_and_reraise_exception(): [ 653.891971] env[61806]: ERROR nova.compute.manager [instance: d5434dd2-206f-4a1e-b97b-bd6f691b6ddf] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 653.892299] env[61806]: ERROR nova.compute.manager [instance: d5434dd2-206f-4a1e-b97b-bd6f691b6ddf] self.force_reraise() [ 653.892299] env[61806]: ERROR nova.compute.manager [instance: d5434dd2-206f-4a1e-b97b-bd6f691b6ddf] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 653.892299] env[61806]: ERROR nova.compute.manager [instance: d5434dd2-206f-4a1e-b97b-bd6f691b6ddf] raise self.value [ 653.892299] env[61806]: ERROR nova.compute.manager [instance: d5434dd2-206f-4a1e-b97b-bd6f691b6ddf] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 653.892299] env[61806]: ERROR nova.compute.manager [instance: d5434dd2-206f-4a1e-b97b-bd6f691b6ddf] updated_port = self._update_port( [ 653.892299] env[61806]: ERROR nova.compute.manager [instance: d5434dd2-206f-4a1e-b97b-bd6f691b6ddf] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 653.892299] env[61806]: ERROR nova.compute.manager [instance: d5434dd2-206f-4a1e-b97b-bd6f691b6ddf] _ensure_no_port_binding_failure(port) [ 653.892299] env[61806]: ERROR nova.compute.manager [instance: d5434dd2-206f-4a1e-b97b-bd6f691b6ddf] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 653.892299] env[61806]: ERROR nova.compute.manager [instance: d5434dd2-206f-4a1e-b97b-bd6f691b6ddf] raise exception.PortBindingFailed(port_id=port['id']) [ 653.892299] env[61806]: ERROR nova.compute.manager [instance: d5434dd2-206f-4a1e-b97b-bd6f691b6ddf] nova.exception.PortBindingFailed: Binding failed for port 6b28a2b6-fc7b-4db9-84db-18b0d6bbe032, please check neutron logs for more information. [ 653.892299] env[61806]: ERROR nova.compute.manager [instance: d5434dd2-206f-4a1e-b97b-bd6f691b6ddf] [ 653.892299] env[61806]: INFO nova.compute.manager [None req-a93ef3a4-6e83-4dfe-9292-bac21f14d678 tempest-ServerExternalEventsTest-1847900285 tempest-ServerExternalEventsTest-1847900285-project-member] [instance: d5434dd2-206f-4a1e-b97b-bd6f691b6ddf] Terminating instance [ 653.893612] env[61806]: DEBUG nova.virt.vmwareapi.images [None req-4803fdfc-9054-49cf-8525-e438672eeedd tempest-ServersAdmin275Test-1945525728 tempest-ServersAdmin275Test-1945525728-project-member] [instance: 9db80c3c-e2c3-4034-a79a-d7be924c4ff6] Downloaded image file data ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa to vmware_temp/7890e5b2-17c3-4ad0-b191-91ffb399ba28/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa/tmp-sparse.vmdk on the data store datastore2 {{(pid=61806) fetch_image /opt/stack/nova/nova/virt/vmwareapi/images.py:258}} [ 653.895480] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-4803fdfc-9054-49cf-8525-e438672eeedd tempest-ServersAdmin275Test-1945525728 tempest-ServersAdmin275Test-1945525728-project-member] [instance: 9db80c3c-e2c3-4034-a79a-d7be924c4ff6] Caching image {{(pid=61806) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:638}} [ 653.895714] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [None req-4803fdfc-9054-49cf-8525-e438672eeedd tempest-ServersAdmin275Test-1945525728 tempest-ServersAdmin275Test-1945525728-project-member] Copying Virtual Disk [datastore2] vmware_temp/7890e5b2-17c3-4ad0-b191-91ffb399ba28/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa/tmp-sparse.vmdk to [datastore2] vmware_temp/7890e5b2-17c3-4ad0-b191-91ffb399ba28/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa.vmdk {{(pid=61806) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 653.896226] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-81d52e2c-4183-4d2b-8429-01f7e459ef4d {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 653.898510] env[61806]: DEBUG oslo_concurrency.lockutils [None req-a93ef3a4-6e83-4dfe-9292-bac21f14d678 tempest-ServerExternalEventsTest-1847900285 tempest-ServerExternalEventsTest-1847900285-project-member] Acquiring lock "refresh_cache-d5434dd2-206f-4a1e-b97b-bd6f691b6ddf" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 653.903028] env[61806]: DEBUG oslo_vmware.api [None req-4803fdfc-9054-49cf-8525-e438672eeedd tempest-ServersAdmin275Test-1945525728 tempest-ServersAdmin275Test-1945525728-project-member] Waiting for the task: (returnval){ [ 653.903028] env[61806]: value = "task-1294469" [ 653.903028] env[61806]: _type = "Task" [ 653.903028] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 653.911323] env[61806]: DEBUG oslo_vmware.api [None req-4803fdfc-9054-49cf-8525-e438672eeedd tempest-ServersAdmin275Test-1945525728 tempest-ServersAdmin275Test-1945525728-project-member] Task: {'id': task-1294469, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 654.114387] env[61806]: DEBUG nova.network.neutron [req-ae408ac2-8895-4a7d-a83d-e973c892c14c req-1c1cd9a1-48df-40a2-9c05-8f95fa9eca04 service nova] [instance: d5434dd2-206f-4a1e-b97b-bd6f691b6ddf] Instance cache missing network info. {{(pid=61806) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 654.180101] env[61806]: DEBUG nova.network.neutron [req-ae408ac2-8895-4a7d-a83d-e973c892c14c req-1c1cd9a1-48df-40a2-9c05-8f95fa9eca04 service nova] [instance: d5434dd2-206f-4a1e-b97b-bd6f691b6ddf] Updating instance_info_cache with network_info: [] {{(pid=61806) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 654.350704] env[61806]: DEBUG nova.network.neutron [None req-3e50272a-5153-4e11-beee-d774a917f2df tempest-ServerPasswordTestJSON-1741413915 tempest-ServerPasswordTestJSON-1741413915-project-member] [instance: 0d757385-ff1c-49c9-825f-f9d542ae074f] Instance cache missing network info. {{(pid=61806) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 654.416435] env[61806]: DEBUG oslo_vmware.api [None req-4803fdfc-9054-49cf-8525-e438672eeedd tempest-ServersAdmin275Test-1945525728 tempest-ServersAdmin275Test-1945525728-project-member] Task: {'id': task-1294469, 'name': CopyVirtualDisk_Task} progress is 100%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 654.479034] env[61806]: DEBUG nova.network.neutron [None req-3e50272a-5153-4e11-beee-d774a917f2df tempest-ServerPasswordTestJSON-1741413915 tempest-ServerPasswordTestJSON-1741413915-project-member] [instance: 0d757385-ff1c-49c9-825f-f9d542ae074f] Updating instance_info_cache with network_info: [] {{(pid=61806) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 654.684597] env[61806]: DEBUG oslo_concurrency.lockutils [req-ae408ac2-8895-4a7d-a83d-e973c892c14c req-1c1cd9a1-48df-40a2-9c05-8f95fa9eca04 service nova] Releasing lock "refresh_cache-d5434dd2-206f-4a1e-b97b-bd6f691b6ddf" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 654.684597] env[61806]: DEBUG oslo_concurrency.lockutils [None req-a93ef3a4-6e83-4dfe-9292-bac21f14d678 tempest-ServerExternalEventsTest-1847900285 tempest-ServerExternalEventsTest-1847900285-project-member] Acquired lock "refresh_cache-d5434dd2-206f-4a1e-b97b-bd6f691b6ddf" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 654.684713] env[61806]: DEBUG nova.network.neutron [None req-a93ef3a4-6e83-4dfe-9292-bac21f14d678 tempest-ServerExternalEventsTest-1847900285 tempest-ServerExternalEventsTest-1847900285-project-member] [instance: d5434dd2-206f-4a1e-b97b-bd6f691b6ddf] Building network info cache for instance {{(pid=61806) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 654.779232] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1ad1c292-654c-40c8-af0a-6d096c9fc85b {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 654.786894] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ebab0e20-f53f-49c2-bae9-6ca7eccdd8fd {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 654.819905] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-60ab5a9a-d164-4b8d-aa5a-cb8810fabf06 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 654.828627] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b333feda-7236-4f3f-a32f-719ce04bf7a6 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 654.842117] env[61806]: DEBUG nova.compute.provider_tree [None req-f8b2f38e-90f5-4b03-89bc-ca1926adb40c tempest-FloatingIPsAssociationTestJSON-35718468 tempest-FloatingIPsAssociationTestJSON-35718468-project-member] Updating inventory in ProviderTree for provider a2858be1-fd22-4e08-979e-87ad25293407 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 138, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61806) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 654.870867] env[61806]: DEBUG oslo_concurrency.lockutils [None req-b3f33a6f-532a-458d-aec6-6ec54b0cadd5 tempest-ServerMetadataTestJSON-316746592 tempest-ServerMetadataTestJSON-316746592-project-member] Acquiring lock "4184e263-6f56-4bc1-99b2-a2c460531516" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 654.871138] env[61806]: DEBUG oslo_concurrency.lockutils [None req-b3f33a6f-532a-458d-aec6-6ec54b0cadd5 tempest-ServerMetadataTestJSON-316746592 tempest-ServerMetadataTestJSON-316746592-project-member] Lock "4184e263-6f56-4bc1-99b2-a2c460531516" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 654.912961] env[61806]: DEBUG oslo_vmware.api [None req-4803fdfc-9054-49cf-8525-e438672eeedd tempest-ServersAdmin275Test-1945525728 tempest-ServersAdmin275Test-1945525728-project-member] Task: {'id': task-1294469, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.701121} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 654.913328] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [None req-4803fdfc-9054-49cf-8525-e438672eeedd tempest-ServersAdmin275Test-1945525728 tempest-ServersAdmin275Test-1945525728-project-member] Copied Virtual Disk [datastore2] vmware_temp/7890e5b2-17c3-4ad0-b191-91ffb399ba28/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa/tmp-sparse.vmdk to [datastore2] vmware_temp/7890e5b2-17c3-4ad0-b191-91ffb399ba28/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa.vmdk {{(pid=61806) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 654.913534] env[61806]: DEBUG nova.virt.vmwareapi.ds_util [None req-4803fdfc-9054-49cf-8525-e438672eeedd tempest-ServersAdmin275Test-1945525728 tempest-ServersAdmin275Test-1945525728-project-member] Deleting the datastore file [datastore2] vmware_temp/7890e5b2-17c3-4ad0-b191-91ffb399ba28/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa/tmp-sparse.vmdk {{(pid=61806) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 654.913794] env[61806]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-ea0c271b-8231-402d-8d6e-b6e1d532eabd {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 654.919983] env[61806]: DEBUG oslo_vmware.api [None req-4803fdfc-9054-49cf-8525-e438672eeedd tempest-ServersAdmin275Test-1945525728 tempest-ServersAdmin275Test-1945525728-project-member] Waiting for the task: (returnval){ [ 654.919983] env[61806]: value = "task-1294470" [ 654.919983] env[61806]: _type = "Task" [ 654.919983] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 654.928363] env[61806]: DEBUG oslo_vmware.api [None req-4803fdfc-9054-49cf-8525-e438672eeedd tempest-ServersAdmin275Test-1945525728 tempest-ServersAdmin275Test-1945525728-project-member] Task: {'id': task-1294470, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 654.984371] env[61806]: DEBUG oslo_concurrency.lockutils [None req-3e50272a-5153-4e11-beee-d774a917f2df tempest-ServerPasswordTestJSON-1741413915 tempest-ServerPasswordTestJSON-1741413915-project-member] Releasing lock "refresh_cache-0d757385-ff1c-49c9-825f-f9d542ae074f" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 654.984649] env[61806]: DEBUG nova.compute.manager [None req-3e50272a-5153-4e11-beee-d774a917f2df tempest-ServerPasswordTestJSON-1741413915 tempest-ServerPasswordTestJSON-1741413915-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61806) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 654.984841] env[61806]: DEBUG nova.compute.manager [None req-3e50272a-5153-4e11-beee-d774a917f2df tempest-ServerPasswordTestJSON-1741413915 tempest-ServerPasswordTestJSON-1741413915-project-member] [instance: 0d757385-ff1c-49c9-825f-f9d542ae074f] Deallocating network for instance {{(pid=61806) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 654.985015] env[61806]: DEBUG nova.network.neutron [None req-3e50272a-5153-4e11-beee-d774a917f2df tempest-ServerPasswordTestJSON-1741413915 tempest-ServerPasswordTestJSON-1741413915-project-member] [instance: 0d757385-ff1c-49c9-825f-f9d542ae074f] deallocate_for_instance() {{(pid=61806) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 655.001230] env[61806]: DEBUG nova.network.neutron [None req-3e50272a-5153-4e11-beee-d774a917f2df tempest-ServerPasswordTestJSON-1741413915 tempest-ServerPasswordTestJSON-1741413915-project-member] [instance: 0d757385-ff1c-49c9-825f-f9d542ae074f] Instance cache missing network info. {{(pid=61806) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 655.203762] env[61806]: DEBUG nova.network.neutron [None req-a93ef3a4-6e83-4dfe-9292-bac21f14d678 tempest-ServerExternalEventsTest-1847900285 tempest-ServerExternalEventsTest-1847900285-project-member] [instance: d5434dd2-206f-4a1e-b97b-bd6f691b6ddf] Instance cache missing network info. {{(pid=61806) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 655.255903] env[61806]: DEBUG nova.network.neutron [None req-a93ef3a4-6e83-4dfe-9292-bac21f14d678 tempest-ServerExternalEventsTest-1847900285 tempest-ServerExternalEventsTest-1847900285-project-member] [instance: d5434dd2-206f-4a1e-b97b-bd6f691b6ddf] Updating instance_info_cache with network_info: [] {{(pid=61806) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 655.370946] env[61806]: ERROR nova.scheduler.client.report [None req-f8b2f38e-90f5-4b03-89bc-ca1926adb40c tempest-FloatingIPsAssociationTestJSON-35718468 tempest-FloatingIPsAssociationTestJSON-35718468-project-member] [req-f2f5c074-7224-4bd7-bb59-72667e144c65] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 138, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID a2858be1-fd22-4e08-979e-87ad25293407. Got 409: {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-f2f5c074-7224-4bd7-bb59-72667e144c65"}]}: nova.exception.PortBindingFailed: Binding failed for port 3eb55d85-873d-4f86-bf99-5f748157a7af, please check neutron logs for more information. [ 655.391674] env[61806]: DEBUG nova.scheduler.client.report [None req-f8b2f38e-90f5-4b03-89bc-ca1926adb40c tempest-FloatingIPsAssociationTestJSON-35718468 tempest-FloatingIPsAssociationTestJSON-35718468-project-member] Refreshing inventories for resource provider a2858be1-fd22-4e08-979e-87ad25293407 {{(pid=61806) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:818}} [ 655.409220] env[61806]: DEBUG nova.scheduler.client.report [None req-f8b2f38e-90f5-4b03-89bc-ca1926adb40c tempest-FloatingIPsAssociationTestJSON-35718468 tempest-FloatingIPsAssociationTestJSON-35718468-project-member] Updating ProviderTree inventory for provider a2858be1-fd22-4e08-979e-87ad25293407 from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 139, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61806) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:782}} [ 655.409449] env[61806]: DEBUG nova.compute.provider_tree [None req-f8b2f38e-90f5-4b03-89bc-ca1926adb40c tempest-FloatingIPsAssociationTestJSON-35718468 tempest-FloatingIPsAssociationTestJSON-35718468-project-member] Updating inventory in ProviderTree for provider a2858be1-fd22-4e08-979e-87ad25293407 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 139, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61806) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 655.422229] env[61806]: DEBUG nova.scheduler.client.report [None req-f8b2f38e-90f5-4b03-89bc-ca1926adb40c tempest-FloatingIPsAssociationTestJSON-35718468 tempest-FloatingIPsAssociationTestJSON-35718468-project-member] Refreshing aggregate associations for resource provider a2858be1-fd22-4e08-979e-87ad25293407, aggregates: None {{(pid=61806) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:827}} [ 655.434435] env[61806]: DEBUG oslo_vmware.api [None req-4803fdfc-9054-49cf-8525-e438672eeedd tempest-ServersAdmin275Test-1945525728 tempest-ServersAdmin275Test-1945525728-project-member] Task: {'id': task-1294470, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.02186} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 655.435345] env[61806]: DEBUG nova.virt.vmwareapi.ds_util [None req-4803fdfc-9054-49cf-8525-e438672eeedd tempest-ServersAdmin275Test-1945525728 tempest-ServersAdmin275Test-1945525728-project-member] Deleted the datastore file {{(pid=61806) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 655.435672] env[61806]: DEBUG nova.virt.vmwareapi.ds_util [None req-4803fdfc-9054-49cf-8525-e438672eeedd tempest-ServersAdmin275Test-1945525728 tempest-ServersAdmin275Test-1945525728-project-member] Moving file from [datastore2] vmware_temp/7890e5b2-17c3-4ad0-b191-91ffb399ba28/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa to [datastore2] devstack-image-cache_base/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa. {{(pid=61806) file_move /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:334}} [ 655.435947] env[61806]: DEBUG oslo_vmware.service [-] Invoking FileManager.MoveDatastoreFile_Task with opID=oslo.vmware-0e5297c8-e8b3-4ade-b4c2-057995151a7f {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 655.442680] env[61806]: DEBUG oslo_vmware.api [None req-4803fdfc-9054-49cf-8525-e438672eeedd tempest-ServersAdmin275Test-1945525728 tempest-ServersAdmin275Test-1945525728-project-member] Waiting for the task: (returnval){ [ 655.442680] env[61806]: value = "task-1294471" [ 655.442680] env[61806]: _type = "Task" [ 655.442680] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 655.452617] env[61806]: DEBUG oslo_vmware.api [None req-4803fdfc-9054-49cf-8525-e438672eeedd tempest-ServersAdmin275Test-1945525728 tempest-ServersAdmin275Test-1945525728-project-member] Task: {'id': task-1294471, 'name': MoveDatastoreFile_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 655.457017] env[61806]: DEBUG nova.scheduler.client.report [None req-f8b2f38e-90f5-4b03-89bc-ca1926adb40c tempest-FloatingIPsAssociationTestJSON-35718468 tempest-FloatingIPsAssociationTestJSON-35718468-project-member] Refreshing trait associations for resource provider a2858be1-fd22-4e08-979e-87ad25293407, traits: COMPUTE_IMAGE_TYPE_ISO,COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_SAME_HOST_COLD_MIGRATE,COMPUTE_NET_ATTACH_INTERFACE,COMPUTE_NODE {{(pid=61806) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:839}} [ 655.505033] env[61806]: DEBUG nova.network.neutron [None req-3e50272a-5153-4e11-beee-d774a917f2df tempest-ServerPasswordTestJSON-1741413915 tempest-ServerPasswordTestJSON-1741413915-project-member] [instance: 0d757385-ff1c-49c9-825f-f9d542ae074f] Updating instance_info_cache with network_info: [] {{(pid=61806) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 655.622442] env[61806]: DEBUG nova.compute.manager [req-456ca6fe-cf3a-4446-bcc6-d74e97b6097a req-5e62fc99-0b39-43b1-8fe0-e45dfad9195b service nova] [instance: d5434dd2-206f-4a1e-b97b-bd6f691b6ddf] Received event network-vif-deleted-6b28a2b6-fc7b-4db9-84db-18b0d6bbe032 {{(pid=61806) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 655.760246] env[61806]: DEBUG oslo_concurrency.lockutils [None req-a93ef3a4-6e83-4dfe-9292-bac21f14d678 tempest-ServerExternalEventsTest-1847900285 tempest-ServerExternalEventsTest-1847900285-project-member] Releasing lock "refresh_cache-d5434dd2-206f-4a1e-b97b-bd6f691b6ddf" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 655.760653] env[61806]: DEBUG nova.compute.manager [None req-a93ef3a4-6e83-4dfe-9292-bac21f14d678 tempest-ServerExternalEventsTest-1847900285 tempest-ServerExternalEventsTest-1847900285-project-member] [instance: d5434dd2-206f-4a1e-b97b-bd6f691b6ddf] Start destroying the instance on the hypervisor. {{(pid=61806) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 655.760843] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-a93ef3a4-6e83-4dfe-9292-bac21f14d678 tempest-ServerExternalEventsTest-1847900285 tempest-ServerExternalEventsTest-1847900285-project-member] [instance: d5434dd2-206f-4a1e-b97b-bd6f691b6ddf] Destroying instance {{(pid=61806) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 655.761320] env[61806]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-1cf24736-6384-45c3-b231-1a4b36263c27 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 655.770515] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f3cc3556-87e7-4afd-bb46-95dceddb7a18 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 655.796345] env[61806]: WARNING nova.virt.vmwareapi.vmops [None req-a93ef3a4-6e83-4dfe-9292-bac21f14d678 tempest-ServerExternalEventsTest-1847900285 tempest-ServerExternalEventsTest-1847900285-project-member] [instance: d5434dd2-206f-4a1e-b97b-bd6f691b6ddf] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance d5434dd2-206f-4a1e-b97b-bd6f691b6ddf could not be found. [ 655.796538] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-a93ef3a4-6e83-4dfe-9292-bac21f14d678 tempest-ServerExternalEventsTest-1847900285 tempest-ServerExternalEventsTest-1847900285-project-member] [instance: d5434dd2-206f-4a1e-b97b-bd6f691b6ddf] Instance destroyed {{(pid=61806) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 655.796702] env[61806]: INFO nova.compute.manager [None req-a93ef3a4-6e83-4dfe-9292-bac21f14d678 tempest-ServerExternalEventsTest-1847900285 tempest-ServerExternalEventsTest-1847900285-project-member] [instance: d5434dd2-206f-4a1e-b97b-bd6f691b6ddf] Took 0.04 seconds to destroy the instance on the hypervisor. [ 655.796948] env[61806]: DEBUG oslo.service.loopingcall [None req-a93ef3a4-6e83-4dfe-9292-bac21f14d678 tempest-ServerExternalEventsTest-1847900285 tempest-ServerExternalEventsTest-1847900285-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61806) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 655.797363] env[61806]: DEBUG nova.compute.manager [-] [instance: d5434dd2-206f-4a1e-b97b-bd6f691b6ddf] Deallocating network for instance {{(pid=61806) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 655.797447] env[61806]: DEBUG nova.network.neutron [-] [instance: d5434dd2-206f-4a1e-b97b-bd6f691b6ddf] deallocate_for_instance() {{(pid=61806) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 655.815374] env[61806]: DEBUG nova.network.neutron [-] [instance: d5434dd2-206f-4a1e-b97b-bd6f691b6ddf] Instance cache missing network info. {{(pid=61806) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 655.879207] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-99469c10-8c9d-4585-a8e5-c739434a06c6 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 655.890557] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c78f4f26-f058-476e-8152-9163a364ecc7 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 655.922911] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3907e8ce-7c3a-44b0-b118-b3f831ff9041 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 655.930497] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cfee947a-e6e0-4882-8146-489a75cd7566 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 655.943819] env[61806]: DEBUG nova.compute.provider_tree [None req-f8b2f38e-90f5-4b03-89bc-ca1926adb40c tempest-FloatingIPsAssociationTestJSON-35718468 tempest-FloatingIPsAssociationTestJSON-35718468-project-member] Updating inventory in ProviderTree for provider a2858be1-fd22-4e08-979e-87ad25293407 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 138, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61806) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 655.956021] env[61806]: DEBUG oslo_vmware.api [None req-4803fdfc-9054-49cf-8525-e438672eeedd tempest-ServersAdmin275Test-1945525728 tempest-ServersAdmin275Test-1945525728-project-member] Task: {'id': task-1294471, 'name': MoveDatastoreFile_Task, 'duration_secs': 0.023671} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 655.956021] env[61806]: DEBUG nova.virt.vmwareapi.ds_util [None req-4803fdfc-9054-49cf-8525-e438672eeedd tempest-ServersAdmin275Test-1945525728 tempest-ServersAdmin275Test-1945525728-project-member] File moved {{(pid=61806) file_move /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:346}} [ 655.956021] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-4803fdfc-9054-49cf-8525-e438672eeedd tempest-ServersAdmin275Test-1945525728 tempest-ServersAdmin275Test-1945525728-project-member] [instance: 9db80c3c-e2c3-4034-a79a-d7be924c4ff6] Cleaning up location [datastore2] vmware_temp/7890e5b2-17c3-4ad0-b191-91ffb399ba28 {{(pid=61806) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:640}} [ 655.956021] env[61806]: DEBUG nova.virt.vmwareapi.ds_util [None req-4803fdfc-9054-49cf-8525-e438672eeedd tempest-ServersAdmin275Test-1945525728 tempest-ServersAdmin275Test-1945525728-project-member] Deleting the datastore file [datastore2] vmware_temp/7890e5b2-17c3-4ad0-b191-91ffb399ba28 {{(pid=61806) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 655.956021] env[61806]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-31dfb8cd-7638-4730-80bb-02747cf0df0a {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 655.960908] env[61806]: DEBUG oslo_vmware.api [None req-4803fdfc-9054-49cf-8525-e438672eeedd tempest-ServersAdmin275Test-1945525728 tempest-ServersAdmin275Test-1945525728-project-member] Waiting for the task: (returnval){ [ 655.960908] env[61806]: value = "task-1294472" [ 655.960908] env[61806]: _type = "Task" [ 655.960908] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 655.968947] env[61806]: DEBUG oslo_vmware.api [None req-4803fdfc-9054-49cf-8525-e438672eeedd tempest-ServersAdmin275Test-1945525728 tempest-ServersAdmin275Test-1945525728-project-member] Task: {'id': task-1294472, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 656.009538] env[61806]: INFO nova.compute.manager [None req-3e50272a-5153-4e11-beee-d774a917f2df tempest-ServerPasswordTestJSON-1741413915 tempest-ServerPasswordTestJSON-1741413915-project-member] [instance: 0d757385-ff1c-49c9-825f-f9d542ae074f] Took 1.02 seconds to deallocate network for instance. [ 656.317880] env[61806]: DEBUG nova.network.neutron [-] [instance: d5434dd2-206f-4a1e-b97b-bd6f691b6ddf] Updating instance_info_cache with network_info: [] {{(pid=61806) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 656.471140] env[61806]: DEBUG oslo_vmware.api [None req-4803fdfc-9054-49cf-8525-e438672eeedd tempest-ServersAdmin275Test-1945525728 tempest-ServersAdmin275Test-1945525728-project-member] Task: {'id': task-1294472, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.026278} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 656.471378] env[61806]: DEBUG nova.virt.vmwareapi.ds_util [None req-4803fdfc-9054-49cf-8525-e438672eeedd tempest-ServersAdmin275Test-1945525728 tempest-ServersAdmin275Test-1945525728-project-member] Deleted the datastore file {{(pid=61806) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 656.471931] env[61806]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-433fcb83-93c4-45df-8ab8-49706d3b883a {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 656.477034] env[61806]: DEBUG oslo_vmware.api [None req-4803fdfc-9054-49cf-8525-e438672eeedd tempest-ServersAdmin275Test-1945525728 tempest-ServersAdmin275Test-1945525728-project-member] Waiting for the task: (returnval){ [ 656.477034] env[61806]: value = "session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]5232901e-2d85-9f3b-87a7-7d7cbb9f974e" [ 656.477034] env[61806]: _type = "Task" [ 656.477034] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 656.480686] env[61806]: DEBUG nova.scheduler.client.report [None req-f8b2f38e-90f5-4b03-89bc-ca1926adb40c tempest-FloatingIPsAssociationTestJSON-35718468 tempest-FloatingIPsAssociationTestJSON-35718468-project-member] Updated inventory for provider a2858be1-fd22-4e08-979e-87ad25293407 with generation 61 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 138, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61806) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:971}} [ 656.480860] env[61806]: DEBUG nova.compute.provider_tree [None req-f8b2f38e-90f5-4b03-89bc-ca1926adb40c tempest-FloatingIPsAssociationTestJSON-35718468 tempest-FloatingIPsAssociationTestJSON-35718468-project-member] Updating resource provider a2858be1-fd22-4e08-979e-87ad25293407 generation from 61 to 62 during operation: update_inventory {{(pid=61806) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 656.481055] env[61806]: DEBUG nova.compute.provider_tree [None req-f8b2f38e-90f5-4b03-89bc-ca1926adb40c tempest-FloatingIPsAssociationTestJSON-35718468 tempest-FloatingIPsAssociationTestJSON-35718468-project-member] Updating inventory in ProviderTree for provider a2858be1-fd22-4e08-979e-87ad25293407 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 138, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61806) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 656.489024] env[61806]: DEBUG oslo_vmware.api [None req-4803fdfc-9054-49cf-8525-e438672eeedd tempest-ServersAdmin275Test-1945525728 tempest-ServersAdmin275Test-1945525728-project-member] Task: {'id': session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]5232901e-2d85-9f3b-87a7-7d7cbb9f974e, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 656.820699] env[61806]: INFO nova.compute.manager [-] [instance: d5434dd2-206f-4a1e-b97b-bd6f691b6ddf] Took 1.02 seconds to deallocate network for instance. [ 656.823148] env[61806]: DEBUG nova.compute.claims [None req-a93ef3a4-6e83-4dfe-9292-bac21f14d678 tempest-ServerExternalEventsTest-1847900285 tempest-ServerExternalEventsTest-1847900285-project-member] [instance: d5434dd2-206f-4a1e-b97b-bd6f691b6ddf] Aborting claim: {{(pid=61806) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 656.823332] env[61806]: DEBUG oslo_concurrency.lockutils [None req-a93ef3a4-6e83-4dfe-9292-bac21f14d678 tempest-ServerExternalEventsTest-1847900285 tempest-ServerExternalEventsTest-1847900285-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 656.987247] env[61806]: DEBUG oslo_vmware.api [None req-4803fdfc-9054-49cf-8525-e438672eeedd tempest-ServersAdmin275Test-1945525728 tempest-ServersAdmin275Test-1945525728-project-member] Task: {'id': session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]5232901e-2d85-9f3b-87a7-7d7cbb9f974e, 'name': SearchDatastore_Task, 'duration_secs': 0.010067} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 656.987521] env[61806]: DEBUG oslo_concurrency.lockutils [None req-4803fdfc-9054-49cf-8525-e438672eeedd tempest-ServersAdmin275Test-1945525728 tempest-ServersAdmin275Test-1945525728-project-member] Releasing lock "[datastore2] devstack-image-cache_base/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa.vmdk" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 656.987772] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [None req-4803fdfc-9054-49cf-8525-e438672eeedd tempest-ServersAdmin275Test-1945525728 tempest-ServersAdmin275Test-1945525728-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa.vmdk to [datastore2] 9db80c3c-e2c3-4034-a79a-d7be924c4ff6/9db80c3c-e2c3-4034-a79a-d7be924c4ff6.vmdk {{(pid=61806) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 656.988075] env[61806]: DEBUG oslo_concurrency.lockutils [None req-fe158576-4234-4a6b-b48d-5ca8bb977c40 tempest-ServerShowV254Test-608938657 tempest-ServerShowV254Test-608938657-project-member] Acquired lock "[datastore2] devstack-image-cache_base/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa.vmdk" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 656.988237] env[61806]: DEBUG nova.virt.vmwareapi.ds_util [None req-fe158576-4234-4a6b-b48d-5ca8bb977c40 tempest-ServerShowV254Test-608938657 tempest-ServerShowV254Test-608938657-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61806) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 656.988450] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-64716749-e677-4d0a-be38-a1a8ffd29689 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 656.990272] env[61806]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-b097117b-4273-48ed-a4cb-a797ee137f5c {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 656.992477] env[61806]: DEBUG oslo_concurrency.lockutils [None req-f8b2f38e-90f5-4b03-89bc-ca1926adb40c tempest-FloatingIPsAssociationTestJSON-35718468 tempest-FloatingIPsAssociationTestJSON-35718468-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 3.170s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 656.993068] env[61806]: ERROR nova.compute.manager [None req-f8b2f38e-90f5-4b03-89bc-ca1926adb40c tempest-FloatingIPsAssociationTestJSON-35718468 tempest-FloatingIPsAssociationTestJSON-35718468-project-member] [instance: 88cb0961-ad34-4ffc-8fb2-6034f08ee740] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 3eb55d85-873d-4f86-bf99-5f748157a7af, please check neutron logs for more information. [ 656.993068] env[61806]: ERROR nova.compute.manager [instance: 88cb0961-ad34-4ffc-8fb2-6034f08ee740] Traceback (most recent call last): [ 656.993068] env[61806]: ERROR nova.compute.manager [instance: 88cb0961-ad34-4ffc-8fb2-6034f08ee740] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 656.993068] env[61806]: ERROR nova.compute.manager [instance: 88cb0961-ad34-4ffc-8fb2-6034f08ee740] self.driver.spawn(context, instance, image_meta, [ 656.993068] env[61806]: ERROR nova.compute.manager [instance: 88cb0961-ad34-4ffc-8fb2-6034f08ee740] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 656.993068] env[61806]: ERROR nova.compute.manager [instance: 88cb0961-ad34-4ffc-8fb2-6034f08ee740] self._vmops.spawn(context, instance, image_meta, injected_files, [ 656.993068] env[61806]: ERROR nova.compute.manager [instance: 88cb0961-ad34-4ffc-8fb2-6034f08ee740] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 656.993068] env[61806]: ERROR nova.compute.manager [instance: 88cb0961-ad34-4ffc-8fb2-6034f08ee740] vm_ref = self.build_virtual_machine(instance, [ 656.993068] env[61806]: ERROR nova.compute.manager [instance: 88cb0961-ad34-4ffc-8fb2-6034f08ee740] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 656.993068] env[61806]: ERROR nova.compute.manager [instance: 88cb0961-ad34-4ffc-8fb2-6034f08ee740] vif_infos = vmwarevif.get_vif_info(self._session, [ 656.993068] env[61806]: ERROR nova.compute.manager [instance: 88cb0961-ad34-4ffc-8fb2-6034f08ee740] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 656.993420] env[61806]: ERROR nova.compute.manager [instance: 88cb0961-ad34-4ffc-8fb2-6034f08ee740] for vif in network_info: [ 656.993420] env[61806]: ERROR nova.compute.manager [instance: 88cb0961-ad34-4ffc-8fb2-6034f08ee740] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 656.993420] env[61806]: ERROR nova.compute.manager [instance: 88cb0961-ad34-4ffc-8fb2-6034f08ee740] return self._sync_wrapper(fn, *args, **kwargs) [ 656.993420] env[61806]: ERROR nova.compute.manager [instance: 88cb0961-ad34-4ffc-8fb2-6034f08ee740] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 656.993420] env[61806]: ERROR nova.compute.manager [instance: 88cb0961-ad34-4ffc-8fb2-6034f08ee740] self.wait() [ 656.993420] env[61806]: ERROR nova.compute.manager [instance: 88cb0961-ad34-4ffc-8fb2-6034f08ee740] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 656.993420] env[61806]: ERROR nova.compute.manager [instance: 88cb0961-ad34-4ffc-8fb2-6034f08ee740] self[:] = self._gt.wait() [ 656.993420] env[61806]: ERROR nova.compute.manager [instance: 88cb0961-ad34-4ffc-8fb2-6034f08ee740] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 656.993420] env[61806]: ERROR nova.compute.manager [instance: 88cb0961-ad34-4ffc-8fb2-6034f08ee740] return self._exit_event.wait() [ 656.993420] env[61806]: ERROR nova.compute.manager [instance: 88cb0961-ad34-4ffc-8fb2-6034f08ee740] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 656.993420] env[61806]: ERROR nova.compute.manager [instance: 88cb0961-ad34-4ffc-8fb2-6034f08ee740] result = hub.switch() [ 656.993420] env[61806]: ERROR nova.compute.manager [instance: 88cb0961-ad34-4ffc-8fb2-6034f08ee740] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 656.993420] env[61806]: ERROR nova.compute.manager [instance: 88cb0961-ad34-4ffc-8fb2-6034f08ee740] return self.greenlet.switch() [ 656.993915] env[61806]: ERROR nova.compute.manager [instance: 88cb0961-ad34-4ffc-8fb2-6034f08ee740] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 656.993915] env[61806]: ERROR nova.compute.manager [instance: 88cb0961-ad34-4ffc-8fb2-6034f08ee740] result = function(*args, **kwargs) [ 656.993915] env[61806]: ERROR nova.compute.manager [instance: 88cb0961-ad34-4ffc-8fb2-6034f08ee740] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 656.993915] env[61806]: ERROR nova.compute.manager [instance: 88cb0961-ad34-4ffc-8fb2-6034f08ee740] return func(*args, **kwargs) [ 656.993915] env[61806]: ERROR nova.compute.manager [instance: 88cb0961-ad34-4ffc-8fb2-6034f08ee740] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 656.993915] env[61806]: ERROR nova.compute.manager [instance: 88cb0961-ad34-4ffc-8fb2-6034f08ee740] raise e [ 656.993915] env[61806]: ERROR nova.compute.manager [instance: 88cb0961-ad34-4ffc-8fb2-6034f08ee740] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 656.993915] env[61806]: ERROR nova.compute.manager [instance: 88cb0961-ad34-4ffc-8fb2-6034f08ee740] nwinfo = self.network_api.allocate_for_instance( [ 656.993915] env[61806]: ERROR nova.compute.manager [instance: 88cb0961-ad34-4ffc-8fb2-6034f08ee740] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 656.993915] env[61806]: ERROR nova.compute.manager [instance: 88cb0961-ad34-4ffc-8fb2-6034f08ee740] created_port_ids = self._update_ports_for_instance( [ 656.993915] env[61806]: ERROR nova.compute.manager [instance: 88cb0961-ad34-4ffc-8fb2-6034f08ee740] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 656.993915] env[61806]: ERROR nova.compute.manager [instance: 88cb0961-ad34-4ffc-8fb2-6034f08ee740] with excutils.save_and_reraise_exception(): [ 656.993915] env[61806]: ERROR nova.compute.manager [instance: 88cb0961-ad34-4ffc-8fb2-6034f08ee740] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 656.994294] env[61806]: ERROR nova.compute.manager [instance: 88cb0961-ad34-4ffc-8fb2-6034f08ee740] self.force_reraise() [ 656.994294] env[61806]: ERROR nova.compute.manager [instance: 88cb0961-ad34-4ffc-8fb2-6034f08ee740] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 656.994294] env[61806]: ERROR nova.compute.manager [instance: 88cb0961-ad34-4ffc-8fb2-6034f08ee740] raise self.value [ 656.994294] env[61806]: ERROR nova.compute.manager [instance: 88cb0961-ad34-4ffc-8fb2-6034f08ee740] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 656.994294] env[61806]: ERROR nova.compute.manager [instance: 88cb0961-ad34-4ffc-8fb2-6034f08ee740] updated_port = self._update_port( [ 656.994294] env[61806]: ERROR nova.compute.manager [instance: 88cb0961-ad34-4ffc-8fb2-6034f08ee740] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 656.994294] env[61806]: ERROR nova.compute.manager [instance: 88cb0961-ad34-4ffc-8fb2-6034f08ee740] _ensure_no_port_binding_failure(port) [ 656.994294] env[61806]: ERROR nova.compute.manager [instance: 88cb0961-ad34-4ffc-8fb2-6034f08ee740] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 656.994294] env[61806]: ERROR nova.compute.manager [instance: 88cb0961-ad34-4ffc-8fb2-6034f08ee740] raise exception.PortBindingFailed(port_id=port['id']) [ 656.994294] env[61806]: ERROR nova.compute.manager [instance: 88cb0961-ad34-4ffc-8fb2-6034f08ee740] nova.exception.PortBindingFailed: Binding failed for port 3eb55d85-873d-4f86-bf99-5f748157a7af, please check neutron logs for more information. [ 656.994294] env[61806]: ERROR nova.compute.manager [instance: 88cb0961-ad34-4ffc-8fb2-6034f08ee740] [ 656.994703] env[61806]: DEBUG nova.compute.utils [None req-f8b2f38e-90f5-4b03-89bc-ca1926adb40c tempest-FloatingIPsAssociationTestJSON-35718468 tempest-FloatingIPsAssociationTestJSON-35718468-project-member] [instance: 88cb0961-ad34-4ffc-8fb2-6034f08ee740] Binding failed for port 3eb55d85-873d-4f86-bf99-5f748157a7af, please check neutron logs for more information. {{(pid=61806) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 656.995011] env[61806]: DEBUG oslo_concurrency.lockutils [None req-b4447f02-a3c7-4b6e-89ba-6bc34ec03205 tempest-InstanceActionsTestJSON-949444324 tempest-InstanceActionsTestJSON-949444324-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 17.076s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 656.996265] env[61806]: INFO nova.compute.claims [None req-b4447f02-a3c7-4b6e-89ba-6bc34ec03205 tempest-InstanceActionsTestJSON-949444324 tempest-InstanceActionsTestJSON-949444324-project-member] [instance: 2681766c-0468-4170-8c51-03389e2bd1ab] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 656.999224] env[61806]: DEBUG nova.compute.manager [None req-f8b2f38e-90f5-4b03-89bc-ca1926adb40c tempest-FloatingIPsAssociationTestJSON-35718468 tempest-FloatingIPsAssociationTestJSON-35718468-project-member] [instance: 88cb0961-ad34-4ffc-8fb2-6034f08ee740] Build of instance 88cb0961-ad34-4ffc-8fb2-6034f08ee740 was re-scheduled: Binding failed for port 3eb55d85-873d-4f86-bf99-5f748157a7af, please check neutron logs for more information. {{(pid=61806) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 656.999663] env[61806]: DEBUG nova.compute.manager [None req-f8b2f38e-90f5-4b03-89bc-ca1926adb40c tempest-FloatingIPsAssociationTestJSON-35718468 tempest-FloatingIPsAssociationTestJSON-35718468-project-member] [instance: 88cb0961-ad34-4ffc-8fb2-6034f08ee740] Unplugging VIFs for instance {{(pid=61806) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 656.999899] env[61806]: DEBUG oslo_concurrency.lockutils [None req-f8b2f38e-90f5-4b03-89bc-ca1926adb40c tempest-FloatingIPsAssociationTestJSON-35718468 tempest-FloatingIPsAssociationTestJSON-35718468-project-member] Acquiring lock "refresh_cache-88cb0961-ad34-4ffc-8fb2-6034f08ee740" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 657.000059] env[61806]: DEBUG oslo_concurrency.lockutils [None req-f8b2f38e-90f5-4b03-89bc-ca1926adb40c tempest-FloatingIPsAssociationTestJSON-35718468 tempest-FloatingIPsAssociationTestJSON-35718468-project-member] Acquired lock "refresh_cache-88cb0961-ad34-4ffc-8fb2-6034f08ee740" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 657.000221] env[61806]: DEBUG nova.network.neutron [None req-f8b2f38e-90f5-4b03-89bc-ca1926adb40c tempest-FloatingIPsAssociationTestJSON-35718468 tempest-FloatingIPsAssociationTestJSON-35718468-project-member] [instance: 88cb0961-ad34-4ffc-8fb2-6034f08ee740] Building network info cache for instance {{(pid=61806) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 657.003084] env[61806]: DEBUG oslo_vmware.api [None req-4803fdfc-9054-49cf-8525-e438672eeedd tempest-ServersAdmin275Test-1945525728 tempest-ServersAdmin275Test-1945525728-project-member] Waiting for the task: (returnval){ [ 657.003084] env[61806]: value = "task-1294473" [ 657.003084] env[61806]: _type = "Task" [ 657.003084] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 657.007953] env[61806]: DEBUG nova.virt.vmwareapi.ds_util [None req-fe158576-4234-4a6b-b48d-5ca8bb977c40 tempest-ServerShowV254Test-608938657 tempest-ServerShowV254Test-608938657-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61806) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 657.008103] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-fe158576-4234-4a6b-b48d-5ca8bb977c40 tempest-ServerShowV254Test-608938657 tempest-ServerShowV254Test-608938657-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61806) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 657.013202] env[61806]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-cd89c2df-a26e-4123-89aa-4ea3122180e9 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 657.015569] env[61806]: DEBUG oslo_vmware.api [None req-4803fdfc-9054-49cf-8525-e438672eeedd tempest-ServersAdmin275Test-1945525728 tempest-ServersAdmin275Test-1945525728-project-member] Task: {'id': task-1294473, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 657.020776] env[61806]: DEBUG oslo_vmware.api [None req-fe158576-4234-4a6b-b48d-5ca8bb977c40 tempest-ServerShowV254Test-608938657 tempest-ServerShowV254Test-608938657-project-member] Waiting for the task: (returnval){ [ 657.020776] env[61806]: value = "session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]5202fbd4-ad65-5d1d-582d-22f82c0aa763" [ 657.020776] env[61806]: _type = "Task" [ 657.020776] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 657.029215] env[61806]: DEBUG oslo_vmware.api [None req-fe158576-4234-4a6b-b48d-5ca8bb977c40 tempest-ServerShowV254Test-608938657 tempest-ServerShowV254Test-608938657-project-member] Task: {'id': session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]5202fbd4-ad65-5d1d-582d-22f82c0aa763, 'name': SearchDatastore_Task, 'duration_secs': 0.007879} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 657.030014] env[61806]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-49f61800-06ac-430f-9c90-df84a493f108 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 657.035430] env[61806]: DEBUG oslo_vmware.api [None req-fe158576-4234-4a6b-b48d-5ca8bb977c40 tempest-ServerShowV254Test-608938657 tempest-ServerShowV254Test-608938657-project-member] Waiting for the task: (returnval){ [ 657.035430] env[61806]: value = "session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]52b3cc7d-a410-7d03-b56a-f7f1cd76fc05" [ 657.035430] env[61806]: _type = "Task" [ 657.035430] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 657.041028] env[61806]: INFO nova.scheduler.client.report [None req-3e50272a-5153-4e11-beee-d774a917f2df tempest-ServerPasswordTestJSON-1741413915 tempest-ServerPasswordTestJSON-1741413915-project-member] Deleted allocations for instance 0d757385-ff1c-49c9-825f-f9d542ae074f [ 657.049422] env[61806]: DEBUG oslo_vmware.api [None req-fe158576-4234-4a6b-b48d-5ca8bb977c40 tempest-ServerShowV254Test-608938657 tempest-ServerShowV254Test-608938657-project-member] Task: {'id': session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]52b3cc7d-a410-7d03-b56a-f7f1cd76fc05, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 657.515946] env[61806]: DEBUG oslo_vmware.api [None req-4803fdfc-9054-49cf-8525-e438672eeedd tempest-ServersAdmin275Test-1945525728 tempest-ServersAdmin275Test-1945525728-project-member] Task: {'id': task-1294473, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.474372} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 657.515946] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [None req-4803fdfc-9054-49cf-8525-e438672eeedd tempest-ServersAdmin275Test-1945525728 tempest-ServersAdmin275Test-1945525728-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa.vmdk to [datastore2] 9db80c3c-e2c3-4034-a79a-d7be924c4ff6/9db80c3c-e2c3-4034-a79a-d7be924c4ff6.vmdk {{(pid=61806) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 657.515946] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-4803fdfc-9054-49cf-8525-e438672eeedd tempest-ServersAdmin275Test-1945525728 tempest-ServersAdmin275Test-1945525728-project-member] [instance: 9db80c3c-e2c3-4034-a79a-d7be924c4ff6] Extending root virtual disk to 1048576 {{(pid=61806) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 657.516334] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-a827d699-70db-4d96-b659-ef952c3a8c17 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 657.523234] env[61806]: DEBUG oslo_vmware.api [None req-4803fdfc-9054-49cf-8525-e438672eeedd tempest-ServersAdmin275Test-1945525728 tempest-ServersAdmin275Test-1945525728-project-member] Waiting for the task: (returnval){ [ 657.523234] env[61806]: value = "task-1294474" [ 657.523234] env[61806]: _type = "Task" [ 657.523234] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 657.530907] env[61806]: DEBUG oslo_vmware.api [None req-4803fdfc-9054-49cf-8525-e438672eeedd tempest-ServersAdmin275Test-1945525728 tempest-ServersAdmin275Test-1945525728-project-member] Task: {'id': task-1294474, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 657.533022] env[61806]: DEBUG nova.network.neutron [None req-f8b2f38e-90f5-4b03-89bc-ca1926adb40c tempest-FloatingIPsAssociationTestJSON-35718468 tempest-FloatingIPsAssociationTestJSON-35718468-project-member] [instance: 88cb0961-ad34-4ffc-8fb2-6034f08ee740] Instance cache missing network info. {{(pid=61806) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 657.543854] env[61806]: DEBUG oslo_vmware.api [None req-fe158576-4234-4a6b-b48d-5ca8bb977c40 tempest-ServerShowV254Test-608938657 tempest-ServerShowV254Test-608938657-project-member] Task: {'id': session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]52b3cc7d-a410-7d03-b56a-f7f1cd76fc05, 'name': SearchDatastore_Task, 'duration_secs': 0.008564} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 657.544099] env[61806]: DEBUG oslo_concurrency.lockutils [None req-fe158576-4234-4a6b-b48d-5ca8bb977c40 tempest-ServerShowV254Test-608938657 tempest-ServerShowV254Test-608938657-project-member] Releasing lock "[datastore2] devstack-image-cache_base/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa.vmdk" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 657.544349] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [None req-fe158576-4234-4a6b-b48d-5ca8bb977c40 tempest-ServerShowV254Test-608938657 tempest-ServerShowV254Test-608938657-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa.vmdk to [datastore2] 8016bafc-b287-4df1-bd4b-a00210177a86/8016bafc-b287-4df1-bd4b-a00210177a86.vmdk {{(pid=61806) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 657.544592] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-1298afc9-ab67-40ad-86e6-ff69436ec331 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 657.550223] env[61806]: DEBUG oslo_vmware.api [None req-fe158576-4234-4a6b-b48d-5ca8bb977c40 tempest-ServerShowV254Test-608938657 tempest-ServerShowV254Test-608938657-project-member] Waiting for the task: (returnval){ [ 657.550223] env[61806]: value = "task-1294475" [ 657.550223] env[61806]: _type = "Task" [ 657.550223] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 657.553539] env[61806]: DEBUG oslo_concurrency.lockutils [None req-3e50272a-5153-4e11-beee-d774a917f2df tempest-ServerPasswordTestJSON-1741413915 tempest-ServerPasswordTestJSON-1741413915-project-member] Lock "0d757385-ff1c-49c9-825f-f9d542ae074f" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 102.223s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 657.559053] env[61806]: DEBUG oslo_vmware.api [None req-fe158576-4234-4a6b-b48d-5ca8bb977c40 tempest-ServerShowV254Test-608938657 tempest-ServerShowV254Test-608938657-project-member] Task: {'id': task-1294475, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 657.686404] env[61806]: DEBUG nova.network.neutron [None req-f8b2f38e-90f5-4b03-89bc-ca1926adb40c tempest-FloatingIPsAssociationTestJSON-35718468 tempest-FloatingIPsAssociationTestJSON-35718468-project-member] [instance: 88cb0961-ad34-4ffc-8fb2-6034f08ee740] Updating instance_info_cache with network_info: [] {{(pid=61806) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 658.037024] env[61806]: DEBUG oslo_vmware.api [None req-4803fdfc-9054-49cf-8525-e438672eeedd tempest-ServersAdmin275Test-1945525728 tempest-ServersAdmin275Test-1945525728-project-member] Task: {'id': task-1294474, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.060258} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 658.037336] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-4803fdfc-9054-49cf-8525-e438672eeedd tempest-ServersAdmin275Test-1945525728 tempest-ServersAdmin275Test-1945525728-project-member] [instance: 9db80c3c-e2c3-4034-a79a-d7be924c4ff6] Extended root virtual disk {{(pid=61806) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 658.038132] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b3868281-8c22-4f40-ae3a-c9fc2d1e593c {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 658.060363] env[61806]: DEBUG nova.virt.vmwareapi.volumeops [None req-4803fdfc-9054-49cf-8525-e438672eeedd tempest-ServersAdmin275Test-1945525728 tempest-ServersAdmin275Test-1945525728-project-member] [instance: 9db80c3c-e2c3-4034-a79a-d7be924c4ff6] Reconfiguring VM instance instance-0000001f to attach disk [datastore2] 9db80c3c-e2c3-4034-a79a-d7be924c4ff6/9db80c3c-e2c3-4034-a79a-d7be924c4ff6.vmdk or device None with type sparse {{(pid=61806) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 658.065970] env[61806]: DEBUG nova.compute.manager [None req-2bf81f39-4b8d-4851-a793-760af008ffe2 tempest-ImagesOneServerTestJSON-1003762139 tempest-ImagesOneServerTestJSON-1003762139-project-member] [instance: 294f0a37-7589-405d-a324-13099c80f458] Starting instance... {{(pid=61806) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 658.068322] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-9660d15f-d134-4ad0-ba17-bf7160c162f7 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 658.090525] env[61806]: DEBUG oslo_vmware.api [None req-fe158576-4234-4a6b-b48d-5ca8bb977c40 tempest-ServerShowV254Test-608938657 tempest-ServerShowV254Test-608938657-project-member] Task: {'id': task-1294475, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.473102} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 658.091864] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [None req-fe158576-4234-4a6b-b48d-5ca8bb977c40 tempest-ServerShowV254Test-608938657 tempest-ServerShowV254Test-608938657-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa.vmdk to [datastore2] 8016bafc-b287-4df1-bd4b-a00210177a86/8016bafc-b287-4df1-bd4b-a00210177a86.vmdk {{(pid=61806) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 658.092127] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-fe158576-4234-4a6b-b48d-5ca8bb977c40 tempest-ServerShowV254Test-608938657 tempest-ServerShowV254Test-608938657-project-member] [instance: 8016bafc-b287-4df1-bd4b-a00210177a86] Extending root virtual disk to 1048576 {{(pid=61806) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 658.092461] env[61806]: DEBUG oslo_vmware.api [None req-4803fdfc-9054-49cf-8525-e438672eeedd tempest-ServersAdmin275Test-1945525728 tempest-ServersAdmin275Test-1945525728-project-member] Waiting for the task: (returnval){ [ 658.092461] env[61806]: value = "task-1294476" [ 658.092461] env[61806]: _type = "Task" [ 658.092461] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 658.095082] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-9f7b7fa4-d532-40db-b444-d20cb0ad9292 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 658.106337] env[61806]: DEBUG oslo_vmware.api [None req-4803fdfc-9054-49cf-8525-e438672eeedd tempest-ServersAdmin275Test-1945525728 tempest-ServersAdmin275Test-1945525728-project-member] Task: {'id': task-1294476, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 658.107752] env[61806]: DEBUG oslo_vmware.api [None req-fe158576-4234-4a6b-b48d-5ca8bb977c40 tempest-ServerShowV254Test-608938657 tempest-ServerShowV254Test-608938657-project-member] Waiting for the task: (returnval){ [ 658.107752] env[61806]: value = "task-1294477" [ 658.107752] env[61806]: _type = "Task" [ 658.107752] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 658.121385] env[61806]: DEBUG oslo_vmware.api [None req-fe158576-4234-4a6b-b48d-5ca8bb977c40 tempest-ServerShowV254Test-608938657 tempest-ServerShowV254Test-608938657-project-member] Task: {'id': task-1294477, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 658.190642] env[61806]: DEBUG oslo_concurrency.lockutils [None req-f8b2f38e-90f5-4b03-89bc-ca1926adb40c tempest-FloatingIPsAssociationTestJSON-35718468 tempest-FloatingIPsAssociationTestJSON-35718468-project-member] Releasing lock "refresh_cache-88cb0961-ad34-4ffc-8fb2-6034f08ee740" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 658.190642] env[61806]: DEBUG nova.compute.manager [None req-f8b2f38e-90f5-4b03-89bc-ca1926adb40c tempest-FloatingIPsAssociationTestJSON-35718468 tempest-FloatingIPsAssociationTestJSON-35718468-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61806) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 658.190642] env[61806]: DEBUG nova.compute.manager [None req-f8b2f38e-90f5-4b03-89bc-ca1926adb40c tempest-FloatingIPsAssociationTestJSON-35718468 tempest-FloatingIPsAssociationTestJSON-35718468-project-member] [instance: 88cb0961-ad34-4ffc-8fb2-6034f08ee740] Deallocating network for instance {{(pid=61806) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 658.190828] env[61806]: DEBUG nova.network.neutron [None req-f8b2f38e-90f5-4b03-89bc-ca1926adb40c tempest-FloatingIPsAssociationTestJSON-35718468 tempest-FloatingIPsAssociationTestJSON-35718468-project-member] [instance: 88cb0961-ad34-4ffc-8fb2-6034f08ee740] deallocate_for_instance() {{(pid=61806) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 658.208289] env[61806]: DEBUG nova.network.neutron [None req-f8b2f38e-90f5-4b03-89bc-ca1926adb40c tempest-FloatingIPsAssociationTestJSON-35718468 tempest-FloatingIPsAssociationTestJSON-35718468-project-member] [instance: 88cb0961-ad34-4ffc-8fb2-6034f08ee740] Instance cache missing network info. {{(pid=61806) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 658.445213] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-933e2e1b-be6a-4b84-a23f-9c7eb470cab7 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 658.453062] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2b6d233b-68cc-468c-b1e6-67c3ab9c1a81 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 658.483194] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f7789967-eecc-4d43-a507-96b30df809f3 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 658.490808] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-39c75af9-657c-4df5-8e7b-fd4efae116a5 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 658.503933] env[61806]: DEBUG nova.compute.provider_tree [None req-b4447f02-a3c7-4b6e-89ba-6bc34ec03205 tempest-InstanceActionsTestJSON-949444324 tempest-InstanceActionsTestJSON-949444324-project-member] Inventory has not changed in ProviderTree for provider: a2858be1-fd22-4e08-979e-87ad25293407 {{(pid=61806) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 658.596864] env[61806]: DEBUG oslo_concurrency.lockutils [None req-2bf81f39-4b8d-4851-a793-760af008ffe2 tempest-ImagesOneServerTestJSON-1003762139 tempest-ImagesOneServerTestJSON-1003762139-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 658.605884] env[61806]: DEBUG oslo_vmware.api [None req-4803fdfc-9054-49cf-8525-e438672eeedd tempest-ServersAdmin275Test-1945525728 tempest-ServersAdmin275Test-1945525728-project-member] Task: {'id': task-1294476, 'name': ReconfigVM_Task, 'duration_secs': 0.315842} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 658.606166] env[61806]: DEBUG nova.virt.vmwareapi.volumeops [None req-4803fdfc-9054-49cf-8525-e438672eeedd tempest-ServersAdmin275Test-1945525728 tempest-ServersAdmin275Test-1945525728-project-member] [instance: 9db80c3c-e2c3-4034-a79a-d7be924c4ff6] Reconfigured VM instance instance-0000001f to attach disk [datastore2] 9db80c3c-e2c3-4034-a79a-d7be924c4ff6/9db80c3c-e2c3-4034-a79a-d7be924c4ff6.vmdk or device None with type sparse {{(pid=61806) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 658.606900] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-d6737f04-8548-4125-942a-bb57c45ea8df {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 658.618762] env[61806]: DEBUG oslo_vmware.api [None req-fe158576-4234-4a6b-b48d-5ca8bb977c40 tempest-ServerShowV254Test-608938657 tempest-ServerShowV254Test-608938657-project-member] Task: {'id': task-1294477, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.207413} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 658.620012] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-fe158576-4234-4a6b-b48d-5ca8bb977c40 tempest-ServerShowV254Test-608938657 tempest-ServerShowV254Test-608938657-project-member] [instance: 8016bafc-b287-4df1-bd4b-a00210177a86] Extended root virtual disk {{(pid=61806) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 658.620305] env[61806]: DEBUG oslo_vmware.api [None req-4803fdfc-9054-49cf-8525-e438672eeedd tempest-ServersAdmin275Test-1945525728 tempest-ServersAdmin275Test-1945525728-project-member] Waiting for the task: (returnval){ [ 658.620305] env[61806]: value = "task-1294478" [ 658.620305] env[61806]: _type = "Task" [ 658.620305] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 658.620990] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bee76d34-65a9-412a-832a-c8fb61bbabe7 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 658.645627] env[61806]: DEBUG nova.virt.vmwareapi.volumeops [None req-fe158576-4234-4a6b-b48d-5ca8bb977c40 tempest-ServerShowV254Test-608938657 tempest-ServerShowV254Test-608938657-project-member] [instance: 8016bafc-b287-4df1-bd4b-a00210177a86] Reconfiguring VM instance instance-0000001c to attach disk [datastore2] 8016bafc-b287-4df1-bd4b-a00210177a86/8016bafc-b287-4df1-bd4b-a00210177a86.vmdk or device None with type sparse {{(pid=61806) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 658.645958] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-46eb7050-b0b6-4c5d-b654-2e8e9b616cfe {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 658.666062] env[61806]: DEBUG oslo_vmware.api [None req-fe158576-4234-4a6b-b48d-5ca8bb977c40 tempest-ServerShowV254Test-608938657 tempest-ServerShowV254Test-608938657-project-member] Waiting for the task: (returnval){ [ 658.666062] env[61806]: value = "task-1294479" [ 658.666062] env[61806]: _type = "Task" [ 658.666062] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 658.675540] env[61806]: DEBUG oslo_vmware.api [None req-fe158576-4234-4a6b-b48d-5ca8bb977c40 tempest-ServerShowV254Test-608938657 tempest-ServerShowV254Test-608938657-project-member] Task: {'id': task-1294479, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 658.710238] env[61806]: DEBUG nova.network.neutron [None req-f8b2f38e-90f5-4b03-89bc-ca1926adb40c tempest-FloatingIPsAssociationTestJSON-35718468 tempest-FloatingIPsAssociationTestJSON-35718468-project-member] [instance: 88cb0961-ad34-4ffc-8fb2-6034f08ee740] Updating instance_info_cache with network_info: [] {{(pid=61806) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 659.006826] env[61806]: DEBUG nova.scheduler.client.report [None req-b4447f02-a3c7-4b6e-89ba-6bc34ec03205 tempest-InstanceActionsTestJSON-949444324 tempest-InstanceActionsTestJSON-949444324-project-member] Inventory has not changed for provider a2858be1-fd22-4e08-979e-87ad25293407 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 138, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61806) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 659.136290] env[61806]: DEBUG oslo_vmware.api [None req-4803fdfc-9054-49cf-8525-e438672eeedd tempest-ServersAdmin275Test-1945525728 tempest-ServersAdmin275Test-1945525728-project-member] Task: {'id': task-1294478, 'name': Rename_Task, 'duration_secs': 0.349313} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 659.136577] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [None req-4803fdfc-9054-49cf-8525-e438672eeedd tempest-ServersAdmin275Test-1945525728 tempest-ServersAdmin275Test-1945525728-project-member] [instance: 9db80c3c-e2c3-4034-a79a-d7be924c4ff6] Powering on the VM {{(pid=61806) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 659.136825] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-7fbf00fb-c000-4e69-9818-3ff3c41a2741 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 659.143282] env[61806]: DEBUG oslo_vmware.api [None req-4803fdfc-9054-49cf-8525-e438672eeedd tempest-ServersAdmin275Test-1945525728 tempest-ServersAdmin275Test-1945525728-project-member] Waiting for the task: (returnval){ [ 659.143282] env[61806]: value = "task-1294480" [ 659.143282] env[61806]: _type = "Task" [ 659.143282] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 659.152995] env[61806]: DEBUG oslo_vmware.api [None req-4803fdfc-9054-49cf-8525-e438672eeedd tempest-ServersAdmin275Test-1945525728 tempest-ServersAdmin275Test-1945525728-project-member] Task: {'id': task-1294480, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 659.181462] env[61806]: DEBUG oslo_vmware.api [None req-fe158576-4234-4a6b-b48d-5ca8bb977c40 tempest-ServerShowV254Test-608938657 tempest-ServerShowV254Test-608938657-project-member] Task: {'id': task-1294479, 'name': ReconfigVM_Task, 'duration_secs': 0.318931} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 659.181802] env[61806]: DEBUG nova.virt.vmwareapi.volumeops [None req-fe158576-4234-4a6b-b48d-5ca8bb977c40 tempest-ServerShowV254Test-608938657 tempest-ServerShowV254Test-608938657-project-member] [instance: 8016bafc-b287-4df1-bd4b-a00210177a86] Reconfigured VM instance instance-0000001c to attach disk [datastore2] 8016bafc-b287-4df1-bd4b-a00210177a86/8016bafc-b287-4df1-bd4b-a00210177a86.vmdk or device None with type sparse {{(pid=61806) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 659.182438] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-bf8607d4-12e5-4b9f-ba6f-ab3ca2ad071c {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 659.188592] env[61806]: DEBUG oslo_vmware.api [None req-fe158576-4234-4a6b-b48d-5ca8bb977c40 tempest-ServerShowV254Test-608938657 tempest-ServerShowV254Test-608938657-project-member] Waiting for the task: (returnval){ [ 659.188592] env[61806]: value = "task-1294481" [ 659.188592] env[61806]: _type = "Task" [ 659.188592] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 659.197839] env[61806]: DEBUG oslo_vmware.api [None req-fe158576-4234-4a6b-b48d-5ca8bb977c40 tempest-ServerShowV254Test-608938657 tempest-ServerShowV254Test-608938657-project-member] Task: {'id': task-1294481, 'name': Rename_Task} progress is 5%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 659.212720] env[61806]: INFO nova.compute.manager [None req-f8b2f38e-90f5-4b03-89bc-ca1926adb40c tempest-FloatingIPsAssociationTestJSON-35718468 tempest-FloatingIPsAssociationTestJSON-35718468-project-member] [instance: 88cb0961-ad34-4ffc-8fb2-6034f08ee740] Took 1.02 seconds to deallocate network for instance. [ 659.513242] env[61806]: DEBUG oslo_concurrency.lockutils [None req-b4447f02-a3c7-4b6e-89ba-6bc34ec03205 tempest-InstanceActionsTestJSON-949444324 tempest-InstanceActionsTestJSON-949444324-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.518s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 659.513797] env[61806]: DEBUG nova.compute.manager [None req-b4447f02-a3c7-4b6e-89ba-6bc34ec03205 tempest-InstanceActionsTestJSON-949444324 tempest-InstanceActionsTestJSON-949444324-project-member] [instance: 2681766c-0468-4170-8c51-03389e2bd1ab] Start building networks asynchronously for instance. {{(pid=61806) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 659.517237] env[61806]: DEBUG oslo_concurrency.lockutils [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 16.770s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 659.653806] env[61806]: DEBUG oslo_vmware.api [None req-4803fdfc-9054-49cf-8525-e438672eeedd tempest-ServersAdmin275Test-1945525728 tempest-ServersAdmin275Test-1945525728-project-member] Task: {'id': task-1294480, 'name': PowerOnVM_Task} progress is 88%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 659.698216] env[61806]: DEBUG oslo_vmware.api [None req-fe158576-4234-4a6b-b48d-5ca8bb977c40 tempest-ServerShowV254Test-608938657 tempest-ServerShowV254Test-608938657-project-member] Task: {'id': task-1294481, 'name': Rename_Task, 'duration_secs': 0.149276} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 659.698499] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [None req-fe158576-4234-4a6b-b48d-5ca8bb977c40 tempest-ServerShowV254Test-608938657 tempest-ServerShowV254Test-608938657-project-member] [instance: 8016bafc-b287-4df1-bd4b-a00210177a86] Powering on the VM {{(pid=61806) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 659.698752] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-d19be341-0ae3-42b4-8b13-cd4f9a2498c3 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 659.705215] env[61806]: DEBUG oslo_vmware.api [None req-fe158576-4234-4a6b-b48d-5ca8bb977c40 tempest-ServerShowV254Test-608938657 tempest-ServerShowV254Test-608938657-project-member] Waiting for the task: (returnval){ [ 659.705215] env[61806]: value = "task-1294482" [ 659.705215] env[61806]: _type = "Task" [ 659.705215] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 659.714030] env[61806]: DEBUG oslo_vmware.api [None req-fe158576-4234-4a6b-b48d-5ca8bb977c40 tempest-ServerShowV254Test-608938657 tempest-ServerShowV254Test-608938657-project-member] Task: {'id': task-1294482, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 660.019736] env[61806]: DEBUG nova.compute.utils [None req-b4447f02-a3c7-4b6e-89ba-6bc34ec03205 tempest-InstanceActionsTestJSON-949444324 tempest-InstanceActionsTestJSON-949444324-project-member] Using /dev/sd instead of None {{(pid=61806) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 660.020612] env[61806]: DEBUG nova.compute.manager [None req-b4447f02-a3c7-4b6e-89ba-6bc34ec03205 tempest-InstanceActionsTestJSON-949444324 tempest-InstanceActionsTestJSON-949444324-project-member] [instance: 2681766c-0468-4170-8c51-03389e2bd1ab] Allocating IP information in the background. {{(pid=61806) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 660.020612] env[61806]: DEBUG nova.network.neutron [None req-b4447f02-a3c7-4b6e-89ba-6bc34ec03205 tempest-InstanceActionsTestJSON-949444324 tempest-InstanceActionsTestJSON-949444324-project-member] [instance: 2681766c-0468-4170-8c51-03389e2bd1ab] allocate_for_instance() {{(pid=61806) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 660.083261] env[61806]: DEBUG nova.policy [None req-b4447f02-a3c7-4b6e-89ba-6bc34ec03205 tempest-InstanceActionsTestJSON-949444324 tempest-InstanceActionsTestJSON-949444324-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '51c81a0812ce47cfa2979a8cec1828c8', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '619bbad56cc54f9db021bdd77a315322', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61806) authorize /opt/stack/nova/nova/policy.py:201}} [ 660.154216] env[61806]: DEBUG oslo_vmware.api [None req-4803fdfc-9054-49cf-8525-e438672eeedd tempest-ServersAdmin275Test-1945525728 tempest-ServersAdmin275Test-1945525728-project-member] Task: {'id': task-1294480, 'name': PowerOnVM_Task, 'duration_secs': 0.582543} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 660.154488] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [None req-4803fdfc-9054-49cf-8525-e438672eeedd tempest-ServersAdmin275Test-1945525728 tempest-ServersAdmin275Test-1945525728-project-member] [instance: 9db80c3c-e2c3-4034-a79a-d7be924c4ff6] Powered on the VM {{(pid=61806) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 660.154677] env[61806]: INFO nova.compute.manager [None req-4803fdfc-9054-49cf-8525-e438672eeedd tempest-ServersAdmin275Test-1945525728 tempest-ServersAdmin275Test-1945525728-project-member] [instance: 9db80c3c-e2c3-4034-a79a-d7be924c4ff6] Took 8.92 seconds to spawn the instance on the hypervisor. [ 660.154849] env[61806]: DEBUG nova.compute.manager [None req-4803fdfc-9054-49cf-8525-e438672eeedd tempest-ServersAdmin275Test-1945525728 tempest-ServersAdmin275Test-1945525728-project-member] [instance: 9db80c3c-e2c3-4034-a79a-d7be924c4ff6] Checking state {{(pid=61806) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 660.155702] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8364f35a-79da-43c6-99c0-ad9288ff6eb1 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 660.216673] env[61806]: DEBUG oslo_vmware.api [None req-fe158576-4234-4a6b-b48d-5ca8bb977c40 tempest-ServerShowV254Test-608938657 tempest-ServerShowV254Test-608938657-project-member] Task: {'id': task-1294482, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 660.252999] env[61806]: INFO nova.scheduler.client.report [None req-f8b2f38e-90f5-4b03-89bc-ca1926adb40c tempest-FloatingIPsAssociationTestJSON-35718468 tempest-FloatingIPsAssociationTestJSON-35718468-project-member] Deleted allocations for instance 88cb0961-ad34-4ffc-8fb2-6034f08ee740 [ 660.428325] env[61806]: DEBUG nova.network.neutron [None req-b4447f02-a3c7-4b6e-89ba-6bc34ec03205 tempest-InstanceActionsTestJSON-949444324 tempest-InstanceActionsTestJSON-949444324-project-member] [instance: 2681766c-0468-4170-8c51-03389e2bd1ab] Successfully created port: 59f4de97-5864-4576-82b3-a14d8b5224cd {{(pid=61806) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 660.527468] env[61806]: DEBUG nova.compute.manager [None req-b4447f02-a3c7-4b6e-89ba-6bc34ec03205 tempest-InstanceActionsTestJSON-949444324 tempest-InstanceActionsTestJSON-949444324-project-member] [instance: 2681766c-0468-4170-8c51-03389e2bd1ab] Start building block device mappings for instance. {{(pid=61806) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 660.559144] env[61806]: DEBUG nova.compute.resource_tracker [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Instance 8016bafc-b287-4df1-bd4b-a00210177a86 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61806) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 660.559250] env[61806]: DEBUG nova.compute.resource_tracker [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Instance e66c921c-320e-4347-96c9-27aa80ba7f41 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61806) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 660.560130] env[61806]: DEBUG nova.compute.resource_tracker [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Instance 56d26d08-3c6d-45e8-b238-6dbca6024561 actively managed on this compute host and has allocations in placement: {'resources': {'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61806) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 660.560130] env[61806]: DEBUG nova.compute.resource_tracker [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Instance 9db80c3c-e2c3-4034-a79a-d7be924c4ff6 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61806) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 660.560130] env[61806]: DEBUG nova.compute.resource_tracker [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Instance d5434dd2-206f-4a1e-b97b-bd6f691b6ddf actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61806) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 660.560130] env[61806]: DEBUG nova.compute.resource_tracker [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Instance 2681766c-0468-4170-8c51-03389e2bd1ab actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61806) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 660.674032] env[61806]: INFO nova.compute.manager [None req-4803fdfc-9054-49cf-8525-e438672eeedd tempest-ServersAdmin275Test-1945525728 tempest-ServersAdmin275Test-1945525728-project-member] [instance: 9db80c3c-e2c3-4034-a79a-d7be924c4ff6] Took 29.94 seconds to build instance. [ 660.721789] env[61806]: DEBUG oslo_vmware.api [None req-fe158576-4234-4a6b-b48d-5ca8bb977c40 tempest-ServerShowV254Test-608938657 tempest-ServerShowV254Test-608938657-project-member] Task: {'id': task-1294482, 'name': PowerOnVM_Task} progress is 100%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 660.763663] env[61806]: DEBUG oslo_concurrency.lockutils [None req-f8b2f38e-90f5-4b03-89bc-ca1926adb40c tempest-FloatingIPsAssociationTestJSON-35718468 tempest-FloatingIPsAssociationTestJSON-35718468-project-member] Lock "88cb0961-ad34-4ffc-8fb2-6034f08ee740" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 104.186s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 661.067940] env[61806]: DEBUG nova.compute.resource_tracker [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Instance 8d0cdc81-56ed-4a7a-8221-30bb906e16b2 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61806) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 661.107990] env[61806]: DEBUG oslo_concurrency.lockutils [None req-3d5b4b69-10ec-4d1a-aa3d-4dddf4155c99 tempest-ServerShowV247Test-1218443059 tempest-ServerShowV247Test-1218443059-project-member] Acquiring lock "7c3fd2ff-ebd5-454f-a743-7fbae1088941" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 661.108857] env[61806]: DEBUG oslo_concurrency.lockutils [None req-3d5b4b69-10ec-4d1a-aa3d-4dddf4155c99 tempest-ServerShowV247Test-1218443059 tempest-ServerShowV247Test-1218443059-project-member] Lock "7c3fd2ff-ebd5-454f-a743-7fbae1088941" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 661.137629] env[61806]: INFO nova.compute.manager [None req-6f137976-e023-4899-ab65-afdf6c94ed87 tempest-ServersAdmin275Test-1945525728 tempest-ServersAdmin275Test-1945525728-project-member] [instance: 9db80c3c-e2c3-4034-a79a-d7be924c4ff6] Rebuilding instance [ 661.178017] env[61806]: DEBUG oslo_concurrency.lockutils [None req-4803fdfc-9054-49cf-8525-e438672eeedd tempest-ServersAdmin275Test-1945525728 tempest-ServersAdmin275Test-1945525728-project-member] Lock "9db80c3c-e2c3-4034-a79a-d7be924c4ff6" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 99.013s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 661.180909] env[61806]: DEBUG nova.compute.manager [None req-6f137976-e023-4899-ab65-afdf6c94ed87 tempest-ServersAdmin275Test-1945525728 tempest-ServersAdmin275Test-1945525728-project-member] [instance: 9db80c3c-e2c3-4034-a79a-d7be924c4ff6] Checking state {{(pid=61806) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 661.181768] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fd61379a-9c4c-46bc-90ed-a589f9cd07c3 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 661.217955] env[61806]: DEBUG oslo_vmware.api [None req-fe158576-4234-4a6b-b48d-5ca8bb977c40 tempest-ServerShowV254Test-608938657 tempest-ServerShowV254Test-608938657-project-member] Task: {'id': task-1294482, 'name': PowerOnVM_Task, 'duration_secs': 1.031404} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 661.218110] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [None req-fe158576-4234-4a6b-b48d-5ca8bb977c40 tempest-ServerShowV254Test-608938657 tempest-ServerShowV254Test-608938657-project-member] [instance: 8016bafc-b287-4df1-bd4b-a00210177a86] Powered on the VM {{(pid=61806) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 661.218859] env[61806]: DEBUG nova.compute.manager [None req-fe158576-4234-4a6b-b48d-5ca8bb977c40 tempest-ServerShowV254Test-608938657 tempest-ServerShowV254Test-608938657-project-member] [instance: 8016bafc-b287-4df1-bd4b-a00210177a86] Checking state {{(pid=61806) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 661.221027] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-01d240ff-bc7e-42bc-9620-f2ac7a8ea8a6 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 661.266570] env[61806]: DEBUG nova.compute.manager [None req-0f0a69c6-2310-4272-abc1-520940b4bcf0 tempest-ServersTestFqdnHostnames-356624376 tempest-ServersTestFqdnHostnames-356624376-project-member] [instance: 46c34ed0-9bac-4803-8d5c-3a3346a64117] Starting instance... {{(pid=61806) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 661.424162] env[61806]: DEBUG nova.compute.manager [req-5fff640f-3b88-4e78-b13b-0713fc58a713 req-22c1dd97-3ae6-4a65-8027-989de20291a4 service nova] [instance: 2681766c-0468-4170-8c51-03389e2bd1ab] Received event network-changed-59f4de97-5864-4576-82b3-a14d8b5224cd {{(pid=61806) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 661.424470] env[61806]: DEBUG nova.compute.manager [req-5fff640f-3b88-4e78-b13b-0713fc58a713 req-22c1dd97-3ae6-4a65-8027-989de20291a4 service nova] [instance: 2681766c-0468-4170-8c51-03389e2bd1ab] Refreshing instance network info cache due to event network-changed-59f4de97-5864-4576-82b3-a14d8b5224cd. {{(pid=61806) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 661.424585] env[61806]: DEBUG oslo_concurrency.lockutils [req-5fff640f-3b88-4e78-b13b-0713fc58a713 req-22c1dd97-3ae6-4a65-8027-989de20291a4 service nova] Acquiring lock "refresh_cache-2681766c-0468-4170-8c51-03389e2bd1ab" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 661.424732] env[61806]: DEBUG oslo_concurrency.lockutils [req-5fff640f-3b88-4e78-b13b-0713fc58a713 req-22c1dd97-3ae6-4a65-8027-989de20291a4 service nova] Acquired lock "refresh_cache-2681766c-0468-4170-8c51-03389e2bd1ab" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 661.424893] env[61806]: DEBUG nova.network.neutron [req-5fff640f-3b88-4e78-b13b-0713fc58a713 req-22c1dd97-3ae6-4a65-8027-989de20291a4 service nova] [instance: 2681766c-0468-4170-8c51-03389e2bd1ab] Refreshing network info cache for port 59f4de97-5864-4576-82b3-a14d8b5224cd {{(pid=61806) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 661.450514] env[61806]: DEBUG oslo_concurrency.lockutils [None req-6a30dd3c-3e08-4933-b4cc-6fae094056e7 tempest-ServerShowV247Test-1218443059 tempest-ServerShowV247Test-1218443059-project-member] Acquiring lock "4febd093-9f2b-494e-b175-e4693b0e3e0d" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 661.450897] env[61806]: DEBUG oslo_concurrency.lockutils [None req-6a30dd3c-3e08-4933-b4cc-6fae094056e7 tempest-ServerShowV247Test-1218443059 tempest-ServerShowV247Test-1218443059-project-member] Lock "4febd093-9f2b-494e-b175-e4693b0e3e0d" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 661.540661] env[61806]: DEBUG nova.compute.manager [None req-b4447f02-a3c7-4b6e-89ba-6bc34ec03205 tempest-InstanceActionsTestJSON-949444324 tempest-InstanceActionsTestJSON-949444324-project-member] [instance: 2681766c-0468-4170-8c51-03389e2bd1ab] Start spawning the instance on the hypervisor. {{(pid=61806) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 661.565202] env[61806]: ERROR nova.compute.manager [None req-b4447f02-a3c7-4b6e-89ba-6bc34ec03205 tempest-InstanceActionsTestJSON-949444324 tempest-InstanceActionsTestJSON-949444324-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 59f4de97-5864-4576-82b3-a14d8b5224cd, please check neutron logs for more information. [ 661.565202] env[61806]: ERROR nova.compute.manager Traceback (most recent call last): [ 661.565202] env[61806]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 661.565202] env[61806]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 661.565202] env[61806]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 661.565202] env[61806]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 661.565202] env[61806]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 661.565202] env[61806]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 661.565202] env[61806]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 661.565202] env[61806]: ERROR nova.compute.manager self.force_reraise() [ 661.565202] env[61806]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 661.565202] env[61806]: ERROR nova.compute.manager raise self.value [ 661.565202] env[61806]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 661.565202] env[61806]: ERROR nova.compute.manager updated_port = self._update_port( [ 661.565202] env[61806]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 661.565202] env[61806]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 661.565916] env[61806]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 661.565916] env[61806]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 661.565916] env[61806]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 59f4de97-5864-4576-82b3-a14d8b5224cd, please check neutron logs for more information. [ 661.565916] env[61806]: ERROR nova.compute.manager [ 661.565916] env[61806]: Traceback (most recent call last): [ 661.565916] env[61806]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 661.565916] env[61806]: listener.cb(fileno) [ 661.565916] env[61806]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 661.565916] env[61806]: result = function(*args, **kwargs) [ 661.565916] env[61806]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 661.565916] env[61806]: return func(*args, **kwargs) [ 661.565916] env[61806]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 661.565916] env[61806]: raise e [ 661.565916] env[61806]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 661.565916] env[61806]: nwinfo = self.network_api.allocate_for_instance( [ 661.565916] env[61806]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 661.565916] env[61806]: created_port_ids = self._update_ports_for_instance( [ 661.565916] env[61806]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 661.565916] env[61806]: with excutils.save_and_reraise_exception(): [ 661.565916] env[61806]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 661.565916] env[61806]: self.force_reraise() [ 661.565916] env[61806]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 661.565916] env[61806]: raise self.value [ 661.565916] env[61806]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 661.565916] env[61806]: updated_port = self._update_port( [ 661.565916] env[61806]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 661.565916] env[61806]: _ensure_no_port_binding_failure(port) [ 661.565916] env[61806]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 661.565916] env[61806]: raise exception.PortBindingFailed(port_id=port['id']) [ 661.567239] env[61806]: nova.exception.PortBindingFailed: Binding failed for port 59f4de97-5864-4576-82b3-a14d8b5224cd, please check neutron logs for more information. [ 661.567239] env[61806]: Removing descriptor: 14 [ 661.570894] env[61806]: DEBUG nova.compute.resource_tracker [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Instance b7d55e87-95db-4c45-9417-bbece4fbb98d has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61806) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 661.576392] env[61806]: DEBUG nova.virt.hardware [None req-b4447f02-a3c7-4b6e-89ba-6bc34ec03205 tempest-InstanceActionsTestJSON-949444324 tempest-InstanceActionsTestJSON-949444324-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-15T18:05:10Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-15T18:04:52Z,direct_url=,disk_format='vmdk',id=ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='84ba83fce2c349a6988173c5d6fc3b07',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-15T18:04:53Z,virtual_size=,visibility=), allow threads: False {{(pid=61806) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 661.576564] env[61806]: DEBUG nova.virt.hardware [None req-b4447f02-a3c7-4b6e-89ba-6bc34ec03205 tempest-InstanceActionsTestJSON-949444324 tempest-InstanceActionsTestJSON-949444324-project-member] Flavor limits 0:0:0 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 661.576709] env[61806]: DEBUG nova.virt.hardware [None req-b4447f02-a3c7-4b6e-89ba-6bc34ec03205 tempest-InstanceActionsTestJSON-949444324 tempest-InstanceActionsTestJSON-949444324-project-member] Image limits 0:0:0 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 661.576948] env[61806]: DEBUG nova.virt.hardware [None req-b4447f02-a3c7-4b6e-89ba-6bc34ec03205 tempest-InstanceActionsTestJSON-949444324 tempest-InstanceActionsTestJSON-949444324-project-member] Flavor pref 0:0:0 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 661.579918] env[61806]: DEBUG nova.virt.hardware [None req-b4447f02-a3c7-4b6e-89ba-6bc34ec03205 tempest-InstanceActionsTestJSON-949444324 tempest-InstanceActionsTestJSON-949444324-project-member] Image pref 0:0:0 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 661.579918] env[61806]: DEBUG nova.virt.hardware [None req-b4447f02-a3c7-4b6e-89ba-6bc34ec03205 tempest-InstanceActionsTestJSON-949444324 tempest-InstanceActionsTestJSON-949444324-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 661.579918] env[61806]: DEBUG nova.virt.hardware [None req-b4447f02-a3c7-4b6e-89ba-6bc34ec03205 tempest-InstanceActionsTestJSON-949444324 tempest-InstanceActionsTestJSON-949444324-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61806) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 661.579918] env[61806]: DEBUG nova.virt.hardware [None req-b4447f02-a3c7-4b6e-89ba-6bc34ec03205 tempest-InstanceActionsTestJSON-949444324 tempest-InstanceActionsTestJSON-949444324-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61806) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 661.579918] env[61806]: DEBUG nova.virt.hardware [None req-b4447f02-a3c7-4b6e-89ba-6bc34ec03205 tempest-InstanceActionsTestJSON-949444324 tempest-InstanceActionsTestJSON-949444324-project-member] Got 1 possible topologies {{(pid=61806) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 661.580142] env[61806]: DEBUG nova.virt.hardware [None req-b4447f02-a3c7-4b6e-89ba-6bc34ec03205 tempest-InstanceActionsTestJSON-949444324 tempest-InstanceActionsTestJSON-949444324-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61806) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 661.580142] env[61806]: DEBUG nova.virt.hardware [None req-b4447f02-a3c7-4b6e-89ba-6bc34ec03205 tempest-InstanceActionsTestJSON-949444324 tempest-InstanceActionsTestJSON-949444324-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61806) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 661.580450] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-886b4c7b-fe20-41f4-8739-d57ab759d6f9 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 661.590839] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d2e2e3d3-96bf-48bd-ad03-eb5c3a223e45 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 661.611169] env[61806]: ERROR nova.compute.manager [None req-b4447f02-a3c7-4b6e-89ba-6bc34ec03205 tempest-InstanceActionsTestJSON-949444324 tempest-InstanceActionsTestJSON-949444324-project-member] [instance: 2681766c-0468-4170-8c51-03389e2bd1ab] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 59f4de97-5864-4576-82b3-a14d8b5224cd, please check neutron logs for more information. [ 661.611169] env[61806]: ERROR nova.compute.manager [instance: 2681766c-0468-4170-8c51-03389e2bd1ab] Traceback (most recent call last): [ 661.611169] env[61806]: ERROR nova.compute.manager [instance: 2681766c-0468-4170-8c51-03389e2bd1ab] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 661.611169] env[61806]: ERROR nova.compute.manager [instance: 2681766c-0468-4170-8c51-03389e2bd1ab] yield resources [ 661.611169] env[61806]: ERROR nova.compute.manager [instance: 2681766c-0468-4170-8c51-03389e2bd1ab] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 661.611169] env[61806]: ERROR nova.compute.manager [instance: 2681766c-0468-4170-8c51-03389e2bd1ab] self.driver.spawn(context, instance, image_meta, [ 661.611169] env[61806]: ERROR nova.compute.manager [instance: 2681766c-0468-4170-8c51-03389e2bd1ab] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 661.611169] env[61806]: ERROR nova.compute.manager [instance: 2681766c-0468-4170-8c51-03389e2bd1ab] self._vmops.spawn(context, instance, image_meta, injected_files, [ 661.611169] env[61806]: ERROR nova.compute.manager [instance: 2681766c-0468-4170-8c51-03389e2bd1ab] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 661.611169] env[61806]: ERROR nova.compute.manager [instance: 2681766c-0468-4170-8c51-03389e2bd1ab] vm_ref = self.build_virtual_machine(instance, [ 661.611169] env[61806]: ERROR nova.compute.manager [instance: 2681766c-0468-4170-8c51-03389e2bd1ab] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 661.611545] env[61806]: ERROR nova.compute.manager [instance: 2681766c-0468-4170-8c51-03389e2bd1ab] vif_infos = vmwarevif.get_vif_info(self._session, [ 661.611545] env[61806]: ERROR nova.compute.manager [instance: 2681766c-0468-4170-8c51-03389e2bd1ab] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 661.611545] env[61806]: ERROR nova.compute.manager [instance: 2681766c-0468-4170-8c51-03389e2bd1ab] for vif in network_info: [ 661.611545] env[61806]: ERROR nova.compute.manager [instance: 2681766c-0468-4170-8c51-03389e2bd1ab] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 661.611545] env[61806]: ERROR nova.compute.manager [instance: 2681766c-0468-4170-8c51-03389e2bd1ab] return self._sync_wrapper(fn, *args, **kwargs) [ 661.611545] env[61806]: ERROR nova.compute.manager [instance: 2681766c-0468-4170-8c51-03389e2bd1ab] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 661.611545] env[61806]: ERROR nova.compute.manager [instance: 2681766c-0468-4170-8c51-03389e2bd1ab] self.wait() [ 661.611545] env[61806]: ERROR nova.compute.manager [instance: 2681766c-0468-4170-8c51-03389e2bd1ab] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 661.611545] env[61806]: ERROR nova.compute.manager [instance: 2681766c-0468-4170-8c51-03389e2bd1ab] self[:] = self._gt.wait() [ 661.611545] env[61806]: ERROR nova.compute.manager [instance: 2681766c-0468-4170-8c51-03389e2bd1ab] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 661.611545] env[61806]: ERROR nova.compute.manager [instance: 2681766c-0468-4170-8c51-03389e2bd1ab] return self._exit_event.wait() [ 661.611545] env[61806]: ERROR nova.compute.manager [instance: 2681766c-0468-4170-8c51-03389e2bd1ab] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 661.611545] env[61806]: ERROR nova.compute.manager [instance: 2681766c-0468-4170-8c51-03389e2bd1ab] current.throw(*self._exc) [ 661.611932] env[61806]: ERROR nova.compute.manager [instance: 2681766c-0468-4170-8c51-03389e2bd1ab] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 661.611932] env[61806]: ERROR nova.compute.manager [instance: 2681766c-0468-4170-8c51-03389e2bd1ab] result = function(*args, **kwargs) [ 661.611932] env[61806]: ERROR nova.compute.manager [instance: 2681766c-0468-4170-8c51-03389e2bd1ab] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 661.611932] env[61806]: ERROR nova.compute.manager [instance: 2681766c-0468-4170-8c51-03389e2bd1ab] return func(*args, **kwargs) [ 661.611932] env[61806]: ERROR nova.compute.manager [instance: 2681766c-0468-4170-8c51-03389e2bd1ab] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 661.611932] env[61806]: ERROR nova.compute.manager [instance: 2681766c-0468-4170-8c51-03389e2bd1ab] raise e [ 661.611932] env[61806]: ERROR nova.compute.manager [instance: 2681766c-0468-4170-8c51-03389e2bd1ab] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 661.611932] env[61806]: ERROR nova.compute.manager [instance: 2681766c-0468-4170-8c51-03389e2bd1ab] nwinfo = self.network_api.allocate_for_instance( [ 661.611932] env[61806]: ERROR nova.compute.manager [instance: 2681766c-0468-4170-8c51-03389e2bd1ab] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 661.611932] env[61806]: ERROR nova.compute.manager [instance: 2681766c-0468-4170-8c51-03389e2bd1ab] created_port_ids = self._update_ports_for_instance( [ 661.611932] env[61806]: ERROR nova.compute.manager [instance: 2681766c-0468-4170-8c51-03389e2bd1ab] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 661.611932] env[61806]: ERROR nova.compute.manager [instance: 2681766c-0468-4170-8c51-03389e2bd1ab] with excutils.save_and_reraise_exception(): [ 661.611932] env[61806]: ERROR nova.compute.manager [instance: 2681766c-0468-4170-8c51-03389e2bd1ab] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 661.612287] env[61806]: ERROR nova.compute.manager [instance: 2681766c-0468-4170-8c51-03389e2bd1ab] self.force_reraise() [ 661.612287] env[61806]: ERROR nova.compute.manager [instance: 2681766c-0468-4170-8c51-03389e2bd1ab] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 661.612287] env[61806]: ERROR nova.compute.manager [instance: 2681766c-0468-4170-8c51-03389e2bd1ab] raise self.value [ 661.612287] env[61806]: ERROR nova.compute.manager [instance: 2681766c-0468-4170-8c51-03389e2bd1ab] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 661.612287] env[61806]: ERROR nova.compute.manager [instance: 2681766c-0468-4170-8c51-03389e2bd1ab] updated_port = self._update_port( [ 661.612287] env[61806]: ERROR nova.compute.manager [instance: 2681766c-0468-4170-8c51-03389e2bd1ab] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 661.612287] env[61806]: ERROR nova.compute.manager [instance: 2681766c-0468-4170-8c51-03389e2bd1ab] _ensure_no_port_binding_failure(port) [ 661.612287] env[61806]: ERROR nova.compute.manager [instance: 2681766c-0468-4170-8c51-03389e2bd1ab] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 661.612287] env[61806]: ERROR nova.compute.manager [instance: 2681766c-0468-4170-8c51-03389e2bd1ab] raise exception.PortBindingFailed(port_id=port['id']) [ 661.612287] env[61806]: ERROR nova.compute.manager [instance: 2681766c-0468-4170-8c51-03389e2bd1ab] nova.exception.PortBindingFailed: Binding failed for port 59f4de97-5864-4576-82b3-a14d8b5224cd, please check neutron logs for more information. [ 661.612287] env[61806]: ERROR nova.compute.manager [instance: 2681766c-0468-4170-8c51-03389e2bd1ab] [ 661.612287] env[61806]: INFO nova.compute.manager [None req-b4447f02-a3c7-4b6e-89ba-6bc34ec03205 tempest-InstanceActionsTestJSON-949444324 tempest-InstanceActionsTestJSON-949444324-project-member] [instance: 2681766c-0468-4170-8c51-03389e2bd1ab] Terminating instance [ 661.622022] env[61806]: DEBUG oslo_concurrency.lockutils [None req-b4447f02-a3c7-4b6e-89ba-6bc34ec03205 tempest-InstanceActionsTestJSON-949444324 tempest-InstanceActionsTestJSON-949444324-project-member] Acquiring lock "refresh_cache-2681766c-0468-4170-8c51-03389e2bd1ab" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 661.685043] env[61806]: DEBUG nova.compute.manager [None req-05ef4399-8310-428c-ba55-af11abf4e445 tempest-SecurityGroupsTestJSON-1213485581 tempest-SecurityGroupsTestJSON-1213485581-project-member] [instance: 294adb58-f23e-4510-a25a-de6b909e3189] Starting instance... {{(pid=61806) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 661.695335] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [None req-6f137976-e023-4899-ab65-afdf6c94ed87 tempest-ServersAdmin275Test-1945525728 tempest-ServersAdmin275Test-1945525728-project-member] [instance: 9db80c3c-e2c3-4034-a79a-d7be924c4ff6] Powering off the VM {{(pid=61806) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 661.699134] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-fdae51c8-1ae3-4f6d-a028-afeb1cec1b85 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 661.703962] env[61806]: DEBUG oslo_vmware.api [None req-6f137976-e023-4899-ab65-afdf6c94ed87 tempest-ServersAdmin275Test-1945525728 tempest-ServersAdmin275Test-1945525728-project-member] Waiting for the task: (returnval){ [ 661.703962] env[61806]: value = "task-1294483" [ 661.703962] env[61806]: _type = "Task" [ 661.703962] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 661.718358] env[61806]: DEBUG oslo_vmware.api [None req-6f137976-e023-4899-ab65-afdf6c94ed87 tempest-ServersAdmin275Test-1945525728 tempest-ServersAdmin275Test-1945525728-project-member] Task: {'id': task-1294483, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 661.739132] env[61806]: DEBUG oslo_concurrency.lockutils [None req-fe158576-4234-4a6b-b48d-5ca8bb977c40 tempest-ServerShowV254Test-608938657 tempest-ServerShowV254Test-608938657-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 661.791495] env[61806]: DEBUG oslo_concurrency.lockutils [None req-0f0a69c6-2310-4272-abc1-520940b4bcf0 tempest-ServersTestFqdnHostnames-356624376 tempest-ServersTestFqdnHostnames-356624376-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 661.946129] env[61806]: DEBUG nova.network.neutron [req-5fff640f-3b88-4e78-b13b-0713fc58a713 req-22c1dd97-3ae6-4a65-8027-989de20291a4 service nova] [instance: 2681766c-0468-4170-8c51-03389e2bd1ab] Instance cache missing network info. {{(pid=61806) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 662.055715] env[61806]: DEBUG nova.network.neutron [req-5fff640f-3b88-4e78-b13b-0713fc58a713 req-22c1dd97-3ae6-4a65-8027-989de20291a4 service nova] [instance: 2681766c-0468-4170-8c51-03389e2bd1ab] Updating instance_info_cache with network_info: [] {{(pid=61806) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 662.075036] env[61806]: DEBUG nova.compute.resource_tracker [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Instance eca2e3da-b38c-482e-b9b6-4ad75d83caa1 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61806) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 662.119587] env[61806]: DEBUG oslo_concurrency.lockutils [None req-6dff00ec-9995-4616-9033-47ec4f3592c9 tempest-ServerShowV254Test-608938657 tempest-ServerShowV254Test-608938657-project-member] Acquiring lock "8016bafc-b287-4df1-bd4b-a00210177a86" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 662.119939] env[61806]: DEBUG oslo_concurrency.lockutils [None req-6dff00ec-9995-4616-9033-47ec4f3592c9 tempest-ServerShowV254Test-608938657 tempest-ServerShowV254Test-608938657-project-member] Lock "8016bafc-b287-4df1-bd4b-a00210177a86" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 662.120186] env[61806]: DEBUG oslo_concurrency.lockutils [None req-6dff00ec-9995-4616-9033-47ec4f3592c9 tempest-ServerShowV254Test-608938657 tempest-ServerShowV254Test-608938657-project-member] Acquiring lock "8016bafc-b287-4df1-bd4b-a00210177a86-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 662.120379] env[61806]: DEBUG oslo_concurrency.lockutils [None req-6dff00ec-9995-4616-9033-47ec4f3592c9 tempest-ServerShowV254Test-608938657 tempest-ServerShowV254Test-608938657-project-member] Lock "8016bafc-b287-4df1-bd4b-a00210177a86-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 662.120741] env[61806]: DEBUG oslo_concurrency.lockutils [None req-6dff00ec-9995-4616-9033-47ec4f3592c9 tempest-ServerShowV254Test-608938657 tempest-ServerShowV254Test-608938657-project-member] Lock "8016bafc-b287-4df1-bd4b-a00210177a86-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 662.123029] env[61806]: INFO nova.compute.manager [None req-6dff00ec-9995-4616-9033-47ec4f3592c9 tempest-ServerShowV254Test-608938657 tempest-ServerShowV254Test-608938657-project-member] [instance: 8016bafc-b287-4df1-bd4b-a00210177a86] Terminating instance [ 662.125029] env[61806]: DEBUG oslo_concurrency.lockutils [None req-6dff00ec-9995-4616-9033-47ec4f3592c9 tempest-ServerShowV254Test-608938657 tempest-ServerShowV254Test-608938657-project-member] Acquiring lock "refresh_cache-8016bafc-b287-4df1-bd4b-a00210177a86" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 662.125029] env[61806]: DEBUG oslo_concurrency.lockutils [None req-6dff00ec-9995-4616-9033-47ec4f3592c9 tempest-ServerShowV254Test-608938657 tempest-ServerShowV254Test-608938657-project-member] Acquired lock "refresh_cache-8016bafc-b287-4df1-bd4b-a00210177a86" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 662.125158] env[61806]: DEBUG nova.network.neutron [None req-6dff00ec-9995-4616-9033-47ec4f3592c9 tempest-ServerShowV254Test-608938657 tempest-ServerShowV254Test-608938657-project-member] [instance: 8016bafc-b287-4df1-bd4b-a00210177a86] Building network info cache for instance {{(pid=61806) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 662.211236] env[61806]: DEBUG oslo_concurrency.lockutils [None req-05ef4399-8310-428c-ba55-af11abf4e445 tempest-SecurityGroupsTestJSON-1213485581 tempest-SecurityGroupsTestJSON-1213485581-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 662.216925] env[61806]: DEBUG oslo_vmware.api [None req-6f137976-e023-4899-ab65-afdf6c94ed87 tempest-ServersAdmin275Test-1945525728 tempest-ServersAdmin275Test-1945525728-project-member] Task: {'id': task-1294483, 'name': PowerOffVM_Task, 'duration_secs': 0.22723} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 662.217217] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [None req-6f137976-e023-4899-ab65-afdf6c94ed87 tempest-ServersAdmin275Test-1945525728 tempest-ServersAdmin275Test-1945525728-project-member] [instance: 9db80c3c-e2c3-4034-a79a-d7be924c4ff6] Powered off the VM {{(pid=61806) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 662.217441] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-6f137976-e023-4899-ab65-afdf6c94ed87 tempest-ServersAdmin275Test-1945525728 tempest-ServersAdmin275Test-1945525728-project-member] [instance: 9db80c3c-e2c3-4034-a79a-d7be924c4ff6] Destroying instance {{(pid=61806) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 662.218195] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fa32b31f-7cc8-46f5-9c6e-c9b63682e17d {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 662.224688] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-6f137976-e023-4899-ab65-afdf6c94ed87 tempest-ServersAdmin275Test-1945525728 tempest-ServersAdmin275Test-1945525728-project-member] [instance: 9db80c3c-e2c3-4034-a79a-d7be924c4ff6] Unregistering the VM {{(pid=61806) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 662.224924] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-8b7d014c-0414-42fc-86fb-13699970a01a {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 662.250150] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-6f137976-e023-4899-ab65-afdf6c94ed87 tempest-ServersAdmin275Test-1945525728 tempest-ServersAdmin275Test-1945525728-project-member] [instance: 9db80c3c-e2c3-4034-a79a-d7be924c4ff6] Unregistered the VM {{(pid=61806) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 662.250375] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-6f137976-e023-4899-ab65-afdf6c94ed87 tempest-ServersAdmin275Test-1945525728 tempest-ServersAdmin275Test-1945525728-project-member] [instance: 9db80c3c-e2c3-4034-a79a-d7be924c4ff6] Deleting contents of the VM from datastore datastore2 {{(pid=61806) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 662.250707] env[61806]: DEBUG nova.virt.vmwareapi.ds_util [None req-6f137976-e023-4899-ab65-afdf6c94ed87 tempest-ServersAdmin275Test-1945525728 tempest-ServersAdmin275Test-1945525728-project-member] Deleting the datastore file [datastore2] 9db80c3c-e2c3-4034-a79a-d7be924c4ff6 {{(pid=61806) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 662.251062] env[61806]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-97e40761-4671-4032-b9a0-04160862d1e8 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 662.257816] env[61806]: DEBUG oslo_vmware.api [None req-6f137976-e023-4899-ab65-afdf6c94ed87 tempest-ServersAdmin275Test-1945525728 tempest-ServersAdmin275Test-1945525728-project-member] Waiting for the task: (returnval){ [ 662.257816] env[61806]: value = "task-1294485" [ 662.257816] env[61806]: _type = "Task" [ 662.257816] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 662.269359] env[61806]: DEBUG oslo_vmware.api [None req-6f137976-e023-4899-ab65-afdf6c94ed87 tempest-ServersAdmin275Test-1945525728 tempest-ServersAdmin275Test-1945525728-project-member] Task: {'id': task-1294485, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 662.563128] env[61806]: DEBUG oslo_concurrency.lockutils [req-5fff640f-3b88-4e78-b13b-0713fc58a713 req-22c1dd97-3ae6-4a65-8027-989de20291a4 service nova] Releasing lock "refresh_cache-2681766c-0468-4170-8c51-03389e2bd1ab" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 662.564044] env[61806]: DEBUG oslo_concurrency.lockutils [None req-b4447f02-a3c7-4b6e-89ba-6bc34ec03205 tempest-InstanceActionsTestJSON-949444324 tempest-InstanceActionsTestJSON-949444324-project-member] Acquired lock "refresh_cache-2681766c-0468-4170-8c51-03389e2bd1ab" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 662.564044] env[61806]: DEBUG nova.network.neutron [None req-b4447f02-a3c7-4b6e-89ba-6bc34ec03205 tempest-InstanceActionsTestJSON-949444324 tempest-InstanceActionsTestJSON-949444324-project-member] [instance: 2681766c-0468-4170-8c51-03389e2bd1ab] Building network info cache for instance {{(pid=61806) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 662.578051] env[61806]: DEBUG nova.compute.resource_tracker [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Instance 294f0a37-7589-405d-a324-13099c80f458 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61806) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 662.647685] env[61806]: DEBUG nova.network.neutron [None req-6dff00ec-9995-4616-9033-47ec4f3592c9 tempest-ServerShowV254Test-608938657 tempest-ServerShowV254Test-608938657-project-member] [instance: 8016bafc-b287-4df1-bd4b-a00210177a86] Instance cache missing network info. {{(pid=61806) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 662.768698] env[61806]: DEBUG oslo_vmware.api [None req-6f137976-e023-4899-ab65-afdf6c94ed87 tempest-ServersAdmin275Test-1945525728 tempest-ServersAdmin275Test-1945525728-project-member] Task: {'id': task-1294485, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.098935} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 662.768698] env[61806]: DEBUG nova.virt.vmwareapi.ds_util [None req-6f137976-e023-4899-ab65-afdf6c94ed87 tempest-ServersAdmin275Test-1945525728 tempest-ServersAdmin275Test-1945525728-project-member] Deleted the datastore file {{(pid=61806) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 662.768698] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-6f137976-e023-4899-ab65-afdf6c94ed87 tempest-ServersAdmin275Test-1945525728 tempest-ServersAdmin275Test-1945525728-project-member] [instance: 9db80c3c-e2c3-4034-a79a-d7be924c4ff6] Deleted contents of the VM from datastore datastore2 {{(pid=61806) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 662.768698] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-6f137976-e023-4899-ab65-afdf6c94ed87 tempest-ServersAdmin275Test-1945525728 tempest-ServersAdmin275Test-1945525728-project-member] [instance: 9db80c3c-e2c3-4034-a79a-d7be924c4ff6] Instance destroyed {{(pid=61806) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 662.906776] env[61806]: DEBUG nova.network.neutron [None req-6dff00ec-9995-4616-9033-47ec4f3592c9 tempest-ServerShowV254Test-608938657 tempest-ServerShowV254Test-608938657-project-member] [instance: 8016bafc-b287-4df1-bd4b-a00210177a86] Updating instance_info_cache with network_info: [] {{(pid=61806) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 663.080311] env[61806]: DEBUG nova.compute.resource_tracker [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Instance 46c34ed0-9bac-4803-8d5c-3a3346a64117 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61806) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 663.087367] env[61806]: DEBUG nova.network.neutron [None req-b4447f02-a3c7-4b6e-89ba-6bc34ec03205 tempest-InstanceActionsTestJSON-949444324 tempest-InstanceActionsTestJSON-949444324-project-member] [instance: 2681766c-0468-4170-8c51-03389e2bd1ab] Instance cache missing network info. {{(pid=61806) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 663.174624] env[61806]: DEBUG nova.network.neutron [None req-b4447f02-a3c7-4b6e-89ba-6bc34ec03205 tempest-InstanceActionsTestJSON-949444324 tempest-InstanceActionsTestJSON-949444324-project-member] [instance: 2681766c-0468-4170-8c51-03389e2bd1ab] Updating instance_info_cache with network_info: [] {{(pid=61806) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 663.411025] env[61806]: DEBUG oslo_concurrency.lockutils [None req-6dff00ec-9995-4616-9033-47ec4f3592c9 tempest-ServerShowV254Test-608938657 tempest-ServerShowV254Test-608938657-project-member] Releasing lock "refresh_cache-8016bafc-b287-4df1-bd4b-a00210177a86" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 663.411476] env[61806]: DEBUG nova.compute.manager [None req-6dff00ec-9995-4616-9033-47ec4f3592c9 tempest-ServerShowV254Test-608938657 tempest-ServerShowV254Test-608938657-project-member] [instance: 8016bafc-b287-4df1-bd4b-a00210177a86] Start destroying the instance on the hypervisor. {{(pid=61806) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 663.411819] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-6dff00ec-9995-4616-9033-47ec4f3592c9 tempest-ServerShowV254Test-608938657 tempest-ServerShowV254Test-608938657-project-member] [instance: 8016bafc-b287-4df1-bd4b-a00210177a86] Destroying instance {{(pid=61806) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 663.412584] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-df9e526a-2e7e-4d5a-acc5-9f294df423ee {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 663.420224] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [None req-6dff00ec-9995-4616-9033-47ec4f3592c9 tempest-ServerShowV254Test-608938657 tempest-ServerShowV254Test-608938657-project-member] [instance: 8016bafc-b287-4df1-bd4b-a00210177a86] Powering off the VM {{(pid=61806) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 663.420524] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-25503c41-d935-4102-a423-3f2bb25274ba {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 663.426978] env[61806]: DEBUG oslo_vmware.api [None req-6dff00ec-9995-4616-9033-47ec4f3592c9 tempest-ServerShowV254Test-608938657 tempest-ServerShowV254Test-608938657-project-member] Waiting for the task: (returnval){ [ 663.426978] env[61806]: value = "task-1294486" [ 663.426978] env[61806]: _type = "Task" [ 663.426978] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 663.437304] env[61806]: DEBUG oslo_vmware.api [None req-6dff00ec-9995-4616-9033-47ec4f3592c9 tempest-ServerShowV254Test-608938657 tempest-ServerShowV254Test-608938657-project-member] Task: {'id': task-1294486, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 663.486154] env[61806]: DEBUG nova.compute.manager [req-a4aebd36-10c0-41f4-be16-f38a6319ecaf req-46b90844-32b2-4d65-b1f9-f52907ebc99a service nova] [instance: 2681766c-0468-4170-8c51-03389e2bd1ab] Received event network-vif-deleted-59f4de97-5864-4576-82b3-a14d8b5224cd {{(pid=61806) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 663.583927] env[61806]: DEBUG nova.compute.resource_tracker [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Instance 294adb58-f23e-4510-a25a-de6b909e3189 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61806) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 663.678420] env[61806]: DEBUG oslo_concurrency.lockutils [None req-b4447f02-a3c7-4b6e-89ba-6bc34ec03205 tempest-InstanceActionsTestJSON-949444324 tempest-InstanceActionsTestJSON-949444324-project-member] Releasing lock "refresh_cache-2681766c-0468-4170-8c51-03389e2bd1ab" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 663.678962] env[61806]: DEBUG nova.compute.manager [None req-b4447f02-a3c7-4b6e-89ba-6bc34ec03205 tempest-InstanceActionsTestJSON-949444324 tempest-InstanceActionsTestJSON-949444324-project-member] [instance: 2681766c-0468-4170-8c51-03389e2bd1ab] Start destroying the instance on the hypervisor. {{(pid=61806) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 663.679231] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-b4447f02-a3c7-4b6e-89ba-6bc34ec03205 tempest-InstanceActionsTestJSON-949444324 tempest-InstanceActionsTestJSON-949444324-project-member] [instance: 2681766c-0468-4170-8c51-03389e2bd1ab] Destroying instance {{(pid=61806) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 663.679482] env[61806]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-f1102700-4c6e-4fa8-9a08-b8c37ba815f5 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 663.688085] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a02bff63-c62a-4c47-b329-efe92f4d1279 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 663.709072] env[61806]: WARNING nova.virt.vmwareapi.vmops [None req-b4447f02-a3c7-4b6e-89ba-6bc34ec03205 tempest-InstanceActionsTestJSON-949444324 tempest-InstanceActionsTestJSON-949444324-project-member] [instance: 2681766c-0468-4170-8c51-03389e2bd1ab] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 2681766c-0468-4170-8c51-03389e2bd1ab could not be found. [ 663.709468] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-b4447f02-a3c7-4b6e-89ba-6bc34ec03205 tempest-InstanceActionsTestJSON-949444324 tempest-InstanceActionsTestJSON-949444324-project-member] [instance: 2681766c-0468-4170-8c51-03389e2bd1ab] Instance destroyed {{(pid=61806) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 663.709468] env[61806]: INFO nova.compute.manager [None req-b4447f02-a3c7-4b6e-89ba-6bc34ec03205 tempest-InstanceActionsTestJSON-949444324 tempest-InstanceActionsTestJSON-949444324-project-member] [instance: 2681766c-0468-4170-8c51-03389e2bd1ab] Took 0.03 seconds to destroy the instance on the hypervisor. [ 663.709683] env[61806]: DEBUG oslo.service.loopingcall [None req-b4447f02-a3c7-4b6e-89ba-6bc34ec03205 tempest-InstanceActionsTestJSON-949444324 tempest-InstanceActionsTestJSON-949444324-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61806) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 663.709900] env[61806]: DEBUG nova.compute.manager [-] [instance: 2681766c-0468-4170-8c51-03389e2bd1ab] Deallocating network for instance {{(pid=61806) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 663.709992] env[61806]: DEBUG nova.network.neutron [-] [instance: 2681766c-0468-4170-8c51-03389e2bd1ab] deallocate_for_instance() {{(pid=61806) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 663.724558] env[61806]: DEBUG nova.network.neutron [-] [instance: 2681766c-0468-4170-8c51-03389e2bd1ab] Instance cache missing network info. {{(pid=61806) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 663.803079] env[61806]: DEBUG nova.virt.hardware [None req-6f137976-e023-4899-ab65-afdf6c94ed87 tempest-ServersAdmin275Test-1945525728 tempest-ServersAdmin275Test-1945525728-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-15T18:05:10Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-15T18:04:52Z,direct_url=,disk_format='vmdk',id=ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='84ba83fce2c349a6988173c5d6fc3b07',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-15T18:04:53Z,virtual_size=,visibility=), allow threads: False {{(pid=61806) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 663.803367] env[61806]: DEBUG nova.virt.hardware [None req-6f137976-e023-4899-ab65-afdf6c94ed87 tempest-ServersAdmin275Test-1945525728 tempest-ServersAdmin275Test-1945525728-project-member] Flavor limits 0:0:0 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 663.803485] env[61806]: DEBUG nova.virt.hardware [None req-6f137976-e023-4899-ab65-afdf6c94ed87 tempest-ServersAdmin275Test-1945525728 tempest-ServersAdmin275Test-1945525728-project-member] Image limits 0:0:0 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 663.803664] env[61806]: DEBUG nova.virt.hardware [None req-6f137976-e023-4899-ab65-afdf6c94ed87 tempest-ServersAdmin275Test-1945525728 tempest-ServersAdmin275Test-1945525728-project-member] Flavor pref 0:0:0 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 663.803813] env[61806]: DEBUG nova.virt.hardware [None req-6f137976-e023-4899-ab65-afdf6c94ed87 tempest-ServersAdmin275Test-1945525728 tempest-ServersAdmin275Test-1945525728-project-member] Image pref 0:0:0 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 663.803957] env[61806]: DEBUG nova.virt.hardware [None req-6f137976-e023-4899-ab65-afdf6c94ed87 tempest-ServersAdmin275Test-1945525728 tempest-ServersAdmin275Test-1945525728-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 663.804182] env[61806]: DEBUG nova.virt.hardware [None req-6f137976-e023-4899-ab65-afdf6c94ed87 tempest-ServersAdmin275Test-1945525728 tempest-ServersAdmin275Test-1945525728-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61806) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 663.804340] env[61806]: DEBUG nova.virt.hardware [None req-6f137976-e023-4899-ab65-afdf6c94ed87 tempest-ServersAdmin275Test-1945525728 tempest-ServersAdmin275Test-1945525728-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61806) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 663.804507] env[61806]: DEBUG nova.virt.hardware [None req-6f137976-e023-4899-ab65-afdf6c94ed87 tempest-ServersAdmin275Test-1945525728 tempest-ServersAdmin275Test-1945525728-project-member] Got 1 possible topologies {{(pid=61806) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 663.804667] env[61806]: DEBUG nova.virt.hardware [None req-6f137976-e023-4899-ab65-afdf6c94ed87 tempest-ServersAdmin275Test-1945525728 tempest-ServersAdmin275Test-1945525728-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61806) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 663.804835] env[61806]: DEBUG nova.virt.hardware [None req-6f137976-e023-4899-ab65-afdf6c94ed87 tempest-ServersAdmin275Test-1945525728 tempest-ServersAdmin275Test-1945525728-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61806) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 663.805672] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c4b3700f-f974-495f-988e-e2a2d923690a {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 663.812894] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-84ba5047-4041-4ee3-9c04-fca8e2b30646 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 663.825952] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-6f137976-e023-4899-ab65-afdf6c94ed87 tempest-ServersAdmin275Test-1945525728 tempest-ServersAdmin275Test-1945525728-project-member] [instance: 9db80c3c-e2c3-4034-a79a-d7be924c4ff6] Instance VIF info [] {{(pid=61806) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 663.831329] env[61806]: DEBUG oslo.service.loopingcall [None req-6f137976-e023-4899-ab65-afdf6c94ed87 tempest-ServersAdmin275Test-1945525728 tempest-ServersAdmin275Test-1945525728-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61806) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 663.831548] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 9db80c3c-e2c3-4034-a79a-d7be924c4ff6] Creating VM on the ESX host {{(pid=61806) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 663.831739] env[61806]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-c2f0a53f-896c-4b62-9876-3dfec3e9349b {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 663.847449] env[61806]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 663.847449] env[61806]: value = "task-1294487" [ 663.847449] env[61806]: _type = "Task" [ 663.847449] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 663.854959] env[61806]: DEBUG oslo_vmware.api [-] Task: {'id': task-1294487, 'name': CreateVM_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 663.938335] env[61806]: DEBUG oslo_vmware.api [None req-6dff00ec-9995-4616-9033-47ec4f3592c9 tempest-ServerShowV254Test-608938657 tempest-ServerShowV254Test-608938657-project-member] Task: {'id': task-1294486, 'name': PowerOffVM_Task, 'duration_secs': 0.188959} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 663.938716] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [None req-6dff00ec-9995-4616-9033-47ec4f3592c9 tempest-ServerShowV254Test-608938657 tempest-ServerShowV254Test-608938657-project-member] [instance: 8016bafc-b287-4df1-bd4b-a00210177a86] Powered off the VM {{(pid=61806) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 663.938995] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-6dff00ec-9995-4616-9033-47ec4f3592c9 tempest-ServerShowV254Test-608938657 tempest-ServerShowV254Test-608938657-project-member] [instance: 8016bafc-b287-4df1-bd4b-a00210177a86] Unregistering the VM {{(pid=61806) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 663.939416] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-7577ee47-00b4-4a9e-87d6-3c471ec898c3 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 663.964183] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-6dff00ec-9995-4616-9033-47ec4f3592c9 tempest-ServerShowV254Test-608938657 tempest-ServerShowV254Test-608938657-project-member] [instance: 8016bafc-b287-4df1-bd4b-a00210177a86] Unregistered the VM {{(pid=61806) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 663.964426] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-6dff00ec-9995-4616-9033-47ec4f3592c9 tempest-ServerShowV254Test-608938657 tempest-ServerShowV254Test-608938657-project-member] [instance: 8016bafc-b287-4df1-bd4b-a00210177a86] Deleting contents of the VM from datastore datastore2 {{(pid=61806) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 663.964637] env[61806]: DEBUG nova.virt.vmwareapi.ds_util [None req-6dff00ec-9995-4616-9033-47ec4f3592c9 tempest-ServerShowV254Test-608938657 tempest-ServerShowV254Test-608938657-project-member] Deleting the datastore file [datastore2] 8016bafc-b287-4df1-bd4b-a00210177a86 {{(pid=61806) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 663.964949] env[61806]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-ba601424-3f42-4d91-9fcd-c2465a97bb02 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 663.972938] env[61806]: DEBUG oslo_vmware.api [None req-6dff00ec-9995-4616-9033-47ec4f3592c9 tempest-ServerShowV254Test-608938657 tempest-ServerShowV254Test-608938657-project-member] Waiting for the task: (returnval){ [ 663.972938] env[61806]: value = "task-1294489" [ 663.972938] env[61806]: _type = "Task" [ 663.972938] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 663.981887] env[61806]: DEBUG oslo_vmware.api [None req-6dff00ec-9995-4616-9033-47ec4f3592c9 tempest-ServerShowV254Test-608938657 tempest-ServerShowV254Test-608938657-project-member] Task: {'id': task-1294489, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 664.086876] env[61806]: DEBUG nova.compute.resource_tracker [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Instance 7ec2497a-7329-4ed4-99ab-dec3d1e8117c has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61806) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 664.226933] env[61806]: DEBUG nova.network.neutron [-] [instance: 2681766c-0468-4170-8c51-03389e2bd1ab] Updating instance_info_cache with network_info: [] {{(pid=61806) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 664.358916] env[61806]: DEBUG oslo_vmware.api [-] Task: {'id': task-1294487, 'name': CreateVM_Task, 'duration_secs': 0.285258} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 664.359106] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 9db80c3c-e2c3-4034-a79a-d7be924c4ff6] Created VM on the ESX host {{(pid=61806) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 664.359500] env[61806]: DEBUG oslo_concurrency.lockutils [None req-6f137976-e023-4899-ab65-afdf6c94ed87 tempest-ServersAdmin275Test-1945525728 tempest-ServersAdmin275Test-1945525728-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 664.359660] env[61806]: DEBUG oslo_concurrency.lockutils [None req-6f137976-e023-4899-ab65-afdf6c94ed87 tempest-ServersAdmin275Test-1945525728 tempest-ServersAdmin275Test-1945525728-project-member] Acquired lock "[datastore1] devstack-image-cache_base/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 664.360035] env[61806]: DEBUG oslo_concurrency.lockutils [None req-6f137976-e023-4899-ab65-afdf6c94ed87 tempest-ServersAdmin275Test-1945525728 tempest-ServersAdmin275Test-1945525728-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 664.360343] env[61806]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-5ba97fff-a882-4978-8346-13d72166371d {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 664.364623] env[61806]: DEBUG oslo_vmware.api [None req-6f137976-e023-4899-ab65-afdf6c94ed87 tempest-ServersAdmin275Test-1945525728 tempest-ServersAdmin275Test-1945525728-project-member] Waiting for the task: (returnval){ [ 664.364623] env[61806]: value = "session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]52b925b9-51d1-5b86-ed86-fce7fdd30e98" [ 664.364623] env[61806]: _type = "Task" [ 664.364623] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 664.372043] env[61806]: DEBUG oslo_vmware.api [None req-6f137976-e023-4899-ab65-afdf6c94ed87 tempest-ServersAdmin275Test-1945525728 tempest-ServersAdmin275Test-1945525728-project-member] Task: {'id': session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]52b925b9-51d1-5b86-ed86-fce7fdd30e98, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 664.482396] env[61806]: DEBUG oslo_vmware.api [None req-6dff00ec-9995-4616-9033-47ec4f3592c9 tempest-ServerShowV254Test-608938657 tempest-ServerShowV254Test-608938657-project-member] Task: {'id': task-1294489, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.171579} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 664.482655] env[61806]: DEBUG nova.virt.vmwareapi.ds_util [None req-6dff00ec-9995-4616-9033-47ec4f3592c9 tempest-ServerShowV254Test-608938657 tempest-ServerShowV254Test-608938657-project-member] Deleted the datastore file {{(pid=61806) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 664.482842] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-6dff00ec-9995-4616-9033-47ec4f3592c9 tempest-ServerShowV254Test-608938657 tempest-ServerShowV254Test-608938657-project-member] [instance: 8016bafc-b287-4df1-bd4b-a00210177a86] Deleted contents of the VM from datastore datastore2 {{(pid=61806) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 664.483025] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-6dff00ec-9995-4616-9033-47ec4f3592c9 tempest-ServerShowV254Test-608938657 tempest-ServerShowV254Test-608938657-project-member] [instance: 8016bafc-b287-4df1-bd4b-a00210177a86] Instance destroyed {{(pid=61806) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 664.483207] env[61806]: INFO nova.compute.manager [None req-6dff00ec-9995-4616-9033-47ec4f3592c9 tempest-ServerShowV254Test-608938657 tempest-ServerShowV254Test-608938657-project-member] [instance: 8016bafc-b287-4df1-bd4b-a00210177a86] Took 1.07 seconds to destroy the instance on the hypervisor. [ 664.483445] env[61806]: DEBUG oslo.service.loopingcall [None req-6dff00ec-9995-4616-9033-47ec4f3592c9 tempest-ServerShowV254Test-608938657 tempest-ServerShowV254Test-608938657-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61806) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 664.483632] env[61806]: DEBUG nova.compute.manager [-] [instance: 8016bafc-b287-4df1-bd4b-a00210177a86] Deallocating network for instance {{(pid=61806) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 664.483726] env[61806]: DEBUG nova.network.neutron [-] [instance: 8016bafc-b287-4df1-bd4b-a00210177a86] deallocate_for_instance() {{(pid=61806) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 664.505925] env[61806]: DEBUG nova.network.neutron [-] [instance: 8016bafc-b287-4df1-bd4b-a00210177a86] Instance cache missing network info. {{(pid=61806) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 664.591111] env[61806]: DEBUG nova.compute.resource_tracker [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Instance 7ea919b6-12f7-4f4e-a123-413044bbbffc has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61806) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 664.729661] env[61806]: INFO nova.compute.manager [-] [instance: 2681766c-0468-4170-8c51-03389e2bd1ab] Took 1.02 seconds to deallocate network for instance. [ 664.732026] env[61806]: DEBUG nova.compute.claims [None req-b4447f02-a3c7-4b6e-89ba-6bc34ec03205 tempest-InstanceActionsTestJSON-949444324 tempest-InstanceActionsTestJSON-949444324-project-member] [instance: 2681766c-0468-4170-8c51-03389e2bd1ab] Aborting claim: {{(pid=61806) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 664.732259] env[61806]: DEBUG oslo_concurrency.lockutils [None req-b4447f02-a3c7-4b6e-89ba-6bc34ec03205 tempest-InstanceActionsTestJSON-949444324 tempest-InstanceActionsTestJSON-949444324-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 664.876633] env[61806]: DEBUG oslo_vmware.api [None req-6f137976-e023-4899-ab65-afdf6c94ed87 tempest-ServersAdmin275Test-1945525728 tempest-ServersAdmin275Test-1945525728-project-member] Task: {'id': session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]52b925b9-51d1-5b86-ed86-fce7fdd30e98, 'name': SearchDatastore_Task, 'duration_secs': 0.009422} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 664.876633] env[61806]: DEBUG oslo_concurrency.lockutils [None req-6f137976-e023-4899-ab65-afdf6c94ed87 tempest-ServersAdmin275Test-1945525728 tempest-ServersAdmin275Test-1945525728-project-member] Releasing lock "[datastore1] devstack-image-cache_base/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 664.876633] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-6f137976-e023-4899-ab65-afdf6c94ed87 tempest-ServersAdmin275Test-1945525728 tempest-ServersAdmin275Test-1945525728-project-member] [instance: 9db80c3c-e2c3-4034-a79a-d7be924c4ff6] Processing image ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa {{(pid=61806) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 664.876633] env[61806]: DEBUG oslo_concurrency.lockutils [None req-6f137976-e023-4899-ab65-afdf6c94ed87 tempest-ServersAdmin275Test-1945525728 tempest-ServersAdmin275Test-1945525728-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa.vmdk" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 664.877050] env[61806]: DEBUG oslo_concurrency.lockutils [None req-6f137976-e023-4899-ab65-afdf6c94ed87 tempest-ServersAdmin275Test-1945525728 tempest-ServersAdmin275Test-1945525728-project-member] Acquired lock "[datastore1] devstack-image-cache_base/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa.vmdk" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 664.877050] env[61806]: DEBUG nova.virt.vmwareapi.ds_util [None req-6f137976-e023-4899-ab65-afdf6c94ed87 tempest-ServersAdmin275Test-1945525728 tempest-ServersAdmin275Test-1945525728-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61806) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 664.877050] env[61806]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-c2965eab-e65f-4b72-b1f7-3a6fae3edc7a {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 664.884578] env[61806]: DEBUG nova.virt.vmwareapi.ds_util [None req-6f137976-e023-4899-ab65-afdf6c94ed87 tempest-ServersAdmin275Test-1945525728 tempest-ServersAdmin275Test-1945525728-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61806) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 664.884759] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-6f137976-e023-4899-ab65-afdf6c94ed87 tempest-ServersAdmin275Test-1945525728 tempest-ServersAdmin275Test-1945525728-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61806) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 664.885460] env[61806]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-484cf4c0-17f2-4926-9be3-d73b59954155 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 664.890555] env[61806]: DEBUG oslo_vmware.api [None req-6f137976-e023-4899-ab65-afdf6c94ed87 tempest-ServersAdmin275Test-1945525728 tempest-ServersAdmin275Test-1945525728-project-member] Waiting for the task: (returnval){ [ 664.890555] env[61806]: value = "session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]529d1d25-1d8d-c6ac-e8f6-e0b9142c595c" [ 664.890555] env[61806]: _type = "Task" [ 664.890555] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 664.897842] env[61806]: DEBUG oslo_vmware.api [None req-6f137976-e023-4899-ab65-afdf6c94ed87 tempest-ServersAdmin275Test-1945525728 tempest-ServersAdmin275Test-1945525728-project-member] Task: {'id': session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]529d1d25-1d8d-c6ac-e8f6-e0b9142c595c, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 665.008609] env[61806]: DEBUG nova.network.neutron [-] [instance: 8016bafc-b287-4df1-bd4b-a00210177a86] Updating instance_info_cache with network_info: [] {{(pid=61806) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 665.094596] env[61806]: DEBUG nova.compute.resource_tracker [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Instance 7211d568-cc68-4e50-80ad-b78878d1deab has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61806) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 665.403611] env[61806]: DEBUG oslo_vmware.api [None req-6f137976-e023-4899-ab65-afdf6c94ed87 tempest-ServersAdmin275Test-1945525728 tempest-ServersAdmin275Test-1945525728-project-member] Task: {'id': session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]529d1d25-1d8d-c6ac-e8f6-e0b9142c595c, 'name': SearchDatastore_Task, 'duration_secs': 0.008764} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 665.404610] env[61806]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f9433312-1b9b-4afc-a1ea-ae93b0fbcffd {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 665.409803] env[61806]: DEBUG oslo_vmware.api [None req-6f137976-e023-4899-ab65-afdf6c94ed87 tempest-ServersAdmin275Test-1945525728 tempest-ServersAdmin275Test-1945525728-project-member] Waiting for the task: (returnval){ [ 665.409803] env[61806]: value = "session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]52bc284e-a7eb-5e49-8b78-b0f24874ae4f" [ 665.409803] env[61806]: _type = "Task" [ 665.409803] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 665.417742] env[61806]: DEBUG oslo_vmware.api [None req-6f137976-e023-4899-ab65-afdf6c94ed87 tempest-ServersAdmin275Test-1945525728 tempest-ServersAdmin275Test-1945525728-project-member] Task: {'id': session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]52bc284e-a7eb-5e49-8b78-b0f24874ae4f, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 665.513647] env[61806]: INFO nova.compute.manager [-] [instance: 8016bafc-b287-4df1-bd4b-a00210177a86] Took 1.03 seconds to deallocate network for instance. [ 665.597841] env[61806]: DEBUG nova.compute.resource_tracker [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Instance 5b22c2fc-525c-481b-b84e-e0a7f68f633e has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61806) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 665.925405] env[61806]: DEBUG oslo_vmware.api [None req-6f137976-e023-4899-ab65-afdf6c94ed87 tempest-ServersAdmin275Test-1945525728 tempest-ServersAdmin275Test-1945525728-project-member] Task: {'id': session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]52bc284e-a7eb-5e49-8b78-b0f24874ae4f, 'name': SearchDatastore_Task, 'duration_secs': 0.016012} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 665.925698] env[61806]: DEBUG oslo_concurrency.lockutils [None req-6f137976-e023-4899-ab65-afdf6c94ed87 tempest-ServersAdmin275Test-1945525728 tempest-ServersAdmin275Test-1945525728-project-member] Releasing lock "[datastore1] devstack-image-cache_base/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa.vmdk" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 665.925950] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [None req-6f137976-e023-4899-ab65-afdf6c94ed87 tempest-ServersAdmin275Test-1945525728 tempest-ServersAdmin275Test-1945525728-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa.vmdk to [datastore1] 9db80c3c-e2c3-4034-a79a-d7be924c4ff6/9db80c3c-e2c3-4034-a79a-d7be924c4ff6.vmdk {{(pid=61806) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 665.926516] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-7c1a4993-9b53-475b-8997-ce60f72fcc36 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 665.933279] env[61806]: DEBUG oslo_vmware.api [None req-6f137976-e023-4899-ab65-afdf6c94ed87 tempest-ServersAdmin275Test-1945525728 tempest-ServersAdmin275Test-1945525728-project-member] Waiting for the task: (returnval){ [ 665.933279] env[61806]: value = "task-1294490" [ 665.933279] env[61806]: _type = "Task" [ 665.933279] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 665.945019] env[61806]: DEBUG oslo_vmware.api [None req-6f137976-e023-4899-ab65-afdf6c94ed87 tempest-ServersAdmin275Test-1945525728 tempest-ServersAdmin275Test-1945525728-project-member] Task: {'id': task-1294490, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 666.019636] env[61806]: DEBUG oslo_concurrency.lockutils [None req-6dff00ec-9995-4616-9033-47ec4f3592c9 tempest-ServerShowV254Test-608938657 tempest-ServerShowV254Test-608938657-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 666.103761] env[61806]: DEBUG nova.compute.resource_tracker [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Instance 15476ac0-289a-4e04-aa9d-4244c658e962 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61806) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 666.445668] env[61806]: DEBUG oslo_vmware.api [None req-6f137976-e023-4899-ab65-afdf6c94ed87 tempest-ServersAdmin275Test-1945525728 tempest-ServersAdmin275Test-1945525728-project-member] Task: {'id': task-1294490, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.509829} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 666.445933] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [None req-6f137976-e023-4899-ab65-afdf6c94ed87 tempest-ServersAdmin275Test-1945525728 tempest-ServersAdmin275Test-1945525728-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa.vmdk to [datastore1] 9db80c3c-e2c3-4034-a79a-d7be924c4ff6/9db80c3c-e2c3-4034-a79a-d7be924c4ff6.vmdk {{(pid=61806) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 666.446155] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-6f137976-e023-4899-ab65-afdf6c94ed87 tempest-ServersAdmin275Test-1945525728 tempest-ServersAdmin275Test-1945525728-project-member] [instance: 9db80c3c-e2c3-4034-a79a-d7be924c4ff6] Extending root virtual disk to 1048576 {{(pid=61806) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 666.446401] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-d67b4157-658d-452f-9e6c-70ebeac12176 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 666.453734] env[61806]: DEBUG oslo_vmware.api [None req-6f137976-e023-4899-ab65-afdf6c94ed87 tempest-ServersAdmin275Test-1945525728 tempest-ServersAdmin275Test-1945525728-project-member] Waiting for the task: (returnval){ [ 666.453734] env[61806]: value = "task-1294491" [ 666.453734] env[61806]: _type = "Task" [ 666.453734] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 666.463313] env[61806]: DEBUG oslo_vmware.api [None req-6f137976-e023-4899-ab65-afdf6c94ed87 tempest-ServersAdmin275Test-1945525728 tempest-ServersAdmin275Test-1945525728-project-member] Task: {'id': task-1294491, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 666.606431] env[61806]: DEBUG nova.compute.resource_tracker [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Instance 949c1050-5003-4519-b24c-8904de21a676 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61806) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 666.964292] env[61806]: DEBUG oslo_vmware.api [None req-6f137976-e023-4899-ab65-afdf6c94ed87 tempest-ServersAdmin275Test-1945525728 tempest-ServersAdmin275Test-1945525728-project-member] Task: {'id': task-1294491, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.070026} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 666.964996] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-6f137976-e023-4899-ab65-afdf6c94ed87 tempest-ServersAdmin275Test-1945525728 tempest-ServersAdmin275Test-1945525728-project-member] [instance: 9db80c3c-e2c3-4034-a79a-d7be924c4ff6] Extended root virtual disk {{(pid=61806) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 666.965898] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-687e0e8b-74ab-49fa-a84a-7f7fbbdeba1b {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 666.986690] env[61806]: DEBUG nova.virt.vmwareapi.volumeops [None req-6f137976-e023-4899-ab65-afdf6c94ed87 tempest-ServersAdmin275Test-1945525728 tempest-ServersAdmin275Test-1945525728-project-member] [instance: 9db80c3c-e2c3-4034-a79a-d7be924c4ff6] Reconfiguring VM instance instance-0000001f to attach disk [datastore1] 9db80c3c-e2c3-4034-a79a-d7be924c4ff6/9db80c3c-e2c3-4034-a79a-d7be924c4ff6.vmdk or device None with type sparse {{(pid=61806) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 666.986986] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-af1c087a-5850-434b-9471-b646c41bbcc6 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 667.007127] env[61806]: DEBUG oslo_vmware.api [None req-6f137976-e023-4899-ab65-afdf6c94ed87 tempest-ServersAdmin275Test-1945525728 tempest-ServersAdmin275Test-1945525728-project-member] Waiting for the task: (returnval){ [ 667.007127] env[61806]: value = "task-1294492" [ 667.007127] env[61806]: _type = "Task" [ 667.007127] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 667.014877] env[61806]: DEBUG oslo_vmware.api [None req-6f137976-e023-4899-ab65-afdf6c94ed87 tempest-ServersAdmin275Test-1945525728 tempest-ServersAdmin275Test-1945525728-project-member] Task: {'id': task-1294492, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 667.109449] env[61806]: DEBUG nova.compute.resource_tracker [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Instance f2446d04-15b8-4b3d-8932-c37df67b0f8f has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61806) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 667.516928] env[61806]: DEBUG oslo_vmware.api [None req-6f137976-e023-4899-ab65-afdf6c94ed87 tempest-ServersAdmin275Test-1945525728 tempest-ServersAdmin275Test-1945525728-project-member] Task: {'id': task-1294492, 'name': ReconfigVM_Task, 'duration_secs': 0.293453} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 667.517294] env[61806]: DEBUG nova.virt.vmwareapi.volumeops [None req-6f137976-e023-4899-ab65-afdf6c94ed87 tempest-ServersAdmin275Test-1945525728 tempest-ServersAdmin275Test-1945525728-project-member] [instance: 9db80c3c-e2c3-4034-a79a-d7be924c4ff6] Reconfigured VM instance instance-0000001f to attach disk [datastore1] 9db80c3c-e2c3-4034-a79a-d7be924c4ff6/9db80c3c-e2c3-4034-a79a-d7be924c4ff6.vmdk or device None with type sparse {{(pid=61806) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 667.517904] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-8b910449-61eb-490e-9155-a4f6b389216b {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 667.525036] env[61806]: DEBUG oslo_vmware.api [None req-6f137976-e023-4899-ab65-afdf6c94ed87 tempest-ServersAdmin275Test-1945525728 tempest-ServersAdmin275Test-1945525728-project-member] Waiting for the task: (returnval){ [ 667.525036] env[61806]: value = "task-1294493" [ 667.525036] env[61806]: _type = "Task" [ 667.525036] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 667.532392] env[61806]: DEBUG oslo_vmware.api [None req-6f137976-e023-4899-ab65-afdf6c94ed87 tempest-ServersAdmin275Test-1945525728 tempest-ServersAdmin275Test-1945525728-project-member] Task: {'id': task-1294493, 'name': Rename_Task} progress is 5%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 667.613989] env[61806]: DEBUG nova.compute.resource_tracker [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Instance ea8d39a1-066d-4eb4-888d-776b0e45f684 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61806) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 668.038462] env[61806]: DEBUG oslo_vmware.api [None req-6f137976-e023-4899-ab65-afdf6c94ed87 tempest-ServersAdmin275Test-1945525728 tempest-ServersAdmin275Test-1945525728-project-member] Task: {'id': task-1294493, 'name': Rename_Task, 'duration_secs': 0.133349} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 668.038462] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [None req-6f137976-e023-4899-ab65-afdf6c94ed87 tempest-ServersAdmin275Test-1945525728 tempest-ServersAdmin275Test-1945525728-project-member] [instance: 9db80c3c-e2c3-4034-a79a-d7be924c4ff6] Powering on the VM {{(pid=61806) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 668.038462] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-de9f5717-a944-4a2e-8456-6985c45d0c0c {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 668.043315] env[61806]: DEBUG oslo_vmware.api [None req-6f137976-e023-4899-ab65-afdf6c94ed87 tempest-ServersAdmin275Test-1945525728 tempest-ServersAdmin275Test-1945525728-project-member] Waiting for the task: (returnval){ [ 668.043315] env[61806]: value = "task-1294494" [ 668.043315] env[61806]: _type = "Task" [ 668.043315] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 668.051611] env[61806]: DEBUG oslo_vmware.api [None req-6f137976-e023-4899-ab65-afdf6c94ed87 tempest-ServersAdmin275Test-1945525728 tempest-ServersAdmin275Test-1945525728-project-member] Task: {'id': task-1294494, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 668.115829] env[61806]: DEBUG nova.compute.resource_tracker [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Instance 198dc1a4-050d-47b2-8bc4-30af5534a789 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61806) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 668.558137] env[61806]: DEBUG oslo_vmware.api [None req-6f137976-e023-4899-ab65-afdf6c94ed87 tempest-ServersAdmin275Test-1945525728 tempest-ServersAdmin275Test-1945525728-project-member] Task: {'id': task-1294494, 'name': PowerOnVM_Task, 'duration_secs': 0.415458} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 668.558485] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [None req-6f137976-e023-4899-ab65-afdf6c94ed87 tempest-ServersAdmin275Test-1945525728 tempest-ServersAdmin275Test-1945525728-project-member] [instance: 9db80c3c-e2c3-4034-a79a-d7be924c4ff6] Powered on the VM {{(pid=61806) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 668.558694] env[61806]: DEBUG nova.compute.manager [None req-6f137976-e023-4899-ab65-afdf6c94ed87 tempest-ServersAdmin275Test-1945525728 tempest-ServersAdmin275Test-1945525728-project-member] [instance: 9db80c3c-e2c3-4034-a79a-d7be924c4ff6] Checking state {{(pid=61806) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 668.559546] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d79214e0-cea9-4956-b7ed-8c841ebd9739 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 668.619379] env[61806]: DEBUG nova.compute.resource_tracker [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Instance 37c07b8d-e2a0-4aa6-b7ba-194feed837e5 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61806) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 669.084521] env[61806]: DEBUG oslo_concurrency.lockutils [None req-6f137976-e023-4899-ab65-afdf6c94ed87 tempest-ServersAdmin275Test-1945525728 tempest-ServersAdmin275Test-1945525728-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 669.123420] env[61806]: DEBUG nova.compute.resource_tracker [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Instance c61df5d3-7f43-48d4-a2c0-a8972372c0ba has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61806) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 669.629909] env[61806]: DEBUG nova.compute.resource_tracker [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Instance 24c83131-1451-429f-84fe-dcd070898d65 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61806) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 670.133675] env[61806]: DEBUG nova.compute.resource_tracker [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Instance c25ddf7c-d0cc-4b73-96da-1dcd6012f072 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61806) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 670.226894] env[61806]: INFO nova.compute.manager [None req-c9d6a9df-bcc9-4e08-88fa-de36a15b3a76 tempest-ServersAdmin275Test-894191221 tempest-ServersAdmin275Test-894191221-project-admin] [instance: 9db80c3c-e2c3-4034-a79a-d7be924c4ff6] Rebuilding instance [ 670.274582] env[61806]: DEBUG nova.compute.manager [None req-c9d6a9df-bcc9-4e08-88fa-de36a15b3a76 tempest-ServersAdmin275Test-894191221 tempest-ServersAdmin275Test-894191221-project-admin] [instance: 9db80c3c-e2c3-4034-a79a-d7be924c4ff6] Checking state {{(pid=61806) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 670.275598] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c6f8c53d-7af6-4ff5-9250-912417e0d3e4 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 670.636654] env[61806]: DEBUG nova.compute.resource_tracker [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Instance e0ef0a35-82a5-495b-9d5c-5805e8306390 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61806) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 670.788503] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [None req-c9d6a9df-bcc9-4e08-88fa-de36a15b3a76 tempest-ServersAdmin275Test-894191221 tempest-ServersAdmin275Test-894191221-project-admin] [instance: 9db80c3c-e2c3-4034-a79a-d7be924c4ff6] Powering off the VM {{(pid=61806) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 670.788783] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-5b27ca3e-1129-4d03-8158-30787ea2e877 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 670.796748] env[61806]: DEBUG oslo_vmware.api [None req-c9d6a9df-bcc9-4e08-88fa-de36a15b3a76 tempest-ServersAdmin275Test-894191221 tempest-ServersAdmin275Test-894191221-project-admin] Waiting for the task: (returnval){ [ 670.796748] env[61806]: value = "task-1294495" [ 670.796748] env[61806]: _type = "Task" [ 670.796748] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 670.805494] env[61806]: DEBUG oslo_vmware.api [None req-c9d6a9df-bcc9-4e08-88fa-de36a15b3a76 tempest-ServersAdmin275Test-894191221 tempest-ServersAdmin275Test-894191221-project-admin] Task: {'id': task-1294495, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 671.140503] env[61806]: DEBUG nova.compute.resource_tracker [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Instance 2a3f81e3-b84f-4370-a598-2f6a607c3d47 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61806) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 671.306791] env[61806]: DEBUG oslo_vmware.api [None req-c9d6a9df-bcc9-4e08-88fa-de36a15b3a76 tempest-ServersAdmin275Test-894191221 tempest-ServersAdmin275Test-894191221-project-admin] Task: {'id': task-1294495, 'name': PowerOffVM_Task, 'duration_secs': 0.191597} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 671.307193] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [None req-c9d6a9df-bcc9-4e08-88fa-de36a15b3a76 tempest-ServersAdmin275Test-894191221 tempest-ServersAdmin275Test-894191221-project-admin] [instance: 9db80c3c-e2c3-4034-a79a-d7be924c4ff6] Powered off the VM {{(pid=61806) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 671.307480] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-c9d6a9df-bcc9-4e08-88fa-de36a15b3a76 tempest-ServersAdmin275Test-894191221 tempest-ServersAdmin275Test-894191221-project-admin] [instance: 9db80c3c-e2c3-4034-a79a-d7be924c4ff6] Destroying instance {{(pid=61806) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 671.308354] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e2bd6a67-796d-490c-9af6-be25251a2c68 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 671.315352] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-c9d6a9df-bcc9-4e08-88fa-de36a15b3a76 tempest-ServersAdmin275Test-894191221 tempest-ServersAdmin275Test-894191221-project-admin] [instance: 9db80c3c-e2c3-4034-a79a-d7be924c4ff6] Unregistering the VM {{(pid=61806) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 671.315574] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-5e60460d-4275-4145-aa71-6e6db25a5441 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 671.346026] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-c9d6a9df-bcc9-4e08-88fa-de36a15b3a76 tempest-ServersAdmin275Test-894191221 tempest-ServersAdmin275Test-894191221-project-admin] [instance: 9db80c3c-e2c3-4034-a79a-d7be924c4ff6] Unregistered the VM {{(pid=61806) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 671.346259] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-c9d6a9df-bcc9-4e08-88fa-de36a15b3a76 tempest-ServersAdmin275Test-894191221 tempest-ServersAdmin275Test-894191221-project-admin] [instance: 9db80c3c-e2c3-4034-a79a-d7be924c4ff6] Deleting contents of the VM from datastore datastore1 {{(pid=61806) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 671.346443] env[61806]: DEBUG nova.virt.vmwareapi.ds_util [None req-c9d6a9df-bcc9-4e08-88fa-de36a15b3a76 tempest-ServersAdmin275Test-894191221 tempest-ServersAdmin275Test-894191221-project-admin] Deleting the datastore file [datastore1] 9db80c3c-e2c3-4034-a79a-d7be924c4ff6 {{(pid=61806) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 671.346700] env[61806]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-ecd7484c-1625-42df-83f9-6f33be4a1e05 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 671.353321] env[61806]: DEBUG oslo_vmware.api [None req-c9d6a9df-bcc9-4e08-88fa-de36a15b3a76 tempest-ServersAdmin275Test-894191221 tempest-ServersAdmin275Test-894191221-project-admin] Waiting for the task: (returnval){ [ 671.353321] env[61806]: value = "task-1294497" [ 671.353321] env[61806]: _type = "Task" [ 671.353321] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 671.361100] env[61806]: DEBUG oslo_vmware.api [None req-c9d6a9df-bcc9-4e08-88fa-de36a15b3a76 tempest-ServersAdmin275Test-894191221 tempest-ServersAdmin275Test-894191221-project-admin] Task: {'id': task-1294497, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 671.643240] env[61806]: DEBUG nova.compute.resource_tracker [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Instance 4184e263-6f56-4bc1-99b2-a2c460531516 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61806) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 671.644014] env[61806]: DEBUG nova.compute.resource_tracker [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Total usable vcpus: 48, total allocated vcpus: 6 {{(pid=61806) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 671.644223] env[61806]: DEBUG nova.compute.resource_tracker [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=1664MB phys_disk=200GB used_disk=5GB total_vcpus=48 used_vcpus=6 pci_stats=[] {{(pid=61806) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 671.863854] env[61806]: DEBUG oslo_vmware.api [None req-c9d6a9df-bcc9-4e08-88fa-de36a15b3a76 tempest-ServersAdmin275Test-894191221 tempest-ServersAdmin275Test-894191221-project-admin] Task: {'id': task-1294497, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.111041} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 671.866118] env[61806]: DEBUG nova.virt.vmwareapi.ds_util [None req-c9d6a9df-bcc9-4e08-88fa-de36a15b3a76 tempest-ServersAdmin275Test-894191221 tempest-ServersAdmin275Test-894191221-project-admin] Deleted the datastore file {{(pid=61806) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 671.866322] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-c9d6a9df-bcc9-4e08-88fa-de36a15b3a76 tempest-ServersAdmin275Test-894191221 tempest-ServersAdmin275Test-894191221-project-admin] [instance: 9db80c3c-e2c3-4034-a79a-d7be924c4ff6] Deleted contents of the VM from datastore datastore1 {{(pid=61806) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 671.866494] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-c9d6a9df-bcc9-4e08-88fa-de36a15b3a76 tempest-ServersAdmin275Test-894191221 tempest-ServersAdmin275Test-894191221-project-admin] [instance: 9db80c3c-e2c3-4034-a79a-d7be924c4ff6] Instance destroyed {{(pid=61806) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 672.013221] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c7d38c1b-6d9e-4d64-a219-75ceb5a258f8 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 672.021214] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-61de68dd-aed2-48e5-803f-7bbef1f6a84e {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 672.053642] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-851242f8-d1ff-46df-9f62-9d0ec1565ab9 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 672.060390] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bc080d96-a8f5-466f-9a70-263f4bfb8b83 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 672.073371] env[61806]: DEBUG nova.compute.provider_tree [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Updating inventory in ProviderTree for provider a2858be1-fd22-4e08-979e-87ad25293407 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 139, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61806) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 672.594203] env[61806]: ERROR nova.scheduler.client.report [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] [req-11b76a63-100b-400f-9920-a93a340367df] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 139, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID a2858be1-fd22-4e08-979e-87ad25293407. Got 409: {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-11b76a63-100b-400f-9920-a93a340367df"}]} [ 672.612995] env[61806]: DEBUG nova.scheduler.client.report [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Refreshing inventories for resource provider a2858be1-fd22-4e08-979e-87ad25293407 {{(pid=61806) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:818}} [ 672.628620] env[61806]: DEBUG nova.scheduler.client.report [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Updating ProviderTree inventory for provider a2858be1-fd22-4e08-979e-87ad25293407 from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 138, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61806) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:782}} [ 672.628810] env[61806]: DEBUG nova.compute.provider_tree [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Updating inventory in ProviderTree for provider a2858be1-fd22-4e08-979e-87ad25293407 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 138, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61806) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 672.640547] env[61806]: DEBUG nova.scheduler.client.report [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Refreshing aggregate associations for resource provider a2858be1-fd22-4e08-979e-87ad25293407, aggregates: None {{(pid=61806) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:827}} [ 672.657654] env[61806]: DEBUG nova.scheduler.client.report [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Refreshing trait associations for resource provider a2858be1-fd22-4e08-979e-87ad25293407, traits: COMPUTE_IMAGE_TYPE_ISO,COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_SAME_HOST_COLD_MIGRATE,COMPUTE_NET_ATTACH_INTERFACE,COMPUTE_NODE {{(pid=61806) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:839}} [ 672.898105] env[61806]: DEBUG nova.virt.hardware [None req-c9d6a9df-bcc9-4e08-88fa-de36a15b3a76 tempest-ServersAdmin275Test-894191221 tempest-ServersAdmin275Test-894191221-project-admin] Getting desirable topologies for flavor Flavor(created_at=2024-10-15T18:05:10Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-15T18:04:52Z,direct_url=,disk_format='vmdk',id=ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='84ba83fce2c349a6988173c5d6fc3b07',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-15T18:04:53Z,virtual_size=,visibility=), allow threads: False {{(pid=61806) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 672.898364] env[61806]: DEBUG nova.virt.hardware [None req-c9d6a9df-bcc9-4e08-88fa-de36a15b3a76 tempest-ServersAdmin275Test-894191221 tempest-ServersAdmin275Test-894191221-project-admin] Flavor limits 0:0:0 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 672.898548] env[61806]: DEBUG nova.virt.hardware [None req-c9d6a9df-bcc9-4e08-88fa-de36a15b3a76 tempest-ServersAdmin275Test-894191221 tempest-ServersAdmin275Test-894191221-project-admin] Image limits 0:0:0 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 672.898768] env[61806]: DEBUG nova.virt.hardware [None req-c9d6a9df-bcc9-4e08-88fa-de36a15b3a76 tempest-ServersAdmin275Test-894191221 tempest-ServersAdmin275Test-894191221-project-admin] Flavor pref 0:0:0 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 672.898921] env[61806]: DEBUG nova.virt.hardware [None req-c9d6a9df-bcc9-4e08-88fa-de36a15b3a76 tempest-ServersAdmin275Test-894191221 tempest-ServersAdmin275Test-894191221-project-admin] Image pref 0:0:0 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 672.899081] env[61806]: DEBUG nova.virt.hardware [None req-c9d6a9df-bcc9-4e08-88fa-de36a15b3a76 tempest-ServersAdmin275Test-894191221 tempest-ServersAdmin275Test-894191221-project-admin] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 672.899289] env[61806]: DEBUG nova.virt.hardware [None req-c9d6a9df-bcc9-4e08-88fa-de36a15b3a76 tempest-ServersAdmin275Test-894191221 tempest-ServersAdmin275Test-894191221-project-admin] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61806) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 672.899889] env[61806]: DEBUG nova.virt.hardware [None req-c9d6a9df-bcc9-4e08-88fa-de36a15b3a76 tempest-ServersAdmin275Test-894191221 tempest-ServersAdmin275Test-894191221-project-admin] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61806) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 672.899889] env[61806]: DEBUG nova.virt.hardware [None req-c9d6a9df-bcc9-4e08-88fa-de36a15b3a76 tempest-ServersAdmin275Test-894191221 tempest-ServersAdmin275Test-894191221-project-admin] Got 1 possible topologies {{(pid=61806) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 672.899889] env[61806]: DEBUG nova.virt.hardware [None req-c9d6a9df-bcc9-4e08-88fa-de36a15b3a76 tempest-ServersAdmin275Test-894191221 tempest-ServersAdmin275Test-894191221-project-admin] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61806) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 672.900060] env[61806]: DEBUG nova.virt.hardware [None req-c9d6a9df-bcc9-4e08-88fa-de36a15b3a76 tempest-ServersAdmin275Test-894191221 tempest-ServersAdmin275Test-894191221-project-admin] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61806) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 672.900797] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-af6b3519-7289-4b05-9a66-616b0d099b71 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 672.910850] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e62fb544-a78d-4cf8-a909-fce831d74021 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 672.927313] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-c9d6a9df-bcc9-4e08-88fa-de36a15b3a76 tempest-ServersAdmin275Test-894191221 tempest-ServersAdmin275Test-894191221-project-admin] [instance: 9db80c3c-e2c3-4034-a79a-d7be924c4ff6] Instance VIF info [] {{(pid=61806) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 672.932870] env[61806]: DEBUG oslo.service.loopingcall [None req-c9d6a9df-bcc9-4e08-88fa-de36a15b3a76 tempest-ServersAdmin275Test-894191221 tempest-ServersAdmin275Test-894191221-project-admin] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61806) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 672.935872] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 9db80c3c-e2c3-4034-a79a-d7be924c4ff6] Creating VM on the ESX host {{(pid=61806) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 672.936473] env[61806]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-8dda1935-2a99-4dfd-8467-f599f36fe696 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 672.955881] env[61806]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 672.955881] env[61806]: value = "task-1294498" [ 672.955881] env[61806]: _type = "Task" [ 672.955881] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 672.964642] env[61806]: DEBUG oslo_vmware.api [-] Task: {'id': task-1294498, 'name': CreateVM_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 673.025890] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fc576b76-9a6c-42c3-a0de-a996f25032a4 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 673.032973] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cc667509-a5f9-4651-92ab-36a067e2e616 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 673.062803] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-87be724c-5708-463c-82e5-50f19d0c3070 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 673.070046] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5a94076f-a4e1-4e2f-974f-95c1cf86a4a6 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 673.083532] env[61806]: DEBUG nova.compute.provider_tree [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Updating inventory in ProviderTree for provider a2858be1-fd22-4e08-979e-87ad25293407 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 139, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61806) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 673.465774] env[61806]: DEBUG oslo_vmware.api [-] Task: {'id': task-1294498, 'name': CreateVM_Task, 'duration_secs': 0.252289} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 673.465954] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 9db80c3c-e2c3-4034-a79a-d7be924c4ff6] Created VM on the ESX host {{(pid=61806) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 673.466899] env[61806]: DEBUG oslo_concurrency.lockutils [None req-c9d6a9df-bcc9-4e08-88fa-de36a15b3a76 tempest-ServersAdmin275Test-894191221 tempest-ServersAdmin275Test-894191221-project-admin] Acquiring lock "[datastore1] devstack-image-cache_base/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 673.467118] env[61806]: DEBUG oslo_concurrency.lockutils [None req-c9d6a9df-bcc9-4e08-88fa-de36a15b3a76 tempest-ServersAdmin275Test-894191221 tempest-ServersAdmin275Test-894191221-project-admin] Acquired lock "[datastore1] devstack-image-cache_base/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 673.467437] env[61806]: DEBUG oslo_concurrency.lockutils [None req-c9d6a9df-bcc9-4e08-88fa-de36a15b3a76 tempest-ServersAdmin275Test-894191221 tempest-ServersAdmin275Test-894191221-project-admin] Acquired external semaphore "[datastore1] devstack-image-cache_base/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 673.467699] env[61806]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-73167cb2-286d-4539-8d07-6648ffe883c0 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 673.472470] env[61806]: DEBUG oslo_vmware.api [None req-c9d6a9df-bcc9-4e08-88fa-de36a15b3a76 tempest-ServersAdmin275Test-894191221 tempest-ServersAdmin275Test-894191221-project-admin] Waiting for the task: (returnval){ [ 673.472470] env[61806]: value = "session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]5225076c-2e8e-05b5-178d-8570fa484375" [ 673.472470] env[61806]: _type = "Task" [ 673.472470] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 673.480667] env[61806]: DEBUG oslo_vmware.api [None req-c9d6a9df-bcc9-4e08-88fa-de36a15b3a76 tempest-ServersAdmin275Test-894191221 tempest-ServersAdmin275Test-894191221-project-admin] Task: {'id': session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]5225076c-2e8e-05b5-178d-8570fa484375, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 673.615397] env[61806]: DEBUG nova.scheduler.client.report [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Updated inventory for provider a2858be1-fd22-4e08-979e-87ad25293407 with generation 64 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 139, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61806) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:971}} [ 673.615688] env[61806]: DEBUG nova.compute.provider_tree [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Updating resource provider a2858be1-fd22-4e08-979e-87ad25293407 generation from 64 to 65 during operation: update_inventory {{(pid=61806) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 673.615781] env[61806]: DEBUG nova.compute.provider_tree [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Updating inventory in ProviderTree for provider a2858be1-fd22-4e08-979e-87ad25293407 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 139, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61806) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 673.983156] env[61806]: DEBUG oslo_vmware.api [None req-c9d6a9df-bcc9-4e08-88fa-de36a15b3a76 tempest-ServersAdmin275Test-894191221 tempest-ServersAdmin275Test-894191221-project-admin] Task: {'id': session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]5225076c-2e8e-05b5-178d-8570fa484375, 'name': SearchDatastore_Task, 'duration_secs': 0.011327} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 673.983472] env[61806]: DEBUG oslo_concurrency.lockutils [None req-c9d6a9df-bcc9-4e08-88fa-de36a15b3a76 tempest-ServersAdmin275Test-894191221 tempest-ServersAdmin275Test-894191221-project-admin] Releasing lock "[datastore1] devstack-image-cache_base/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 673.983729] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-c9d6a9df-bcc9-4e08-88fa-de36a15b3a76 tempest-ServersAdmin275Test-894191221 tempest-ServersAdmin275Test-894191221-project-admin] [instance: 9db80c3c-e2c3-4034-a79a-d7be924c4ff6] Processing image ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa {{(pid=61806) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 673.983996] env[61806]: DEBUG oslo_concurrency.lockutils [None req-c9d6a9df-bcc9-4e08-88fa-de36a15b3a76 tempest-ServersAdmin275Test-894191221 tempest-ServersAdmin275Test-894191221-project-admin] Acquiring lock "[datastore1] devstack-image-cache_base/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa.vmdk" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 673.984179] env[61806]: DEBUG oslo_concurrency.lockutils [None req-c9d6a9df-bcc9-4e08-88fa-de36a15b3a76 tempest-ServersAdmin275Test-894191221 tempest-ServersAdmin275Test-894191221-project-admin] Acquired lock "[datastore1] devstack-image-cache_base/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa.vmdk" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 673.984363] env[61806]: DEBUG nova.virt.vmwareapi.ds_util [None req-c9d6a9df-bcc9-4e08-88fa-de36a15b3a76 tempest-ServersAdmin275Test-894191221 tempest-ServersAdmin275Test-894191221-project-admin] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61806) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 673.984636] env[61806]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-fae777e9-a524-42a6-8ecf-b5976ef3b39b {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 673.992218] env[61806]: DEBUG nova.virt.vmwareapi.ds_util [None req-c9d6a9df-bcc9-4e08-88fa-de36a15b3a76 tempest-ServersAdmin275Test-894191221 tempest-ServersAdmin275Test-894191221-project-admin] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61806) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 673.992397] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-c9d6a9df-bcc9-4e08-88fa-de36a15b3a76 tempest-ServersAdmin275Test-894191221 tempest-ServersAdmin275Test-894191221-project-admin] Folder [datastore1] devstack-image-cache_base created. {{(pid=61806) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 673.993132] env[61806]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-904fd341-2838-4113-ae1a-461577e87f65 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 673.998363] env[61806]: DEBUG oslo_vmware.api [None req-c9d6a9df-bcc9-4e08-88fa-de36a15b3a76 tempest-ServersAdmin275Test-894191221 tempest-ServersAdmin275Test-894191221-project-admin] Waiting for the task: (returnval){ [ 673.998363] env[61806]: value = "session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]52fa187e-f855-5c6b-cb2f-e27c8c54a3f6" [ 673.998363] env[61806]: _type = "Task" [ 673.998363] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 674.005317] env[61806]: DEBUG oslo_vmware.api [None req-c9d6a9df-bcc9-4e08-88fa-de36a15b3a76 tempest-ServersAdmin275Test-894191221 tempest-ServersAdmin275Test-894191221-project-admin] Task: {'id': session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]52fa187e-f855-5c6b-cb2f-e27c8c54a3f6, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 674.121091] env[61806]: DEBUG nova.compute.resource_tracker [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=61806) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 674.121091] env[61806]: DEBUG oslo_concurrency.lockutils [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 14.603s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 674.121091] env[61806]: DEBUG oslo_concurrency.lockutils [None req-88cd9ecf-9bda-478c-b5e6-5a21edd9271f tempest-MigrationsAdminTest-470098049 tempest-MigrationsAdminTest-470098049-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 29.092s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 674.508807] env[61806]: DEBUG oslo_vmware.api [None req-c9d6a9df-bcc9-4e08-88fa-de36a15b3a76 tempest-ServersAdmin275Test-894191221 tempest-ServersAdmin275Test-894191221-project-admin] Task: {'id': session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]52fa187e-f855-5c6b-cb2f-e27c8c54a3f6, 'name': SearchDatastore_Task, 'duration_secs': 0.009434} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 674.509557] env[61806]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c1f75385-52b6-4db3-8dc1-58f7d1b8629c {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 674.515963] env[61806]: DEBUG oslo_vmware.api [None req-c9d6a9df-bcc9-4e08-88fa-de36a15b3a76 tempest-ServersAdmin275Test-894191221 tempest-ServersAdmin275Test-894191221-project-admin] Waiting for the task: (returnval){ [ 674.515963] env[61806]: value = "session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]52d3059a-f6b4-7628-00b7-5a783e01b5bc" [ 674.515963] env[61806]: _type = "Task" [ 674.515963] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 674.523618] env[61806]: DEBUG oslo_vmware.api [None req-c9d6a9df-bcc9-4e08-88fa-de36a15b3a76 tempest-ServersAdmin275Test-894191221 tempest-ServersAdmin275Test-894191221-project-admin] Task: {'id': session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]52d3059a-f6b4-7628-00b7-5a783e01b5bc, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 674.954503] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-53e6b0e0-7172-44d6-8bde-2c30cadef346 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 674.962181] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3685dd71-1240-4693-a012-4986b48195dd {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 674.990932] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d8f434c7-ac36-43f9-9819-8fdf3623717f {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 674.997818] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ad2d1138-9258-4a78-a411-ec34cb0d39eb {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 675.011525] env[61806]: DEBUG nova.compute.provider_tree [None req-88cd9ecf-9bda-478c-b5e6-5a21edd9271f tempest-MigrationsAdminTest-470098049 tempest-MigrationsAdminTest-470098049-project-member] Inventory has not changed in ProviderTree for provider: a2858be1-fd22-4e08-979e-87ad25293407 {{(pid=61806) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 675.025410] env[61806]: DEBUG oslo_vmware.api [None req-c9d6a9df-bcc9-4e08-88fa-de36a15b3a76 tempest-ServersAdmin275Test-894191221 tempest-ServersAdmin275Test-894191221-project-admin] Task: {'id': session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]52d3059a-f6b4-7628-00b7-5a783e01b5bc, 'name': SearchDatastore_Task, 'duration_secs': 0.009323} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 675.025647] env[61806]: DEBUG oslo_concurrency.lockutils [None req-c9d6a9df-bcc9-4e08-88fa-de36a15b3a76 tempest-ServersAdmin275Test-894191221 tempest-ServersAdmin275Test-894191221-project-admin] Releasing lock "[datastore1] devstack-image-cache_base/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa.vmdk" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 675.025889] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [None req-c9d6a9df-bcc9-4e08-88fa-de36a15b3a76 tempest-ServersAdmin275Test-894191221 tempest-ServersAdmin275Test-894191221-project-admin] Copying Virtual Disk [datastore1] devstack-image-cache_base/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa.vmdk to [datastore1] 9db80c3c-e2c3-4034-a79a-d7be924c4ff6/9db80c3c-e2c3-4034-a79a-d7be924c4ff6.vmdk {{(pid=61806) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 675.026129] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-8a169832-adc6-473c-aef4-029145f02b63 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 675.032578] env[61806]: DEBUG oslo_vmware.api [None req-c9d6a9df-bcc9-4e08-88fa-de36a15b3a76 tempest-ServersAdmin275Test-894191221 tempest-ServersAdmin275Test-894191221-project-admin] Waiting for the task: (returnval){ [ 675.032578] env[61806]: value = "task-1294499" [ 675.032578] env[61806]: _type = "Task" [ 675.032578] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 675.039755] env[61806]: DEBUG oslo_vmware.api [None req-c9d6a9df-bcc9-4e08-88fa-de36a15b3a76 tempest-ServersAdmin275Test-894191221 tempest-ServersAdmin275Test-894191221-project-admin] Task: {'id': task-1294499, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 675.514662] env[61806]: DEBUG nova.scheduler.client.report [None req-88cd9ecf-9bda-478c-b5e6-5a21edd9271f tempest-MigrationsAdminTest-470098049 tempest-MigrationsAdminTest-470098049-project-member] Inventory has not changed for provider a2858be1-fd22-4e08-979e-87ad25293407 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 139, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61806) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 675.542930] env[61806]: DEBUG oslo_vmware.api [None req-c9d6a9df-bcc9-4e08-88fa-de36a15b3a76 tempest-ServersAdmin275Test-894191221 tempest-ServersAdmin275Test-894191221-project-admin] Task: {'id': task-1294499, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.477872} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 675.543278] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [None req-c9d6a9df-bcc9-4e08-88fa-de36a15b3a76 tempest-ServersAdmin275Test-894191221 tempest-ServersAdmin275Test-894191221-project-admin] Copied Virtual Disk [datastore1] devstack-image-cache_base/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa.vmdk to [datastore1] 9db80c3c-e2c3-4034-a79a-d7be924c4ff6/9db80c3c-e2c3-4034-a79a-d7be924c4ff6.vmdk {{(pid=61806) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 675.543499] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-c9d6a9df-bcc9-4e08-88fa-de36a15b3a76 tempest-ServersAdmin275Test-894191221 tempest-ServersAdmin275Test-894191221-project-admin] [instance: 9db80c3c-e2c3-4034-a79a-d7be924c4ff6] Extending root virtual disk to 1048576 {{(pid=61806) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 675.543748] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-b4614ba1-49f3-405f-9633-ee71d0523945 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 675.550508] env[61806]: DEBUG oslo_vmware.api [None req-c9d6a9df-bcc9-4e08-88fa-de36a15b3a76 tempest-ServersAdmin275Test-894191221 tempest-ServersAdmin275Test-894191221-project-admin] Waiting for the task: (returnval){ [ 675.550508] env[61806]: value = "task-1294500" [ 675.550508] env[61806]: _type = "Task" [ 675.550508] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 675.558323] env[61806]: DEBUG oslo_vmware.api [None req-c9d6a9df-bcc9-4e08-88fa-de36a15b3a76 tempest-ServersAdmin275Test-894191221 tempest-ServersAdmin275Test-894191221-project-admin] Task: {'id': task-1294500, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 676.020089] env[61806]: DEBUG oslo_concurrency.lockutils [None req-88cd9ecf-9bda-478c-b5e6-5a21edd9271f tempest-MigrationsAdminTest-470098049 tempest-MigrationsAdminTest-470098049-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.899s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 676.020843] env[61806]: ERROR nova.compute.manager [None req-88cd9ecf-9bda-478c-b5e6-5a21edd9271f tempest-MigrationsAdminTest-470098049 tempest-MigrationsAdminTest-470098049-project-member] [instance: e66c921c-320e-4347-96c9-27aa80ba7f41] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port e5cb5794-df67-4846-a386-fcf7657ff23f, please check neutron logs for more information. [ 676.020843] env[61806]: ERROR nova.compute.manager [instance: e66c921c-320e-4347-96c9-27aa80ba7f41] Traceback (most recent call last): [ 676.020843] env[61806]: ERROR nova.compute.manager [instance: e66c921c-320e-4347-96c9-27aa80ba7f41] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 676.020843] env[61806]: ERROR nova.compute.manager [instance: e66c921c-320e-4347-96c9-27aa80ba7f41] self.driver.spawn(context, instance, image_meta, [ 676.020843] env[61806]: ERROR nova.compute.manager [instance: e66c921c-320e-4347-96c9-27aa80ba7f41] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 676.020843] env[61806]: ERROR nova.compute.manager [instance: e66c921c-320e-4347-96c9-27aa80ba7f41] self._vmops.spawn(context, instance, image_meta, injected_files, [ 676.020843] env[61806]: ERROR nova.compute.manager [instance: e66c921c-320e-4347-96c9-27aa80ba7f41] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 676.020843] env[61806]: ERROR nova.compute.manager [instance: e66c921c-320e-4347-96c9-27aa80ba7f41] vm_ref = self.build_virtual_machine(instance, [ 676.020843] env[61806]: ERROR nova.compute.manager [instance: e66c921c-320e-4347-96c9-27aa80ba7f41] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 676.020843] env[61806]: ERROR nova.compute.manager [instance: e66c921c-320e-4347-96c9-27aa80ba7f41] vif_infos = vmwarevif.get_vif_info(self._session, [ 676.020843] env[61806]: ERROR nova.compute.manager [instance: e66c921c-320e-4347-96c9-27aa80ba7f41] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 676.021233] env[61806]: ERROR nova.compute.manager [instance: e66c921c-320e-4347-96c9-27aa80ba7f41] for vif in network_info: [ 676.021233] env[61806]: ERROR nova.compute.manager [instance: e66c921c-320e-4347-96c9-27aa80ba7f41] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 676.021233] env[61806]: ERROR nova.compute.manager [instance: e66c921c-320e-4347-96c9-27aa80ba7f41] return self._sync_wrapper(fn, *args, **kwargs) [ 676.021233] env[61806]: ERROR nova.compute.manager [instance: e66c921c-320e-4347-96c9-27aa80ba7f41] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 676.021233] env[61806]: ERROR nova.compute.manager [instance: e66c921c-320e-4347-96c9-27aa80ba7f41] self.wait() [ 676.021233] env[61806]: ERROR nova.compute.manager [instance: e66c921c-320e-4347-96c9-27aa80ba7f41] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 676.021233] env[61806]: ERROR nova.compute.manager [instance: e66c921c-320e-4347-96c9-27aa80ba7f41] self[:] = self._gt.wait() [ 676.021233] env[61806]: ERROR nova.compute.manager [instance: e66c921c-320e-4347-96c9-27aa80ba7f41] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 676.021233] env[61806]: ERROR nova.compute.manager [instance: e66c921c-320e-4347-96c9-27aa80ba7f41] return self._exit_event.wait() [ 676.021233] env[61806]: ERROR nova.compute.manager [instance: e66c921c-320e-4347-96c9-27aa80ba7f41] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 676.021233] env[61806]: ERROR nova.compute.manager [instance: e66c921c-320e-4347-96c9-27aa80ba7f41] current.throw(*self._exc) [ 676.021233] env[61806]: ERROR nova.compute.manager [instance: e66c921c-320e-4347-96c9-27aa80ba7f41] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 676.021233] env[61806]: ERROR nova.compute.manager [instance: e66c921c-320e-4347-96c9-27aa80ba7f41] result = function(*args, **kwargs) [ 676.021674] env[61806]: ERROR nova.compute.manager [instance: e66c921c-320e-4347-96c9-27aa80ba7f41] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 676.021674] env[61806]: ERROR nova.compute.manager [instance: e66c921c-320e-4347-96c9-27aa80ba7f41] return func(*args, **kwargs) [ 676.021674] env[61806]: ERROR nova.compute.manager [instance: e66c921c-320e-4347-96c9-27aa80ba7f41] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 676.021674] env[61806]: ERROR nova.compute.manager [instance: e66c921c-320e-4347-96c9-27aa80ba7f41] raise e [ 676.021674] env[61806]: ERROR nova.compute.manager [instance: e66c921c-320e-4347-96c9-27aa80ba7f41] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 676.021674] env[61806]: ERROR nova.compute.manager [instance: e66c921c-320e-4347-96c9-27aa80ba7f41] nwinfo = self.network_api.allocate_for_instance( [ 676.021674] env[61806]: ERROR nova.compute.manager [instance: e66c921c-320e-4347-96c9-27aa80ba7f41] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 676.021674] env[61806]: ERROR nova.compute.manager [instance: e66c921c-320e-4347-96c9-27aa80ba7f41] created_port_ids = self._update_ports_for_instance( [ 676.021674] env[61806]: ERROR nova.compute.manager [instance: e66c921c-320e-4347-96c9-27aa80ba7f41] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 676.021674] env[61806]: ERROR nova.compute.manager [instance: e66c921c-320e-4347-96c9-27aa80ba7f41] with excutils.save_and_reraise_exception(): [ 676.021674] env[61806]: ERROR nova.compute.manager [instance: e66c921c-320e-4347-96c9-27aa80ba7f41] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 676.021674] env[61806]: ERROR nova.compute.manager [instance: e66c921c-320e-4347-96c9-27aa80ba7f41] self.force_reraise() [ 676.021674] env[61806]: ERROR nova.compute.manager [instance: e66c921c-320e-4347-96c9-27aa80ba7f41] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 676.022075] env[61806]: ERROR nova.compute.manager [instance: e66c921c-320e-4347-96c9-27aa80ba7f41] raise self.value [ 676.022075] env[61806]: ERROR nova.compute.manager [instance: e66c921c-320e-4347-96c9-27aa80ba7f41] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 676.022075] env[61806]: ERROR nova.compute.manager [instance: e66c921c-320e-4347-96c9-27aa80ba7f41] updated_port = self._update_port( [ 676.022075] env[61806]: ERROR nova.compute.manager [instance: e66c921c-320e-4347-96c9-27aa80ba7f41] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 676.022075] env[61806]: ERROR nova.compute.manager [instance: e66c921c-320e-4347-96c9-27aa80ba7f41] _ensure_no_port_binding_failure(port) [ 676.022075] env[61806]: ERROR nova.compute.manager [instance: e66c921c-320e-4347-96c9-27aa80ba7f41] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 676.022075] env[61806]: ERROR nova.compute.manager [instance: e66c921c-320e-4347-96c9-27aa80ba7f41] raise exception.PortBindingFailed(port_id=port['id']) [ 676.022075] env[61806]: ERROR nova.compute.manager [instance: e66c921c-320e-4347-96c9-27aa80ba7f41] nova.exception.PortBindingFailed: Binding failed for port e5cb5794-df67-4846-a386-fcf7657ff23f, please check neutron logs for more information. [ 676.022075] env[61806]: ERROR nova.compute.manager [instance: e66c921c-320e-4347-96c9-27aa80ba7f41] [ 676.022075] env[61806]: DEBUG nova.compute.utils [None req-88cd9ecf-9bda-478c-b5e6-5a21edd9271f tempest-MigrationsAdminTest-470098049 tempest-MigrationsAdminTest-470098049-project-member] [instance: e66c921c-320e-4347-96c9-27aa80ba7f41] Binding failed for port e5cb5794-df67-4846-a386-fcf7657ff23f, please check neutron logs for more information. {{(pid=61806) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 676.023359] env[61806]: DEBUG oslo_concurrency.lockutils [None req-7bb45eee-0cbe-4995-9898-530114f6f104 tempest-ImagesOneServerNegativeTestJSON-606465252 tempest-ImagesOneServerNegativeTestJSON-606465252-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 26.652s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 676.024790] env[61806]: INFO nova.compute.claims [None req-7bb45eee-0cbe-4995-9898-530114f6f104 tempest-ImagesOneServerNegativeTestJSON-606465252 tempest-ImagesOneServerNegativeTestJSON-606465252-project-member] [instance: 8d0cdc81-56ed-4a7a-8221-30bb906e16b2] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 676.027574] env[61806]: DEBUG nova.compute.manager [None req-88cd9ecf-9bda-478c-b5e6-5a21edd9271f tempest-MigrationsAdminTest-470098049 tempest-MigrationsAdminTest-470098049-project-member] [instance: e66c921c-320e-4347-96c9-27aa80ba7f41] Build of instance e66c921c-320e-4347-96c9-27aa80ba7f41 was re-scheduled: Binding failed for port e5cb5794-df67-4846-a386-fcf7657ff23f, please check neutron logs for more information. {{(pid=61806) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 676.027830] env[61806]: DEBUG nova.compute.manager [None req-88cd9ecf-9bda-478c-b5e6-5a21edd9271f tempest-MigrationsAdminTest-470098049 tempest-MigrationsAdminTest-470098049-project-member] [instance: e66c921c-320e-4347-96c9-27aa80ba7f41] Unplugging VIFs for instance {{(pid=61806) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 676.028069] env[61806]: DEBUG oslo_concurrency.lockutils [None req-88cd9ecf-9bda-478c-b5e6-5a21edd9271f tempest-MigrationsAdminTest-470098049 tempest-MigrationsAdminTest-470098049-project-member] Acquiring lock "refresh_cache-e66c921c-320e-4347-96c9-27aa80ba7f41" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 676.028224] env[61806]: DEBUG oslo_concurrency.lockutils [None req-88cd9ecf-9bda-478c-b5e6-5a21edd9271f tempest-MigrationsAdminTest-470098049 tempest-MigrationsAdminTest-470098049-project-member] Acquired lock "refresh_cache-e66c921c-320e-4347-96c9-27aa80ba7f41" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 676.028385] env[61806]: DEBUG nova.network.neutron [None req-88cd9ecf-9bda-478c-b5e6-5a21edd9271f tempest-MigrationsAdminTest-470098049 tempest-MigrationsAdminTest-470098049-project-member] [instance: e66c921c-320e-4347-96c9-27aa80ba7f41] Building network info cache for instance {{(pid=61806) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 676.060501] env[61806]: DEBUG oslo_vmware.api [None req-c9d6a9df-bcc9-4e08-88fa-de36a15b3a76 tempest-ServersAdmin275Test-894191221 tempest-ServersAdmin275Test-894191221-project-admin] Task: {'id': task-1294500, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.081097} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 676.060771] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-c9d6a9df-bcc9-4e08-88fa-de36a15b3a76 tempest-ServersAdmin275Test-894191221 tempest-ServersAdmin275Test-894191221-project-admin] [instance: 9db80c3c-e2c3-4034-a79a-d7be924c4ff6] Extended root virtual disk {{(pid=61806) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 676.061541] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e0e5ca1c-a27f-459b-95ca-6b515bb4227a {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 676.081369] env[61806]: DEBUG nova.virt.vmwareapi.volumeops [None req-c9d6a9df-bcc9-4e08-88fa-de36a15b3a76 tempest-ServersAdmin275Test-894191221 tempest-ServersAdmin275Test-894191221-project-admin] [instance: 9db80c3c-e2c3-4034-a79a-d7be924c4ff6] Reconfiguring VM instance instance-0000001f to attach disk [datastore1] 9db80c3c-e2c3-4034-a79a-d7be924c4ff6/9db80c3c-e2c3-4034-a79a-d7be924c4ff6.vmdk or device None with type sparse {{(pid=61806) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 676.081822] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-d20cd166-7710-4459-a072-eafa1654b33d {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 676.100701] env[61806]: DEBUG oslo_vmware.api [None req-c9d6a9df-bcc9-4e08-88fa-de36a15b3a76 tempest-ServersAdmin275Test-894191221 tempest-ServersAdmin275Test-894191221-project-admin] Waiting for the task: (returnval){ [ 676.100701] env[61806]: value = "task-1294501" [ 676.100701] env[61806]: _type = "Task" [ 676.100701] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 676.108083] env[61806]: DEBUG oslo_vmware.api [None req-c9d6a9df-bcc9-4e08-88fa-de36a15b3a76 tempest-ServersAdmin275Test-894191221 tempest-ServersAdmin275Test-894191221-project-admin] Task: {'id': task-1294501, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 676.548352] env[61806]: DEBUG nova.network.neutron [None req-88cd9ecf-9bda-478c-b5e6-5a21edd9271f tempest-MigrationsAdminTest-470098049 tempest-MigrationsAdminTest-470098049-project-member] [instance: e66c921c-320e-4347-96c9-27aa80ba7f41] Instance cache missing network info. {{(pid=61806) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 676.598569] env[61806]: DEBUG nova.network.neutron [None req-88cd9ecf-9bda-478c-b5e6-5a21edd9271f tempest-MigrationsAdminTest-470098049 tempest-MigrationsAdminTest-470098049-project-member] [instance: e66c921c-320e-4347-96c9-27aa80ba7f41] Updating instance_info_cache with network_info: [] {{(pid=61806) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 676.611279] env[61806]: DEBUG oslo_vmware.api [None req-c9d6a9df-bcc9-4e08-88fa-de36a15b3a76 tempest-ServersAdmin275Test-894191221 tempest-ServersAdmin275Test-894191221-project-admin] Task: {'id': task-1294501, 'name': ReconfigVM_Task, 'duration_secs': 0.290626} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 676.611543] env[61806]: DEBUG nova.virt.vmwareapi.volumeops [None req-c9d6a9df-bcc9-4e08-88fa-de36a15b3a76 tempest-ServersAdmin275Test-894191221 tempest-ServersAdmin275Test-894191221-project-admin] [instance: 9db80c3c-e2c3-4034-a79a-d7be924c4ff6] Reconfigured VM instance instance-0000001f to attach disk [datastore1] 9db80c3c-e2c3-4034-a79a-d7be924c4ff6/9db80c3c-e2c3-4034-a79a-d7be924c4ff6.vmdk or device None with type sparse {{(pid=61806) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 676.612150] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-73d3d486-ec4d-42a2-860d-2c6c77a0559b {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 676.619097] env[61806]: DEBUG oslo_vmware.api [None req-c9d6a9df-bcc9-4e08-88fa-de36a15b3a76 tempest-ServersAdmin275Test-894191221 tempest-ServersAdmin275Test-894191221-project-admin] Waiting for the task: (returnval){ [ 676.619097] env[61806]: value = "task-1294502" [ 676.619097] env[61806]: _type = "Task" [ 676.619097] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 676.628376] env[61806]: DEBUG oslo_vmware.api [None req-c9d6a9df-bcc9-4e08-88fa-de36a15b3a76 tempest-ServersAdmin275Test-894191221 tempest-ServersAdmin275Test-894191221-project-admin] Task: {'id': task-1294502, 'name': Rename_Task} progress is 5%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 677.101546] env[61806]: DEBUG oslo_concurrency.lockutils [None req-88cd9ecf-9bda-478c-b5e6-5a21edd9271f tempest-MigrationsAdminTest-470098049 tempest-MigrationsAdminTest-470098049-project-member] Releasing lock "refresh_cache-e66c921c-320e-4347-96c9-27aa80ba7f41" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 677.101944] env[61806]: DEBUG nova.compute.manager [None req-88cd9ecf-9bda-478c-b5e6-5a21edd9271f tempest-MigrationsAdminTest-470098049 tempest-MigrationsAdminTest-470098049-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61806) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 677.101944] env[61806]: DEBUG nova.compute.manager [None req-88cd9ecf-9bda-478c-b5e6-5a21edd9271f tempest-MigrationsAdminTest-470098049 tempest-MigrationsAdminTest-470098049-project-member] [instance: e66c921c-320e-4347-96c9-27aa80ba7f41] Deallocating network for instance {{(pid=61806) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 677.102363] env[61806]: DEBUG nova.network.neutron [None req-88cd9ecf-9bda-478c-b5e6-5a21edd9271f tempest-MigrationsAdminTest-470098049 tempest-MigrationsAdminTest-470098049-project-member] [instance: e66c921c-320e-4347-96c9-27aa80ba7f41] deallocate_for_instance() {{(pid=61806) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 677.126503] env[61806]: DEBUG nova.network.neutron [None req-88cd9ecf-9bda-478c-b5e6-5a21edd9271f tempest-MigrationsAdminTest-470098049 tempest-MigrationsAdminTest-470098049-project-member] [instance: e66c921c-320e-4347-96c9-27aa80ba7f41] Instance cache missing network info. {{(pid=61806) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 677.133947] env[61806]: DEBUG oslo_vmware.api [None req-c9d6a9df-bcc9-4e08-88fa-de36a15b3a76 tempest-ServersAdmin275Test-894191221 tempest-ServersAdmin275Test-894191221-project-admin] Task: {'id': task-1294502, 'name': Rename_Task, 'duration_secs': 0.13267} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 677.133947] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [None req-c9d6a9df-bcc9-4e08-88fa-de36a15b3a76 tempest-ServersAdmin275Test-894191221 tempest-ServersAdmin275Test-894191221-project-admin] [instance: 9db80c3c-e2c3-4034-a79a-d7be924c4ff6] Powering on the VM {{(pid=61806) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 677.133947] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-1b46c717-0145-4399-9d1d-e4d34fa8a1ac {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 677.141485] env[61806]: DEBUG oslo_vmware.api [None req-c9d6a9df-bcc9-4e08-88fa-de36a15b3a76 tempest-ServersAdmin275Test-894191221 tempest-ServersAdmin275Test-894191221-project-admin] Waiting for the task: (returnval){ [ 677.141485] env[61806]: value = "task-1294503" [ 677.141485] env[61806]: _type = "Task" [ 677.141485] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 677.153572] env[61806]: DEBUG oslo_vmware.api [None req-c9d6a9df-bcc9-4e08-88fa-de36a15b3a76 tempest-ServersAdmin275Test-894191221 tempest-ServersAdmin275Test-894191221-project-admin] Task: {'id': task-1294503, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 677.432787] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-106c137e-eb15-481f-b907-b15dbdaf1a1a {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 677.440384] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-94d9b7e7-1ac3-4bee-a7e5-c66ce4495fbe {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 677.470746] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7a60ec10-eae2-43c6-bd93-fe7086bf1208 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 677.478194] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d88af050-414f-4c98-bc22-865095d05d29 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 677.491172] env[61806]: DEBUG nova.compute.provider_tree [None req-7bb45eee-0cbe-4995-9898-530114f6f104 tempest-ImagesOneServerNegativeTestJSON-606465252 tempest-ImagesOneServerNegativeTestJSON-606465252-project-member] Updating inventory in ProviderTree for provider a2858be1-fd22-4e08-979e-87ad25293407 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 138, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61806) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 677.629133] env[61806]: DEBUG nova.network.neutron [None req-88cd9ecf-9bda-478c-b5e6-5a21edd9271f tempest-MigrationsAdminTest-470098049 tempest-MigrationsAdminTest-470098049-project-member] [instance: e66c921c-320e-4347-96c9-27aa80ba7f41] Updating instance_info_cache with network_info: [] {{(pid=61806) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 677.650735] env[61806]: DEBUG oslo_vmware.api [None req-c9d6a9df-bcc9-4e08-88fa-de36a15b3a76 tempest-ServersAdmin275Test-894191221 tempest-ServersAdmin275Test-894191221-project-admin] Task: {'id': task-1294503, 'name': PowerOnVM_Task, 'duration_secs': 0.442808} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 677.650923] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [None req-c9d6a9df-bcc9-4e08-88fa-de36a15b3a76 tempest-ServersAdmin275Test-894191221 tempest-ServersAdmin275Test-894191221-project-admin] [instance: 9db80c3c-e2c3-4034-a79a-d7be924c4ff6] Powered on the VM {{(pid=61806) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 677.651157] env[61806]: DEBUG nova.compute.manager [None req-c9d6a9df-bcc9-4e08-88fa-de36a15b3a76 tempest-ServersAdmin275Test-894191221 tempest-ServersAdmin275Test-894191221-project-admin] [instance: 9db80c3c-e2c3-4034-a79a-d7be924c4ff6] Checking state {{(pid=61806) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 677.651860] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7c094a65-7536-4f7a-be1e-3988481edd97 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 678.022689] env[61806]: DEBUG nova.scheduler.client.report [None req-7bb45eee-0cbe-4995-9898-530114f6f104 tempest-ImagesOneServerNegativeTestJSON-606465252 tempest-ImagesOneServerNegativeTestJSON-606465252-project-member] Updated inventory for provider a2858be1-fd22-4e08-979e-87ad25293407 with generation 65 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 138, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61806) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:971}} [ 678.022953] env[61806]: DEBUG nova.compute.provider_tree [None req-7bb45eee-0cbe-4995-9898-530114f6f104 tempest-ImagesOneServerNegativeTestJSON-606465252 tempest-ImagesOneServerNegativeTestJSON-606465252-project-member] Updating resource provider a2858be1-fd22-4e08-979e-87ad25293407 generation from 65 to 66 during operation: update_inventory {{(pid=61806) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 678.023148] env[61806]: DEBUG nova.compute.provider_tree [None req-7bb45eee-0cbe-4995-9898-530114f6f104 tempest-ImagesOneServerNegativeTestJSON-606465252 tempest-ImagesOneServerNegativeTestJSON-606465252-project-member] Updating inventory in ProviderTree for provider a2858be1-fd22-4e08-979e-87ad25293407 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 138, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61806) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 678.132230] env[61806]: INFO nova.compute.manager [None req-88cd9ecf-9bda-478c-b5e6-5a21edd9271f tempest-MigrationsAdminTest-470098049 tempest-MigrationsAdminTest-470098049-project-member] [instance: e66c921c-320e-4347-96c9-27aa80ba7f41] Took 1.03 seconds to deallocate network for instance. [ 678.168905] env[61806]: DEBUG oslo_concurrency.lockutils [None req-c9d6a9df-bcc9-4e08-88fa-de36a15b3a76 tempest-ServersAdmin275Test-894191221 tempest-ServersAdmin275Test-894191221-project-admin] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 678.527943] env[61806]: DEBUG oslo_concurrency.lockutils [None req-7bb45eee-0cbe-4995-9898-530114f6f104 tempest-ImagesOneServerNegativeTestJSON-606465252 tempest-ImagesOneServerNegativeTestJSON-606465252-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.505s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 678.528504] env[61806]: DEBUG nova.compute.manager [None req-7bb45eee-0cbe-4995-9898-530114f6f104 tempest-ImagesOneServerNegativeTestJSON-606465252 tempest-ImagesOneServerNegativeTestJSON-606465252-project-member] [instance: 8d0cdc81-56ed-4a7a-8221-30bb906e16b2] Start building networks asynchronously for instance. {{(pid=61806) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 678.531067] env[61806]: DEBUG oslo_concurrency.lockutils [None req-44eba412-ca3f-4386-8c36-f42e9734628d tempest-ServersTestBootFromVolume-1815739798 tempest-ServersTestBootFromVolume-1815739798-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 28.591s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 679.036097] env[61806]: DEBUG nova.compute.utils [None req-7bb45eee-0cbe-4995-9898-530114f6f104 tempest-ImagesOneServerNegativeTestJSON-606465252 tempest-ImagesOneServerNegativeTestJSON-606465252-project-member] Using /dev/sd instead of None {{(pid=61806) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 679.041040] env[61806]: DEBUG nova.compute.manager [None req-7bb45eee-0cbe-4995-9898-530114f6f104 tempest-ImagesOneServerNegativeTestJSON-606465252 tempest-ImagesOneServerNegativeTestJSON-606465252-project-member] [instance: 8d0cdc81-56ed-4a7a-8221-30bb906e16b2] Allocating IP information in the background. {{(pid=61806) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 679.041040] env[61806]: DEBUG nova.network.neutron [None req-7bb45eee-0cbe-4995-9898-530114f6f104 tempest-ImagesOneServerNegativeTestJSON-606465252 tempest-ImagesOneServerNegativeTestJSON-606465252-project-member] [instance: 8d0cdc81-56ed-4a7a-8221-30bb906e16b2] allocate_for_instance() {{(pid=61806) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 679.097409] env[61806]: DEBUG nova.policy [None req-7bb45eee-0cbe-4995-9898-530114f6f104 tempest-ImagesOneServerNegativeTestJSON-606465252 tempest-ImagesOneServerNegativeTestJSON-606465252-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '4921c17d6e1b4f668ad289f4e76dd1a1', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'e56eacced7bc4d3787462c9f1a1f053c', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61806) authorize /opt/stack/nova/nova/policy.py:201}} [ 679.105774] env[61806]: DEBUG oslo_concurrency.lockutils [None req-6168822a-f60c-43e0-b0ac-2eb07b4977dc tempest-ServersAdmin275Test-1945525728 tempest-ServersAdmin275Test-1945525728-project-member] Acquiring lock "9db80c3c-e2c3-4034-a79a-d7be924c4ff6" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 679.106009] env[61806]: DEBUG oslo_concurrency.lockutils [None req-6168822a-f60c-43e0-b0ac-2eb07b4977dc tempest-ServersAdmin275Test-1945525728 tempest-ServersAdmin275Test-1945525728-project-member] Lock "9db80c3c-e2c3-4034-a79a-d7be924c4ff6" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 679.106219] env[61806]: DEBUG oslo_concurrency.lockutils [None req-6168822a-f60c-43e0-b0ac-2eb07b4977dc tempest-ServersAdmin275Test-1945525728 tempest-ServersAdmin275Test-1945525728-project-member] Acquiring lock "9db80c3c-e2c3-4034-a79a-d7be924c4ff6-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 679.106597] env[61806]: DEBUG oslo_concurrency.lockutils [None req-6168822a-f60c-43e0-b0ac-2eb07b4977dc tempest-ServersAdmin275Test-1945525728 tempest-ServersAdmin275Test-1945525728-project-member] Lock "9db80c3c-e2c3-4034-a79a-d7be924c4ff6-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 679.106597] env[61806]: DEBUG oslo_concurrency.lockutils [None req-6168822a-f60c-43e0-b0ac-2eb07b4977dc tempest-ServersAdmin275Test-1945525728 tempest-ServersAdmin275Test-1945525728-project-member] Lock "9db80c3c-e2c3-4034-a79a-d7be924c4ff6-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 679.108617] env[61806]: INFO nova.compute.manager [None req-6168822a-f60c-43e0-b0ac-2eb07b4977dc tempest-ServersAdmin275Test-1945525728 tempest-ServersAdmin275Test-1945525728-project-member] [instance: 9db80c3c-e2c3-4034-a79a-d7be924c4ff6] Terminating instance [ 679.112692] env[61806]: DEBUG oslo_concurrency.lockutils [None req-6168822a-f60c-43e0-b0ac-2eb07b4977dc tempest-ServersAdmin275Test-1945525728 tempest-ServersAdmin275Test-1945525728-project-member] Acquiring lock "refresh_cache-9db80c3c-e2c3-4034-a79a-d7be924c4ff6" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 679.112849] env[61806]: DEBUG oslo_concurrency.lockutils [None req-6168822a-f60c-43e0-b0ac-2eb07b4977dc tempest-ServersAdmin275Test-1945525728 tempest-ServersAdmin275Test-1945525728-project-member] Acquired lock "refresh_cache-9db80c3c-e2c3-4034-a79a-d7be924c4ff6" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 679.113552] env[61806]: DEBUG nova.network.neutron [None req-6168822a-f60c-43e0-b0ac-2eb07b4977dc tempest-ServersAdmin275Test-1945525728 tempest-ServersAdmin275Test-1945525728-project-member] [instance: 9db80c3c-e2c3-4034-a79a-d7be924c4ff6] Building network info cache for instance {{(pid=61806) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 679.168463] env[61806]: INFO nova.scheduler.client.report [None req-88cd9ecf-9bda-478c-b5e6-5a21edd9271f tempest-MigrationsAdminTest-470098049 tempest-MigrationsAdminTest-470098049-project-member] Deleted allocations for instance e66c921c-320e-4347-96c9-27aa80ba7f41 [ 679.408111] env[61806]: DEBUG nova.network.neutron [None req-7bb45eee-0cbe-4995-9898-530114f6f104 tempest-ImagesOneServerNegativeTestJSON-606465252 tempest-ImagesOneServerNegativeTestJSON-606465252-project-member] [instance: 8d0cdc81-56ed-4a7a-8221-30bb906e16b2] Successfully created port: 4cd363a6-ef40-4b4b-909f-85a169e5c84e {{(pid=61806) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 679.431573] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-18e38202-747e-463b-a934-57b5eaed3161 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 679.439547] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c61504c5-d1ff-49eb-a780-6481809a55f3 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 679.470705] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b7db329c-052e-456c-9482-9976530b2023 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 679.477935] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f9b124bf-3c58-4f2e-99ee-66d01a620901 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 679.490947] env[61806]: DEBUG nova.compute.provider_tree [None req-44eba412-ca3f-4386-8c36-f42e9734628d tempest-ServersTestBootFromVolume-1815739798 tempest-ServersTestBootFromVolume-1815739798-project-member] Inventory has not changed in ProviderTree for provider: a2858be1-fd22-4e08-979e-87ad25293407 {{(pid=61806) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 679.541291] env[61806]: DEBUG nova.compute.manager [None req-7bb45eee-0cbe-4995-9898-530114f6f104 tempest-ImagesOneServerNegativeTestJSON-606465252 tempest-ImagesOneServerNegativeTestJSON-606465252-project-member] [instance: 8d0cdc81-56ed-4a7a-8221-30bb906e16b2] Start building block device mappings for instance. {{(pid=61806) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 679.634083] env[61806]: DEBUG nova.network.neutron [None req-6168822a-f60c-43e0-b0ac-2eb07b4977dc tempest-ServersAdmin275Test-1945525728 tempest-ServersAdmin275Test-1945525728-project-member] [instance: 9db80c3c-e2c3-4034-a79a-d7be924c4ff6] Instance cache missing network info. {{(pid=61806) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 679.678260] env[61806]: DEBUG oslo_concurrency.lockutils [None req-88cd9ecf-9bda-478c-b5e6-5a21edd9271f tempest-MigrationsAdminTest-470098049 tempest-MigrationsAdminTest-470098049-project-member] Lock "e66c921c-320e-4347-96c9-27aa80ba7f41" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 119.069s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 679.698834] env[61806]: DEBUG nova.network.neutron [None req-6168822a-f60c-43e0-b0ac-2eb07b4977dc tempest-ServersAdmin275Test-1945525728 tempest-ServersAdmin275Test-1945525728-project-member] [instance: 9db80c3c-e2c3-4034-a79a-d7be924c4ff6] Updating instance_info_cache with network_info: [] {{(pid=61806) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 679.993791] env[61806]: DEBUG nova.scheduler.client.report [None req-44eba412-ca3f-4386-8c36-f42e9734628d tempest-ServersTestBootFromVolume-1815739798 tempest-ServersTestBootFromVolume-1815739798-project-member] Inventory has not changed for provider a2858be1-fd22-4e08-979e-87ad25293407 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 138, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61806) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 680.181581] env[61806]: DEBUG nova.compute.manager [None req-fcbc4cf1-fc11-42b1-9d49-ef556b52782e tempest-ServerRescueNegativeTestJSON-1483258717 tempest-ServerRescueNegativeTestJSON-1483258717-project-member] [instance: 7ec2497a-7329-4ed4-99ab-dec3d1e8117c] Starting instance... {{(pid=61806) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 680.201333] env[61806]: DEBUG oslo_concurrency.lockutils [None req-6168822a-f60c-43e0-b0ac-2eb07b4977dc tempest-ServersAdmin275Test-1945525728 tempest-ServersAdmin275Test-1945525728-project-member] Releasing lock "refresh_cache-9db80c3c-e2c3-4034-a79a-d7be924c4ff6" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 680.201805] env[61806]: DEBUG nova.compute.manager [None req-6168822a-f60c-43e0-b0ac-2eb07b4977dc tempest-ServersAdmin275Test-1945525728 tempest-ServersAdmin275Test-1945525728-project-member] [instance: 9db80c3c-e2c3-4034-a79a-d7be924c4ff6] Start destroying the instance on the hypervisor. {{(pid=61806) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 680.202009] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-6168822a-f60c-43e0-b0ac-2eb07b4977dc tempest-ServersAdmin275Test-1945525728 tempest-ServersAdmin275Test-1945525728-project-member] [instance: 9db80c3c-e2c3-4034-a79a-d7be924c4ff6] Destroying instance {{(pid=61806) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 680.204993] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cc33ac20-ba28-4411-9243-4bc5cf02cd65 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 680.214184] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [None req-6168822a-f60c-43e0-b0ac-2eb07b4977dc tempest-ServersAdmin275Test-1945525728 tempest-ServersAdmin275Test-1945525728-project-member] [instance: 9db80c3c-e2c3-4034-a79a-d7be924c4ff6] Powering off the VM {{(pid=61806) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 680.216332] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-7359e25f-e741-4ada-845e-31d4b98c6caa {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 680.223236] env[61806]: DEBUG oslo_vmware.api [None req-6168822a-f60c-43e0-b0ac-2eb07b4977dc tempest-ServersAdmin275Test-1945525728 tempest-ServersAdmin275Test-1945525728-project-member] Waiting for the task: (returnval){ [ 680.223236] env[61806]: value = "task-1294504" [ 680.223236] env[61806]: _type = "Task" [ 680.223236] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 680.229364] env[61806]: DEBUG nova.compute.manager [req-cbaf3ea0-ff4b-4786-9320-9f666d26dae5 req-baa15dc2-6b2c-4a0b-89c1-594406be1d08 service nova] [instance: 8d0cdc81-56ed-4a7a-8221-30bb906e16b2] Received event network-changed-4cd363a6-ef40-4b4b-909f-85a169e5c84e {{(pid=61806) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 680.229364] env[61806]: DEBUG nova.compute.manager [req-cbaf3ea0-ff4b-4786-9320-9f666d26dae5 req-baa15dc2-6b2c-4a0b-89c1-594406be1d08 service nova] [instance: 8d0cdc81-56ed-4a7a-8221-30bb906e16b2] Refreshing instance network info cache due to event network-changed-4cd363a6-ef40-4b4b-909f-85a169e5c84e. {{(pid=61806) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 680.229638] env[61806]: DEBUG oslo_concurrency.lockutils [req-cbaf3ea0-ff4b-4786-9320-9f666d26dae5 req-baa15dc2-6b2c-4a0b-89c1-594406be1d08 service nova] Acquiring lock "refresh_cache-8d0cdc81-56ed-4a7a-8221-30bb906e16b2" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 680.229708] env[61806]: DEBUG oslo_concurrency.lockutils [req-cbaf3ea0-ff4b-4786-9320-9f666d26dae5 req-baa15dc2-6b2c-4a0b-89c1-594406be1d08 service nova] Acquired lock "refresh_cache-8d0cdc81-56ed-4a7a-8221-30bb906e16b2" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 680.229844] env[61806]: DEBUG nova.network.neutron [req-cbaf3ea0-ff4b-4786-9320-9f666d26dae5 req-baa15dc2-6b2c-4a0b-89c1-594406be1d08 service nova] [instance: 8d0cdc81-56ed-4a7a-8221-30bb906e16b2] Refreshing network info cache for port 4cd363a6-ef40-4b4b-909f-85a169e5c84e {{(pid=61806) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 680.236632] env[61806]: DEBUG oslo_vmware.api [None req-6168822a-f60c-43e0-b0ac-2eb07b4977dc tempest-ServersAdmin275Test-1945525728 tempest-ServersAdmin275Test-1945525728-project-member] Task: {'id': task-1294504, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 680.430449] env[61806]: ERROR nova.compute.manager [None req-7bb45eee-0cbe-4995-9898-530114f6f104 tempest-ImagesOneServerNegativeTestJSON-606465252 tempest-ImagesOneServerNegativeTestJSON-606465252-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 4cd363a6-ef40-4b4b-909f-85a169e5c84e, please check neutron logs for more information. [ 680.430449] env[61806]: ERROR nova.compute.manager Traceback (most recent call last): [ 680.430449] env[61806]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 680.430449] env[61806]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 680.430449] env[61806]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 680.430449] env[61806]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 680.430449] env[61806]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 680.430449] env[61806]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 680.430449] env[61806]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 680.430449] env[61806]: ERROR nova.compute.manager self.force_reraise() [ 680.430449] env[61806]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 680.430449] env[61806]: ERROR nova.compute.manager raise self.value [ 680.430449] env[61806]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 680.430449] env[61806]: ERROR nova.compute.manager updated_port = self._update_port( [ 680.430449] env[61806]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 680.430449] env[61806]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 680.431156] env[61806]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 680.431156] env[61806]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 680.431156] env[61806]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 4cd363a6-ef40-4b4b-909f-85a169e5c84e, please check neutron logs for more information. [ 680.431156] env[61806]: ERROR nova.compute.manager [ 680.431156] env[61806]: Traceback (most recent call last): [ 680.431156] env[61806]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 680.431156] env[61806]: listener.cb(fileno) [ 680.431156] env[61806]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 680.431156] env[61806]: result = function(*args, **kwargs) [ 680.431156] env[61806]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 680.431156] env[61806]: return func(*args, **kwargs) [ 680.431156] env[61806]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 680.431156] env[61806]: raise e [ 680.431156] env[61806]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 680.431156] env[61806]: nwinfo = self.network_api.allocate_for_instance( [ 680.431156] env[61806]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 680.431156] env[61806]: created_port_ids = self._update_ports_for_instance( [ 680.431156] env[61806]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 680.431156] env[61806]: with excutils.save_and_reraise_exception(): [ 680.431156] env[61806]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 680.431156] env[61806]: self.force_reraise() [ 680.431156] env[61806]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 680.431156] env[61806]: raise self.value [ 680.431156] env[61806]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 680.431156] env[61806]: updated_port = self._update_port( [ 680.431156] env[61806]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 680.431156] env[61806]: _ensure_no_port_binding_failure(port) [ 680.431156] env[61806]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 680.431156] env[61806]: raise exception.PortBindingFailed(port_id=port['id']) [ 680.431986] env[61806]: nova.exception.PortBindingFailed: Binding failed for port 4cd363a6-ef40-4b4b-909f-85a169e5c84e, please check neutron logs for more information. [ 680.431986] env[61806]: Removing descriptor: 14 [ 680.502147] env[61806]: DEBUG oslo_concurrency.lockutils [None req-44eba412-ca3f-4386-8c36-f42e9734628d tempest-ServersTestBootFromVolume-1815739798 tempest-ServersTestBootFromVolume-1815739798-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.969s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 680.502147] env[61806]: ERROR nova.compute.manager [None req-44eba412-ca3f-4386-8c36-f42e9734628d tempest-ServersTestBootFromVolume-1815739798 tempest-ServersTestBootFromVolume-1815739798-project-member] [instance: 56d26d08-3c6d-45e8-b238-6dbca6024561] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 1570039f-d77a-49a9-a526-42d234745ef6, please check neutron logs for more information. [ 680.502147] env[61806]: ERROR nova.compute.manager [instance: 56d26d08-3c6d-45e8-b238-6dbca6024561] Traceback (most recent call last): [ 680.502147] env[61806]: ERROR nova.compute.manager [instance: 56d26d08-3c6d-45e8-b238-6dbca6024561] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 680.502147] env[61806]: ERROR nova.compute.manager [instance: 56d26d08-3c6d-45e8-b238-6dbca6024561] self.driver.spawn(context, instance, image_meta, [ 680.502147] env[61806]: ERROR nova.compute.manager [instance: 56d26d08-3c6d-45e8-b238-6dbca6024561] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 680.502147] env[61806]: ERROR nova.compute.manager [instance: 56d26d08-3c6d-45e8-b238-6dbca6024561] self._vmops.spawn(context, instance, image_meta, injected_files, [ 680.502147] env[61806]: ERROR nova.compute.manager [instance: 56d26d08-3c6d-45e8-b238-6dbca6024561] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 680.502147] env[61806]: ERROR nova.compute.manager [instance: 56d26d08-3c6d-45e8-b238-6dbca6024561] vm_ref = self.build_virtual_machine(instance, [ 680.502422] env[61806]: ERROR nova.compute.manager [instance: 56d26d08-3c6d-45e8-b238-6dbca6024561] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 680.502422] env[61806]: ERROR nova.compute.manager [instance: 56d26d08-3c6d-45e8-b238-6dbca6024561] vif_infos = vmwarevif.get_vif_info(self._session, [ 680.502422] env[61806]: ERROR nova.compute.manager [instance: 56d26d08-3c6d-45e8-b238-6dbca6024561] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 680.502422] env[61806]: ERROR nova.compute.manager [instance: 56d26d08-3c6d-45e8-b238-6dbca6024561] for vif in network_info: [ 680.502422] env[61806]: ERROR nova.compute.manager [instance: 56d26d08-3c6d-45e8-b238-6dbca6024561] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 680.502422] env[61806]: ERROR nova.compute.manager [instance: 56d26d08-3c6d-45e8-b238-6dbca6024561] return self._sync_wrapper(fn, *args, **kwargs) [ 680.502422] env[61806]: ERROR nova.compute.manager [instance: 56d26d08-3c6d-45e8-b238-6dbca6024561] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 680.502422] env[61806]: ERROR nova.compute.manager [instance: 56d26d08-3c6d-45e8-b238-6dbca6024561] self.wait() [ 680.502422] env[61806]: ERROR nova.compute.manager [instance: 56d26d08-3c6d-45e8-b238-6dbca6024561] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 680.502422] env[61806]: ERROR nova.compute.manager [instance: 56d26d08-3c6d-45e8-b238-6dbca6024561] self[:] = self._gt.wait() [ 680.502422] env[61806]: ERROR nova.compute.manager [instance: 56d26d08-3c6d-45e8-b238-6dbca6024561] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 680.502422] env[61806]: ERROR nova.compute.manager [instance: 56d26d08-3c6d-45e8-b238-6dbca6024561] return self._exit_event.wait() [ 680.502422] env[61806]: ERROR nova.compute.manager [instance: 56d26d08-3c6d-45e8-b238-6dbca6024561] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 680.502856] env[61806]: ERROR nova.compute.manager [instance: 56d26d08-3c6d-45e8-b238-6dbca6024561] current.throw(*self._exc) [ 680.502856] env[61806]: ERROR nova.compute.manager [instance: 56d26d08-3c6d-45e8-b238-6dbca6024561] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 680.502856] env[61806]: ERROR nova.compute.manager [instance: 56d26d08-3c6d-45e8-b238-6dbca6024561] result = function(*args, **kwargs) [ 680.502856] env[61806]: ERROR nova.compute.manager [instance: 56d26d08-3c6d-45e8-b238-6dbca6024561] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 680.502856] env[61806]: ERROR nova.compute.manager [instance: 56d26d08-3c6d-45e8-b238-6dbca6024561] return func(*args, **kwargs) [ 680.502856] env[61806]: ERROR nova.compute.manager [instance: 56d26d08-3c6d-45e8-b238-6dbca6024561] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 680.502856] env[61806]: ERROR nova.compute.manager [instance: 56d26d08-3c6d-45e8-b238-6dbca6024561] raise e [ 680.502856] env[61806]: ERROR nova.compute.manager [instance: 56d26d08-3c6d-45e8-b238-6dbca6024561] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 680.502856] env[61806]: ERROR nova.compute.manager [instance: 56d26d08-3c6d-45e8-b238-6dbca6024561] nwinfo = self.network_api.allocate_for_instance( [ 680.502856] env[61806]: ERROR nova.compute.manager [instance: 56d26d08-3c6d-45e8-b238-6dbca6024561] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 680.502856] env[61806]: ERROR nova.compute.manager [instance: 56d26d08-3c6d-45e8-b238-6dbca6024561] created_port_ids = self._update_ports_for_instance( [ 680.502856] env[61806]: ERROR nova.compute.manager [instance: 56d26d08-3c6d-45e8-b238-6dbca6024561] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 680.502856] env[61806]: ERROR nova.compute.manager [instance: 56d26d08-3c6d-45e8-b238-6dbca6024561] with excutils.save_and_reraise_exception(): [ 680.503269] env[61806]: ERROR nova.compute.manager [instance: 56d26d08-3c6d-45e8-b238-6dbca6024561] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 680.503269] env[61806]: ERROR nova.compute.manager [instance: 56d26d08-3c6d-45e8-b238-6dbca6024561] self.force_reraise() [ 680.503269] env[61806]: ERROR nova.compute.manager [instance: 56d26d08-3c6d-45e8-b238-6dbca6024561] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 680.503269] env[61806]: ERROR nova.compute.manager [instance: 56d26d08-3c6d-45e8-b238-6dbca6024561] raise self.value [ 680.503269] env[61806]: ERROR nova.compute.manager [instance: 56d26d08-3c6d-45e8-b238-6dbca6024561] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 680.503269] env[61806]: ERROR nova.compute.manager [instance: 56d26d08-3c6d-45e8-b238-6dbca6024561] updated_port = self._update_port( [ 680.503269] env[61806]: ERROR nova.compute.manager [instance: 56d26d08-3c6d-45e8-b238-6dbca6024561] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 680.503269] env[61806]: ERROR nova.compute.manager [instance: 56d26d08-3c6d-45e8-b238-6dbca6024561] _ensure_no_port_binding_failure(port) [ 680.503269] env[61806]: ERROR nova.compute.manager [instance: 56d26d08-3c6d-45e8-b238-6dbca6024561] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 680.503269] env[61806]: ERROR nova.compute.manager [instance: 56d26d08-3c6d-45e8-b238-6dbca6024561] raise exception.PortBindingFailed(port_id=port['id']) [ 680.503269] env[61806]: ERROR nova.compute.manager [instance: 56d26d08-3c6d-45e8-b238-6dbca6024561] nova.exception.PortBindingFailed: Binding failed for port 1570039f-d77a-49a9-a526-42d234745ef6, please check neutron logs for more information. [ 680.503269] env[61806]: ERROR nova.compute.manager [instance: 56d26d08-3c6d-45e8-b238-6dbca6024561] [ 680.503621] env[61806]: DEBUG nova.compute.utils [None req-44eba412-ca3f-4386-8c36-f42e9734628d tempest-ServersTestBootFromVolume-1815739798 tempest-ServersTestBootFromVolume-1815739798-project-member] [instance: 56d26d08-3c6d-45e8-b238-6dbca6024561] Binding failed for port 1570039f-d77a-49a9-a526-42d234745ef6, please check neutron logs for more information. {{(pid=61806) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 680.503621] env[61806]: DEBUG nova.compute.manager [None req-44eba412-ca3f-4386-8c36-f42e9734628d tempest-ServersTestBootFromVolume-1815739798 tempest-ServersTestBootFromVolume-1815739798-project-member] [instance: 56d26d08-3c6d-45e8-b238-6dbca6024561] Build of instance 56d26d08-3c6d-45e8-b238-6dbca6024561 was re-scheduled: Binding failed for port 1570039f-d77a-49a9-a526-42d234745ef6, please check neutron logs for more information. {{(pid=61806) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 680.503860] env[61806]: DEBUG nova.compute.manager [None req-44eba412-ca3f-4386-8c36-f42e9734628d tempest-ServersTestBootFromVolume-1815739798 tempest-ServersTestBootFromVolume-1815739798-project-member] [instance: 56d26d08-3c6d-45e8-b238-6dbca6024561] Unplugging VIFs for instance {{(pid=61806) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 680.504104] env[61806]: DEBUG oslo_concurrency.lockutils [None req-44eba412-ca3f-4386-8c36-f42e9734628d tempest-ServersTestBootFromVolume-1815739798 tempest-ServersTestBootFromVolume-1815739798-project-member] Acquiring lock "refresh_cache-56d26d08-3c6d-45e8-b238-6dbca6024561" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 680.504256] env[61806]: DEBUG oslo_concurrency.lockutils [None req-44eba412-ca3f-4386-8c36-f42e9734628d tempest-ServersTestBootFromVolume-1815739798 tempest-ServersTestBootFromVolume-1815739798-project-member] Acquired lock "refresh_cache-56d26d08-3c6d-45e8-b238-6dbca6024561" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 680.504526] env[61806]: DEBUG nova.network.neutron [None req-44eba412-ca3f-4386-8c36-f42e9734628d tempest-ServersTestBootFromVolume-1815739798 tempest-ServersTestBootFromVolume-1815739798-project-member] [instance: 56d26d08-3c6d-45e8-b238-6dbca6024561] Building network info cache for instance {{(pid=61806) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 680.505963] env[61806]: DEBUG oslo_concurrency.lockutils [None req-865353ee-538a-40fc-bfba-502aba0b6a20 tempest-VolumesAdminNegativeTest-1387177911 tempest-VolumesAdminNegativeTest-1387177911-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 29.834s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 680.509147] env[61806]: INFO nova.compute.claims [None req-865353ee-538a-40fc-bfba-502aba0b6a20 tempest-VolumesAdminNegativeTest-1387177911 tempest-VolumesAdminNegativeTest-1387177911-project-member] [instance: b7d55e87-95db-4c45-9417-bbece4fbb98d] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 680.552411] env[61806]: DEBUG nova.compute.manager [None req-7bb45eee-0cbe-4995-9898-530114f6f104 tempest-ImagesOneServerNegativeTestJSON-606465252 tempest-ImagesOneServerNegativeTestJSON-606465252-project-member] [instance: 8d0cdc81-56ed-4a7a-8221-30bb906e16b2] Start spawning the instance on the hypervisor. {{(pid=61806) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 680.576875] env[61806]: DEBUG nova.virt.hardware [None req-7bb45eee-0cbe-4995-9898-530114f6f104 tempest-ImagesOneServerNegativeTestJSON-606465252 tempest-ImagesOneServerNegativeTestJSON-606465252-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-15T18:05:10Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-15T18:04:52Z,direct_url=,disk_format='vmdk',id=ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='84ba83fce2c349a6988173c5d6fc3b07',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-15T18:04:53Z,virtual_size=,visibility=), allow threads: False {{(pid=61806) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 680.577197] env[61806]: DEBUG nova.virt.hardware [None req-7bb45eee-0cbe-4995-9898-530114f6f104 tempest-ImagesOneServerNegativeTestJSON-606465252 tempest-ImagesOneServerNegativeTestJSON-606465252-project-member] Flavor limits 0:0:0 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 680.577362] env[61806]: DEBUG nova.virt.hardware [None req-7bb45eee-0cbe-4995-9898-530114f6f104 tempest-ImagesOneServerNegativeTestJSON-606465252 tempest-ImagesOneServerNegativeTestJSON-606465252-project-member] Image limits 0:0:0 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 680.578103] env[61806]: DEBUG nova.virt.hardware [None req-7bb45eee-0cbe-4995-9898-530114f6f104 tempest-ImagesOneServerNegativeTestJSON-606465252 tempest-ImagesOneServerNegativeTestJSON-606465252-project-member] Flavor pref 0:0:0 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 680.578103] env[61806]: DEBUG nova.virt.hardware [None req-7bb45eee-0cbe-4995-9898-530114f6f104 tempest-ImagesOneServerNegativeTestJSON-606465252 tempest-ImagesOneServerNegativeTestJSON-606465252-project-member] Image pref 0:0:0 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 680.578103] env[61806]: DEBUG nova.virt.hardware [None req-7bb45eee-0cbe-4995-9898-530114f6f104 tempest-ImagesOneServerNegativeTestJSON-606465252 tempest-ImagesOneServerNegativeTestJSON-606465252-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 680.578103] env[61806]: DEBUG nova.virt.hardware [None req-7bb45eee-0cbe-4995-9898-530114f6f104 tempest-ImagesOneServerNegativeTestJSON-606465252 tempest-ImagesOneServerNegativeTestJSON-606465252-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61806) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 680.578319] env[61806]: DEBUG nova.virt.hardware [None req-7bb45eee-0cbe-4995-9898-530114f6f104 tempest-ImagesOneServerNegativeTestJSON-606465252 tempest-ImagesOneServerNegativeTestJSON-606465252-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61806) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 680.578502] env[61806]: DEBUG nova.virt.hardware [None req-7bb45eee-0cbe-4995-9898-530114f6f104 tempest-ImagesOneServerNegativeTestJSON-606465252 tempest-ImagesOneServerNegativeTestJSON-606465252-project-member] Got 1 possible topologies {{(pid=61806) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 680.578711] env[61806]: DEBUG nova.virt.hardware [None req-7bb45eee-0cbe-4995-9898-530114f6f104 tempest-ImagesOneServerNegativeTestJSON-606465252 tempest-ImagesOneServerNegativeTestJSON-606465252-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61806) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 680.578902] env[61806]: DEBUG nova.virt.hardware [None req-7bb45eee-0cbe-4995-9898-530114f6f104 tempest-ImagesOneServerNegativeTestJSON-606465252 tempest-ImagesOneServerNegativeTestJSON-606465252-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61806) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 680.579765] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-da313745-bf9f-4eda-a944-5301b22fc9cc {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 680.588490] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5c2ccfa6-d132-4d7f-8792-ddbd89986d12 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 680.603485] env[61806]: ERROR nova.compute.manager [None req-7bb45eee-0cbe-4995-9898-530114f6f104 tempest-ImagesOneServerNegativeTestJSON-606465252 tempest-ImagesOneServerNegativeTestJSON-606465252-project-member] [instance: 8d0cdc81-56ed-4a7a-8221-30bb906e16b2] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 4cd363a6-ef40-4b4b-909f-85a169e5c84e, please check neutron logs for more information. [ 680.603485] env[61806]: ERROR nova.compute.manager [instance: 8d0cdc81-56ed-4a7a-8221-30bb906e16b2] Traceback (most recent call last): [ 680.603485] env[61806]: ERROR nova.compute.manager [instance: 8d0cdc81-56ed-4a7a-8221-30bb906e16b2] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 680.603485] env[61806]: ERROR nova.compute.manager [instance: 8d0cdc81-56ed-4a7a-8221-30bb906e16b2] yield resources [ 680.603485] env[61806]: ERROR nova.compute.manager [instance: 8d0cdc81-56ed-4a7a-8221-30bb906e16b2] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 680.603485] env[61806]: ERROR nova.compute.manager [instance: 8d0cdc81-56ed-4a7a-8221-30bb906e16b2] self.driver.spawn(context, instance, image_meta, [ 680.603485] env[61806]: ERROR nova.compute.manager [instance: 8d0cdc81-56ed-4a7a-8221-30bb906e16b2] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 680.603485] env[61806]: ERROR nova.compute.manager [instance: 8d0cdc81-56ed-4a7a-8221-30bb906e16b2] self._vmops.spawn(context, instance, image_meta, injected_files, [ 680.603485] env[61806]: ERROR nova.compute.manager [instance: 8d0cdc81-56ed-4a7a-8221-30bb906e16b2] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 680.603485] env[61806]: ERROR nova.compute.manager [instance: 8d0cdc81-56ed-4a7a-8221-30bb906e16b2] vm_ref = self.build_virtual_machine(instance, [ 680.603485] env[61806]: ERROR nova.compute.manager [instance: 8d0cdc81-56ed-4a7a-8221-30bb906e16b2] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 680.603770] env[61806]: ERROR nova.compute.manager [instance: 8d0cdc81-56ed-4a7a-8221-30bb906e16b2] vif_infos = vmwarevif.get_vif_info(self._session, [ 680.603770] env[61806]: ERROR nova.compute.manager [instance: 8d0cdc81-56ed-4a7a-8221-30bb906e16b2] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 680.603770] env[61806]: ERROR nova.compute.manager [instance: 8d0cdc81-56ed-4a7a-8221-30bb906e16b2] for vif in network_info: [ 680.603770] env[61806]: ERROR nova.compute.manager [instance: 8d0cdc81-56ed-4a7a-8221-30bb906e16b2] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 680.603770] env[61806]: ERROR nova.compute.manager [instance: 8d0cdc81-56ed-4a7a-8221-30bb906e16b2] return self._sync_wrapper(fn, *args, **kwargs) [ 680.603770] env[61806]: ERROR nova.compute.manager [instance: 8d0cdc81-56ed-4a7a-8221-30bb906e16b2] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 680.603770] env[61806]: ERROR nova.compute.manager [instance: 8d0cdc81-56ed-4a7a-8221-30bb906e16b2] self.wait() [ 680.603770] env[61806]: ERROR nova.compute.manager [instance: 8d0cdc81-56ed-4a7a-8221-30bb906e16b2] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 680.603770] env[61806]: ERROR nova.compute.manager [instance: 8d0cdc81-56ed-4a7a-8221-30bb906e16b2] self[:] = self._gt.wait() [ 680.603770] env[61806]: ERROR nova.compute.manager [instance: 8d0cdc81-56ed-4a7a-8221-30bb906e16b2] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 680.603770] env[61806]: ERROR nova.compute.manager [instance: 8d0cdc81-56ed-4a7a-8221-30bb906e16b2] return self._exit_event.wait() [ 680.603770] env[61806]: ERROR nova.compute.manager [instance: 8d0cdc81-56ed-4a7a-8221-30bb906e16b2] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 680.603770] env[61806]: ERROR nova.compute.manager [instance: 8d0cdc81-56ed-4a7a-8221-30bb906e16b2] current.throw(*self._exc) [ 680.604212] env[61806]: ERROR nova.compute.manager [instance: 8d0cdc81-56ed-4a7a-8221-30bb906e16b2] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 680.604212] env[61806]: ERROR nova.compute.manager [instance: 8d0cdc81-56ed-4a7a-8221-30bb906e16b2] result = function(*args, **kwargs) [ 680.604212] env[61806]: ERROR nova.compute.manager [instance: 8d0cdc81-56ed-4a7a-8221-30bb906e16b2] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 680.604212] env[61806]: ERROR nova.compute.manager [instance: 8d0cdc81-56ed-4a7a-8221-30bb906e16b2] return func(*args, **kwargs) [ 680.604212] env[61806]: ERROR nova.compute.manager [instance: 8d0cdc81-56ed-4a7a-8221-30bb906e16b2] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 680.604212] env[61806]: ERROR nova.compute.manager [instance: 8d0cdc81-56ed-4a7a-8221-30bb906e16b2] raise e [ 680.604212] env[61806]: ERROR nova.compute.manager [instance: 8d0cdc81-56ed-4a7a-8221-30bb906e16b2] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 680.604212] env[61806]: ERROR nova.compute.manager [instance: 8d0cdc81-56ed-4a7a-8221-30bb906e16b2] nwinfo = self.network_api.allocate_for_instance( [ 680.604212] env[61806]: ERROR nova.compute.manager [instance: 8d0cdc81-56ed-4a7a-8221-30bb906e16b2] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 680.604212] env[61806]: ERROR nova.compute.manager [instance: 8d0cdc81-56ed-4a7a-8221-30bb906e16b2] created_port_ids = self._update_ports_for_instance( [ 680.604212] env[61806]: ERROR nova.compute.manager [instance: 8d0cdc81-56ed-4a7a-8221-30bb906e16b2] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 680.604212] env[61806]: ERROR nova.compute.manager [instance: 8d0cdc81-56ed-4a7a-8221-30bb906e16b2] with excutils.save_and_reraise_exception(): [ 680.604212] env[61806]: ERROR nova.compute.manager [instance: 8d0cdc81-56ed-4a7a-8221-30bb906e16b2] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 680.604654] env[61806]: ERROR nova.compute.manager [instance: 8d0cdc81-56ed-4a7a-8221-30bb906e16b2] self.force_reraise() [ 680.604654] env[61806]: ERROR nova.compute.manager [instance: 8d0cdc81-56ed-4a7a-8221-30bb906e16b2] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 680.604654] env[61806]: ERROR nova.compute.manager [instance: 8d0cdc81-56ed-4a7a-8221-30bb906e16b2] raise self.value [ 680.604654] env[61806]: ERROR nova.compute.manager [instance: 8d0cdc81-56ed-4a7a-8221-30bb906e16b2] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 680.604654] env[61806]: ERROR nova.compute.manager [instance: 8d0cdc81-56ed-4a7a-8221-30bb906e16b2] updated_port = self._update_port( [ 680.604654] env[61806]: ERROR nova.compute.manager [instance: 8d0cdc81-56ed-4a7a-8221-30bb906e16b2] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 680.604654] env[61806]: ERROR nova.compute.manager [instance: 8d0cdc81-56ed-4a7a-8221-30bb906e16b2] _ensure_no_port_binding_failure(port) [ 680.604654] env[61806]: ERROR nova.compute.manager [instance: 8d0cdc81-56ed-4a7a-8221-30bb906e16b2] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 680.604654] env[61806]: ERROR nova.compute.manager [instance: 8d0cdc81-56ed-4a7a-8221-30bb906e16b2] raise exception.PortBindingFailed(port_id=port['id']) [ 680.604654] env[61806]: ERROR nova.compute.manager [instance: 8d0cdc81-56ed-4a7a-8221-30bb906e16b2] nova.exception.PortBindingFailed: Binding failed for port 4cd363a6-ef40-4b4b-909f-85a169e5c84e, please check neutron logs for more information. [ 680.604654] env[61806]: ERROR nova.compute.manager [instance: 8d0cdc81-56ed-4a7a-8221-30bb906e16b2] [ 680.604654] env[61806]: INFO nova.compute.manager [None req-7bb45eee-0cbe-4995-9898-530114f6f104 tempest-ImagesOneServerNegativeTestJSON-606465252 tempest-ImagesOneServerNegativeTestJSON-606465252-project-member] [instance: 8d0cdc81-56ed-4a7a-8221-30bb906e16b2] Terminating instance [ 680.606205] env[61806]: DEBUG oslo_concurrency.lockutils [None req-7bb45eee-0cbe-4995-9898-530114f6f104 tempest-ImagesOneServerNegativeTestJSON-606465252 tempest-ImagesOneServerNegativeTestJSON-606465252-project-member] Acquiring lock "refresh_cache-8d0cdc81-56ed-4a7a-8221-30bb906e16b2" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 680.707573] env[61806]: DEBUG oslo_concurrency.lockutils [None req-fcbc4cf1-fc11-42b1-9d49-ef556b52782e tempest-ServerRescueNegativeTestJSON-1483258717 tempest-ServerRescueNegativeTestJSON-1483258717-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 680.736117] env[61806]: DEBUG oslo_vmware.api [None req-6168822a-f60c-43e0-b0ac-2eb07b4977dc tempest-ServersAdmin275Test-1945525728 tempest-ServersAdmin275Test-1945525728-project-member] Task: {'id': task-1294504, 'name': PowerOffVM_Task, 'duration_secs': 0.129314} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 680.736398] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [None req-6168822a-f60c-43e0-b0ac-2eb07b4977dc tempest-ServersAdmin275Test-1945525728 tempest-ServersAdmin275Test-1945525728-project-member] [instance: 9db80c3c-e2c3-4034-a79a-d7be924c4ff6] Powered off the VM {{(pid=61806) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 680.736571] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-6168822a-f60c-43e0-b0ac-2eb07b4977dc tempest-ServersAdmin275Test-1945525728 tempest-ServersAdmin275Test-1945525728-project-member] [instance: 9db80c3c-e2c3-4034-a79a-d7be924c4ff6] Unregistering the VM {{(pid=61806) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 680.737217] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-b548acc9-bb22-476a-8536-48b06855bcee {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 680.750729] env[61806]: DEBUG nova.network.neutron [req-cbaf3ea0-ff4b-4786-9320-9f666d26dae5 req-baa15dc2-6b2c-4a0b-89c1-594406be1d08 service nova] [instance: 8d0cdc81-56ed-4a7a-8221-30bb906e16b2] Instance cache missing network info. {{(pid=61806) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 680.761541] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-6168822a-f60c-43e0-b0ac-2eb07b4977dc tempest-ServersAdmin275Test-1945525728 tempest-ServersAdmin275Test-1945525728-project-member] [instance: 9db80c3c-e2c3-4034-a79a-d7be924c4ff6] Unregistered the VM {{(pid=61806) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 680.761766] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-6168822a-f60c-43e0-b0ac-2eb07b4977dc tempest-ServersAdmin275Test-1945525728 tempest-ServersAdmin275Test-1945525728-project-member] [instance: 9db80c3c-e2c3-4034-a79a-d7be924c4ff6] Deleting contents of the VM from datastore datastore1 {{(pid=61806) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 680.762032] env[61806]: DEBUG nova.virt.vmwareapi.ds_util [None req-6168822a-f60c-43e0-b0ac-2eb07b4977dc tempest-ServersAdmin275Test-1945525728 tempest-ServersAdmin275Test-1945525728-project-member] Deleting the datastore file [datastore1] 9db80c3c-e2c3-4034-a79a-d7be924c4ff6 {{(pid=61806) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 680.762297] env[61806]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-a8a5a93f-1573-41ee-84ee-0b8691c3c08e {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 680.769763] env[61806]: DEBUG oslo_vmware.api [None req-6168822a-f60c-43e0-b0ac-2eb07b4977dc tempest-ServersAdmin275Test-1945525728 tempest-ServersAdmin275Test-1945525728-project-member] Waiting for the task: (returnval){ [ 680.769763] env[61806]: value = "task-1294506" [ 680.769763] env[61806]: _type = "Task" [ 680.769763] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 680.778502] env[61806]: DEBUG oslo_vmware.api [None req-6168822a-f60c-43e0-b0ac-2eb07b4977dc tempest-ServersAdmin275Test-1945525728 tempest-ServersAdmin275Test-1945525728-project-member] Task: {'id': task-1294506, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 680.875864] env[61806]: DEBUG nova.network.neutron [req-cbaf3ea0-ff4b-4786-9320-9f666d26dae5 req-baa15dc2-6b2c-4a0b-89c1-594406be1d08 service nova] [instance: 8d0cdc81-56ed-4a7a-8221-30bb906e16b2] Updating instance_info_cache with network_info: [] {{(pid=61806) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 681.027244] env[61806]: DEBUG nova.network.neutron [None req-44eba412-ca3f-4386-8c36-f42e9734628d tempest-ServersTestBootFromVolume-1815739798 tempest-ServersTestBootFromVolume-1815739798-project-member] [instance: 56d26d08-3c6d-45e8-b238-6dbca6024561] Instance cache missing network info. {{(pid=61806) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 681.112186] env[61806]: DEBUG nova.network.neutron [None req-44eba412-ca3f-4386-8c36-f42e9734628d tempest-ServersTestBootFromVolume-1815739798 tempest-ServersTestBootFromVolume-1815739798-project-member] [instance: 56d26d08-3c6d-45e8-b238-6dbca6024561] Updating instance_info_cache with network_info: [] {{(pid=61806) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 681.279874] env[61806]: DEBUG oslo_vmware.api [None req-6168822a-f60c-43e0-b0ac-2eb07b4977dc tempest-ServersAdmin275Test-1945525728 tempest-ServersAdmin275Test-1945525728-project-member] Task: {'id': task-1294506, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.115571} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 681.280138] env[61806]: DEBUG nova.virt.vmwareapi.ds_util [None req-6168822a-f60c-43e0-b0ac-2eb07b4977dc tempest-ServersAdmin275Test-1945525728 tempest-ServersAdmin275Test-1945525728-project-member] Deleted the datastore file {{(pid=61806) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 681.280325] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-6168822a-f60c-43e0-b0ac-2eb07b4977dc tempest-ServersAdmin275Test-1945525728 tempest-ServersAdmin275Test-1945525728-project-member] [instance: 9db80c3c-e2c3-4034-a79a-d7be924c4ff6] Deleted contents of the VM from datastore datastore1 {{(pid=61806) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 681.280500] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-6168822a-f60c-43e0-b0ac-2eb07b4977dc tempest-ServersAdmin275Test-1945525728 tempest-ServersAdmin275Test-1945525728-project-member] [instance: 9db80c3c-e2c3-4034-a79a-d7be924c4ff6] Instance destroyed {{(pid=61806) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 681.280667] env[61806]: INFO nova.compute.manager [None req-6168822a-f60c-43e0-b0ac-2eb07b4977dc tempest-ServersAdmin275Test-1945525728 tempest-ServersAdmin275Test-1945525728-project-member] [instance: 9db80c3c-e2c3-4034-a79a-d7be924c4ff6] Took 1.08 seconds to destroy the instance on the hypervisor. [ 681.280959] env[61806]: DEBUG oslo.service.loopingcall [None req-6168822a-f60c-43e0-b0ac-2eb07b4977dc tempest-ServersAdmin275Test-1945525728 tempest-ServersAdmin275Test-1945525728-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61806) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 681.281161] env[61806]: DEBUG nova.compute.manager [-] [instance: 9db80c3c-e2c3-4034-a79a-d7be924c4ff6] Deallocating network for instance {{(pid=61806) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 681.281259] env[61806]: DEBUG nova.network.neutron [-] [instance: 9db80c3c-e2c3-4034-a79a-d7be924c4ff6] deallocate_for_instance() {{(pid=61806) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 681.380040] env[61806]: DEBUG oslo_concurrency.lockutils [req-cbaf3ea0-ff4b-4786-9320-9f666d26dae5 req-baa15dc2-6b2c-4a0b-89c1-594406be1d08 service nova] Releasing lock "refresh_cache-8d0cdc81-56ed-4a7a-8221-30bb906e16b2" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 681.381783] env[61806]: DEBUG oslo_concurrency.lockutils [None req-7bb45eee-0cbe-4995-9898-530114f6f104 tempest-ImagesOneServerNegativeTestJSON-606465252 tempest-ImagesOneServerNegativeTestJSON-606465252-project-member] Acquired lock "refresh_cache-8d0cdc81-56ed-4a7a-8221-30bb906e16b2" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 681.383014] env[61806]: DEBUG nova.network.neutron [None req-7bb45eee-0cbe-4995-9898-530114f6f104 tempest-ImagesOneServerNegativeTestJSON-606465252 tempest-ImagesOneServerNegativeTestJSON-606465252-project-member] [instance: 8d0cdc81-56ed-4a7a-8221-30bb906e16b2] Building network info cache for instance {{(pid=61806) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 681.451930] env[61806]: DEBUG nova.network.neutron [-] [instance: 9db80c3c-e2c3-4034-a79a-d7be924c4ff6] Instance cache missing network info. {{(pid=61806) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 681.615368] env[61806]: DEBUG oslo_concurrency.lockutils [None req-44eba412-ca3f-4386-8c36-f42e9734628d tempest-ServersTestBootFromVolume-1815739798 tempest-ServersTestBootFromVolume-1815739798-project-member] Releasing lock "refresh_cache-56d26d08-3c6d-45e8-b238-6dbca6024561" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 681.615602] env[61806]: DEBUG nova.compute.manager [None req-44eba412-ca3f-4386-8c36-f42e9734628d tempest-ServersTestBootFromVolume-1815739798 tempest-ServersTestBootFromVolume-1815739798-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61806) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 681.615786] env[61806]: DEBUG nova.compute.manager [None req-44eba412-ca3f-4386-8c36-f42e9734628d tempest-ServersTestBootFromVolume-1815739798 tempest-ServersTestBootFromVolume-1815739798-project-member] [instance: 56d26d08-3c6d-45e8-b238-6dbca6024561] Deallocating network for instance {{(pid=61806) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 681.615957] env[61806]: DEBUG nova.network.neutron [None req-44eba412-ca3f-4386-8c36-f42e9734628d tempest-ServersTestBootFromVolume-1815739798 tempest-ServersTestBootFromVolume-1815739798-project-member] [instance: 56d26d08-3c6d-45e8-b238-6dbca6024561] deallocate_for_instance() {{(pid=61806) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 681.632109] env[61806]: DEBUG nova.network.neutron [None req-44eba412-ca3f-4386-8c36-f42e9734628d tempest-ServersTestBootFromVolume-1815739798 tempest-ServersTestBootFromVolume-1815739798-project-member] [instance: 56d26d08-3c6d-45e8-b238-6dbca6024561] Instance cache missing network info. {{(pid=61806) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 681.854320] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dd12e1fa-1c4c-4a49-987a-26685efedf34 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 681.869662] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c7326c70-8d3b-4c9e-9e60-014c242c5a24 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 681.904119] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f3ae5b20-5df8-405b-80dd-9c563a8b55e7 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 681.911723] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4f8c83b8-905c-4d34-a954-56c0f413b020 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 681.924968] env[61806]: DEBUG nova.compute.provider_tree [None req-865353ee-538a-40fc-bfba-502aba0b6a20 tempest-VolumesAdminNegativeTest-1387177911 tempest-VolumesAdminNegativeTest-1387177911-project-member] Inventory has not changed in ProviderTree for provider: a2858be1-fd22-4e08-979e-87ad25293407 {{(pid=61806) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 681.926771] env[61806]: DEBUG nova.network.neutron [None req-7bb45eee-0cbe-4995-9898-530114f6f104 tempest-ImagesOneServerNegativeTestJSON-606465252 tempest-ImagesOneServerNegativeTestJSON-606465252-project-member] [instance: 8d0cdc81-56ed-4a7a-8221-30bb906e16b2] Instance cache missing network info. {{(pid=61806) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 681.953751] env[61806]: DEBUG nova.network.neutron [-] [instance: 9db80c3c-e2c3-4034-a79a-d7be924c4ff6] Updating instance_info_cache with network_info: [] {{(pid=61806) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 682.026983] env[61806]: DEBUG nova.network.neutron [None req-7bb45eee-0cbe-4995-9898-530114f6f104 tempest-ImagesOneServerNegativeTestJSON-606465252 tempest-ImagesOneServerNegativeTestJSON-606465252-project-member] [instance: 8d0cdc81-56ed-4a7a-8221-30bb906e16b2] Updating instance_info_cache with network_info: [] {{(pid=61806) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 682.135325] env[61806]: DEBUG nova.network.neutron [None req-44eba412-ca3f-4386-8c36-f42e9734628d tempest-ServersTestBootFromVolume-1815739798 tempest-ServersTestBootFromVolume-1815739798-project-member] [instance: 56d26d08-3c6d-45e8-b238-6dbca6024561] Updating instance_info_cache with network_info: [] {{(pid=61806) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 682.254058] env[61806]: DEBUG nova.compute.manager [req-5ff8520f-49b6-4fdb-870f-7bba02a71da8 req-04ae0dcd-7eba-4df1-b8fa-a0b21d8af69d service nova] [instance: 8d0cdc81-56ed-4a7a-8221-30bb906e16b2] Received event network-vif-deleted-4cd363a6-ef40-4b4b-909f-85a169e5c84e {{(pid=61806) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 682.430417] env[61806]: DEBUG nova.scheduler.client.report [None req-865353ee-538a-40fc-bfba-502aba0b6a20 tempest-VolumesAdminNegativeTest-1387177911 tempest-VolumesAdminNegativeTest-1387177911-project-member] Inventory has not changed for provider a2858be1-fd22-4e08-979e-87ad25293407 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 138, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61806) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 682.456434] env[61806]: INFO nova.compute.manager [-] [instance: 9db80c3c-e2c3-4034-a79a-d7be924c4ff6] Took 1.18 seconds to deallocate network for instance. [ 682.470631] env[61806]: DEBUG oslo_concurrency.lockutils [None req-f7f6aee3-3d33-4e4b-b451-fa3af2e320d4 tempest-ServerDiagnosticsV248Test-463927155 tempest-ServerDiagnosticsV248Test-463927155-project-member] Acquiring lock "a934d02d-26aa-4900-b473-a58489e5629e" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 682.470978] env[61806]: DEBUG oslo_concurrency.lockutils [None req-f7f6aee3-3d33-4e4b-b451-fa3af2e320d4 tempest-ServerDiagnosticsV248Test-463927155 tempest-ServerDiagnosticsV248Test-463927155-project-member] Lock "a934d02d-26aa-4900-b473-a58489e5629e" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 682.529819] env[61806]: DEBUG oslo_concurrency.lockutils [None req-7bb45eee-0cbe-4995-9898-530114f6f104 tempest-ImagesOneServerNegativeTestJSON-606465252 tempest-ImagesOneServerNegativeTestJSON-606465252-project-member] Releasing lock "refresh_cache-8d0cdc81-56ed-4a7a-8221-30bb906e16b2" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 682.530274] env[61806]: DEBUG nova.compute.manager [None req-7bb45eee-0cbe-4995-9898-530114f6f104 tempest-ImagesOneServerNegativeTestJSON-606465252 tempest-ImagesOneServerNegativeTestJSON-606465252-project-member] [instance: 8d0cdc81-56ed-4a7a-8221-30bb906e16b2] Start destroying the instance on the hypervisor. {{(pid=61806) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 682.530470] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-7bb45eee-0cbe-4995-9898-530114f6f104 tempest-ImagesOneServerNegativeTestJSON-606465252 tempest-ImagesOneServerNegativeTestJSON-606465252-project-member] [instance: 8d0cdc81-56ed-4a7a-8221-30bb906e16b2] Destroying instance {{(pid=61806) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 682.531820] env[61806]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-ca588ef2-94d0-4fc6-a675-15628652dea6 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 682.540120] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2d0ecc91-8442-4375-8ed9-d5597dca8ea4 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 682.562040] env[61806]: WARNING nova.virt.vmwareapi.vmops [None req-7bb45eee-0cbe-4995-9898-530114f6f104 tempest-ImagesOneServerNegativeTestJSON-606465252 tempest-ImagesOneServerNegativeTestJSON-606465252-project-member] [instance: 8d0cdc81-56ed-4a7a-8221-30bb906e16b2] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 8d0cdc81-56ed-4a7a-8221-30bb906e16b2 could not be found. [ 682.562263] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-7bb45eee-0cbe-4995-9898-530114f6f104 tempest-ImagesOneServerNegativeTestJSON-606465252 tempest-ImagesOneServerNegativeTestJSON-606465252-project-member] [instance: 8d0cdc81-56ed-4a7a-8221-30bb906e16b2] Instance destroyed {{(pid=61806) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 682.562444] env[61806]: INFO nova.compute.manager [None req-7bb45eee-0cbe-4995-9898-530114f6f104 tempest-ImagesOneServerNegativeTestJSON-606465252 tempest-ImagesOneServerNegativeTestJSON-606465252-project-member] [instance: 8d0cdc81-56ed-4a7a-8221-30bb906e16b2] Took 0.03 seconds to destroy the instance on the hypervisor. [ 682.562681] env[61806]: DEBUG oslo.service.loopingcall [None req-7bb45eee-0cbe-4995-9898-530114f6f104 tempest-ImagesOneServerNegativeTestJSON-606465252 tempest-ImagesOneServerNegativeTestJSON-606465252-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61806) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 682.562883] env[61806]: DEBUG nova.compute.manager [-] [instance: 8d0cdc81-56ed-4a7a-8221-30bb906e16b2] Deallocating network for instance {{(pid=61806) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 682.562975] env[61806]: DEBUG nova.network.neutron [-] [instance: 8d0cdc81-56ed-4a7a-8221-30bb906e16b2] deallocate_for_instance() {{(pid=61806) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 682.578131] env[61806]: DEBUG nova.network.neutron [-] [instance: 8d0cdc81-56ed-4a7a-8221-30bb906e16b2] Instance cache missing network info. {{(pid=61806) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 682.639840] env[61806]: INFO nova.compute.manager [None req-44eba412-ca3f-4386-8c36-f42e9734628d tempest-ServersTestBootFromVolume-1815739798 tempest-ServersTestBootFromVolume-1815739798-project-member] [instance: 56d26d08-3c6d-45e8-b238-6dbca6024561] Took 1.02 seconds to deallocate network for instance. [ 682.935222] env[61806]: DEBUG oslo_concurrency.lockutils [None req-865353ee-538a-40fc-bfba-502aba0b6a20 tempest-VolumesAdminNegativeTest-1387177911 tempest-VolumesAdminNegativeTest-1387177911-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.429s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 682.937850] env[61806]: DEBUG nova.compute.manager [None req-865353ee-538a-40fc-bfba-502aba0b6a20 tempest-VolumesAdminNegativeTest-1387177911 tempest-VolumesAdminNegativeTest-1387177911-project-member] [instance: b7d55e87-95db-4c45-9417-bbece4fbb98d] Start building networks asynchronously for instance. {{(pid=61806) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 682.938715] env[61806]: DEBUG oslo_concurrency.lockutils [None req-81a55da1-45c4-45f8-af56-f8f7d424c8aa tempest-ServersTestJSON-430716070 tempest-ServersTestJSON-430716070-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 31.482s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 682.940209] env[61806]: INFO nova.compute.claims [None req-81a55da1-45c4-45f8-af56-f8f7d424c8aa tempest-ServersTestJSON-430716070 tempest-ServersTestJSON-430716070-project-member] [instance: eca2e3da-b38c-482e-b9b6-4ad75d83caa1] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 682.962159] env[61806]: DEBUG oslo_concurrency.lockutils [None req-6168822a-f60c-43e0-b0ac-2eb07b4977dc tempest-ServersAdmin275Test-1945525728 tempest-ServersAdmin275Test-1945525728-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 683.080904] env[61806]: DEBUG nova.network.neutron [-] [instance: 8d0cdc81-56ed-4a7a-8221-30bb906e16b2] Updating instance_info_cache with network_info: [] {{(pid=61806) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 683.445026] env[61806]: DEBUG nova.compute.utils [None req-865353ee-538a-40fc-bfba-502aba0b6a20 tempest-VolumesAdminNegativeTest-1387177911 tempest-VolumesAdminNegativeTest-1387177911-project-member] Using /dev/sd instead of None {{(pid=61806) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 683.448022] env[61806]: DEBUG nova.compute.manager [None req-865353ee-538a-40fc-bfba-502aba0b6a20 tempest-VolumesAdminNegativeTest-1387177911 tempest-VolumesAdminNegativeTest-1387177911-project-member] [instance: b7d55e87-95db-4c45-9417-bbece4fbb98d] Allocating IP information in the background. {{(pid=61806) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 683.448022] env[61806]: DEBUG nova.network.neutron [None req-865353ee-538a-40fc-bfba-502aba0b6a20 tempest-VolumesAdminNegativeTest-1387177911 tempest-VolumesAdminNegativeTest-1387177911-project-member] [instance: b7d55e87-95db-4c45-9417-bbece4fbb98d] allocate_for_instance() {{(pid=61806) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 683.496580] env[61806]: DEBUG nova.policy [None req-865353ee-538a-40fc-bfba-502aba0b6a20 tempest-VolumesAdminNegativeTest-1387177911 tempest-VolumesAdminNegativeTest-1387177911-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '5b2be9100e8f4a1387600ab9b76961c6', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '44b0a78c155d47d98223d2543ac7673b', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61806) authorize /opt/stack/nova/nova/policy.py:201}} [ 683.583212] env[61806]: INFO nova.compute.manager [-] [instance: 8d0cdc81-56ed-4a7a-8221-30bb906e16b2] Took 1.02 seconds to deallocate network for instance. [ 683.585287] env[61806]: DEBUG nova.compute.claims [None req-7bb45eee-0cbe-4995-9898-530114f6f104 tempest-ImagesOneServerNegativeTestJSON-606465252 tempest-ImagesOneServerNegativeTestJSON-606465252-project-member] [instance: 8d0cdc81-56ed-4a7a-8221-30bb906e16b2] Aborting claim: {{(pid=61806) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 683.585467] env[61806]: DEBUG oslo_concurrency.lockutils [None req-7bb45eee-0cbe-4995-9898-530114f6f104 tempest-ImagesOneServerNegativeTestJSON-606465252 tempest-ImagesOneServerNegativeTestJSON-606465252-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 683.670683] env[61806]: INFO nova.scheduler.client.report [None req-44eba412-ca3f-4386-8c36-f42e9734628d tempest-ServersTestBootFromVolume-1815739798 tempest-ServersTestBootFromVolume-1815739798-project-member] Deleted allocations for instance 56d26d08-3c6d-45e8-b238-6dbca6024561 [ 683.803329] env[61806]: DEBUG nova.network.neutron [None req-865353ee-538a-40fc-bfba-502aba0b6a20 tempest-VolumesAdminNegativeTest-1387177911 tempest-VolumesAdminNegativeTest-1387177911-project-member] [instance: b7d55e87-95db-4c45-9417-bbece4fbb98d] Successfully created port: 03736e5c-7351-4a9e-ab04-f54729f3099f {{(pid=61806) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 683.954878] env[61806]: DEBUG nova.compute.manager [None req-865353ee-538a-40fc-bfba-502aba0b6a20 tempest-VolumesAdminNegativeTest-1387177911 tempest-VolumesAdminNegativeTest-1387177911-project-member] [instance: b7d55e87-95db-4c45-9417-bbece4fbb98d] Start building block device mappings for instance. {{(pid=61806) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 684.181420] env[61806]: DEBUG oslo_concurrency.lockutils [None req-44eba412-ca3f-4386-8c36-f42e9734628d tempest-ServersTestBootFromVolume-1815739798 tempest-ServersTestBootFromVolume-1815739798-project-member] Lock "56d26d08-3c6d-45e8-b238-6dbca6024561" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 122.763s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 684.376760] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3862f0b8-9480-435a-a971-312d9cc84e2e {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 684.385137] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-05191d35-b8f8-441f-b784-9165ab8f2bd3 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 684.417808] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b44512b9-6dc2-4d1d-89ed-d4eba9988d17 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 684.426171] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5acd7c68-edf6-47de-8eaa-7866738b94e4 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 684.441239] env[61806]: DEBUG nova.compute.provider_tree [None req-81a55da1-45c4-45f8-af56-f8f7d424c8aa tempest-ServersTestJSON-430716070 tempest-ServersTestJSON-430716070-project-member] Inventory has not changed in ProviderTree for provider: a2858be1-fd22-4e08-979e-87ad25293407 {{(pid=61806) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 684.660757] env[61806]: DEBUG nova.compute.manager [req-b21a30cb-e457-4f7f-8248-c5841198cc64 req-24ed0238-f790-4ae7-aa56-7ef941665829 service nova] [instance: b7d55e87-95db-4c45-9417-bbece4fbb98d] Received event network-changed-03736e5c-7351-4a9e-ab04-f54729f3099f {{(pid=61806) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 684.661105] env[61806]: DEBUG nova.compute.manager [req-b21a30cb-e457-4f7f-8248-c5841198cc64 req-24ed0238-f790-4ae7-aa56-7ef941665829 service nova] [instance: b7d55e87-95db-4c45-9417-bbece4fbb98d] Refreshing instance network info cache due to event network-changed-03736e5c-7351-4a9e-ab04-f54729f3099f. {{(pid=61806) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 684.661178] env[61806]: DEBUG oslo_concurrency.lockutils [req-b21a30cb-e457-4f7f-8248-c5841198cc64 req-24ed0238-f790-4ae7-aa56-7ef941665829 service nova] Acquiring lock "refresh_cache-b7d55e87-95db-4c45-9417-bbece4fbb98d" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 684.661280] env[61806]: DEBUG oslo_concurrency.lockutils [req-b21a30cb-e457-4f7f-8248-c5841198cc64 req-24ed0238-f790-4ae7-aa56-7ef941665829 service nova] Acquired lock "refresh_cache-b7d55e87-95db-4c45-9417-bbece4fbb98d" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 684.661436] env[61806]: DEBUG nova.network.neutron [req-b21a30cb-e457-4f7f-8248-c5841198cc64 req-24ed0238-f790-4ae7-aa56-7ef941665829 service nova] [instance: b7d55e87-95db-4c45-9417-bbece4fbb98d] Refreshing network info cache for port 03736e5c-7351-4a9e-ab04-f54729f3099f {{(pid=61806) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 684.684419] env[61806]: DEBUG nova.compute.manager [None req-6b55cbbf-00f7-4d4d-87f3-a59cc1c553b0 tempest-ServerRescueNegativeTestJSON-1483258717 tempest-ServerRescueNegativeTestJSON-1483258717-project-member] [instance: 7ea919b6-12f7-4f4e-a123-413044bbbffc] Starting instance... {{(pid=61806) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 684.944459] env[61806]: DEBUG nova.scheduler.client.report [None req-81a55da1-45c4-45f8-af56-f8f7d424c8aa tempest-ServersTestJSON-430716070 tempest-ServersTestJSON-430716070-project-member] Inventory has not changed for provider a2858be1-fd22-4e08-979e-87ad25293407 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 138, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61806) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 684.964627] env[61806]: DEBUG nova.compute.manager [None req-865353ee-538a-40fc-bfba-502aba0b6a20 tempest-VolumesAdminNegativeTest-1387177911 tempest-VolumesAdminNegativeTest-1387177911-project-member] [instance: b7d55e87-95db-4c45-9417-bbece4fbb98d] Start spawning the instance on the hypervisor. {{(pid=61806) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 684.974840] env[61806]: ERROR nova.compute.manager [None req-865353ee-538a-40fc-bfba-502aba0b6a20 tempest-VolumesAdminNegativeTest-1387177911 tempest-VolumesAdminNegativeTest-1387177911-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 03736e5c-7351-4a9e-ab04-f54729f3099f, please check neutron logs for more information. [ 684.974840] env[61806]: ERROR nova.compute.manager Traceback (most recent call last): [ 684.974840] env[61806]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 684.974840] env[61806]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 684.974840] env[61806]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 684.974840] env[61806]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 684.974840] env[61806]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 684.974840] env[61806]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 684.974840] env[61806]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 684.974840] env[61806]: ERROR nova.compute.manager self.force_reraise() [ 684.974840] env[61806]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 684.974840] env[61806]: ERROR nova.compute.manager raise self.value [ 684.974840] env[61806]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 684.974840] env[61806]: ERROR nova.compute.manager updated_port = self._update_port( [ 684.974840] env[61806]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 684.974840] env[61806]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 684.975400] env[61806]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 684.975400] env[61806]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 684.975400] env[61806]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 03736e5c-7351-4a9e-ab04-f54729f3099f, please check neutron logs for more information. [ 684.975400] env[61806]: ERROR nova.compute.manager [ 684.975400] env[61806]: Traceback (most recent call last): [ 684.975400] env[61806]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 684.975400] env[61806]: listener.cb(fileno) [ 684.975400] env[61806]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 684.975400] env[61806]: result = function(*args, **kwargs) [ 684.975400] env[61806]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 684.975400] env[61806]: return func(*args, **kwargs) [ 684.975400] env[61806]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 684.975400] env[61806]: raise e [ 684.975400] env[61806]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 684.975400] env[61806]: nwinfo = self.network_api.allocate_for_instance( [ 684.975400] env[61806]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 684.975400] env[61806]: created_port_ids = self._update_ports_for_instance( [ 684.975400] env[61806]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 684.975400] env[61806]: with excutils.save_and_reraise_exception(): [ 684.975400] env[61806]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 684.975400] env[61806]: self.force_reraise() [ 684.975400] env[61806]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 684.975400] env[61806]: raise self.value [ 684.975400] env[61806]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 684.975400] env[61806]: updated_port = self._update_port( [ 684.975400] env[61806]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 684.975400] env[61806]: _ensure_no_port_binding_failure(port) [ 684.975400] env[61806]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 684.975400] env[61806]: raise exception.PortBindingFailed(port_id=port['id']) [ 684.976326] env[61806]: nova.exception.PortBindingFailed: Binding failed for port 03736e5c-7351-4a9e-ab04-f54729f3099f, please check neutron logs for more information. [ 684.976326] env[61806]: Removing descriptor: 14 [ 684.992546] env[61806]: DEBUG nova.virt.hardware [None req-865353ee-538a-40fc-bfba-502aba0b6a20 tempest-VolumesAdminNegativeTest-1387177911 tempest-VolumesAdminNegativeTest-1387177911-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-15T18:05:10Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-15T18:04:52Z,direct_url=,disk_format='vmdk',id=ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='84ba83fce2c349a6988173c5d6fc3b07',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-15T18:04:53Z,virtual_size=,visibility=), allow threads: False {{(pid=61806) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 684.992782] env[61806]: DEBUG nova.virt.hardware [None req-865353ee-538a-40fc-bfba-502aba0b6a20 tempest-VolumesAdminNegativeTest-1387177911 tempest-VolumesAdminNegativeTest-1387177911-project-member] Flavor limits 0:0:0 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 684.992938] env[61806]: DEBUG nova.virt.hardware [None req-865353ee-538a-40fc-bfba-502aba0b6a20 tempest-VolumesAdminNegativeTest-1387177911 tempest-VolumesAdminNegativeTest-1387177911-project-member] Image limits 0:0:0 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 684.993133] env[61806]: DEBUG nova.virt.hardware [None req-865353ee-538a-40fc-bfba-502aba0b6a20 tempest-VolumesAdminNegativeTest-1387177911 tempest-VolumesAdminNegativeTest-1387177911-project-member] Flavor pref 0:0:0 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 684.993280] env[61806]: DEBUG nova.virt.hardware [None req-865353ee-538a-40fc-bfba-502aba0b6a20 tempest-VolumesAdminNegativeTest-1387177911 tempest-VolumesAdminNegativeTest-1387177911-project-member] Image pref 0:0:0 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 684.993423] env[61806]: DEBUG nova.virt.hardware [None req-865353ee-538a-40fc-bfba-502aba0b6a20 tempest-VolumesAdminNegativeTest-1387177911 tempest-VolumesAdminNegativeTest-1387177911-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 684.993625] env[61806]: DEBUG nova.virt.hardware [None req-865353ee-538a-40fc-bfba-502aba0b6a20 tempest-VolumesAdminNegativeTest-1387177911 tempest-VolumesAdminNegativeTest-1387177911-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61806) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 684.993895] env[61806]: DEBUG nova.virt.hardware [None req-865353ee-538a-40fc-bfba-502aba0b6a20 tempest-VolumesAdminNegativeTest-1387177911 tempest-VolumesAdminNegativeTest-1387177911-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61806) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 684.994364] env[61806]: DEBUG nova.virt.hardware [None req-865353ee-538a-40fc-bfba-502aba0b6a20 tempest-VolumesAdminNegativeTest-1387177911 tempest-VolumesAdminNegativeTest-1387177911-project-member] Got 1 possible topologies {{(pid=61806) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 684.994364] env[61806]: DEBUG nova.virt.hardware [None req-865353ee-538a-40fc-bfba-502aba0b6a20 tempest-VolumesAdminNegativeTest-1387177911 tempest-VolumesAdminNegativeTest-1387177911-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61806) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 684.994364] env[61806]: DEBUG nova.virt.hardware [None req-865353ee-538a-40fc-bfba-502aba0b6a20 tempest-VolumesAdminNegativeTest-1387177911 tempest-VolumesAdminNegativeTest-1387177911-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61806) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 684.995221] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f507f096-23d5-4fda-9011-050c4abf719d {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 685.002963] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-50daed63-c261-4b0d-b6e2-f69c3d5fa66c {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 685.016063] env[61806]: ERROR nova.compute.manager [None req-865353ee-538a-40fc-bfba-502aba0b6a20 tempest-VolumesAdminNegativeTest-1387177911 tempest-VolumesAdminNegativeTest-1387177911-project-member] [instance: b7d55e87-95db-4c45-9417-bbece4fbb98d] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 03736e5c-7351-4a9e-ab04-f54729f3099f, please check neutron logs for more information. [ 685.016063] env[61806]: ERROR nova.compute.manager [instance: b7d55e87-95db-4c45-9417-bbece4fbb98d] Traceback (most recent call last): [ 685.016063] env[61806]: ERROR nova.compute.manager [instance: b7d55e87-95db-4c45-9417-bbece4fbb98d] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 685.016063] env[61806]: ERROR nova.compute.manager [instance: b7d55e87-95db-4c45-9417-bbece4fbb98d] yield resources [ 685.016063] env[61806]: ERROR nova.compute.manager [instance: b7d55e87-95db-4c45-9417-bbece4fbb98d] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 685.016063] env[61806]: ERROR nova.compute.manager [instance: b7d55e87-95db-4c45-9417-bbece4fbb98d] self.driver.spawn(context, instance, image_meta, [ 685.016063] env[61806]: ERROR nova.compute.manager [instance: b7d55e87-95db-4c45-9417-bbece4fbb98d] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 685.016063] env[61806]: ERROR nova.compute.manager [instance: b7d55e87-95db-4c45-9417-bbece4fbb98d] self._vmops.spawn(context, instance, image_meta, injected_files, [ 685.016063] env[61806]: ERROR nova.compute.manager [instance: b7d55e87-95db-4c45-9417-bbece4fbb98d] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 685.016063] env[61806]: ERROR nova.compute.manager [instance: b7d55e87-95db-4c45-9417-bbece4fbb98d] vm_ref = self.build_virtual_machine(instance, [ 685.016063] env[61806]: ERROR nova.compute.manager [instance: b7d55e87-95db-4c45-9417-bbece4fbb98d] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 685.016472] env[61806]: ERROR nova.compute.manager [instance: b7d55e87-95db-4c45-9417-bbece4fbb98d] vif_infos = vmwarevif.get_vif_info(self._session, [ 685.016472] env[61806]: ERROR nova.compute.manager [instance: b7d55e87-95db-4c45-9417-bbece4fbb98d] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 685.016472] env[61806]: ERROR nova.compute.manager [instance: b7d55e87-95db-4c45-9417-bbece4fbb98d] for vif in network_info: [ 685.016472] env[61806]: ERROR nova.compute.manager [instance: b7d55e87-95db-4c45-9417-bbece4fbb98d] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 685.016472] env[61806]: ERROR nova.compute.manager [instance: b7d55e87-95db-4c45-9417-bbece4fbb98d] return self._sync_wrapper(fn, *args, **kwargs) [ 685.016472] env[61806]: ERROR nova.compute.manager [instance: b7d55e87-95db-4c45-9417-bbece4fbb98d] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 685.016472] env[61806]: ERROR nova.compute.manager [instance: b7d55e87-95db-4c45-9417-bbece4fbb98d] self.wait() [ 685.016472] env[61806]: ERROR nova.compute.manager [instance: b7d55e87-95db-4c45-9417-bbece4fbb98d] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 685.016472] env[61806]: ERROR nova.compute.manager [instance: b7d55e87-95db-4c45-9417-bbece4fbb98d] self[:] = self._gt.wait() [ 685.016472] env[61806]: ERROR nova.compute.manager [instance: b7d55e87-95db-4c45-9417-bbece4fbb98d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 685.016472] env[61806]: ERROR nova.compute.manager [instance: b7d55e87-95db-4c45-9417-bbece4fbb98d] return self._exit_event.wait() [ 685.016472] env[61806]: ERROR nova.compute.manager [instance: b7d55e87-95db-4c45-9417-bbece4fbb98d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 685.016472] env[61806]: ERROR nova.compute.manager [instance: b7d55e87-95db-4c45-9417-bbece4fbb98d] current.throw(*self._exc) [ 685.016815] env[61806]: ERROR nova.compute.manager [instance: b7d55e87-95db-4c45-9417-bbece4fbb98d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 685.016815] env[61806]: ERROR nova.compute.manager [instance: b7d55e87-95db-4c45-9417-bbece4fbb98d] result = function(*args, **kwargs) [ 685.016815] env[61806]: ERROR nova.compute.manager [instance: b7d55e87-95db-4c45-9417-bbece4fbb98d] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 685.016815] env[61806]: ERROR nova.compute.manager [instance: b7d55e87-95db-4c45-9417-bbece4fbb98d] return func(*args, **kwargs) [ 685.016815] env[61806]: ERROR nova.compute.manager [instance: b7d55e87-95db-4c45-9417-bbece4fbb98d] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 685.016815] env[61806]: ERROR nova.compute.manager [instance: b7d55e87-95db-4c45-9417-bbece4fbb98d] raise e [ 685.016815] env[61806]: ERROR nova.compute.manager [instance: b7d55e87-95db-4c45-9417-bbece4fbb98d] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 685.016815] env[61806]: ERROR nova.compute.manager [instance: b7d55e87-95db-4c45-9417-bbece4fbb98d] nwinfo = self.network_api.allocate_for_instance( [ 685.016815] env[61806]: ERROR nova.compute.manager [instance: b7d55e87-95db-4c45-9417-bbece4fbb98d] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 685.016815] env[61806]: ERROR nova.compute.manager [instance: b7d55e87-95db-4c45-9417-bbece4fbb98d] created_port_ids = self._update_ports_for_instance( [ 685.016815] env[61806]: ERROR nova.compute.manager [instance: b7d55e87-95db-4c45-9417-bbece4fbb98d] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 685.016815] env[61806]: ERROR nova.compute.manager [instance: b7d55e87-95db-4c45-9417-bbece4fbb98d] with excutils.save_and_reraise_exception(): [ 685.016815] env[61806]: ERROR nova.compute.manager [instance: b7d55e87-95db-4c45-9417-bbece4fbb98d] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 685.017204] env[61806]: ERROR nova.compute.manager [instance: b7d55e87-95db-4c45-9417-bbece4fbb98d] self.force_reraise() [ 685.017204] env[61806]: ERROR nova.compute.manager [instance: b7d55e87-95db-4c45-9417-bbece4fbb98d] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 685.017204] env[61806]: ERROR nova.compute.manager [instance: b7d55e87-95db-4c45-9417-bbece4fbb98d] raise self.value [ 685.017204] env[61806]: ERROR nova.compute.manager [instance: b7d55e87-95db-4c45-9417-bbece4fbb98d] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 685.017204] env[61806]: ERROR nova.compute.manager [instance: b7d55e87-95db-4c45-9417-bbece4fbb98d] updated_port = self._update_port( [ 685.017204] env[61806]: ERROR nova.compute.manager [instance: b7d55e87-95db-4c45-9417-bbece4fbb98d] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 685.017204] env[61806]: ERROR nova.compute.manager [instance: b7d55e87-95db-4c45-9417-bbece4fbb98d] _ensure_no_port_binding_failure(port) [ 685.017204] env[61806]: ERROR nova.compute.manager [instance: b7d55e87-95db-4c45-9417-bbece4fbb98d] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 685.017204] env[61806]: ERROR nova.compute.manager [instance: b7d55e87-95db-4c45-9417-bbece4fbb98d] raise exception.PortBindingFailed(port_id=port['id']) [ 685.017204] env[61806]: ERROR nova.compute.manager [instance: b7d55e87-95db-4c45-9417-bbece4fbb98d] nova.exception.PortBindingFailed: Binding failed for port 03736e5c-7351-4a9e-ab04-f54729f3099f, please check neutron logs for more information. [ 685.017204] env[61806]: ERROR nova.compute.manager [instance: b7d55e87-95db-4c45-9417-bbece4fbb98d] [ 685.017204] env[61806]: INFO nova.compute.manager [None req-865353ee-538a-40fc-bfba-502aba0b6a20 tempest-VolumesAdminNegativeTest-1387177911 tempest-VolumesAdminNegativeTest-1387177911-project-member] [instance: b7d55e87-95db-4c45-9417-bbece4fbb98d] Terminating instance [ 685.018311] env[61806]: DEBUG oslo_concurrency.lockutils [None req-865353ee-538a-40fc-bfba-502aba0b6a20 tempest-VolumesAdminNegativeTest-1387177911 tempest-VolumesAdminNegativeTest-1387177911-project-member] Acquiring lock "refresh_cache-b7d55e87-95db-4c45-9417-bbece4fbb98d" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 685.180694] env[61806]: DEBUG nova.network.neutron [req-b21a30cb-e457-4f7f-8248-c5841198cc64 req-24ed0238-f790-4ae7-aa56-7ef941665829 service nova] [instance: b7d55e87-95db-4c45-9417-bbece4fbb98d] Instance cache missing network info. {{(pid=61806) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 685.210115] env[61806]: DEBUG oslo_concurrency.lockutils [None req-6b55cbbf-00f7-4d4d-87f3-a59cc1c553b0 tempest-ServerRescueNegativeTestJSON-1483258717 tempest-ServerRescueNegativeTestJSON-1483258717-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 685.289335] env[61806]: DEBUG nova.network.neutron [req-b21a30cb-e457-4f7f-8248-c5841198cc64 req-24ed0238-f790-4ae7-aa56-7ef941665829 service nova] [instance: b7d55e87-95db-4c45-9417-bbece4fbb98d] Updating instance_info_cache with network_info: [] {{(pid=61806) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 685.449622] env[61806]: DEBUG oslo_concurrency.lockutils [None req-81a55da1-45c4-45f8-af56-f8f7d424c8aa tempest-ServersTestJSON-430716070 tempest-ServersTestJSON-430716070-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.511s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 685.451173] env[61806]: DEBUG nova.compute.manager [None req-81a55da1-45c4-45f8-af56-f8f7d424c8aa tempest-ServersTestJSON-430716070 tempest-ServersTestJSON-430716070-project-member] [instance: eca2e3da-b38c-482e-b9b6-4ad75d83caa1] Start building networks asynchronously for instance. {{(pid=61806) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 685.454393] env[61806]: DEBUG oslo_concurrency.lockutils [None req-a93ef3a4-6e83-4dfe-9292-bac21f14d678 tempest-ServerExternalEventsTest-1847900285 tempest-ServerExternalEventsTest-1847900285-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 28.631s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 685.796233] env[61806]: DEBUG oslo_concurrency.lockutils [req-b21a30cb-e457-4f7f-8248-c5841198cc64 req-24ed0238-f790-4ae7-aa56-7ef941665829 service nova] Releasing lock "refresh_cache-b7d55e87-95db-4c45-9417-bbece4fbb98d" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 685.796689] env[61806]: DEBUG oslo_concurrency.lockutils [None req-865353ee-538a-40fc-bfba-502aba0b6a20 tempest-VolumesAdminNegativeTest-1387177911 tempest-VolumesAdminNegativeTest-1387177911-project-member] Acquired lock "refresh_cache-b7d55e87-95db-4c45-9417-bbece4fbb98d" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 685.796869] env[61806]: DEBUG nova.network.neutron [None req-865353ee-538a-40fc-bfba-502aba0b6a20 tempest-VolumesAdminNegativeTest-1387177911 tempest-VolumesAdminNegativeTest-1387177911-project-member] [instance: b7d55e87-95db-4c45-9417-bbece4fbb98d] Building network info cache for instance {{(pid=61806) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 685.959771] env[61806]: DEBUG nova.compute.utils [None req-81a55da1-45c4-45f8-af56-f8f7d424c8aa tempest-ServersTestJSON-430716070 tempest-ServersTestJSON-430716070-project-member] Using /dev/sd instead of None {{(pid=61806) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 685.967951] env[61806]: DEBUG nova.compute.manager [None req-81a55da1-45c4-45f8-af56-f8f7d424c8aa tempest-ServersTestJSON-430716070 tempest-ServersTestJSON-430716070-project-member] [instance: eca2e3da-b38c-482e-b9b6-4ad75d83caa1] Allocating IP information in the background. {{(pid=61806) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 685.968369] env[61806]: DEBUG nova.network.neutron [None req-81a55da1-45c4-45f8-af56-f8f7d424c8aa tempest-ServersTestJSON-430716070 tempest-ServersTestJSON-430716070-project-member] [instance: eca2e3da-b38c-482e-b9b6-4ad75d83caa1] allocate_for_instance() {{(pid=61806) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 686.035901] env[61806]: DEBUG nova.policy [None req-81a55da1-45c4-45f8-af56-f8f7d424c8aa tempest-ServersTestJSON-430716070 tempest-ServersTestJSON-430716070-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'ca174cdc99b1435b8faa6c9786241f28', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'd201f9468eb549518a8cf54852c06056', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61806) authorize /opt/stack/nova/nova/policy.py:201}} [ 686.388446] env[61806]: DEBUG nova.network.neutron [None req-81a55da1-45c4-45f8-af56-f8f7d424c8aa tempest-ServersTestJSON-430716070 tempest-ServersTestJSON-430716070-project-member] [instance: eca2e3da-b38c-482e-b9b6-4ad75d83caa1] Successfully created port: 007a0ffc-be2b-4f45-b7a2-5df4203fb2da {{(pid=61806) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 686.407773] env[61806]: DEBUG nova.network.neutron [None req-865353ee-538a-40fc-bfba-502aba0b6a20 tempest-VolumesAdminNegativeTest-1387177911 tempest-VolumesAdminNegativeTest-1387177911-project-member] [instance: b7d55e87-95db-4c45-9417-bbece4fbb98d] Instance cache missing network info. {{(pid=61806) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 686.414038] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4f0f3265-071e-4109-b3f1-52d2fe618235 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 686.421610] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c3d9a40e-76cc-4c80-965b-523ab3330710 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 686.457542] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9cb47147-ce5d-4e6e-8424-f3b02b80ab4d {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 686.467255] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f4f83351-5d70-4fba-b95f-a8ab918c1497 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 686.473403] env[61806]: DEBUG nova.compute.manager [None req-81a55da1-45c4-45f8-af56-f8f7d424c8aa tempest-ServersTestJSON-430716070 tempest-ServersTestJSON-430716070-project-member] [instance: eca2e3da-b38c-482e-b9b6-4ad75d83caa1] Start building block device mappings for instance. {{(pid=61806) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 686.485698] env[61806]: DEBUG nova.compute.provider_tree [None req-a93ef3a4-6e83-4dfe-9292-bac21f14d678 tempest-ServerExternalEventsTest-1847900285 tempest-ServerExternalEventsTest-1847900285-project-member] Inventory has not changed in ProviderTree for provider: a2858be1-fd22-4e08-979e-87ad25293407 {{(pid=61806) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 686.553146] env[61806]: DEBUG nova.network.neutron [None req-865353ee-538a-40fc-bfba-502aba0b6a20 tempest-VolumesAdminNegativeTest-1387177911 tempest-VolumesAdminNegativeTest-1387177911-project-member] [instance: b7d55e87-95db-4c45-9417-bbece4fbb98d] Updating instance_info_cache with network_info: [] {{(pid=61806) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 686.694666] env[61806]: DEBUG nova.compute.manager [req-b6b5ea89-71a6-4f00-88c7-98fe9dd1b71d req-f35f6d1a-76d5-416b-9b6a-7ffb319ae1d7 service nova] [instance: b7d55e87-95db-4c45-9417-bbece4fbb98d] Received event network-vif-deleted-03736e5c-7351-4a9e-ab04-f54729f3099f {{(pid=61806) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 686.994017] env[61806]: DEBUG nova.scheduler.client.report [None req-a93ef3a4-6e83-4dfe-9292-bac21f14d678 tempest-ServerExternalEventsTest-1847900285 tempest-ServerExternalEventsTest-1847900285-project-member] Inventory has not changed for provider a2858be1-fd22-4e08-979e-87ad25293407 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 138, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61806) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 687.058261] env[61806]: DEBUG oslo_concurrency.lockutils [None req-865353ee-538a-40fc-bfba-502aba0b6a20 tempest-VolumesAdminNegativeTest-1387177911 tempest-VolumesAdminNegativeTest-1387177911-project-member] Releasing lock "refresh_cache-b7d55e87-95db-4c45-9417-bbece4fbb98d" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 687.058691] env[61806]: DEBUG nova.compute.manager [None req-865353ee-538a-40fc-bfba-502aba0b6a20 tempest-VolumesAdminNegativeTest-1387177911 tempest-VolumesAdminNegativeTest-1387177911-project-member] [instance: b7d55e87-95db-4c45-9417-bbece4fbb98d] Start destroying the instance on the hypervisor. {{(pid=61806) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 687.058886] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-865353ee-538a-40fc-bfba-502aba0b6a20 tempest-VolumesAdminNegativeTest-1387177911 tempest-VolumesAdminNegativeTest-1387177911-project-member] [instance: b7d55e87-95db-4c45-9417-bbece4fbb98d] Destroying instance {{(pid=61806) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 687.059198] env[61806]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-52cde168-05e1-48e1-ab5b-b251e1e86c7f {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 687.068319] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-160b31cd-73b6-43a9-99f1-d7b358bc962c {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 687.090701] env[61806]: WARNING nova.virt.vmwareapi.vmops [None req-865353ee-538a-40fc-bfba-502aba0b6a20 tempest-VolumesAdminNegativeTest-1387177911 tempest-VolumesAdminNegativeTest-1387177911-project-member] [instance: b7d55e87-95db-4c45-9417-bbece4fbb98d] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance b7d55e87-95db-4c45-9417-bbece4fbb98d could not be found. [ 687.090983] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-865353ee-538a-40fc-bfba-502aba0b6a20 tempest-VolumesAdminNegativeTest-1387177911 tempest-VolumesAdminNegativeTest-1387177911-project-member] [instance: b7d55e87-95db-4c45-9417-bbece4fbb98d] Instance destroyed {{(pid=61806) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 687.091209] env[61806]: INFO nova.compute.manager [None req-865353ee-538a-40fc-bfba-502aba0b6a20 tempest-VolumesAdminNegativeTest-1387177911 tempest-VolumesAdminNegativeTest-1387177911-project-member] [instance: b7d55e87-95db-4c45-9417-bbece4fbb98d] Took 0.03 seconds to destroy the instance on the hypervisor. [ 687.091414] env[61806]: DEBUG oslo.service.loopingcall [None req-865353ee-538a-40fc-bfba-502aba0b6a20 tempest-VolumesAdminNegativeTest-1387177911 tempest-VolumesAdminNegativeTest-1387177911-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61806) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 687.091867] env[61806]: DEBUG nova.compute.manager [-] [instance: b7d55e87-95db-4c45-9417-bbece4fbb98d] Deallocating network for instance {{(pid=61806) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 687.091867] env[61806]: DEBUG nova.network.neutron [-] [instance: b7d55e87-95db-4c45-9417-bbece4fbb98d] deallocate_for_instance() {{(pid=61806) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 687.105249] env[61806]: DEBUG nova.network.neutron [-] [instance: b7d55e87-95db-4c45-9417-bbece4fbb98d] Instance cache missing network info. {{(pid=61806) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 687.392397] env[61806]: ERROR nova.compute.manager [None req-81a55da1-45c4-45f8-af56-f8f7d424c8aa tempest-ServersTestJSON-430716070 tempest-ServersTestJSON-430716070-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 007a0ffc-be2b-4f45-b7a2-5df4203fb2da, please check neutron logs for more information. [ 687.392397] env[61806]: ERROR nova.compute.manager Traceback (most recent call last): [ 687.392397] env[61806]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 687.392397] env[61806]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 687.392397] env[61806]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 687.392397] env[61806]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 687.392397] env[61806]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 687.392397] env[61806]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 687.392397] env[61806]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 687.392397] env[61806]: ERROR nova.compute.manager self.force_reraise() [ 687.392397] env[61806]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 687.392397] env[61806]: ERROR nova.compute.manager raise self.value [ 687.392397] env[61806]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 687.392397] env[61806]: ERROR nova.compute.manager updated_port = self._update_port( [ 687.392397] env[61806]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 687.392397] env[61806]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 687.392980] env[61806]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 687.392980] env[61806]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 687.392980] env[61806]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 007a0ffc-be2b-4f45-b7a2-5df4203fb2da, please check neutron logs for more information. [ 687.392980] env[61806]: ERROR nova.compute.manager [ 687.392980] env[61806]: Traceback (most recent call last): [ 687.392980] env[61806]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 687.392980] env[61806]: listener.cb(fileno) [ 687.392980] env[61806]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 687.392980] env[61806]: result = function(*args, **kwargs) [ 687.392980] env[61806]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 687.392980] env[61806]: return func(*args, **kwargs) [ 687.392980] env[61806]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 687.392980] env[61806]: raise e [ 687.392980] env[61806]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 687.392980] env[61806]: nwinfo = self.network_api.allocate_for_instance( [ 687.392980] env[61806]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 687.392980] env[61806]: created_port_ids = self._update_ports_for_instance( [ 687.392980] env[61806]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 687.392980] env[61806]: with excutils.save_and_reraise_exception(): [ 687.392980] env[61806]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 687.392980] env[61806]: self.force_reraise() [ 687.392980] env[61806]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 687.392980] env[61806]: raise self.value [ 687.392980] env[61806]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 687.392980] env[61806]: updated_port = self._update_port( [ 687.392980] env[61806]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 687.392980] env[61806]: _ensure_no_port_binding_failure(port) [ 687.392980] env[61806]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 687.392980] env[61806]: raise exception.PortBindingFailed(port_id=port['id']) [ 687.393877] env[61806]: nova.exception.PortBindingFailed: Binding failed for port 007a0ffc-be2b-4f45-b7a2-5df4203fb2da, please check neutron logs for more information. [ 687.393877] env[61806]: Removing descriptor: 14 [ 687.486218] env[61806]: DEBUG nova.compute.manager [None req-81a55da1-45c4-45f8-af56-f8f7d424c8aa tempest-ServersTestJSON-430716070 tempest-ServersTestJSON-430716070-project-member] [instance: eca2e3da-b38c-482e-b9b6-4ad75d83caa1] Start spawning the instance on the hypervisor. {{(pid=61806) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 687.498230] env[61806]: DEBUG oslo_concurrency.lockutils [None req-a93ef3a4-6e83-4dfe-9292-bac21f14d678 tempest-ServerExternalEventsTest-1847900285 tempest-ServerExternalEventsTest-1847900285-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.044s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 687.498891] env[61806]: ERROR nova.compute.manager [None req-a93ef3a4-6e83-4dfe-9292-bac21f14d678 tempest-ServerExternalEventsTest-1847900285 tempest-ServerExternalEventsTest-1847900285-project-member] [instance: d5434dd2-206f-4a1e-b97b-bd6f691b6ddf] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 6b28a2b6-fc7b-4db9-84db-18b0d6bbe032, please check neutron logs for more information. [ 687.498891] env[61806]: ERROR nova.compute.manager [instance: d5434dd2-206f-4a1e-b97b-bd6f691b6ddf] Traceback (most recent call last): [ 687.498891] env[61806]: ERROR nova.compute.manager [instance: d5434dd2-206f-4a1e-b97b-bd6f691b6ddf] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 687.498891] env[61806]: ERROR nova.compute.manager [instance: d5434dd2-206f-4a1e-b97b-bd6f691b6ddf] self.driver.spawn(context, instance, image_meta, [ 687.498891] env[61806]: ERROR nova.compute.manager [instance: d5434dd2-206f-4a1e-b97b-bd6f691b6ddf] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 687.498891] env[61806]: ERROR nova.compute.manager [instance: d5434dd2-206f-4a1e-b97b-bd6f691b6ddf] self._vmops.spawn(context, instance, image_meta, injected_files, [ 687.498891] env[61806]: ERROR nova.compute.manager [instance: d5434dd2-206f-4a1e-b97b-bd6f691b6ddf] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 687.498891] env[61806]: ERROR nova.compute.manager [instance: d5434dd2-206f-4a1e-b97b-bd6f691b6ddf] vm_ref = self.build_virtual_machine(instance, [ 687.498891] env[61806]: ERROR nova.compute.manager [instance: d5434dd2-206f-4a1e-b97b-bd6f691b6ddf] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 687.498891] env[61806]: ERROR nova.compute.manager [instance: d5434dd2-206f-4a1e-b97b-bd6f691b6ddf] vif_infos = vmwarevif.get_vif_info(self._session, [ 687.498891] env[61806]: ERROR nova.compute.manager [instance: d5434dd2-206f-4a1e-b97b-bd6f691b6ddf] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 687.499454] env[61806]: ERROR nova.compute.manager [instance: d5434dd2-206f-4a1e-b97b-bd6f691b6ddf] for vif in network_info: [ 687.499454] env[61806]: ERROR nova.compute.manager [instance: d5434dd2-206f-4a1e-b97b-bd6f691b6ddf] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 687.499454] env[61806]: ERROR nova.compute.manager [instance: d5434dd2-206f-4a1e-b97b-bd6f691b6ddf] return self._sync_wrapper(fn, *args, **kwargs) [ 687.499454] env[61806]: ERROR nova.compute.manager [instance: d5434dd2-206f-4a1e-b97b-bd6f691b6ddf] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 687.499454] env[61806]: ERROR nova.compute.manager [instance: d5434dd2-206f-4a1e-b97b-bd6f691b6ddf] self.wait() [ 687.499454] env[61806]: ERROR nova.compute.manager [instance: d5434dd2-206f-4a1e-b97b-bd6f691b6ddf] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 687.499454] env[61806]: ERROR nova.compute.manager [instance: d5434dd2-206f-4a1e-b97b-bd6f691b6ddf] self[:] = self._gt.wait() [ 687.499454] env[61806]: ERROR nova.compute.manager [instance: d5434dd2-206f-4a1e-b97b-bd6f691b6ddf] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 687.499454] env[61806]: ERROR nova.compute.manager [instance: d5434dd2-206f-4a1e-b97b-bd6f691b6ddf] return self._exit_event.wait() [ 687.499454] env[61806]: ERROR nova.compute.manager [instance: d5434dd2-206f-4a1e-b97b-bd6f691b6ddf] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 687.499454] env[61806]: ERROR nova.compute.manager [instance: d5434dd2-206f-4a1e-b97b-bd6f691b6ddf] current.throw(*self._exc) [ 687.499454] env[61806]: ERROR nova.compute.manager [instance: d5434dd2-206f-4a1e-b97b-bd6f691b6ddf] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 687.499454] env[61806]: ERROR nova.compute.manager [instance: d5434dd2-206f-4a1e-b97b-bd6f691b6ddf] result = function(*args, **kwargs) [ 687.500390] env[61806]: ERROR nova.compute.manager [instance: d5434dd2-206f-4a1e-b97b-bd6f691b6ddf] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 687.500390] env[61806]: ERROR nova.compute.manager [instance: d5434dd2-206f-4a1e-b97b-bd6f691b6ddf] return func(*args, **kwargs) [ 687.500390] env[61806]: ERROR nova.compute.manager [instance: d5434dd2-206f-4a1e-b97b-bd6f691b6ddf] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 687.500390] env[61806]: ERROR nova.compute.manager [instance: d5434dd2-206f-4a1e-b97b-bd6f691b6ddf] raise e [ 687.500390] env[61806]: ERROR nova.compute.manager [instance: d5434dd2-206f-4a1e-b97b-bd6f691b6ddf] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 687.500390] env[61806]: ERROR nova.compute.manager [instance: d5434dd2-206f-4a1e-b97b-bd6f691b6ddf] nwinfo = self.network_api.allocate_for_instance( [ 687.500390] env[61806]: ERROR nova.compute.manager [instance: d5434dd2-206f-4a1e-b97b-bd6f691b6ddf] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 687.500390] env[61806]: ERROR nova.compute.manager [instance: d5434dd2-206f-4a1e-b97b-bd6f691b6ddf] created_port_ids = self._update_ports_for_instance( [ 687.500390] env[61806]: ERROR nova.compute.manager [instance: d5434dd2-206f-4a1e-b97b-bd6f691b6ddf] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 687.500390] env[61806]: ERROR nova.compute.manager [instance: d5434dd2-206f-4a1e-b97b-bd6f691b6ddf] with excutils.save_and_reraise_exception(): [ 687.500390] env[61806]: ERROR nova.compute.manager [instance: d5434dd2-206f-4a1e-b97b-bd6f691b6ddf] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 687.500390] env[61806]: ERROR nova.compute.manager [instance: d5434dd2-206f-4a1e-b97b-bd6f691b6ddf] self.force_reraise() [ 687.500390] env[61806]: ERROR nova.compute.manager [instance: d5434dd2-206f-4a1e-b97b-bd6f691b6ddf] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 687.501300] env[61806]: ERROR nova.compute.manager [instance: d5434dd2-206f-4a1e-b97b-bd6f691b6ddf] raise self.value [ 687.501300] env[61806]: ERROR nova.compute.manager [instance: d5434dd2-206f-4a1e-b97b-bd6f691b6ddf] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 687.501300] env[61806]: ERROR nova.compute.manager [instance: d5434dd2-206f-4a1e-b97b-bd6f691b6ddf] updated_port = self._update_port( [ 687.501300] env[61806]: ERROR nova.compute.manager [instance: d5434dd2-206f-4a1e-b97b-bd6f691b6ddf] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 687.501300] env[61806]: ERROR nova.compute.manager [instance: d5434dd2-206f-4a1e-b97b-bd6f691b6ddf] _ensure_no_port_binding_failure(port) [ 687.501300] env[61806]: ERROR nova.compute.manager [instance: d5434dd2-206f-4a1e-b97b-bd6f691b6ddf] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 687.501300] env[61806]: ERROR nova.compute.manager [instance: d5434dd2-206f-4a1e-b97b-bd6f691b6ddf] raise exception.PortBindingFailed(port_id=port['id']) [ 687.501300] env[61806]: ERROR nova.compute.manager [instance: d5434dd2-206f-4a1e-b97b-bd6f691b6ddf] nova.exception.PortBindingFailed: Binding failed for port 6b28a2b6-fc7b-4db9-84db-18b0d6bbe032, please check neutron logs for more information. [ 687.501300] env[61806]: ERROR nova.compute.manager [instance: d5434dd2-206f-4a1e-b97b-bd6f691b6ddf] [ 687.501300] env[61806]: DEBUG nova.compute.utils [None req-a93ef3a4-6e83-4dfe-9292-bac21f14d678 tempest-ServerExternalEventsTest-1847900285 tempest-ServerExternalEventsTest-1847900285-project-member] [instance: d5434dd2-206f-4a1e-b97b-bd6f691b6ddf] Binding failed for port 6b28a2b6-fc7b-4db9-84db-18b0d6bbe032, please check neutron logs for more information. {{(pid=61806) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 687.501951] env[61806]: DEBUG oslo_concurrency.lockutils [None req-2bf81f39-4b8d-4851-a793-760af008ffe2 tempest-ImagesOneServerTestJSON-1003762139 tempest-ImagesOneServerTestJSON-1003762139-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 28.904s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 687.502417] env[61806]: INFO nova.compute.claims [None req-2bf81f39-4b8d-4851-a793-760af008ffe2 tempest-ImagesOneServerTestJSON-1003762139 tempest-ImagesOneServerTestJSON-1003762139-project-member] [instance: 294f0a37-7589-405d-a324-13099c80f458] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 687.505105] env[61806]: DEBUG nova.compute.manager [None req-a93ef3a4-6e83-4dfe-9292-bac21f14d678 tempest-ServerExternalEventsTest-1847900285 tempest-ServerExternalEventsTest-1847900285-project-member] [instance: d5434dd2-206f-4a1e-b97b-bd6f691b6ddf] Build of instance d5434dd2-206f-4a1e-b97b-bd6f691b6ddf was re-scheduled: Binding failed for port 6b28a2b6-fc7b-4db9-84db-18b0d6bbe032, please check neutron logs for more information. {{(pid=61806) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 687.505539] env[61806]: DEBUG nova.compute.manager [None req-a93ef3a4-6e83-4dfe-9292-bac21f14d678 tempest-ServerExternalEventsTest-1847900285 tempest-ServerExternalEventsTest-1847900285-project-member] [instance: d5434dd2-206f-4a1e-b97b-bd6f691b6ddf] Unplugging VIFs for instance {{(pid=61806) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 687.505763] env[61806]: DEBUG oslo_concurrency.lockutils [None req-a93ef3a4-6e83-4dfe-9292-bac21f14d678 tempest-ServerExternalEventsTest-1847900285 tempest-ServerExternalEventsTest-1847900285-project-member] Acquiring lock "refresh_cache-d5434dd2-206f-4a1e-b97b-bd6f691b6ddf" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 687.505909] env[61806]: DEBUG oslo_concurrency.lockutils [None req-a93ef3a4-6e83-4dfe-9292-bac21f14d678 tempest-ServerExternalEventsTest-1847900285 tempest-ServerExternalEventsTest-1847900285-project-member] Acquired lock "refresh_cache-d5434dd2-206f-4a1e-b97b-bd6f691b6ddf" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 687.506073] env[61806]: DEBUG nova.network.neutron [None req-a93ef3a4-6e83-4dfe-9292-bac21f14d678 tempest-ServerExternalEventsTest-1847900285 tempest-ServerExternalEventsTest-1847900285-project-member] [instance: d5434dd2-206f-4a1e-b97b-bd6f691b6ddf] Building network info cache for instance {{(pid=61806) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 687.517171] env[61806]: DEBUG nova.virt.hardware [None req-81a55da1-45c4-45f8-af56-f8f7d424c8aa tempest-ServersTestJSON-430716070 tempest-ServersTestJSON-430716070-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-15T18:05:10Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-15T18:04:52Z,direct_url=,disk_format='vmdk',id=ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='84ba83fce2c349a6988173c5d6fc3b07',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-15T18:04:53Z,virtual_size=,visibility=), allow threads: False {{(pid=61806) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 687.517417] env[61806]: DEBUG nova.virt.hardware [None req-81a55da1-45c4-45f8-af56-f8f7d424c8aa tempest-ServersTestJSON-430716070 tempest-ServersTestJSON-430716070-project-member] Flavor limits 0:0:0 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 687.517593] env[61806]: DEBUG nova.virt.hardware [None req-81a55da1-45c4-45f8-af56-f8f7d424c8aa tempest-ServersTestJSON-430716070 tempest-ServersTestJSON-430716070-project-member] Image limits 0:0:0 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 687.517796] env[61806]: DEBUG nova.virt.hardware [None req-81a55da1-45c4-45f8-af56-f8f7d424c8aa tempest-ServersTestJSON-430716070 tempest-ServersTestJSON-430716070-project-member] Flavor pref 0:0:0 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 687.517951] env[61806]: DEBUG nova.virt.hardware [None req-81a55da1-45c4-45f8-af56-f8f7d424c8aa tempest-ServersTestJSON-430716070 tempest-ServersTestJSON-430716070-project-member] Image pref 0:0:0 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 687.518259] env[61806]: DEBUG nova.virt.hardware [None req-81a55da1-45c4-45f8-af56-f8f7d424c8aa tempest-ServersTestJSON-430716070 tempest-ServersTestJSON-430716070-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 687.518491] env[61806]: DEBUG nova.virt.hardware [None req-81a55da1-45c4-45f8-af56-f8f7d424c8aa tempest-ServersTestJSON-430716070 tempest-ServersTestJSON-430716070-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61806) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 687.518655] env[61806]: DEBUG nova.virt.hardware [None req-81a55da1-45c4-45f8-af56-f8f7d424c8aa tempest-ServersTestJSON-430716070 tempest-ServersTestJSON-430716070-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61806) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 687.518824] env[61806]: DEBUG nova.virt.hardware [None req-81a55da1-45c4-45f8-af56-f8f7d424c8aa tempest-ServersTestJSON-430716070 tempest-ServersTestJSON-430716070-project-member] Got 1 possible topologies {{(pid=61806) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 687.518989] env[61806]: DEBUG nova.virt.hardware [None req-81a55da1-45c4-45f8-af56-f8f7d424c8aa tempest-ServersTestJSON-430716070 tempest-ServersTestJSON-430716070-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61806) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 687.519202] env[61806]: DEBUG nova.virt.hardware [None req-81a55da1-45c4-45f8-af56-f8f7d424c8aa tempest-ServersTestJSON-430716070 tempest-ServersTestJSON-430716070-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61806) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 687.520341] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-24f3f97a-f0d2-4ae7-a381-3a7fd672ba63 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 687.530217] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e16a00c6-230e-4cc0-b3f9-009959c9983b {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 687.545482] env[61806]: ERROR nova.compute.manager [None req-81a55da1-45c4-45f8-af56-f8f7d424c8aa tempest-ServersTestJSON-430716070 tempest-ServersTestJSON-430716070-project-member] [instance: eca2e3da-b38c-482e-b9b6-4ad75d83caa1] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 007a0ffc-be2b-4f45-b7a2-5df4203fb2da, please check neutron logs for more information. [ 687.545482] env[61806]: ERROR nova.compute.manager [instance: eca2e3da-b38c-482e-b9b6-4ad75d83caa1] Traceback (most recent call last): [ 687.545482] env[61806]: ERROR nova.compute.manager [instance: eca2e3da-b38c-482e-b9b6-4ad75d83caa1] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 687.545482] env[61806]: ERROR nova.compute.manager [instance: eca2e3da-b38c-482e-b9b6-4ad75d83caa1] yield resources [ 687.545482] env[61806]: ERROR nova.compute.manager [instance: eca2e3da-b38c-482e-b9b6-4ad75d83caa1] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 687.545482] env[61806]: ERROR nova.compute.manager [instance: eca2e3da-b38c-482e-b9b6-4ad75d83caa1] self.driver.spawn(context, instance, image_meta, [ 687.545482] env[61806]: ERROR nova.compute.manager [instance: eca2e3da-b38c-482e-b9b6-4ad75d83caa1] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 687.545482] env[61806]: ERROR nova.compute.manager [instance: eca2e3da-b38c-482e-b9b6-4ad75d83caa1] self._vmops.spawn(context, instance, image_meta, injected_files, [ 687.545482] env[61806]: ERROR nova.compute.manager [instance: eca2e3da-b38c-482e-b9b6-4ad75d83caa1] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 687.545482] env[61806]: ERROR nova.compute.manager [instance: eca2e3da-b38c-482e-b9b6-4ad75d83caa1] vm_ref = self.build_virtual_machine(instance, [ 687.545482] env[61806]: ERROR nova.compute.manager [instance: eca2e3da-b38c-482e-b9b6-4ad75d83caa1] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 687.545870] env[61806]: ERROR nova.compute.manager [instance: eca2e3da-b38c-482e-b9b6-4ad75d83caa1] vif_infos = vmwarevif.get_vif_info(self._session, [ 687.545870] env[61806]: ERROR nova.compute.manager [instance: eca2e3da-b38c-482e-b9b6-4ad75d83caa1] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 687.545870] env[61806]: ERROR nova.compute.manager [instance: eca2e3da-b38c-482e-b9b6-4ad75d83caa1] for vif in network_info: [ 687.545870] env[61806]: ERROR nova.compute.manager [instance: eca2e3da-b38c-482e-b9b6-4ad75d83caa1] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 687.545870] env[61806]: ERROR nova.compute.manager [instance: eca2e3da-b38c-482e-b9b6-4ad75d83caa1] return self._sync_wrapper(fn, *args, **kwargs) [ 687.545870] env[61806]: ERROR nova.compute.manager [instance: eca2e3da-b38c-482e-b9b6-4ad75d83caa1] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 687.545870] env[61806]: ERROR nova.compute.manager [instance: eca2e3da-b38c-482e-b9b6-4ad75d83caa1] self.wait() [ 687.545870] env[61806]: ERROR nova.compute.manager [instance: eca2e3da-b38c-482e-b9b6-4ad75d83caa1] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 687.545870] env[61806]: ERROR nova.compute.manager [instance: eca2e3da-b38c-482e-b9b6-4ad75d83caa1] self[:] = self._gt.wait() [ 687.545870] env[61806]: ERROR nova.compute.manager [instance: eca2e3da-b38c-482e-b9b6-4ad75d83caa1] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 687.545870] env[61806]: ERROR nova.compute.manager [instance: eca2e3da-b38c-482e-b9b6-4ad75d83caa1] return self._exit_event.wait() [ 687.545870] env[61806]: ERROR nova.compute.manager [instance: eca2e3da-b38c-482e-b9b6-4ad75d83caa1] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 687.545870] env[61806]: ERROR nova.compute.manager [instance: eca2e3da-b38c-482e-b9b6-4ad75d83caa1] current.throw(*self._exc) [ 687.546292] env[61806]: ERROR nova.compute.manager [instance: eca2e3da-b38c-482e-b9b6-4ad75d83caa1] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 687.546292] env[61806]: ERROR nova.compute.manager [instance: eca2e3da-b38c-482e-b9b6-4ad75d83caa1] result = function(*args, **kwargs) [ 687.546292] env[61806]: ERROR nova.compute.manager [instance: eca2e3da-b38c-482e-b9b6-4ad75d83caa1] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 687.546292] env[61806]: ERROR nova.compute.manager [instance: eca2e3da-b38c-482e-b9b6-4ad75d83caa1] return func(*args, **kwargs) [ 687.546292] env[61806]: ERROR nova.compute.manager [instance: eca2e3da-b38c-482e-b9b6-4ad75d83caa1] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 687.546292] env[61806]: ERROR nova.compute.manager [instance: eca2e3da-b38c-482e-b9b6-4ad75d83caa1] raise e [ 687.546292] env[61806]: ERROR nova.compute.manager [instance: eca2e3da-b38c-482e-b9b6-4ad75d83caa1] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 687.546292] env[61806]: ERROR nova.compute.manager [instance: eca2e3da-b38c-482e-b9b6-4ad75d83caa1] nwinfo = self.network_api.allocate_for_instance( [ 687.546292] env[61806]: ERROR nova.compute.manager [instance: eca2e3da-b38c-482e-b9b6-4ad75d83caa1] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 687.546292] env[61806]: ERROR nova.compute.manager [instance: eca2e3da-b38c-482e-b9b6-4ad75d83caa1] created_port_ids = self._update_ports_for_instance( [ 687.546292] env[61806]: ERROR nova.compute.manager [instance: eca2e3da-b38c-482e-b9b6-4ad75d83caa1] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 687.546292] env[61806]: ERROR nova.compute.manager [instance: eca2e3da-b38c-482e-b9b6-4ad75d83caa1] with excutils.save_and_reraise_exception(): [ 687.546292] env[61806]: ERROR nova.compute.manager [instance: eca2e3da-b38c-482e-b9b6-4ad75d83caa1] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 687.546684] env[61806]: ERROR nova.compute.manager [instance: eca2e3da-b38c-482e-b9b6-4ad75d83caa1] self.force_reraise() [ 687.546684] env[61806]: ERROR nova.compute.manager [instance: eca2e3da-b38c-482e-b9b6-4ad75d83caa1] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 687.546684] env[61806]: ERROR nova.compute.manager [instance: eca2e3da-b38c-482e-b9b6-4ad75d83caa1] raise self.value [ 687.546684] env[61806]: ERROR nova.compute.manager [instance: eca2e3da-b38c-482e-b9b6-4ad75d83caa1] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 687.546684] env[61806]: ERROR nova.compute.manager [instance: eca2e3da-b38c-482e-b9b6-4ad75d83caa1] updated_port = self._update_port( [ 687.546684] env[61806]: ERROR nova.compute.manager [instance: eca2e3da-b38c-482e-b9b6-4ad75d83caa1] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 687.546684] env[61806]: ERROR nova.compute.manager [instance: eca2e3da-b38c-482e-b9b6-4ad75d83caa1] _ensure_no_port_binding_failure(port) [ 687.546684] env[61806]: ERROR nova.compute.manager [instance: eca2e3da-b38c-482e-b9b6-4ad75d83caa1] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 687.546684] env[61806]: ERROR nova.compute.manager [instance: eca2e3da-b38c-482e-b9b6-4ad75d83caa1] raise exception.PortBindingFailed(port_id=port['id']) [ 687.546684] env[61806]: ERROR nova.compute.manager [instance: eca2e3da-b38c-482e-b9b6-4ad75d83caa1] nova.exception.PortBindingFailed: Binding failed for port 007a0ffc-be2b-4f45-b7a2-5df4203fb2da, please check neutron logs for more information. [ 687.546684] env[61806]: ERROR nova.compute.manager [instance: eca2e3da-b38c-482e-b9b6-4ad75d83caa1] [ 687.546684] env[61806]: INFO nova.compute.manager [None req-81a55da1-45c4-45f8-af56-f8f7d424c8aa tempest-ServersTestJSON-430716070 tempest-ServersTestJSON-430716070-project-member] [instance: eca2e3da-b38c-482e-b9b6-4ad75d83caa1] Terminating instance [ 687.548513] env[61806]: DEBUG oslo_concurrency.lockutils [None req-81a55da1-45c4-45f8-af56-f8f7d424c8aa tempest-ServersTestJSON-430716070 tempest-ServersTestJSON-430716070-project-member] Acquiring lock "refresh_cache-eca2e3da-b38c-482e-b9b6-4ad75d83caa1" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 687.548694] env[61806]: DEBUG oslo_concurrency.lockutils [None req-81a55da1-45c4-45f8-af56-f8f7d424c8aa tempest-ServersTestJSON-430716070 tempest-ServersTestJSON-430716070-project-member] Acquired lock "refresh_cache-eca2e3da-b38c-482e-b9b6-4ad75d83caa1" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 687.548877] env[61806]: DEBUG nova.network.neutron [None req-81a55da1-45c4-45f8-af56-f8f7d424c8aa tempest-ServersTestJSON-430716070 tempest-ServersTestJSON-430716070-project-member] [instance: eca2e3da-b38c-482e-b9b6-4ad75d83caa1] Building network info cache for instance {{(pid=61806) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 687.607523] env[61806]: DEBUG nova.network.neutron [-] [instance: b7d55e87-95db-4c45-9417-bbece4fbb98d] Updating instance_info_cache with network_info: [] {{(pid=61806) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 688.046443] env[61806]: DEBUG nova.network.neutron [None req-a93ef3a4-6e83-4dfe-9292-bac21f14d678 tempest-ServerExternalEventsTest-1847900285 tempest-ServerExternalEventsTest-1847900285-project-member] [instance: d5434dd2-206f-4a1e-b97b-bd6f691b6ddf] Instance cache missing network info. {{(pid=61806) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 688.084773] env[61806]: DEBUG nova.network.neutron [None req-81a55da1-45c4-45f8-af56-f8f7d424c8aa tempest-ServersTestJSON-430716070 tempest-ServersTestJSON-430716070-project-member] [instance: eca2e3da-b38c-482e-b9b6-4ad75d83caa1] Instance cache missing network info. {{(pid=61806) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 688.095058] env[61806]: DEBUG nova.network.neutron [None req-a93ef3a4-6e83-4dfe-9292-bac21f14d678 tempest-ServerExternalEventsTest-1847900285 tempest-ServerExternalEventsTest-1847900285-project-member] [instance: d5434dd2-206f-4a1e-b97b-bd6f691b6ddf] Updating instance_info_cache with network_info: [] {{(pid=61806) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 688.110188] env[61806]: INFO nova.compute.manager [-] [instance: b7d55e87-95db-4c45-9417-bbece4fbb98d] Took 1.02 seconds to deallocate network for instance. [ 688.112606] env[61806]: DEBUG nova.compute.claims [None req-865353ee-538a-40fc-bfba-502aba0b6a20 tempest-VolumesAdminNegativeTest-1387177911 tempest-VolumesAdminNegativeTest-1387177911-project-member] [instance: b7d55e87-95db-4c45-9417-bbece4fbb98d] Aborting claim: {{(pid=61806) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 688.112831] env[61806]: DEBUG oslo_concurrency.lockutils [None req-865353ee-538a-40fc-bfba-502aba0b6a20 tempest-VolumesAdminNegativeTest-1387177911 tempest-VolumesAdminNegativeTest-1387177911-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 688.211728] env[61806]: DEBUG nova.network.neutron [None req-81a55da1-45c4-45f8-af56-f8f7d424c8aa tempest-ServersTestJSON-430716070 tempest-ServersTestJSON-430716070-project-member] [instance: eca2e3da-b38c-482e-b9b6-4ad75d83caa1] Updating instance_info_cache with network_info: [] {{(pid=61806) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 688.599514] env[61806]: DEBUG oslo_concurrency.lockutils [None req-a93ef3a4-6e83-4dfe-9292-bac21f14d678 tempest-ServerExternalEventsTest-1847900285 tempest-ServerExternalEventsTest-1847900285-project-member] Releasing lock "refresh_cache-d5434dd2-206f-4a1e-b97b-bd6f691b6ddf" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 688.600884] env[61806]: DEBUG nova.compute.manager [None req-a93ef3a4-6e83-4dfe-9292-bac21f14d678 tempest-ServerExternalEventsTest-1847900285 tempest-ServerExternalEventsTest-1847900285-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61806) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 688.600884] env[61806]: DEBUG nova.compute.manager [None req-a93ef3a4-6e83-4dfe-9292-bac21f14d678 tempest-ServerExternalEventsTest-1847900285 tempest-ServerExternalEventsTest-1847900285-project-member] [instance: d5434dd2-206f-4a1e-b97b-bd6f691b6ddf] Deallocating network for instance {{(pid=61806) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 688.600884] env[61806]: DEBUG nova.network.neutron [None req-a93ef3a4-6e83-4dfe-9292-bac21f14d678 tempest-ServerExternalEventsTest-1847900285 tempest-ServerExternalEventsTest-1847900285-project-member] [instance: d5434dd2-206f-4a1e-b97b-bd6f691b6ddf] deallocate_for_instance() {{(pid=61806) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 688.621144] env[61806]: DEBUG nova.network.neutron [None req-a93ef3a4-6e83-4dfe-9292-bac21f14d678 tempest-ServerExternalEventsTest-1847900285 tempest-ServerExternalEventsTest-1847900285-project-member] [instance: d5434dd2-206f-4a1e-b97b-bd6f691b6ddf] Instance cache missing network info. {{(pid=61806) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 688.715990] env[61806]: DEBUG oslo_concurrency.lockutils [None req-81a55da1-45c4-45f8-af56-f8f7d424c8aa tempest-ServersTestJSON-430716070 tempest-ServersTestJSON-430716070-project-member] Releasing lock "refresh_cache-eca2e3da-b38c-482e-b9b6-4ad75d83caa1" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 688.716350] env[61806]: DEBUG nova.compute.manager [None req-81a55da1-45c4-45f8-af56-f8f7d424c8aa tempest-ServersTestJSON-430716070 tempest-ServersTestJSON-430716070-project-member] [instance: eca2e3da-b38c-482e-b9b6-4ad75d83caa1] Start destroying the instance on the hypervisor. {{(pid=61806) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 688.716546] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-81a55da1-45c4-45f8-af56-f8f7d424c8aa tempest-ServersTestJSON-430716070 tempest-ServersTestJSON-430716070-project-member] [instance: eca2e3da-b38c-482e-b9b6-4ad75d83caa1] Destroying instance {{(pid=61806) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 688.717070] env[61806]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-a2c78bfa-7f09-43d2-a7eb-fa4e2c64dc40 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 688.726627] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1b68c8ec-9aec-4c7d-b03a-bda05a884bfe {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 688.756999] env[61806]: WARNING nova.virt.vmwareapi.vmops [None req-81a55da1-45c4-45f8-af56-f8f7d424c8aa tempest-ServersTestJSON-430716070 tempest-ServersTestJSON-430716070-project-member] [instance: eca2e3da-b38c-482e-b9b6-4ad75d83caa1] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance eca2e3da-b38c-482e-b9b6-4ad75d83caa1 could not be found. [ 688.757246] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-81a55da1-45c4-45f8-af56-f8f7d424c8aa tempest-ServersTestJSON-430716070 tempest-ServersTestJSON-430716070-project-member] [instance: eca2e3da-b38c-482e-b9b6-4ad75d83caa1] Instance destroyed {{(pid=61806) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 688.757425] env[61806]: INFO nova.compute.manager [None req-81a55da1-45c4-45f8-af56-f8f7d424c8aa tempest-ServersTestJSON-430716070 tempest-ServersTestJSON-430716070-project-member] [instance: eca2e3da-b38c-482e-b9b6-4ad75d83caa1] Took 0.04 seconds to destroy the instance on the hypervisor. [ 688.757670] env[61806]: DEBUG oslo.service.loopingcall [None req-81a55da1-45c4-45f8-af56-f8f7d424c8aa tempest-ServersTestJSON-430716070 tempest-ServersTestJSON-430716070-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61806) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 688.757891] env[61806]: DEBUG nova.compute.manager [-] [instance: eca2e3da-b38c-482e-b9b6-4ad75d83caa1] Deallocating network for instance {{(pid=61806) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 688.757983] env[61806]: DEBUG nova.network.neutron [-] [instance: eca2e3da-b38c-482e-b9b6-4ad75d83caa1] deallocate_for_instance() {{(pid=61806) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 688.785170] env[61806]: DEBUG nova.network.neutron [-] [instance: eca2e3da-b38c-482e-b9b6-4ad75d83caa1] Instance cache missing network info. {{(pid=61806) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 688.800975] env[61806]: DEBUG nova.compute.manager [req-e14255a6-de4a-4d1e-8caf-05232e03aee8 req-498d69b4-cccb-4353-9ff4-73af3abcffe2 service nova] [instance: eca2e3da-b38c-482e-b9b6-4ad75d83caa1] Received event network-changed-007a0ffc-be2b-4f45-b7a2-5df4203fb2da {{(pid=61806) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 688.801324] env[61806]: DEBUG nova.compute.manager [req-e14255a6-de4a-4d1e-8caf-05232e03aee8 req-498d69b4-cccb-4353-9ff4-73af3abcffe2 service nova] [instance: eca2e3da-b38c-482e-b9b6-4ad75d83caa1] Refreshing instance network info cache due to event network-changed-007a0ffc-be2b-4f45-b7a2-5df4203fb2da. {{(pid=61806) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 688.801513] env[61806]: DEBUG oslo_concurrency.lockutils [req-e14255a6-de4a-4d1e-8caf-05232e03aee8 req-498d69b4-cccb-4353-9ff4-73af3abcffe2 service nova] Acquiring lock "refresh_cache-eca2e3da-b38c-482e-b9b6-4ad75d83caa1" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 688.801657] env[61806]: DEBUG oslo_concurrency.lockutils [req-e14255a6-de4a-4d1e-8caf-05232e03aee8 req-498d69b4-cccb-4353-9ff4-73af3abcffe2 service nova] Acquired lock "refresh_cache-eca2e3da-b38c-482e-b9b6-4ad75d83caa1" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 688.801836] env[61806]: DEBUG nova.network.neutron [req-e14255a6-de4a-4d1e-8caf-05232e03aee8 req-498d69b4-cccb-4353-9ff4-73af3abcffe2 service nova] [instance: eca2e3da-b38c-482e-b9b6-4ad75d83caa1] Refreshing network info cache for port 007a0ffc-be2b-4f45-b7a2-5df4203fb2da {{(pid=61806) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 688.976315] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fd271b04-885b-4ab0-963b-cf58c744b7d7 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 688.984526] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aad3df59-eb26-4a8a-bf8b-dc35b6859aa1 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 689.016161] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-59991213-be78-4842-ad8a-e6b98d79730c {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 689.024279] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7ab863fb-9f8d-4563-907d-4bd25eaa6154 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 689.037924] env[61806]: DEBUG nova.compute.provider_tree [None req-2bf81f39-4b8d-4851-a793-760af008ffe2 tempest-ImagesOneServerTestJSON-1003762139 tempest-ImagesOneServerTestJSON-1003762139-project-member] Updating inventory in ProviderTree for provider a2858be1-fd22-4e08-979e-87ad25293407 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 139, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61806) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 689.128838] env[61806]: DEBUG nova.network.neutron [None req-a93ef3a4-6e83-4dfe-9292-bac21f14d678 tempest-ServerExternalEventsTest-1847900285 tempest-ServerExternalEventsTest-1847900285-project-member] [instance: d5434dd2-206f-4a1e-b97b-bd6f691b6ddf] Updating instance_info_cache with network_info: [] {{(pid=61806) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 689.290249] env[61806]: DEBUG nova.network.neutron [-] [instance: eca2e3da-b38c-482e-b9b6-4ad75d83caa1] Updating instance_info_cache with network_info: [] {{(pid=61806) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 689.320923] env[61806]: DEBUG nova.network.neutron [req-e14255a6-de4a-4d1e-8caf-05232e03aee8 req-498d69b4-cccb-4353-9ff4-73af3abcffe2 service nova] [instance: eca2e3da-b38c-482e-b9b6-4ad75d83caa1] Instance cache missing network info. {{(pid=61806) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 689.423754] env[61806]: DEBUG nova.network.neutron [req-e14255a6-de4a-4d1e-8caf-05232e03aee8 req-498d69b4-cccb-4353-9ff4-73af3abcffe2 service nova] [instance: eca2e3da-b38c-482e-b9b6-4ad75d83caa1] Updating instance_info_cache with network_info: [] {{(pid=61806) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 689.562356] env[61806]: ERROR nova.scheduler.client.report [None req-2bf81f39-4b8d-4851-a793-760af008ffe2 tempest-ImagesOneServerTestJSON-1003762139 tempest-ImagesOneServerTestJSON-1003762139-project-member] [req-bcc01685-0be1-4159-ac05-8cc953066034] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 139, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID a2858be1-fd22-4e08-979e-87ad25293407. Got 409: {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-bcc01685-0be1-4159-ac05-8cc953066034"}]} [ 689.579335] env[61806]: DEBUG nova.scheduler.client.report [None req-2bf81f39-4b8d-4851-a793-760af008ffe2 tempest-ImagesOneServerTestJSON-1003762139 tempest-ImagesOneServerTestJSON-1003762139-project-member] Refreshing inventories for resource provider a2858be1-fd22-4e08-979e-87ad25293407 {{(pid=61806) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:818}} [ 689.595568] env[61806]: DEBUG nova.scheduler.client.report [None req-2bf81f39-4b8d-4851-a793-760af008ffe2 tempest-ImagesOneServerTestJSON-1003762139 tempest-ImagesOneServerTestJSON-1003762139-project-member] Updating ProviderTree inventory for provider a2858be1-fd22-4e08-979e-87ad25293407 from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 138, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61806) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:782}} [ 689.595858] env[61806]: DEBUG nova.compute.provider_tree [None req-2bf81f39-4b8d-4851-a793-760af008ffe2 tempest-ImagesOneServerTestJSON-1003762139 tempest-ImagesOneServerTestJSON-1003762139-project-member] Updating inventory in ProviderTree for provider a2858be1-fd22-4e08-979e-87ad25293407 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 138, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61806) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 689.609167] env[61806]: DEBUG nova.scheduler.client.report [None req-2bf81f39-4b8d-4851-a793-760af008ffe2 tempest-ImagesOneServerTestJSON-1003762139 tempest-ImagesOneServerTestJSON-1003762139-project-member] Refreshing aggregate associations for resource provider a2858be1-fd22-4e08-979e-87ad25293407, aggregates: None {{(pid=61806) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:827}} [ 689.631145] env[61806]: DEBUG nova.scheduler.client.report [None req-2bf81f39-4b8d-4851-a793-760af008ffe2 tempest-ImagesOneServerTestJSON-1003762139 tempest-ImagesOneServerTestJSON-1003762139-project-member] Refreshing trait associations for resource provider a2858be1-fd22-4e08-979e-87ad25293407, traits: COMPUTE_IMAGE_TYPE_ISO,COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_SAME_HOST_COLD_MIGRATE,COMPUTE_NET_ATTACH_INTERFACE,COMPUTE_NODE {{(pid=61806) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:839}} [ 689.633699] env[61806]: INFO nova.compute.manager [None req-a93ef3a4-6e83-4dfe-9292-bac21f14d678 tempest-ServerExternalEventsTest-1847900285 tempest-ServerExternalEventsTest-1847900285-project-member] [instance: d5434dd2-206f-4a1e-b97b-bd6f691b6ddf] Took 1.03 seconds to deallocate network for instance. [ 689.798418] env[61806]: INFO nova.compute.manager [-] [instance: eca2e3da-b38c-482e-b9b6-4ad75d83caa1] Took 1.04 seconds to deallocate network for instance. [ 689.799571] env[61806]: DEBUG nova.compute.claims [None req-81a55da1-45c4-45f8-af56-f8f7d424c8aa tempest-ServersTestJSON-430716070 tempest-ServersTestJSON-430716070-project-member] [instance: eca2e3da-b38c-482e-b9b6-4ad75d83caa1] Aborting claim: {{(pid=61806) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 689.799738] env[61806]: DEBUG oslo_concurrency.lockutils [None req-81a55da1-45c4-45f8-af56-f8f7d424c8aa tempest-ServersTestJSON-430716070 tempest-ServersTestJSON-430716070-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 689.926196] env[61806]: DEBUG oslo_concurrency.lockutils [req-e14255a6-de4a-4d1e-8caf-05232e03aee8 req-498d69b4-cccb-4353-9ff4-73af3abcffe2 service nova] Releasing lock "refresh_cache-eca2e3da-b38c-482e-b9b6-4ad75d83caa1" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 689.926460] env[61806]: DEBUG nova.compute.manager [req-e14255a6-de4a-4d1e-8caf-05232e03aee8 req-498d69b4-cccb-4353-9ff4-73af3abcffe2 service nova] [instance: eca2e3da-b38c-482e-b9b6-4ad75d83caa1] Received event network-vif-deleted-007a0ffc-be2b-4f45-b7a2-5df4203fb2da {{(pid=61806) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 690.107615] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f2718245-a5cc-4872-9e72-5e8b7629fffa {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 690.115394] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-49c58045-b260-4817-9d39-74f65a691d3a {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 690.152371] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d3883c2b-b9a3-4cc8-8c6b-01fee9033a57 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 690.160062] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-454739cd-404d-4141-ba08-4c66c9252170 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 690.177141] env[61806]: DEBUG nova.compute.provider_tree [None req-2bf81f39-4b8d-4851-a793-760af008ffe2 tempest-ImagesOneServerTestJSON-1003762139 tempest-ImagesOneServerTestJSON-1003762139-project-member] Updating inventory in ProviderTree for provider a2858be1-fd22-4e08-979e-87ad25293407 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 139, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61806) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 690.685776] env[61806]: INFO nova.scheduler.client.report [None req-a93ef3a4-6e83-4dfe-9292-bac21f14d678 tempest-ServerExternalEventsTest-1847900285 tempest-ServerExternalEventsTest-1847900285-project-member] Deleted allocations for instance d5434dd2-206f-4a1e-b97b-bd6f691b6ddf [ 690.714149] env[61806]: DEBUG nova.scheduler.client.report [None req-2bf81f39-4b8d-4851-a793-760af008ffe2 tempest-ImagesOneServerTestJSON-1003762139 tempest-ImagesOneServerTestJSON-1003762139-project-member] Updated inventory for provider a2858be1-fd22-4e08-979e-87ad25293407 with generation 67 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 139, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61806) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:971}} [ 690.714469] env[61806]: DEBUG nova.compute.provider_tree [None req-2bf81f39-4b8d-4851-a793-760af008ffe2 tempest-ImagesOneServerTestJSON-1003762139 tempest-ImagesOneServerTestJSON-1003762139-project-member] Updating resource provider a2858be1-fd22-4e08-979e-87ad25293407 generation from 67 to 68 during operation: update_inventory {{(pid=61806) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 690.714689] env[61806]: DEBUG nova.compute.provider_tree [None req-2bf81f39-4b8d-4851-a793-760af008ffe2 tempest-ImagesOneServerTestJSON-1003762139 tempest-ImagesOneServerTestJSON-1003762139-project-member] Updating inventory in ProviderTree for provider a2858be1-fd22-4e08-979e-87ad25293407 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 139, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61806) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 691.198781] env[61806]: DEBUG oslo_concurrency.lockutils [None req-a93ef3a4-6e83-4dfe-9292-bac21f14d678 tempest-ServerExternalEventsTest-1847900285 tempest-ServerExternalEventsTest-1847900285-project-member] Lock "d5434dd2-206f-4a1e-b97b-bd6f691b6ddf" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 129.004s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 691.221187] env[61806]: DEBUG oslo_concurrency.lockutils [None req-2bf81f39-4b8d-4851-a793-760af008ffe2 tempest-ImagesOneServerTestJSON-1003762139 tempest-ImagesOneServerTestJSON-1003762139-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 3.720s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 691.221561] env[61806]: DEBUG nova.compute.manager [None req-2bf81f39-4b8d-4851-a793-760af008ffe2 tempest-ImagesOneServerTestJSON-1003762139 tempest-ImagesOneServerTestJSON-1003762139-project-member] [instance: 294f0a37-7589-405d-a324-13099c80f458] Start building networks asynchronously for instance. {{(pid=61806) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 691.225830] env[61806]: DEBUG oslo_concurrency.lockutils [None req-fe158576-4234-4a6b-b48d-5ca8bb977c40 tempest-ServerShowV254Test-608938657 tempest-ServerShowV254Test-608938657-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: waited 29.486s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 691.225830] env[61806]: DEBUG nova.objects.instance [None req-fe158576-4234-4a6b-b48d-5ca8bb977c40 tempest-ServerShowV254Test-608938657 tempest-ServerShowV254Test-608938657-project-member] [instance: 8016bafc-b287-4df1-bd4b-a00210177a86] Trying to apply a migration context that does not seem to be set for this instance {{(pid=61806) apply_migration_context /opt/stack/nova/nova/objects/instance.py:1067}} [ 691.702463] env[61806]: DEBUG nova.compute.manager [None req-3c3981d2-fa25-4f4b-87d4-a5411e67199e tempest-MultipleCreateTestJSON-1589015462 tempest-MultipleCreateTestJSON-1589015462-project-member] [instance: 7211d568-cc68-4e50-80ad-b78878d1deab] Starting instance... {{(pid=61806) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 691.726710] env[61806]: DEBUG nova.compute.utils [None req-2bf81f39-4b8d-4851-a793-760af008ffe2 tempest-ImagesOneServerTestJSON-1003762139 tempest-ImagesOneServerTestJSON-1003762139-project-member] Using /dev/sd instead of None {{(pid=61806) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 691.729351] env[61806]: DEBUG nova.compute.manager [None req-2bf81f39-4b8d-4851-a793-760af008ffe2 tempest-ImagesOneServerTestJSON-1003762139 tempest-ImagesOneServerTestJSON-1003762139-project-member] [instance: 294f0a37-7589-405d-a324-13099c80f458] Allocating IP information in the background. {{(pid=61806) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 691.729351] env[61806]: DEBUG nova.network.neutron [None req-2bf81f39-4b8d-4851-a793-760af008ffe2 tempest-ImagesOneServerTestJSON-1003762139 tempest-ImagesOneServerTestJSON-1003762139-project-member] [instance: 294f0a37-7589-405d-a324-13099c80f458] allocate_for_instance() {{(pid=61806) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 691.774880] env[61806]: DEBUG nova.policy [None req-2bf81f39-4b8d-4851-a793-760af008ffe2 tempest-ImagesOneServerTestJSON-1003762139 tempest-ImagesOneServerTestJSON-1003762139-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '0652be6fca6c48c98b60a855e1d75bab', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '005973038085400890c87f440c40f130', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61806) authorize /opt/stack/nova/nova/policy.py:201}} [ 692.096746] env[61806]: DEBUG nova.network.neutron [None req-2bf81f39-4b8d-4851-a793-760af008ffe2 tempest-ImagesOneServerTestJSON-1003762139 tempest-ImagesOneServerTestJSON-1003762139-project-member] [instance: 294f0a37-7589-405d-a324-13099c80f458] Successfully created port: b9eb0672-5644-46ba-9b38-f6a90aef16ee {{(pid=61806) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 692.232228] env[61806]: DEBUG oslo_concurrency.lockutils [None req-3c3981d2-fa25-4f4b-87d4-a5411e67199e tempest-MultipleCreateTestJSON-1589015462 tempest-MultipleCreateTestJSON-1589015462-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 692.233935] env[61806]: DEBUG nova.compute.manager [None req-2bf81f39-4b8d-4851-a793-760af008ffe2 tempest-ImagesOneServerTestJSON-1003762139 tempest-ImagesOneServerTestJSON-1003762139-project-member] [instance: 294f0a37-7589-405d-a324-13099c80f458] Start building block device mappings for instance. {{(pid=61806) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 692.238700] env[61806]: DEBUG oslo_concurrency.lockutils [None req-fe158576-4234-4a6b-b48d-5ca8bb977c40 tempest-ServerShowV254Test-608938657 tempest-ServerShowV254Test-608938657-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: held 1.014s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 692.239783] env[61806]: DEBUG oslo_concurrency.lockutils [None req-0f0a69c6-2310-4272-abc1-520940b4bcf0 tempest-ServersTestFqdnHostnames-356624376 tempest-ServersTestFqdnHostnames-356624376-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 30.449s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 692.243019] env[61806]: INFO nova.compute.claims [None req-0f0a69c6-2310-4272-abc1-520940b4bcf0 tempest-ServersTestFqdnHostnames-356624376 tempest-ServersTestFqdnHostnames-356624376-project-member] [instance: 46c34ed0-9bac-4803-8d5c-3a3346a64117] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 693.178144] env[61806]: DEBUG nova.compute.manager [req-fd0dd2dd-b65d-46c7-a54a-93e1e23414a8 req-4c2a51ee-3e1e-402b-af52-73b621db82a2 service nova] [instance: 294f0a37-7589-405d-a324-13099c80f458] Received event network-changed-b9eb0672-5644-46ba-9b38-f6a90aef16ee {{(pid=61806) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 693.178375] env[61806]: DEBUG nova.compute.manager [req-fd0dd2dd-b65d-46c7-a54a-93e1e23414a8 req-4c2a51ee-3e1e-402b-af52-73b621db82a2 service nova] [instance: 294f0a37-7589-405d-a324-13099c80f458] Refreshing instance network info cache due to event network-changed-b9eb0672-5644-46ba-9b38-f6a90aef16ee. {{(pid=61806) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 693.178598] env[61806]: DEBUG oslo_concurrency.lockutils [req-fd0dd2dd-b65d-46c7-a54a-93e1e23414a8 req-4c2a51ee-3e1e-402b-af52-73b621db82a2 service nova] Acquiring lock "refresh_cache-294f0a37-7589-405d-a324-13099c80f458" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 693.178723] env[61806]: DEBUG oslo_concurrency.lockutils [req-fd0dd2dd-b65d-46c7-a54a-93e1e23414a8 req-4c2a51ee-3e1e-402b-af52-73b621db82a2 service nova] Acquired lock "refresh_cache-294f0a37-7589-405d-a324-13099c80f458" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 693.178909] env[61806]: DEBUG nova.network.neutron [req-fd0dd2dd-b65d-46c7-a54a-93e1e23414a8 req-4c2a51ee-3e1e-402b-af52-73b621db82a2 service nova] [instance: 294f0a37-7589-405d-a324-13099c80f458] Refreshing network info cache for port b9eb0672-5644-46ba-9b38-f6a90aef16ee {{(pid=61806) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 693.248248] env[61806]: DEBUG nova.compute.manager [None req-2bf81f39-4b8d-4851-a793-760af008ffe2 tempest-ImagesOneServerTestJSON-1003762139 tempest-ImagesOneServerTestJSON-1003762139-project-member] [instance: 294f0a37-7589-405d-a324-13099c80f458] Start spawning the instance on the hypervisor. {{(pid=61806) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 693.282454] env[61806]: DEBUG nova.virt.hardware [None req-2bf81f39-4b8d-4851-a793-760af008ffe2 tempest-ImagesOneServerTestJSON-1003762139 tempest-ImagesOneServerTestJSON-1003762139-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-15T18:05:10Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-15T18:04:52Z,direct_url=,disk_format='vmdk',id=ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='84ba83fce2c349a6988173c5d6fc3b07',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-15T18:04:53Z,virtual_size=,visibility=), allow threads: False {{(pid=61806) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 693.282883] env[61806]: DEBUG nova.virt.hardware [None req-2bf81f39-4b8d-4851-a793-760af008ffe2 tempest-ImagesOneServerTestJSON-1003762139 tempest-ImagesOneServerTestJSON-1003762139-project-member] Flavor limits 0:0:0 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 693.283098] env[61806]: DEBUG nova.virt.hardware [None req-2bf81f39-4b8d-4851-a793-760af008ffe2 tempest-ImagesOneServerTestJSON-1003762139 tempest-ImagesOneServerTestJSON-1003762139-project-member] Image limits 0:0:0 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 693.283295] env[61806]: DEBUG nova.virt.hardware [None req-2bf81f39-4b8d-4851-a793-760af008ffe2 tempest-ImagesOneServerTestJSON-1003762139 tempest-ImagesOneServerTestJSON-1003762139-project-member] Flavor pref 0:0:0 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 693.283437] env[61806]: DEBUG nova.virt.hardware [None req-2bf81f39-4b8d-4851-a793-760af008ffe2 tempest-ImagesOneServerTestJSON-1003762139 tempest-ImagesOneServerTestJSON-1003762139-project-member] Image pref 0:0:0 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 693.283580] env[61806]: DEBUG nova.virt.hardware [None req-2bf81f39-4b8d-4851-a793-760af008ffe2 tempest-ImagesOneServerTestJSON-1003762139 tempest-ImagesOneServerTestJSON-1003762139-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 693.283860] env[61806]: DEBUG nova.virt.hardware [None req-2bf81f39-4b8d-4851-a793-760af008ffe2 tempest-ImagesOneServerTestJSON-1003762139 tempest-ImagesOneServerTestJSON-1003762139-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61806) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 693.283991] env[61806]: DEBUG nova.virt.hardware [None req-2bf81f39-4b8d-4851-a793-760af008ffe2 tempest-ImagesOneServerTestJSON-1003762139 tempest-ImagesOneServerTestJSON-1003762139-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61806) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 693.284229] env[61806]: DEBUG nova.virt.hardware [None req-2bf81f39-4b8d-4851-a793-760af008ffe2 tempest-ImagesOneServerTestJSON-1003762139 tempest-ImagesOneServerTestJSON-1003762139-project-member] Got 1 possible topologies {{(pid=61806) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 693.284433] env[61806]: DEBUG nova.virt.hardware [None req-2bf81f39-4b8d-4851-a793-760af008ffe2 tempest-ImagesOneServerTestJSON-1003762139 tempest-ImagesOneServerTestJSON-1003762139-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61806) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 693.284632] env[61806]: DEBUG nova.virt.hardware [None req-2bf81f39-4b8d-4851-a793-760af008ffe2 tempest-ImagesOneServerTestJSON-1003762139 tempest-ImagesOneServerTestJSON-1003762139-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61806) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 693.285787] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-451e6534-b54b-40e5-b04b-801dfe17ef75 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 693.299233] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f85778f5-0fc1-4974-9f31-cb2c4fa268ce {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 693.348967] env[61806]: ERROR nova.compute.manager [None req-2bf81f39-4b8d-4851-a793-760af008ffe2 tempest-ImagesOneServerTestJSON-1003762139 tempest-ImagesOneServerTestJSON-1003762139-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port b9eb0672-5644-46ba-9b38-f6a90aef16ee, please check neutron logs for more information. [ 693.348967] env[61806]: ERROR nova.compute.manager Traceback (most recent call last): [ 693.348967] env[61806]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 693.348967] env[61806]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 693.348967] env[61806]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 693.348967] env[61806]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 693.348967] env[61806]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 693.348967] env[61806]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 693.348967] env[61806]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 693.348967] env[61806]: ERROR nova.compute.manager self.force_reraise() [ 693.348967] env[61806]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 693.348967] env[61806]: ERROR nova.compute.manager raise self.value [ 693.348967] env[61806]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 693.348967] env[61806]: ERROR nova.compute.manager updated_port = self._update_port( [ 693.348967] env[61806]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 693.348967] env[61806]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 693.349525] env[61806]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 693.349525] env[61806]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 693.349525] env[61806]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port b9eb0672-5644-46ba-9b38-f6a90aef16ee, please check neutron logs for more information. [ 693.349525] env[61806]: ERROR nova.compute.manager [ 693.349525] env[61806]: Traceback (most recent call last): [ 693.349525] env[61806]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 693.349525] env[61806]: listener.cb(fileno) [ 693.349525] env[61806]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 693.349525] env[61806]: result = function(*args, **kwargs) [ 693.349525] env[61806]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 693.349525] env[61806]: return func(*args, **kwargs) [ 693.349525] env[61806]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 693.349525] env[61806]: raise e [ 693.349525] env[61806]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 693.349525] env[61806]: nwinfo = self.network_api.allocate_for_instance( [ 693.349525] env[61806]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 693.349525] env[61806]: created_port_ids = self._update_ports_for_instance( [ 693.349525] env[61806]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 693.349525] env[61806]: with excutils.save_and_reraise_exception(): [ 693.349525] env[61806]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 693.349525] env[61806]: self.force_reraise() [ 693.349525] env[61806]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 693.349525] env[61806]: raise self.value [ 693.349525] env[61806]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 693.349525] env[61806]: updated_port = self._update_port( [ 693.349525] env[61806]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 693.349525] env[61806]: _ensure_no_port_binding_failure(port) [ 693.349525] env[61806]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 693.349525] env[61806]: raise exception.PortBindingFailed(port_id=port['id']) [ 693.350341] env[61806]: nova.exception.PortBindingFailed: Binding failed for port b9eb0672-5644-46ba-9b38-f6a90aef16ee, please check neutron logs for more information. [ 693.350341] env[61806]: Removing descriptor: 16 [ 693.350341] env[61806]: ERROR nova.compute.manager [None req-2bf81f39-4b8d-4851-a793-760af008ffe2 tempest-ImagesOneServerTestJSON-1003762139 tempest-ImagesOneServerTestJSON-1003762139-project-member] [instance: 294f0a37-7589-405d-a324-13099c80f458] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port b9eb0672-5644-46ba-9b38-f6a90aef16ee, please check neutron logs for more information. [ 693.350341] env[61806]: ERROR nova.compute.manager [instance: 294f0a37-7589-405d-a324-13099c80f458] Traceback (most recent call last): [ 693.350341] env[61806]: ERROR nova.compute.manager [instance: 294f0a37-7589-405d-a324-13099c80f458] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 693.350341] env[61806]: ERROR nova.compute.manager [instance: 294f0a37-7589-405d-a324-13099c80f458] yield resources [ 693.350341] env[61806]: ERROR nova.compute.manager [instance: 294f0a37-7589-405d-a324-13099c80f458] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 693.350341] env[61806]: ERROR nova.compute.manager [instance: 294f0a37-7589-405d-a324-13099c80f458] self.driver.spawn(context, instance, image_meta, [ 693.350341] env[61806]: ERROR nova.compute.manager [instance: 294f0a37-7589-405d-a324-13099c80f458] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 693.350341] env[61806]: ERROR nova.compute.manager [instance: 294f0a37-7589-405d-a324-13099c80f458] self._vmops.spawn(context, instance, image_meta, injected_files, [ 693.350341] env[61806]: ERROR nova.compute.manager [instance: 294f0a37-7589-405d-a324-13099c80f458] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 693.350341] env[61806]: ERROR nova.compute.manager [instance: 294f0a37-7589-405d-a324-13099c80f458] vm_ref = self.build_virtual_machine(instance, [ 693.350646] env[61806]: ERROR nova.compute.manager [instance: 294f0a37-7589-405d-a324-13099c80f458] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 693.350646] env[61806]: ERROR nova.compute.manager [instance: 294f0a37-7589-405d-a324-13099c80f458] vif_infos = vmwarevif.get_vif_info(self._session, [ 693.350646] env[61806]: ERROR nova.compute.manager [instance: 294f0a37-7589-405d-a324-13099c80f458] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 693.350646] env[61806]: ERROR nova.compute.manager [instance: 294f0a37-7589-405d-a324-13099c80f458] for vif in network_info: [ 693.350646] env[61806]: ERROR nova.compute.manager [instance: 294f0a37-7589-405d-a324-13099c80f458] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 693.350646] env[61806]: ERROR nova.compute.manager [instance: 294f0a37-7589-405d-a324-13099c80f458] return self._sync_wrapper(fn, *args, **kwargs) [ 693.350646] env[61806]: ERROR nova.compute.manager [instance: 294f0a37-7589-405d-a324-13099c80f458] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 693.350646] env[61806]: ERROR nova.compute.manager [instance: 294f0a37-7589-405d-a324-13099c80f458] self.wait() [ 693.350646] env[61806]: ERROR nova.compute.manager [instance: 294f0a37-7589-405d-a324-13099c80f458] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 693.350646] env[61806]: ERROR nova.compute.manager [instance: 294f0a37-7589-405d-a324-13099c80f458] self[:] = self._gt.wait() [ 693.350646] env[61806]: ERROR nova.compute.manager [instance: 294f0a37-7589-405d-a324-13099c80f458] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 693.350646] env[61806]: ERROR nova.compute.manager [instance: 294f0a37-7589-405d-a324-13099c80f458] return self._exit_event.wait() [ 693.350646] env[61806]: ERROR nova.compute.manager [instance: 294f0a37-7589-405d-a324-13099c80f458] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 693.351022] env[61806]: ERROR nova.compute.manager [instance: 294f0a37-7589-405d-a324-13099c80f458] result = hub.switch() [ 693.351022] env[61806]: ERROR nova.compute.manager [instance: 294f0a37-7589-405d-a324-13099c80f458] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 693.351022] env[61806]: ERROR nova.compute.manager [instance: 294f0a37-7589-405d-a324-13099c80f458] return self.greenlet.switch() [ 693.351022] env[61806]: ERROR nova.compute.manager [instance: 294f0a37-7589-405d-a324-13099c80f458] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 693.351022] env[61806]: ERROR nova.compute.manager [instance: 294f0a37-7589-405d-a324-13099c80f458] result = function(*args, **kwargs) [ 693.351022] env[61806]: ERROR nova.compute.manager [instance: 294f0a37-7589-405d-a324-13099c80f458] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 693.351022] env[61806]: ERROR nova.compute.manager [instance: 294f0a37-7589-405d-a324-13099c80f458] return func(*args, **kwargs) [ 693.351022] env[61806]: ERROR nova.compute.manager [instance: 294f0a37-7589-405d-a324-13099c80f458] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 693.351022] env[61806]: ERROR nova.compute.manager [instance: 294f0a37-7589-405d-a324-13099c80f458] raise e [ 693.351022] env[61806]: ERROR nova.compute.manager [instance: 294f0a37-7589-405d-a324-13099c80f458] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 693.351022] env[61806]: ERROR nova.compute.manager [instance: 294f0a37-7589-405d-a324-13099c80f458] nwinfo = self.network_api.allocate_for_instance( [ 693.351022] env[61806]: ERROR nova.compute.manager [instance: 294f0a37-7589-405d-a324-13099c80f458] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 693.351022] env[61806]: ERROR nova.compute.manager [instance: 294f0a37-7589-405d-a324-13099c80f458] created_port_ids = self._update_ports_for_instance( [ 693.351365] env[61806]: ERROR nova.compute.manager [instance: 294f0a37-7589-405d-a324-13099c80f458] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 693.351365] env[61806]: ERROR nova.compute.manager [instance: 294f0a37-7589-405d-a324-13099c80f458] with excutils.save_and_reraise_exception(): [ 693.351365] env[61806]: ERROR nova.compute.manager [instance: 294f0a37-7589-405d-a324-13099c80f458] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 693.351365] env[61806]: ERROR nova.compute.manager [instance: 294f0a37-7589-405d-a324-13099c80f458] self.force_reraise() [ 693.351365] env[61806]: ERROR nova.compute.manager [instance: 294f0a37-7589-405d-a324-13099c80f458] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 693.351365] env[61806]: ERROR nova.compute.manager [instance: 294f0a37-7589-405d-a324-13099c80f458] raise self.value [ 693.351365] env[61806]: ERROR nova.compute.manager [instance: 294f0a37-7589-405d-a324-13099c80f458] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 693.351365] env[61806]: ERROR nova.compute.manager [instance: 294f0a37-7589-405d-a324-13099c80f458] updated_port = self._update_port( [ 693.351365] env[61806]: ERROR nova.compute.manager [instance: 294f0a37-7589-405d-a324-13099c80f458] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 693.351365] env[61806]: ERROR nova.compute.manager [instance: 294f0a37-7589-405d-a324-13099c80f458] _ensure_no_port_binding_failure(port) [ 693.351365] env[61806]: ERROR nova.compute.manager [instance: 294f0a37-7589-405d-a324-13099c80f458] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 693.351365] env[61806]: ERROR nova.compute.manager [instance: 294f0a37-7589-405d-a324-13099c80f458] raise exception.PortBindingFailed(port_id=port['id']) [ 693.351694] env[61806]: ERROR nova.compute.manager [instance: 294f0a37-7589-405d-a324-13099c80f458] nova.exception.PortBindingFailed: Binding failed for port b9eb0672-5644-46ba-9b38-f6a90aef16ee, please check neutron logs for more information. [ 693.351694] env[61806]: ERROR nova.compute.manager [instance: 294f0a37-7589-405d-a324-13099c80f458] [ 693.351694] env[61806]: INFO nova.compute.manager [None req-2bf81f39-4b8d-4851-a793-760af008ffe2 tempest-ImagesOneServerTestJSON-1003762139 tempest-ImagesOneServerTestJSON-1003762139-project-member] [instance: 294f0a37-7589-405d-a324-13099c80f458] Terminating instance [ 693.354885] env[61806]: DEBUG oslo_concurrency.lockutils [None req-2bf81f39-4b8d-4851-a793-760af008ffe2 tempest-ImagesOneServerTestJSON-1003762139 tempest-ImagesOneServerTestJSON-1003762139-project-member] Acquiring lock "refresh_cache-294f0a37-7589-405d-a324-13099c80f458" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 693.685901] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-edc1faa9-59e3-4778-9919-adaf48a89581 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 693.695073] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b5c3b8dc-9d85-4e97-a905-1fe2363ec583 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 693.728493] env[61806]: DEBUG nova.network.neutron [req-fd0dd2dd-b65d-46c7-a54a-93e1e23414a8 req-4c2a51ee-3e1e-402b-af52-73b621db82a2 service nova] [instance: 294f0a37-7589-405d-a324-13099c80f458] Instance cache missing network info. {{(pid=61806) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 693.731056] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f9c4898e-2c59-482f-96b0-0fc6649fdcbf {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 693.738456] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1dda887f-065b-42e4-ba67-2e0aa6080fb2 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 693.751819] env[61806]: DEBUG nova.compute.provider_tree [None req-0f0a69c6-2310-4272-abc1-520940b4bcf0 tempest-ServersTestFqdnHostnames-356624376 tempest-ServersTestFqdnHostnames-356624376-project-member] Inventory has not changed in ProviderTree for provider: a2858be1-fd22-4e08-979e-87ad25293407 {{(pid=61806) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 693.840364] env[61806]: DEBUG nova.network.neutron [req-fd0dd2dd-b65d-46c7-a54a-93e1e23414a8 req-4c2a51ee-3e1e-402b-af52-73b621db82a2 service nova] [instance: 294f0a37-7589-405d-a324-13099c80f458] Updating instance_info_cache with network_info: [] {{(pid=61806) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 694.254722] env[61806]: DEBUG nova.scheduler.client.report [None req-0f0a69c6-2310-4272-abc1-520940b4bcf0 tempest-ServersTestFqdnHostnames-356624376 tempest-ServersTestFqdnHostnames-356624376-project-member] Inventory has not changed for provider a2858be1-fd22-4e08-979e-87ad25293407 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 139, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61806) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 694.343474] env[61806]: DEBUG oslo_concurrency.lockutils [req-fd0dd2dd-b65d-46c7-a54a-93e1e23414a8 req-4c2a51ee-3e1e-402b-af52-73b621db82a2 service nova] Releasing lock "refresh_cache-294f0a37-7589-405d-a324-13099c80f458" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 694.344770] env[61806]: DEBUG oslo_concurrency.lockutils [None req-2bf81f39-4b8d-4851-a793-760af008ffe2 tempest-ImagesOneServerTestJSON-1003762139 tempest-ImagesOneServerTestJSON-1003762139-project-member] Acquired lock "refresh_cache-294f0a37-7589-405d-a324-13099c80f458" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 694.344770] env[61806]: DEBUG nova.network.neutron [None req-2bf81f39-4b8d-4851-a793-760af008ffe2 tempest-ImagesOneServerTestJSON-1003762139 tempest-ImagesOneServerTestJSON-1003762139-project-member] [instance: 294f0a37-7589-405d-a324-13099c80f458] Building network info cache for instance {{(pid=61806) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 694.760045] env[61806]: DEBUG oslo_concurrency.lockutils [None req-0f0a69c6-2310-4272-abc1-520940b4bcf0 tempest-ServersTestFqdnHostnames-356624376 tempest-ServersTestFqdnHostnames-356624376-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.520s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 694.760573] env[61806]: DEBUG nova.compute.manager [None req-0f0a69c6-2310-4272-abc1-520940b4bcf0 tempest-ServersTestFqdnHostnames-356624376 tempest-ServersTestFqdnHostnames-356624376-project-member] [instance: 46c34ed0-9bac-4803-8d5c-3a3346a64117] Start building networks asynchronously for instance. {{(pid=61806) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 694.763605] env[61806]: DEBUG oslo_concurrency.lockutils [None req-05ef4399-8310-428c-ba55-af11abf4e445 tempest-SecurityGroupsTestJSON-1213485581 tempest-SecurityGroupsTestJSON-1213485581-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 32.553s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 694.764952] env[61806]: INFO nova.compute.claims [None req-05ef4399-8310-428c-ba55-af11abf4e445 tempest-SecurityGroupsTestJSON-1213485581 tempest-SecurityGroupsTestJSON-1213485581-project-member] [instance: 294adb58-f23e-4510-a25a-de6b909e3189] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 694.863906] env[61806]: DEBUG nova.network.neutron [None req-2bf81f39-4b8d-4851-a793-760af008ffe2 tempest-ImagesOneServerTestJSON-1003762139 tempest-ImagesOneServerTestJSON-1003762139-project-member] [instance: 294f0a37-7589-405d-a324-13099c80f458] Instance cache missing network info. {{(pid=61806) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 694.952413] env[61806]: DEBUG nova.network.neutron [None req-2bf81f39-4b8d-4851-a793-760af008ffe2 tempest-ImagesOneServerTestJSON-1003762139 tempest-ImagesOneServerTestJSON-1003762139-project-member] [instance: 294f0a37-7589-405d-a324-13099c80f458] Updating instance_info_cache with network_info: [] {{(pid=61806) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 695.227899] env[61806]: DEBUG nova.compute.manager [req-a833afb2-28f2-4ad9-a971-5bb343f5cb67 req-5d103815-247b-4faa-9b7f-2fe425b94418 service nova] [instance: 294f0a37-7589-405d-a324-13099c80f458] Received event network-vif-deleted-b9eb0672-5644-46ba-9b38-f6a90aef16ee {{(pid=61806) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 695.246337] env[61806]: DEBUG oslo_concurrency.lockutils [None req-38045444-a393-40c6-8f1e-16b0707aea48 tempest-ServerShowV257Test-1159790492 tempest-ServerShowV257Test-1159790492-project-member] Acquiring lock "616ec206-9804-469e-ab5c-41aea7f048aa" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 695.246337] env[61806]: DEBUG oslo_concurrency.lockutils [None req-38045444-a393-40c6-8f1e-16b0707aea48 tempest-ServerShowV257Test-1159790492 tempest-ServerShowV257Test-1159790492-project-member] Lock "616ec206-9804-469e-ab5c-41aea7f048aa" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 695.269745] env[61806]: DEBUG nova.compute.utils [None req-0f0a69c6-2310-4272-abc1-520940b4bcf0 tempest-ServersTestFqdnHostnames-356624376 tempest-ServersTestFqdnHostnames-356624376-project-member] Using /dev/sd instead of None {{(pid=61806) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 695.272810] env[61806]: DEBUG nova.compute.manager [None req-0f0a69c6-2310-4272-abc1-520940b4bcf0 tempest-ServersTestFqdnHostnames-356624376 tempest-ServersTestFqdnHostnames-356624376-project-member] [instance: 46c34ed0-9bac-4803-8d5c-3a3346a64117] Allocating IP information in the background. {{(pid=61806) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 695.272992] env[61806]: DEBUG nova.network.neutron [None req-0f0a69c6-2310-4272-abc1-520940b4bcf0 tempest-ServersTestFqdnHostnames-356624376 tempest-ServersTestFqdnHostnames-356624376-project-member] [instance: 46c34ed0-9bac-4803-8d5c-3a3346a64117] allocate_for_instance() {{(pid=61806) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 695.313629] env[61806]: DEBUG nova.policy [None req-0f0a69c6-2310-4272-abc1-520940b4bcf0 tempest-ServersTestFqdnHostnames-356624376 tempest-ServersTestFqdnHostnames-356624376-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '8c51bf2dcf1c4709b765497e02ac2bc0', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'a20b5be0380c4d888ae676cab8c57b5c', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61806) authorize /opt/stack/nova/nova/policy.py:201}} [ 695.457673] env[61806]: DEBUG oslo_concurrency.lockutils [None req-2bf81f39-4b8d-4851-a793-760af008ffe2 tempest-ImagesOneServerTestJSON-1003762139 tempest-ImagesOneServerTestJSON-1003762139-project-member] Releasing lock "refresh_cache-294f0a37-7589-405d-a324-13099c80f458" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 695.458128] env[61806]: DEBUG nova.compute.manager [None req-2bf81f39-4b8d-4851-a793-760af008ffe2 tempest-ImagesOneServerTestJSON-1003762139 tempest-ImagesOneServerTestJSON-1003762139-project-member] [instance: 294f0a37-7589-405d-a324-13099c80f458] Start destroying the instance on the hypervisor. {{(pid=61806) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 695.458328] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-2bf81f39-4b8d-4851-a793-760af008ffe2 tempest-ImagesOneServerTestJSON-1003762139 tempest-ImagesOneServerTestJSON-1003762139-project-member] [instance: 294f0a37-7589-405d-a324-13099c80f458] Destroying instance {{(pid=61806) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 695.458630] env[61806]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-977b93aa-63cd-4e62-9c63-73d3984ffd56 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 695.471722] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7f78ecf5-7f1a-400b-b3ad-850c9283ba2c {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 695.494274] env[61806]: WARNING nova.virt.vmwareapi.vmops [None req-2bf81f39-4b8d-4851-a793-760af008ffe2 tempest-ImagesOneServerTestJSON-1003762139 tempest-ImagesOneServerTestJSON-1003762139-project-member] [instance: 294f0a37-7589-405d-a324-13099c80f458] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 294f0a37-7589-405d-a324-13099c80f458 could not be found. [ 695.494414] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-2bf81f39-4b8d-4851-a793-760af008ffe2 tempest-ImagesOneServerTestJSON-1003762139 tempest-ImagesOneServerTestJSON-1003762139-project-member] [instance: 294f0a37-7589-405d-a324-13099c80f458] Instance destroyed {{(pid=61806) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 695.494727] env[61806]: INFO nova.compute.manager [None req-2bf81f39-4b8d-4851-a793-760af008ffe2 tempest-ImagesOneServerTestJSON-1003762139 tempest-ImagesOneServerTestJSON-1003762139-project-member] [instance: 294f0a37-7589-405d-a324-13099c80f458] Took 0.04 seconds to destroy the instance on the hypervisor. [ 695.494844] env[61806]: DEBUG oslo.service.loopingcall [None req-2bf81f39-4b8d-4851-a793-760af008ffe2 tempest-ImagesOneServerTestJSON-1003762139 tempest-ImagesOneServerTestJSON-1003762139-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61806) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 695.495102] env[61806]: DEBUG nova.compute.manager [-] [instance: 294f0a37-7589-405d-a324-13099c80f458] Deallocating network for instance {{(pid=61806) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 695.495192] env[61806]: DEBUG nova.network.neutron [-] [instance: 294f0a37-7589-405d-a324-13099c80f458] deallocate_for_instance() {{(pid=61806) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 695.511063] env[61806]: DEBUG nova.network.neutron [-] [instance: 294f0a37-7589-405d-a324-13099c80f458] Instance cache missing network info. {{(pid=61806) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 695.643463] env[61806]: DEBUG nova.network.neutron [None req-0f0a69c6-2310-4272-abc1-520940b4bcf0 tempest-ServersTestFqdnHostnames-356624376 tempest-ServersTestFqdnHostnames-356624376-project-member] [instance: 46c34ed0-9bac-4803-8d5c-3a3346a64117] Successfully created port: b68f927d-fcde-4af6-ba22-2354a0b31a3c {{(pid=61806) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 695.775061] env[61806]: DEBUG nova.compute.manager [None req-0f0a69c6-2310-4272-abc1-520940b4bcf0 tempest-ServersTestFqdnHostnames-356624376 tempest-ServersTestFqdnHostnames-356624376-project-member] [instance: 46c34ed0-9bac-4803-8d5c-3a3346a64117] Start building block device mappings for instance. {{(pid=61806) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 696.013235] env[61806]: DEBUG nova.network.neutron [-] [instance: 294f0a37-7589-405d-a324-13099c80f458] Updating instance_info_cache with network_info: [] {{(pid=61806) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 696.161422] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0a14ec76-63e7-4b8b-a740-f82ccc5f1df9 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 696.171250] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ef919f8a-84a9-479c-954a-2b438b3f6023 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 696.205977] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1d3e7ce9-9aa7-4587-a429-14f58388cd7f {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 696.213381] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d54bca3a-a10d-4150-904a-91bd066297f6 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 696.226149] env[61806]: DEBUG nova.compute.provider_tree [None req-05ef4399-8310-428c-ba55-af11abf4e445 tempest-SecurityGroupsTestJSON-1213485581 tempest-SecurityGroupsTestJSON-1213485581-project-member] Inventory has not changed in ProviderTree for provider: a2858be1-fd22-4e08-979e-87ad25293407 {{(pid=61806) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 696.517276] env[61806]: INFO nova.compute.manager [-] [instance: 294f0a37-7589-405d-a324-13099c80f458] Took 1.02 seconds to deallocate network for instance. [ 696.518357] env[61806]: DEBUG nova.compute.claims [None req-2bf81f39-4b8d-4851-a793-760af008ffe2 tempest-ImagesOneServerTestJSON-1003762139 tempest-ImagesOneServerTestJSON-1003762139-project-member] [instance: 294f0a37-7589-405d-a324-13099c80f458] Aborting claim: {{(pid=61806) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 696.518684] env[61806]: DEBUG oslo_concurrency.lockutils [None req-2bf81f39-4b8d-4851-a793-760af008ffe2 tempest-ImagesOneServerTestJSON-1003762139 tempest-ImagesOneServerTestJSON-1003762139-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 696.621900] env[61806]: ERROR nova.compute.manager [None req-0f0a69c6-2310-4272-abc1-520940b4bcf0 tempest-ServersTestFqdnHostnames-356624376 tempest-ServersTestFqdnHostnames-356624376-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port b68f927d-fcde-4af6-ba22-2354a0b31a3c, please check neutron logs for more information. [ 696.621900] env[61806]: ERROR nova.compute.manager Traceback (most recent call last): [ 696.621900] env[61806]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 696.621900] env[61806]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 696.621900] env[61806]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 696.621900] env[61806]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 696.621900] env[61806]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 696.621900] env[61806]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 696.621900] env[61806]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 696.621900] env[61806]: ERROR nova.compute.manager self.force_reraise() [ 696.621900] env[61806]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 696.621900] env[61806]: ERROR nova.compute.manager raise self.value [ 696.621900] env[61806]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 696.621900] env[61806]: ERROR nova.compute.manager updated_port = self._update_port( [ 696.621900] env[61806]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 696.621900] env[61806]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 696.622425] env[61806]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 696.622425] env[61806]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 696.622425] env[61806]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port b68f927d-fcde-4af6-ba22-2354a0b31a3c, please check neutron logs for more information. [ 696.622425] env[61806]: ERROR nova.compute.manager [ 696.622425] env[61806]: Traceback (most recent call last): [ 696.622425] env[61806]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 696.622425] env[61806]: listener.cb(fileno) [ 696.622425] env[61806]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 696.622425] env[61806]: result = function(*args, **kwargs) [ 696.622425] env[61806]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 696.622425] env[61806]: return func(*args, **kwargs) [ 696.622425] env[61806]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 696.622425] env[61806]: raise e [ 696.622425] env[61806]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 696.622425] env[61806]: nwinfo = self.network_api.allocate_for_instance( [ 696.622425] env[61806]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 696.622425] env[61806]: created_port_ids = self._update_ports_for_instance( [ 696.622425] env[61806]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 696.622425] env[61806]: with excutils.save_and_reraise_exception(): [ 696.622425] env[61806]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 696.622425] env[61806]: self.force_reraise() [ 696.622425] env[61806]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 696.622425] env[61806]: raise self.value [ 696.622425] env[61806]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 696.622425] env[61806]: updated_port = self._update_port( [ 696.622425] env[61806]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 696.622425] env[61806]: _ensure_no_port_binding_failure(port) [ 696.622425] env[61806]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 696.622425] env[61806]: raise exception.PortBindingFailed(port_id=port['id']) [ 696.623272] env[61806]: nova.exception.PortBindingFailed: Binding failed for port b68f927d-fcde-4af6-ba22-2354a0b31a3c, please check neutron logs for more information. [ 696.623272] env[61806]: Removing descriptor: 16 [ 696.729581] env[61806]: DEBUG nova.scheduler.client.report [None req-05ef4399-8310-428c-ba55-af11abf4e445 tempest-SecurityGroupsTestJSON-1213485581 tempest-SecurityGroupsTestJSON-1213485581-project-member] Inventory has not changed for provider a2858be1-fd22-4e08-979e-87ad25293407 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 139, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61806) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 696.789485] env[61806]: DEBUG nova.compute.manager [None req-0f0a69c6-2310-4272-abc1-520940b4bcf0 tempest-ServersTestFqdnHostnames-356624376 tempest-ServersTestFqdnHostnames-356624376-project-member] [instance: 46c34ed0-9bac-4803-8d5c-3a3346a64117] Start spawning the instance on the hypervisor. {{(pid=61806) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 696.814690] env[61806]: DEBUG nova.virt.hardware [None req-0f0a69c6-2310-4272-abc1-520940b4bcf0 tempest-ServersTestFqdnHostnames-356624376 tempest-ServersTestFqdnHostnames-356624376-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-15T18:05:10Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-15T18:04:52Z,direct_url=,disk_format='vmdk',id=ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='84ba83fce2c349a6988173c5d6fc3b07',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-15T18:04:53Z,virtual_size=,visibility=), allow threads: False {{(pid=61806) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 696.814944] env[61806]: DEBUG nova.virt.hardware [None req-0f0a69c6-2310-4272-abc1-520940b4bcf0 tempest-ServersTestFqdnHostnames-356624376 tempest-ServersTestFqdnHostnames-356624376-project-member] Flavor limits 0:0:0 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 696.815138] env[61806]: DEBUG nova.virt.hardware [None req-0f0a69c6-2310-4272-abc1-520940b4bcf0 tempest-ServersTestFqdnHostnames-356624376 tempest-ServersTestFqdnHostnames-356624376-project-member] Image limits 0:0:0 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 696.815286] env[61806]: DEBUG nova.virt.hardware [None req-0f0a69c6-2310-4272-abc1-520940b4bcf0 tempest-ServersTestFqdnHostnames-356624376 tempest-ServersTestFqdnHostnames-356624376-project-member] Flavor pref 0:0:0 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 696.815434] env[61806]: DEBUG nova.virt.hardware [None req-0f0a69c6-2310-4272-abc1-520940b4bcf0 tempest-ServersTestFqdnHostnames-356624376 tempest-ServersTestFqdnHostnames-356624376-project-member] Image pref 0:0:0 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 696.815581] env[61806]: DEBUG nova.virt.hardware [None req-0f0a69c6-2310-4272-abc1-520940b4bcf0 tempest-ServersTestFqdnHostnames-356624376 tempest-ServersTestFqdnHostnames-356624376-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 696.815786] env[61806]: DEBUG nova.virt.hardware [None req-0f0a69c6-2310-4272-abc1-520940b4bcf0 tempest-ServersTestFqdnHostnames-356624376 tempest-ServersTestFqdnHostnames-356624376-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61806) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 696.815946] env[61806]: DEBUG nova.virt.hardware [None req-0f0a69c6-2310-4272-abc1-520940b4bcf0 tempest-ServersTestFqdnHostnames-356624376 tempest-ServersTestFqdnHostnames-356624376-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61806) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 696.816127] env[61806]: DEBUG nova.virt.hardware [None req-0f0a69c6-2310-4272-abc1-520940b4bcf0 tempest-ServersTestFqdnHostnames-356624376 tempest-ServersTestFqdnHostnames-356624376-project-member] Got 1 possible topologies {{(pid=61806) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 696.816291] env[61806]: DEBUG nova.virt.hardware [None req-0f0a69c6-2310-4272-abc1-520940b4bcf0 tempest-ServersTestFqdnHostnames-356624376 tempest-ServersTestFqdnHostnames-356624376-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61806) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 696.816463] env[61806]: DEBUG nova.virt.hardware [None req-0f0a69c6-2310-4272-abc1-520940b4bcf0 tempest-ServersTestFqdnHostnames-356624376 tempest-ServersTestFqdnHostnames-356624376-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61806) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 696.817325] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d782042e-0dca-409e-980c-3f2fa83767d0 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 696.824823] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6d83f427-8692-4c1f-b575-b3f39382cf1d {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 696.839394] env[61806]: ERROR nova.compute.manager [None req-0f0a69c6-2310-4272-abc1-520940b4bcf0 tempest-ServersTestFqdnHostnames-356624376 tempest-ServersTestFqdnHostnames-356624376-project-member] [instance: 46c34ed0-9bac-4803-8d5c-3a3346a64117] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port b68f927d-fcde-4af6-ba22-2354a0b31a3c, please check neutron logs for more information. [ 696.839394] env[61806]: ERROR nova.compute.manager [instance: 46c34ed0-9bac-4803-8d5c-3a3346a64117] Traceback (most recent call last): [ 696.839394] env[61806]: ERROR nova.compute.manager [instance: 46c34ed0-9bac-4803-8d5c-3a3346a64117] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 696.839394] env[61806]: ERROR nova.compute.manager [instance: 46c34ed0-9bac-4803-8d5c-3a3346a64117] yield resources [ 696.839394] env[61806]: ERROR nova.compute.manager [instance: 46c34ed0-9bac-4803-8d5c-3a3346a64117] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 696.839394] env[61806]: ERROR nova.compute.manager [instance: 46c34ed0-9bac-4803-8d5c-3a3346a64117] self.driver.spawn(context, instance, image_meta, [ 696.839394] env[61806]: ERROR nova.compute.manager [instance: 46c34ed0-9bac-4803-8d5c-3a3346a64117] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 696.839394] env[61806]: ERROR nova.compute.manager [instance: 46c34ed0-9bac-4803-8d5c-3a3346a64117] self._vmops.spawn(context, instance, image_meta, injected_files, [ 696.839394] env[61806]: ERROR nova.compute.manager [instance: 46c34ed0-9bac-4803-8d5c-3a3346a64117] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 696.839394] env[61806]: ERROR nova.compute.manager [instance: 46c34ed0-9bac-4803-8d5c-3a3346a64117] vm_ref = self.build_virtual_machine(instance, [ 696.839394] env[61806]: ERROR nova.compute.manager [instance: 46c34ed0-9bac-4803-8d5c-3a3346a64117] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 696.839732] env[61806]: ERROR nova.compute.manager [instance: 46c34ed0-9bac-4803-8d5c-3a3346a64117] vif_infos = vmwarevif.get_vif_info(self._session, [ 696.839732] env[61806]: ERROR nova.compute.manager [instance: 46c34ed0-9bac-4803-8d5c-3a3346a64117] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 696.839732] env[61806]: ERROR nova.compute.manager [instance: 46c34ed0-9bac-4803-8d5c-3a3346a64117] for vif in network_info: [ 696.839732] env[61806]: ERROR nova.compute.manager [instance: 46c34ed0-9bac-4803-8d5c-3a3346a64117] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 696.839732] env[61806]: ERROR nova.compute.manager [instance: 46c34ed0-9bac-4803-8d5c-3a3346a64117] return self._sync_wrapper(fn, *args, **kwargs) [ 696.839732] env[61806]: ERROR nova.compute.manager [instance: 46c34ed0-9bac-4803-8d5c-3a3346a64117] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 696.839732] env[61806]: ERROR nova.compute.manager [instance: 46c34ed0-9bac-4803-8d5c-3a3346a64117] self.wait() [ 696.839732] env[61806]: ERROR nova.compute.manager [instance: 46c34ed0-9bac-4803-8d5c-3a3346a64117] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 696.839732] env[61806]: ERROR nova.compute.manager [instance: 46c34ed0-9bac-4803-8d5c-3a3346a64117] self[:] = self._gt.wait() [ 696.839732] env[61806]: ERROR nova.compute.manager [instance: 46c34ed0-9bac-4803-8d5c-3a3346a64117] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 696.839732] env[61806]: ERROR nova.compute.manager [instance: 46c34ed0-9bac-4803-8d5c-3a3346a64117] return self._exit_event.wait() [ 696.839732] env[61806]: ERROR nova.compute.manager [instance: 46c34ed0-9bac-4803-8d5c-3a3346a64117] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 696.839732] env[61806]: ERROR nova.compute.manager [instance: 46c34ed0-9bac-4803-8d5c-3a3346a64117] current.throw(*self._exc) [ 696.840085] env[61806]: ERROR nova.compute.manager [instance: 46c34ed0-9bac-4803-8d5c-3a3346a64117] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 696.840085] env[61806]: ERROR nova.compute.manager [instance: 46c34ed0-9bac-4803-8d5c-3a3346a64117] result = function(*args, **kwargs) [ 696.840085] env[61806]: ERROR nova.compute.manager [instance: 46c34ed0-9bac-4803-8d5c-3a3346a64117] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 696.840085] env[61806]: ERROR nova.compute.manager [instance: 46c34ed0-9bac-4803-8d5c-3a3346a64117] return func(*args, **kwargs) [ 696.840085] env[61806]: ERROR nova.compute.manager [instance: 46c34ed0-9bac-4803-8d5c-3a3346a64117] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 696.840085] env[61806]: ERROR nova.compute.manager [instance: 46c34ed0-9bac-4803-8d5c-3a3346a64117] raise e [ 696.840085] env[61806]: ERROR nova.compute.manager [instance: 46c34ed0-9bac-4803-8d5c-3a3346a64117] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 696.840085] env[61806]: ERROR nova.compute.manager [instance: 46c34ed0-9bac-4803-8d5c-3a3346a64117] nwinfo = self.network_api.allocate_for_instance( [ 696.840085] env[61806]: ERROR nova.compute.manager [instance: 46c34ed0-9bac-4803-8d5c-3a3346a64117] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 696.840085] env[61806]: ERROR nova.compute.manager [instance: 46c34ed0-9bac-4803-8d5c-3a3346a64117] created_port_ids = self._update_ports_for_instance( [ 696.840085] env[61806]: ERROR nova.compute.manager [instance: 46c34ed0-9bac-4803-8d5c-3a3346a64117] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 696.840085] env[61806]: ERROR nova.compute.manager [instance: 46c34ed0-9bac-4803-8d5c-3a3346a64117] with excutils.save_and_reraise_exception(): [ 696.840085] env[61806]: ERROR nova.compute.manager [instance: 46c34ed0-9bac-4803-8d5c-3a3346a64117] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 696.840439] env[61806]: ERROR nova.compute.manager [instance: 46c34ed0-9bac-4803-8d5c-3a3346a64117] self.force_reraise() [ 696.840439] env[61806]: ERROR nova.compute.manager [instance: 46c34ed0-9bac-4803-8d5c-3a3346a64117] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 696.840439] env[61806]: ERROR nova.compute.manager [instance: 46c34ed0-9bac-4803-8d5c-3a3346a64117] raise self.value [ 696.840439] env[61806]: ERROR nova.compute.manager [instance: 46c34ed0-9bac-4803-8d5c-3a3346a64117] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 696.840439] env[61806]: ERROR nova.compute.manager [instance: 46c34ed0-9bac-4803-8d5c-3a3346a64117] updated_port = self._update_port( [ 696.840439] env[61806]: ERROR nova.compute.manager [instance: 46c34ed0-9bac-4803-8d5c-3a3346a64117] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 696.840439] env[61806]: ERROR nova.compute.manager [instance: 46c34ed0-9bac-4803-8d5c-3a3346a64117] _ensure_no_port_binding_failure(port) [ 696.840439] env[61806]: ERROR nova.compute.manager [instance: 46c34ed0-9bac-4803-8d5c-3a3346a64117] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 696.840439] env[61806]: ERROR nova.compute.manager [instance: 46c34ed0-9bac-4803-8d5c-3a3346a64117] raise exception.PortBindingFailed(port_id=port['id']) [ 696.840439] env[61806]: ERROR nova.compute.manager [instance: 46c34ed0-9bac-4803-8d5c-3a3346a64117] nova.exception.PortBindingFailed: Binding failed for port b68f927d-fcde-4af6-ba22-2354a0b31a3c, please check neutron logs for more information. [ 696.840439] env[61806]: ERROR nova.compute.manager [instance: 46c34ed0-9bac-4803-8d5c-3a3346a64117] [ 696.840439] env[61806]: INFO nova.compute.manager [None req-0f0a69c6-2310-4272-abc1-520940b4bcf0 tempest-ServersTestFqdnHostnames-356624376 tempest-ServersTestFqdnHostnames-356624376-project-member] [instance: 46c34ed0-9bac-4803-8d5c-3a3346a64117] Terminating instance [ 696.841746] env[61806]: DEBUG oslo_concurrency.lockutils [None req-0f0a69c6-2310-4272-abc1-520940b4bcf0 tempest-ServersTestFqdnHostnames-356624376 tempest-ServersTestFqdnHostnames-356624376-project-member] Acquiring lock "refresh_cache-46c34ed0-9bac-4803-8d5c-3a3346a64117" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 696.841913] env[61806]: DEBUG oslo_concurrency.lockutils [None req-0f0a69c6-2310-4272-abc1-520940b4bcf0 tempest-ServersTestFqdnHostnames-356624376 tempest-ServersTestFqdnHostnames-356624376-project-member] Acquired lock "refresh_cache-46c34ed0-9bac-4803-8d5c-3a3346a64117" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 696.842091] env[61806]: DEBUG nova.network.neutron [None req-0f0a69c6-2310-4272-abc1-520940b4bcf0 tempest-ServersTestFqdnHostnames-356624376 tempest-ServersTestFqdnHostnames-356624376-project-member] [instance: 46c34ed0-9bac-4803-8d5c-3a3346a64117] Building network info cache for instance {{(pid=61806) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 697.235612] env[61806]: DEBUG oslo_concurrency.lockutils [None req-05ef4399-8310-428c-ba55-af11abf4e445 tempest-SecurityGroupsTestJSON-1213485581 tempest-SecurityGroupsTestJSON-1213485581-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.472s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 697.236233] env[61806]: DEBUG nova.compute.manager [None req-05ef4399-8310-428c-ba55-af11abf4e445 tempest-SecurityGroupsTestJSON-1213485581 tempest-SecurityGroupsTestJSON-1213485581-project-member] [instance: 294adb58-f23e-4510-a25a-de6b909e3189] Start building networks asynchronously for instance. {{(pid=61806) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 697.239130] env[61806]: DEBUG oslo_concurrency.lockutils [None req-b4447f02-a3c7-4b6e-89ba-6bc34ec03205 tempest-InstanceActionsTestJSON-949444324 tempest-InstanceActionsTestJSON-949444324-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 32.507s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 697.257111] env[61806]: DEBUG nova.compute.manager [req-5883008d-3d17-4285-800a-060272e6677b req-2d27f347-6f0a-45d9-9036-a19bf07e4f80 service nova] [instance: 46c34ed0-9bac-4803-8d5c-3a3346a64117] Received event network-changed-b68f927d-fcde-4af6-ba22-2354a0b31a3c {{(pid=61806) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 697.257317] env[61806]: DEBUG nova.compute.manager [req-5883008d-3d17-4285-800a-060272e6677b req-2d27f347-6f0a-45d9-9036-a19bf07e4f80 service nova] [instance: 46c34ed0-9bac-4803-8d5c-3a3346a64117] Refreshing instance network info cache due to event network-changed-b68f927d-fcde-4af6-ba22-2354a0b31a3c. {{(pid=61806) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 697.257504] env[61806]: DEBUG oslo_concurrency.lockutils [req-5883008d-3d17-4285-800a-060272e6677b req-2d27f347-6f0a-45d9-9036-a19bf07e4f80 service nova] Acquiring lock "refresh_cache-46c34ed0-9bac-4803-8d5c-3a3346a64117" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 697.359526] env[61806]: DEBUG nova.network.neutron [None req-0f0a69c6-2310-4272-abc1-520940b4bcf0 tempest-ServersTestFqdnHostnames-356624376 tempest-ServersTestFqdnHostnames-356624376-project-member] [instance: 46c34ed0-9bac-4803-8d5c-3a3346a64117] Instance cache missing network info. {{(pid=61806) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 697.452613] env[61806]: DEBUG nova.network.neutron [None req-0f0a69c6-2310-4272-abc1-520940b4bcf0 tempest-ServersTestFqdnHostnames-356624376 tempest-ServersTestFqdnHostnames-356624376-project-member] [instance: 46c34ed0-9bac-4803-8d5c-3a3346a64117] Updating instance_info_cache with network_info: [] {{(pid=61806) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 697.745958] env[61806]: DEBUG nova.compute.utils [None req-05ef4399-8310-428c-ba55-af11abf4e445 tempest-SecurityGroupsTestJSON-1213485581 tempest-SecurityGroupsTestJSON-1213485581-project-member] Using /dev/sd instead of None {{(pid=61806) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 697.752019] env[61806]: DEBUG nova.compute.manager [None req-05ef4399-8310-428c-ba55-af11abf4e445 tempest-SecurityGroupsTestJSON-1213485581 tempest-SecurityGroupsTestJSON-1213485581-project-member] [instance: 294adb58-f23e-4510-a25a-de6b909e3189] Allocating IP information in the background. {{(pid=61806) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 697.752019] env[61806]: DEBUG nova.network.neutron [None req-05ef4399-8310-428c-ba55-af11abf4e445 tempest-SecurityGroupsTestJSON-1213485581 tempest-SecurityGroupsTestJSON-1213485581-project-member] [instance: 294adb58-f23e-4510-a25a-de6b909e3189] allocate_for_instance() {{(pid=61806) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 697.805016] env[61806]: DEBUG nova.policy [None req-05ef4399-8310-428c-ba55-af11abf4e445 tempest-SecurityGroupsTestJSON-1213485581 tempest-SecurityGroupsTestJSON-1213485581-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'e32051e5482847ce85b4f9fa6e89f9e1', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'cfce93add38d40d79f294017f52e13d5', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61806) authorize /opt/stack/nova/nova/policy.py:201}} [ 697.955193] env[61806]: DEBUG oslo_concurrency.lockutils [None req-0f0a69c6-2310-4272-abc1-520940b4bcf0 tempest-ServersTestFqdnHostnames-356624376 tempest-ServersTestFqdnHostnames-356624376-project-member] Releasing lock "refresh_cache-46c34ed0-9bac-4803-8d5c-3a3346a64117" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 697.955798] env[61806]: DEBUG nova.compute.manager [None req-0f0a69c6-2310-4272-abc1-520940b4bcf0 tempest-ServersTestFqdnHostnames-356624376 tempest-ServersTestFqdnHostnames-356624376-project-member] [instance: 46c34ed0-9bac-4803-8d5c-3a3346a64117] Start destroying the instance on the hypervisor. {{(pid=61806) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 697.956187] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-0f0a69c6-2310-4272-abc1-520940b4bcf0 tempest-ServersTestFqdnHostnames-356624376 tempest-ServersTestFqdnHostnames-356624376-project-member] [instance: 46c34ed0-9bac-4803-8d5c-3a3346a64117] Destroying instance {{(pid=61806) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 697.956418] env[61806]: DEBUG oslo_concurrency.lockutils [req-5883008d-3d17-4285-800a-060272e6677b req-2d27f347-6f0a-45d9-9036-a19bf07e4f80 service nova] Acquired lock "refresh_cache-46c34ed0-9bac-4803-8d5c-3a3346a64117" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 697.956592] env[61806]: DEBUG nova.network.neutron [req-5883008d-3d17-4285-800a-060272e6677b req-2d27f347-6f0a-45d9-9036-a19bf07e4f80 service nova] [instance: 46c34ed0-9bac-4803-8d5c-3a3346a64117] Refreshing network info cache for port b68f927d-fcde-4af6-ba22-2354a0b31a3c {{(pid=61806) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 697.957884] env[61806]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-762dbaaf-7332-4c8e-9b1b-913c496e47e4 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 697.974365] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-469d4e63-9c01-4344-8a43-264d6b3ff9ca {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 698.003228] env[61806]: WARNING nova.virt.vmwareapi.vmops [None req-0f0a69c6-2310-4272-abc1-520940b4bcf0 tempest-ServersTestFqdnHostnames-356624376 tempest-ServersTestFqdnHostnames-356624376-project-member] [instance: 46c34ed0-9bac-4803-8d5c-3a3346a64117] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 46c34ed0-9bac-4803-8d5c-3a3346a64117 could not be found. [ 698.003228] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-0f0a69c6-2310-4272-abc1-520940b4bcf0 tempest-ServersTestFqdnHostnames-356624376 tempest-ServersTestFqdnHostnames-356624376-project-member] [instance: 46c34ed0-9bac-4803-8d5c-3a3346a64117] Instance destroyed {{(pid=61806) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 698.003228] env[61806]: INFO nova.compute.manager [None req-0f0a69c6-2310-4272-abc1-520940b4bcf0 tempest-ServersTestFqdnHostnames-356624376 tempest-ServersTestFqdnHostnames-356624376-project-member] [instance: 46c34ed0-9bac-4803-8d5c-3a3346a64117] Took 0.05 seconds to destroy the instance on the hypervisor. [ 698.003228] env[61806]: DEBUG oslo.service.loopingcall [None req-0f0a69c6-2310-4272-abc1-520940b4bcf0 tempest-ServersTestFqdnHostnames-356624376 tempest-ServersTestFqdnHostnames-356624376-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61806) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 698.007027] env[61806]: DEBUG nova.compute.manager [-] [instance: 46c34ed0-9bac-4803-8d5c-3a3346a64117] Deallocating network for instance {{(pid=61806) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 698.007027] env[61806]: DEBUG nova.network.neutron [-] [instance: 46c34ed0-9bac-4803-8d5c-3a3346a64117] deallocate_for_instance() {{(pid=61806) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 698.020852] env[61806]: DEBUG nova.network.neutron [-] [instance: 46c34ed0-9bac-4803-8d5c-3a3346a64117] Instance cache missing network info. {{(pid=61806) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 698.104966] env[61806]: DEBUG nova.network.neutron [None req-05ef4399-8310-428c-ba55-af11abf4e445 tempest-SecurityGroupsTestJSON-1213485581 tempest-SecurityGroupsTestJSON-1213485581-project-member] [instance: 294adb58-f23e-4510-a25a-de6b909e3189] Successfully created port: 5fb92c74-bcc1-4423-9d6b-394b8c87bb92 {{(pid=61806) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 698.137542] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a53bb1f8-0e00-466c-bf6a-967216b029ea {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 698.144677] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b633049c-15cc-46f8-a13b-ad44e39bea6a {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 698.175690] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f7d5579f-5905-445b-8447-6bfd176dffab {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 698.182659] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-852f9b55-9759-4575-9122-4dc583efc704 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 698.195306] env[61806]: DEBUG nova.compute.provider_tree [None req-b4447f02-a3c7-4b6e-89ba-6bc34ec03205 tempest-InstanceActionsTestJSON-949444324 tempest-InstanceActionsTestJSON-949444324-project-member] Inventory has not changed in ProviderTree for provider: a2858be1-fd22-4e08-979e-87ad25293407 {{(pid=61806) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 698.250995] env[61806]: DEBUG nova.compute.manager [None req-05ef4399-8310-428c-ba55-af11abf4e445 tempest-SecurityGroupsTestJSON-1213485581 tempest-SecurityGroupsTestJSON-1213485581-project-member] [instance: 294adb58-f23e-4510-a25a-de6b909e3189] Start building block device mappings for instance. {{(pid=61806) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 698.487072] env[61806]: DEBUG nova.network.neutron [req-5883008d-3d17-4285-800a-060272e6677b req-2d27f347-6f0a-45d9-9036-a19bf07e4f80 service nova] [instance: 46c34ed0-9bac-4803-8d5c-3a3346a64117] Instance cache missing network info. {{(pid=61806) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 698.527834] env[61806]: DEBUG nova.network.neutron [-] [instance: 46c34ed0-9bac-4803-8d5c-3a3346a64117] Updating instance_info_cache with network_info: [] {{(pid=61806) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 698.599313] env[61806]: DEBUG nova.network.neutron [req-5883008d-3d17-4285-800a-060272e6677b req-2d27f347-6f0a-45d9-9036-a19bf07e4f80 service nova] [instance: 46c34ed0-9bac-4803-8d5c-3a3346a64117] Updating instance_info_cache with network_info: [] {{(pid=61806) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 698.698041] env[61806]: DEBUG nova.scheduler.client.report [None req-b4447f02-a3c7-4b6e-89ba-6bc34ec03205 tempest-InstanceActionsTestJSON-949444324 tempest-InstanceActionsTestJSON-949444324-project-member] Inventory has not changed for provider a2858be1-fd22-4e08-979e-87ad25293407 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 139, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61806) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 699.033022] env[61806]: INFO nova.compute.manager [-] [instance: 46c34ed0-9bac-4803-8d5c-3a3346a64117] Took 1.02 seconds to deallocate network for instance. [ 699.033296] env[61806]: DEBUG nova.compute.claims [None req-0f0a69c6-2310-4272-abc1-520940b4bcf0 tempest-ServersTestFqdnHostnames-356624376 tempest-ServersTestFqdnHostnames-356624376-project-member] [instance: 46c34ed0-9bac-4803-8d5c-3a3346a64117] Aborting claim: {{(pid=61806) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 699.033421] env[61806]: DEBUG oslo_concurrency.lockutils [None req-0f0a69c6-2310-4272-abc1-520940b4bcf0 tempest-ServersTestFqdnHostnames-356624376 tempest-ServersTestFqdnHostnames-356624376-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 699.052644] env[61806]: ERROR nova.compute.manager [None req-05ef4399-8310-428c-ba55-af11abf4e445 tempest-SecurityGroupsTestJSON-1213485581 tempest-SecurityGroupsTestJSON-1213485581-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 5fb92c74-bcc1-4423-9d6b-394b8c87bb92, please check neutron logs for more information. [ 699.052644] env[61806]: ERROR nova.compute.manager Traceback (most recent call last): [ 699.052644] env[61806]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 699.052644] env[61806]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 699.052644] env[61806]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 699.052644] env[61806]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 699.052644] env[61806]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 699.052644] env[61806]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 699.052644] env[61806]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 699.052644] env[61806]: ERROR nova.compute.manager self.force_reraise() [ 699.052644] env[61806]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 699.052644] env[61806]: ERROR nova.compute.manager raise self.value [ 699.052644] env[61806]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 699.052644] env[61806]: ERROR nova.compute.manager updated_port = self._update_port( [ 699.052644] env[61806]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 699.052644] env[61806]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 699.053500] env[61806]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 699.053500] env[61806]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 699.053500] env[61806]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 5fb92c74-bcc1-4423-9d6b-394b8c87bb92, please check neutron logs for more information. [ 699.053500] env[61806]: ERROR nova.compute.manager [ 699.053500] env[61806]: Traceback (most recent call last): [ 699.053500] env[61806]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 699.053500] env[61806]: listener.cb(fileno) [ 699.053500] env[61806]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 699.053500] env[61806]: result = function(*args, **kwargs) [ 699.053500] env[61806]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 699.053500] env[61806]: return func(*args, **kwargs) [ 699.053500] env[61806]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 699.053500] env[61806]: raise e [ 699.053500] env[61806]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 699.053500] env[61806]: nwinfo = self.network_api.allocate_for_instance( [ 699.053500] env[61806]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 699.053500] env[61806]: created_port_ids = self._update_ports_for_instance( [ 699.053500] env[61806]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 699.053500] env[61806]: with excutils.save_and_reraise_exception(): [ 699.053500] env[61806]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 699.053500] env[61806]: self.force_reraise() [ 699.053500] env[61806]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 699.053500] env[61806]: raise self.value [ 699.053500] env[61806]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 699.053500] env[61806]: updated_port = self._update_port( [ 699.053500] env[61806]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 699.053500] env[61806]: _ensure_no_port_binding_failure(port) [ 699.053500] env[61806]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 699.053500] env[61806]: raise exception.PortBindingFailed(port_id=port['id']) [ 699.054670] env[61806]: nova.exception.PortBindingFailed: Binding failed for port 5fb92c74-bcc1-4423-9d6b-394b8c87bb92, please check neutron logs for more information. [ 699.054670] env[61806]: Removing descriptor: 16 [ 699.101654] env[61806]: DEBUG oslo_concurrency.lockutils [req-5883008d-3d17-4285-800a-060272e6677b req-2d27f347-6f0a-45d9-9036-a19bf07e4f80 service nova] Releasing lock "refresh_cache-46c34ed0-9bac-4803-8d5c-3a3346a64117" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 699.101994] env[61806]: DEBUG nova.compute.manager [req-5883008d-3d17-4285-800a-060272e6677b req-2d27f347-6f0a-45d9-9036-a19bf07e4f80 service nova] [instance: 46c34ed0-9bac-4803-8d5c-3a3346a64117] Received event network-vif-deleted-b68f927d-fcde-4af6-ba22-2354a0b31a3c {{(pid=61806) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 699.203624] env[61806]: DEBUG oslo_concurrency.lockutils [None req-b4447f02-a3c7-4b6e-89ba-6bc34ec03205 tempest-InstanceActionsTestJSON-949444324 tempest-InstanceActionsTestJSON-949444324-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.964s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 699.204267] env[61806]: ERROR nova.compute.manager [None req-b4447f02-a3c7-4b6e-89ba-6bc34ec03205 tempest-InstanceActionsTestJSON-949444324 tempest-InstanceActionsTestJSON-949444324-project-member] [instance: 2681766c-0468-4170-8c51-03389e2bd1ab] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 59f4de97-5864-4576-82b3-a14d8b5224cd, please check neutron logs for more information. [ 699.204267] env[61806]: ERROR nova.compute.manager [instance: 2681766c-0468-4170-8c51-03389e2bd1ab] Traceback (most recent call last): [ 699.204267] env[61806]: ERROR nova.compute.manager [instance: 2681766c-0468-4170-8c51-03389e2bd1ab] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 699.204267] env[61806]: ERROR nova.compute.manager [instance: 2681766c-0468-4170-8c51-03389e2bd1ab] self.driver.spawn(context, instance, image_meta, [ 699.204267] env[61806]: ERROR nova.compute.manager [instance: 2681766c-0468-4170-8c51-03389e2bd1ab] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 699.204267] env[61806]: ERROR nova.compute.manager [instance: 2681766c-0468-4170-8c51-03389e2bd1ab] self._vmops.spawn(context, instance, image_meta, injected_files, [ 699.204267] env[61806]: ERROR nova.compute.manager [instance: 2681766c-0468-4170-8c51-03389e2bd1ab] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 699.204267] env[61806]: ERROR nova.compute.manager [instance: 2681766c-0468-4170-8c51-03389e2bd1ab] vm_ref = self.build_virtual_machine(instance, [ 699.204267] env[61806]: ERROR nova.compute.manager [instance: 2681766c-0468-4170-8c51-03389e2bd1ab] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 699.204267] env[61806]: ERROR nova.compute.manager [instance: 2681766c-0468-4170-8c51-03389e2bd1ab] vif_infos = vmwarevif.get_vif_info(self._session, [ 699.204267] env[61806]: ERROR nova.compute.manager [instance: 2681766c-0468-4170-8c51-03389e2bd1ab] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 699.204604] env[61806]: ERROR nova.compute.manager [instance: 2681766c-0468-4170-8c51-03389e2bd1ab] for vif in network_info: [ 699.204604] env[61806]: ERROR nova.compute.manager [instance: 2681766c-0468-4170-8c51-03389e2bd1ab] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 699.204604] env[61806]: ERROR nova.compute.manager [instance: 2681766c-0468-4170-8c51-03389e2bd1ab] return self._sync_wrapper(fn, *args, **kwargs) [ 699.204604] env[61806]: ERROR nova.compute.manager [instance: 2681766c-0468-4170-8c51-03389e2bd1ab] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 699.204604] env[61806]: ERROR nova.compute.manager [instance: 2681766c-0468-4170-8c51-03389e2bd1ab] self.wait() [ 699.204604] env[61806]: ERROR nova.compute.manager [instance: 2681766c-0468-4170-8c51-03389e2bd1ab] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 699.204604] env[61806]: ERROR nova.compute.manager [instance: 2681766c-0468-4170-8c51-03389e2bd1ab] self[:] = self._gt.wait() [ 699.204604] env[61806]: ERROR nova.compute.manager [instance: 2681766c-0468-4170-8c51-03389e2bd1ab] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 699.204604] env[61806]: ERROR nova.compute.manager [instance: 2681766c-0468-4170-8c51-03389e2bd1ab] return self._exit_event.wait() [ 699.204604] env[61806]: ERROR nova.compute.manager [instance: 2681766c-0468-4170-8c51-03389e2bd1ab] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 699.204604] env[61806]: ERROR nova.compute.manager [instance: 2681766c-0468-4170-8c51-03389e2bd1ab] current.throw(*self._exc) [ 699.204604] env[61806]: ERROR nova.compute.manager [instance: 2681766c-0468-4170-8c51-03389e2bd1ab] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 699.204604] env[61806]: ERROR nova.compute.manager [instance: 2681766c-0468-4170-8c51-03389e2bd1ab] result = function(*args, **kwargs) [ 699.204935] env[61806]: ERROR nova.compute.manager [instance: 2681766c-0468-4170-8c51-03389e2bd1ab] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 699.204935] env[61806]: ERROR nova.compute.manager [instance: 2681766c-0468-4170-8c51-03389e2bd1ab] return func(*args, **kwargs) [ 699.204935] env[61806]: ERROR nova.compute.manager [instance: 2681766c-0468-4170-8c51-03389e2bd1ab] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 699.204935] env[61806]: ERROR nova.compute.manager [instance: 2681766c-0468-4170-8c51-03389e2bd1ab] raise e [ 699.204935] env[61806]: ERROR nova.compute.manager [instance: 2681766c-0468-4170-8c51-03389e2bd1ab] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 699.204935] env[61806]: ERROR nova.compute.manager [instance: 2681766c-0468-4170-8c51-03389e2bd1ab] nwinfo = self.network_api.allocate_for_instance( [ 699.204935] env[61806]: ERROR nova.compute.manager [instance: 2681766c-0468-4170-8c51-03389e2bd1ab] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 699.204935] env[61806]: ERROR nova.compute.manager [instance: 2681766c-0468-4170-8c51-03389e2bd1ab] created_port_ids = self._update_ports_for_instance( [ 699.204935] env[61806]: ERROR nova.compute.manager [instance: 2681766c-0468-4170-8c51-03389e2bd1ab] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 699.204935] env[61806]: ERROR nova.compute.manager [instance: 2681766c-0468-4170-8c51-03389e2bd1ab] with excutils.save_and_reraise_exception(): [ 699.204935] env[61806]: ERROR nova.compute.manager [instance: 2681766c-0468-4170-8c51-03389e2bd1ab] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 699.204935] env[61806]: ERROR nova.compute.manager [instance: 2681766c-0468-4170-8c51-03389e2bd1ab] self.force_reraise() [ 699.204935] env[61806]: ERROR nova.compute.manager [instance: 2681766c-0468-4170-8c51-03389e2bd1ab] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 699.205615] env[61806]: ERROR nova.compute.manager [instance: 2681766c-0468-4170-8c51-03389e2bd1ab] raise self.value [ 699.205615] env[61806]: ERROR nova.compute.manager [instance: 2681766c-0468-4170-8c51-03389e2bd1ab] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 699.205615] env[61806]: ERROR nova.compute.manager [instance: 2681766c-0468-4170-8c51-03389e2bd1ab] updated_port = self._update_port( [ 699.205615] env[61806]: ERROR nova.compute.manager [instance: 2681766c-0468-4170-8c51-03389e2bd1ab] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 699.205615] env[61806]: ERROR nova.compute.manager [instance: 2681766c-0468-4170-8c51-03389e2bd1ab] _ensure_no_port_binding_failure(port) [ 699.205615] env[61806]: ERROR nova.compute.manager [instance: 2681766c-0468-4170-8c51-03389e2bd1ab] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 699.205615] env[61806]: ERROR nova.compute.manager [instance: 2681766c-0468-4170-8c51-03389e2bd1ab] raise exception.PortBindingFailed(port_id=port['id']) [ 699.205615] env[61806]: ERROR nova.compute.manager [instance: 2681766c-0468-4170-8c51-03389e2bd1ab] nova.exception.PortBindingFailed: Binding failed for port 59f4de97-5864-4576-82b3-a14d8b5224cd, please check neutron logs for more information. [ 699.205615] env[61806]: ERROR nova.compute.manager [instance: 2681766c-0468-4170-8c51-03389e2bd1ab] [ 699.205615] env[61806]: DEBUG nova.compute.utils [None req-b4447f02-a3c7-4b6e-89ba-6bc34ec03205 tempest-InstanceActionsTestJSON-949444324 tempest-InstanceActionsTestJSON-949444324-project-member] [instance: 2681766c-0468-4170-8c51-03389e2bd1ab] Binding failed for port 59f4de97-5864-4576-82b3-a14d8b5224cd, please check neutron logs for more information. {{(pid=61806) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 699.206331] env[61806]: DEBUG oslo_concurrency.lockutils [None req-6dff00ec-9995-4616-9033-47ec4f3592c9 tempest-ServerShowV254Test-608938657 tempest-ServerShowV254Test-608938657-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 33.187s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 699.206553] env[61806]: DEBUG nova.objects.instance [None req-6dff00ec-9995-4616-9033-47ec4f3592c9 tempest-ServerShowV254Test-608938657 tempest-ServerShowV254Test-608938657-project-member] Lazy-loading 'resources' on Instance uuid 8016bafc-b287-4df1-bd4b-a00210177a86 {{(pid=61806) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 699.207921] env[61806]: DEBUG nova.compute.manager [None req-b4447f02-a3c7-4b6e-89ba-6bc34ec03205 tempest-InstanceActionsTestJSON-949444324 tempest-InstanceActionsTestJSON-949444324-project-member] [instance: 2681766c-0468-4170-8c51-03389e2bd1ab] Build of instance 2681766c-0468-4170-8c51-03389e2bd1ab was re-scheduled: Binding failed for port 59f4de97-5864-4576-82b3-a14d8b5224cd, please check neutron logs for more information. {{(pid=61806) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 699.208339] env[61806]: DEBUG nova.compute.manager [None req-b4447f02-a3c7-4b6e-89ba-6bc34ec03205 tempest-InstanceActionsTestJSON-949444324 tempest-InstanceActionsTestJSON-949444324-project-member] [instance: 2681766c-0468-4170-8c51-03389e2bd1ab] Unplugging VIFs for instance {{(pid=61806) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 699.208564] env[61806]: DEBUG oslo_concurrency.lockutils [None req-b4447f02-a3c7-4b6e-89ba-6bc34ec03205 tempest-InstanceActionsTestJSON-949444324 tempest-InstanceActionsTestJSON-949444324-project-member] Acquiring lock "refresh_cache-2681766c-0468-4170-8c51-03389e2bd1ab" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 699.208711] env[61806]: DEBUG oslo_concurrency.lockutils [None req-b4447f02-a3c7-4b6e-89ba-6bc34ec03205 tempest-InstanceActionsTestJSON-949444324 tempest-InstanceActionsTestJSON-949444324-project-member] Acquired lock "refresh_cache-2681766c-0468-4170-8c51-03389e2bd1ab" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 699.208906] env[61806]: DEBUG nova.network.neutron [None req-b4447f02-a3c7-4b6e-89ba-6bc34ec03205 tempest-InstanceActionsTestJSON-949444324 tempest-InstanceActionsTestJSON-949444324-project-member] [instance: 2681766c-0468-4170-8c51-03389e2bd1ab] Building network info cache for instance {{(pid=61806) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 699.259545] env[61806]: DEBUG nova.compute.manager [None req-05ef4399-8310-428c-ba55-af11abf4e445 tempest-SecurityGroupsTestJSON-1213485581 tempest-SecurityGroupsTestJSON-1213485581-project-member] [instance: 294adb58-f23e-4510-a25a-de6b909e3189] Start spawning the instance on the hypervisor. {{(pid=61806) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 699.284710] env[61806]: DEBUG nova.virt.hardware [None req-05ef4399-8310-428c-ba55-af11abf4e445 tempest-SecurityGroupsTestJSON-1213485581 tempest-SecurityGroupsTestJSON-1213485581-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-15T18:05:10Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-15T18:04:52Z,direct_url=,disk_format='vmdk',id=ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='84ba83fce2c349a6988173c5d6fc3b07',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-15T18:04:53Z,virtual_size=,visibility=), allow threads: False {{(pid=61806) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 699.284812] env[61806]: DEBUG nova.virt.hardware [None req-05ef4399-8310-428c-ba55-af11abf4e445 tempest-SecurityGroupsTestJSON-1213485581 tempest-SecurityGroupsTestJSON-1213485581-project-member] Flavor limits 0:0:0 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 699.284956] env[61806]: DEBUG nova.virt.hardware [None req-05ef4399-8310-428c-ba55-af11abf4e445 tempest-SecurityGroupsTestJSON-1213485581 tempest-SecurityGroupsTestJSON-1213485581-project-member] Image limits 0:0:0 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 699.285153] env[61806]: DEBUG nova.virt.hardware [None req-05ef4399-8310-428c-ba55-af11abf4e445 tempest-SecurityGroupsTestJSON-1213485581 tempest-SecurityGroupsTestJSON-1213485581-project-member] Flavor pref 0:0:0 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 699.285295] env[61806]: DEBUG nova.virt.hardware [None req-05ef4399-8310-428c-ba55-af11abf4e445 tempest-SecurityGroupsTestJSON-1213485581 tempest-SecurityGroupsTestJSON-1213485581-project-member] Image pref 0:0:0 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 699.285654] env[61806]: DEBUG nova.virt.hardware [None req-05ef4399-8310-428c-ba55-af11abf4e445 tempest-SecurityGroupsTestJSON-1213485581 tempest-SecurityGroupsTestJSON-1213485581-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 699.285654] env[61806]: DEBUG nova.virt.hardware [None req-05ef4399-8310-428c-ba55-af11abf4e445 tempest-SecurityGroupsTestJSON-1213485581 tempest-SecurityGroupsTestJSON-1213485581-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61806) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 699.285881] env[61806]: DEBUG nova.virt.hardware [None req-05ef4399-8310-428c-ba55-af11abf4e445 tempest-SecurityGroupsTestJSON-1213485581 tempest-SecurityGroupsTestJSON-1213485581-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61806) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 699.285969] env[61806]: DEBUG nova.virt.hardware [None req-05ef4399-8310-428c-ba55-af11abf4e445 tempest-SecurityGroupsTestJSON-1213485581 tempest-SecurityGroupsTestJSON-1213485581-project-member] Got 1 possible topologies {{(pid=61806) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 699.286161] env[61806]: DEBUG nova.virt.hardware [None req-05ef4399-8310-428c-ba55-af11abf4e445 tempest-SecurityGroupsTestJSON-1213485581 tempest-SecurityGroupsTestJSON-1213485581-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61806) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 699.286329] env[61806]: DEBUG nova.virt.hardware [None req-05ef4399-8310-428c-ba55-af11abf4e445 tempest-SecurityGroupsTestJSON-1213485581 tempest-SecurityGroupsTestJSON-1213485581-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61806) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 699.287212] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0006eebb-ec21-479c-957b-5457ad8e1bea {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 699.291875] env[61806]: DEBUG nova.compute.manager [req-01e199d9-c3a9-437b-b5a2-6ef8a8e5a794 req-778c220a-5391-40ff-921d-df748c9e28a4 service nova] [instance: 294adb58-f23e-4510-a25a-de6b909e3189] Received event network-changed-5fb92c74-bcc1-4423-9d6b-394b8c87bb92 {{(pid=61806) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 699.292143] env[61806]: DEBUG nova.compute.manager [req-01e199d9-c3a9-437b-b5a2-6ef8a8e5a794 req-778c220a-5391-40ff-921d-df748c9e28a4 service nova] [instance: 294adb58-f23e-4510-a25a-de6b909e3189] Refreshing instance network info cache due to event network-changed-5fb92c74-bcc1-4423-9d6b-394b8c87bb92. {{(pid=61806) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 699.292378] env[61806]: DEBUG oslo_concurrency.lockutils [req-01e199d9-c3a9-437b-b5a2-6ef8a8e5a794 req-778c220a-5391-40ff-921d-df748c9e28a4 service nova] Acquiring lock "refresh_cache-294adb58-f23e-4510-a25a-de6b909e3189" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 699.292522] env[61806]: DEBUG oslo_concurrency.lockutils [req-01e199d9-c3a9-437b-b5a2-6ef8a8e5a794 req-778c220a-5391-40ff-921d-df748c9e28a4 service nova] Acquired lock "refresh_cache-294adb58-f23e-4510-a25a-de6b909e3189" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 699.292686] env[61806]: DEBUG nova.network.neutron [req-01e199d9-c3a9-437b-b5a2-6ef8a8e5a794 req-778c220a-5391-40ff-921d-df748c9e28a4 service nova] [instance: 294adb58-f23e-4510-a25a-de6b909e3189] Refreshing network info cache for port 5fb92c74-bcc1-4423-9d6b-394b8c87bb92 {{(pid=61806) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 699.299758] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8e145db1-2087-4cb6-813c-67f58f4a5ca4 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 699.317761] env[61806]: ERROR nova.compute.manager [None req-05ef4399-8310-428c-ba55-af11abf4e445 tempest-SecurityGroupsTestJSON-1213485581 tempest-SecurityGroupsTestJSON-1213485581-project-member] [instance: 294adb58-f23e-4510-a25a-de6b909e3189] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 5fb92c74-bcc1-4423-9d6b-394b8c87bb92, please check neutron logs for more information. [ 699.317761] env[61806]: ERROR nova.compute.manager [instance: 294adb58-f23e-4510-a25a-de6b909e3189] Traceback (most recent call last): [ 699.317761] env[61806]: ERROR nova.compute.manager [instance: 294adb58-f23e-4510-a25a-de6b909e3189] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 699.317761] env[61806]: ERROR nova.compute.manager [instance: 294adb58-f23e-4510-a25a-de6b909e3189] yield resources [ 699.317761] env[61806]: ERROR nova.compute.manager [instance: 294adb58-f23e-4510-a25a-de6b909e3189] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 699.317761] env[61806]: ERROR nova.compute.manager [instance: 294adb58-f23e-4510-a25a-de6b909e3189] self.driver.spawn(context, instance, image_meta, [ 699.317761] env[61806]: ERROR nova.compute.manager [instance: 294adb58-f23e-4510-a25a-de6b909e3189] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 699.317761] env[61806]: ERROR nova.compute.manager [instance: 294adb58-f23e-4510-a25a-de6b909e3189] self._vmops.spawn(context, instance, image_meta, injected_files, [ 699.317761] env[61806]: ERROR nova.compute.manager [instance: 294adb58-f23e-4510-a25a-de6b909e3189] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 699.317761] env[61806]: ERROR nova.compute.manager [instance: 294adb58-f23e-4510-a25a-de6b909e3189] vm_ref = self.build_virtual_machine(instance, [ 699.317761] env[61806]: ERROR nova.compute.manager [instance: 294adb58-f23e-4510-a25a-de6b909e3189] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 699.318217] env[61806]: ERROR nova.compute.manager [instance: 294adb58-f23e-4510-a25a-de6b909e3189] vif_infos = vmwarevif.get_vif_info(self._session, [ 699.318217] env[61806]: ERROR nova.compute.manager [instance: 294adb58-f23e-4510-a25a-de6b909e3189] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 699.318217] env[61806]: ERROR nova.compute.manager [instance: 294adb58-f23e-4510-a25a-de6b909e3189] for vif in network_info: [ 699.318217] env[61806]: ERROR nova.compute.manager [instance: 294adb58-f23e-4510-a25a-de6b909e3189] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 699.318217] env[61806]: ERROR nova.compute.manager [instance: 294adb58-f23e-4510-a25a-de6b909e3189] return self._sync_wrapper(fn, *args, **kwargs) [ 699.318217] env[61806]: ERROR nova.compute.manager [instance: 294adb58-f23e-4510-a25a-de6b909e3189] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 699.318217] env[61806]: ERROR nova.compute.manager [instance: 294adb58-f23e-4510-a25a-de6b909e3189] self.wait() [ 699.318217] env[61806]: ERROR nova.compute.manager [instance: 294adb58-f23e-4510-a25a-de6b909e3189] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 699.318217] env[61806]: ERROR nova.compute.manager [instance: 294adb58-f23e-4510-a25a-de6b909e3189] self[:] = self._gt.wait() [ 699.318217] env[61806]: ERROR nova.compute.manager [instance: 294adb58-f23e-4510-a25a-de6b909e3189] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 699.318217] env[61806]: ERROR nova.compute.manager [instance: 294adb58-f23e-4510-a25a-de6b909e3189] return self._exit_event.wait() [ 699.318217] env[61806]: ERROR nova.compute.manager [instance: 294adb58-f23e-4510-a25a-de6b909e3189] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 699.318217] env[61806]: ERROR nova.compute.manager [instance: 294adb58-f23e-4510-a25a-de6b909e3189] current.throw(*self._exc) [ 699.318586] env[61806]: ERROR nova.compute.manager [instance: 294adb58-f23e-4510-a25a-de6b909e3189] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 699.318586] env[61806]: ERROR nova.compute.manager [instance: 294adb58-f23e-4510-a25a-de6b909e3189] result = function(*args, **kwargs) [ 699.318586] env[61806]: ERROR nova.compute.manager [instance: 294adb58-f23e-4510-a25a-de6b909e3189] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 699.318586] env[61806]: ERROR nova.compute.manager [instance: 294adb58-f23e-4510-a25a-de6b909e3189] return func(*args, **kwargs) [ 699.318586] env[61806]: ERROR nova.compute.manager [instance: 294adb58-f23e-4510-a25a-de6b909e3189] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 699.318586] env[61806]: ERROR nova.compute.manager [instance: 294adb58-f23e-4510-a25a-de6b909e3189] raise e [ 699.318586] env[61806]: ERROR nova.compute.manager [instance: 294adb58-f23e-4510-a25a-de6b909e3189] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 699.318586] env[61806]: ERROR nova.compute.manager [instance: 294adb58-f23e-4510-a25a-de6b909e3189] nwinfo = self.network_api.allocate_for_instance( [ 699.318586] env[61806]: ERROR nova.compute.manager [instance: 294adb58-f23e-4510-a25a-de6b909e3189] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 699.318586] env[61806]: ERROR nova.compute.manager [instance: 294adb58-f23e-4510-a25a-de6b909e3189] created_port_ids = self._update_ports_for_instance( [ 699.318586] env[61806]: ERROR nova.compute.manager [instance: 294adb58-f23e-4510-a25a-de6b909e3189] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 699.318586] env[61806]: ERROR nova.compute.manager [instance: 294adb58-f23e-4510-a25a-de6b909e3189] with excutils.save_and_reraise_exception(): [ 699.318586] env[61806]: ERROR nova.compute.manager [instance: 294adb58-f23e-4510-a25a-de6b909e3189] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 699.318996] env[61806]: ERROR nova.compute.manager [instance: 294adb58-f23e-4510-a25a-de6b909e3189] self.force_reraise() [ 699.318996] env[61806]: ERROR nova.compute.manager [instance: 294adb58-f23e-4510-a25a-de6b909e3189] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 699.318996] env[61806]: ERROR nova.compute.manager [instance: 294adb58-f23e-4510-a25a-de6b909e3189] raise self.value [ 699.318996] env[61806]: ERROR nova.compute.manager [instance: 294adb58-f23e-4510-a25a-de6b909e3189] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 699.318996] env[61806]: ERROR nova.compute.manager [instance: 294adb58-f23e-4510-a25a-de6b909e3189] updated_port = self._update_port( [ 699.318996] env[61806]: ERROR nova.compute.manager [instance: 294adb58-f23e-4510-a25a-de6b909e3189] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 699.318996] env[61806]: ERROR nova.compute.manager [instance: 294adb58-f23e-4510-a25a-de6b909e3189] _ensure_no_port_binding_failure(port) [ 699.318996] env[61806]: ERROR nova.compute.manager [instance: 294adb58-f23e-4510-a25a-de6b909e3189] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 699.318996] env[61806]: ERROR nova.compute.manager [instance: 294adb58-f23e-4510-a25a-de6b909e3189] raise exception.PortBindingFailed(port_id=port['id']) [ 699.318996] env[61806]: ERROR nova.compute.manager [instance: 294adb58-f23e-4510-a25a-de6b909e3189] nova.exception.PortBindingFailed: Binding failed for port 5fb92c74-bcc1-4423-9d6b-394b8c87bb92, please check neutron logs for more information. [ 699.318996] env[61806]: ERROR nova.compute.manager [instance: 294adb58-f23e-4510-a25a-de6b909e3189] [ 699.318996] env[61806]: INFO nova.compute.manager [None req-05ef4399-8310-428c-ba55-af11abf4e445 tempest-SecurityGroupsTestJSON-1213485581 tempest-SecurityGroupsTestJSON-1213485581-project-member] [instance: 294adb58-f23e-4510-a25a-de6b909e3189] Terminating instance [ 699.320288] env[61806]: DEBUG oslo_concurrency.lockutils [None req-05ef4399-8310-428c-ba55-af11abf4e445 tempest-SecurityGroupsTestJSON-1213485581 tempest-SecurityGroupsTestJSON-1213485581-project-member] Acquiring lock "refresh_cache-294adb58-f23e-4510-a25a-de6b909e3189" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 699.729180] env[61806]: DEBUG nova.network.neutron [None req-b4447f02-a3c7-4b6e-89ba-6bc34ec03205 tempest-InstanceActionsTestJSON-949444324 tempest-InstanceActionsTestJSON-949444324-project-member] [instance: 2681766c-0468-4170-8c51-03389e2bd1ab] Instance cache missing network info. {{(pid=61806) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 699.812639] env[61806]: DEBUG nova.network.neutron [req-01e199d9-c3a9-437b-b5a2-6ef8a8e5a794 req-778c220a-5391-40ff-921d-df748c9e28a4 service nova] [instance: 294adb58-f23e-4510-a25a-de6b909e3189] Instance cache missing network info. {{(pid=61806) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 699.850075] env[61806]: DEBUG nova.network.neutron [None req-b4447f02-a3c7-4b6e-89ba-6bc34ec03205 tempest-InstanceActionsTestJSON-949444324 tempest-InstanceActionsTestJSON-949444324-project-member] [instance: 2681766c-0468-4170-8c51-03389e2bd1ab] Updating instance_info_cache with network_info: [] {{(pid=61806) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 699.934210] env[61806]: DEBUG nova.network.neutron [req-01e199d9-c3a9-437b-b5a2-6ef8a8e5a794 req-778c220a-5391-40ff-921d-df748c9e28a4 service nova] [instance: 294adb58-f23e-4510-a25a-de6b909e3189] Updating instance_info_cache with network_info: [] {{(pid=61806) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 700.103209] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-73a21854-f32c-470e-a4cb-2ec4ca15447c {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 700.110884] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-535a800b-eab9-4cda-a1ab-719552c54056 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 700.139861] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6ee824f0-3a94-4785-a113-b4392fc9ca9d {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 700.147607] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-10a7105d-026e-48a9-ab1d-b26f7b3b666a {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 700.160653] env[61806]: DEBUG nova.compute.provider_tree [None req-6dff00ec-9995-4616-9033-47ec4f3592c9 tempest-ServerShowV254Test-608938657 tempest-ServerShowV254Test-608938657-project-member] Inventory has not changed in ProviderTree for provider: a2858be1-fd22-4e08-979e-87ad25293407 {{(pid=61806) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 700.353739] env[61806]: DEBUG oslo_concurrency.lockutils [None req-b4447f02-a3c7-4b6e-89ba-6bc34ec03205 tempest-InstanceActionsTestJSON-949444324 tempest-InstanceActionsTestJSON-949444324-project-member] Releasing lock "refresh_cache-2681766c-0468-4170-8c51-03389e2bd1ab" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 700.353992] env[61806]: DEBUG nova.compute.manager [None req-b4447f02-a3c7-4b6e-89ba-6bc34ec03205 tempest-InstanceActionsTestJSON-949444324 tempest-InstanceActionsTestJSON-949444324-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61806) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 700.354206] env[61806]: DEBUG nova.compute.manager [None req-b4447f02-a3c7-4b6e-89ba-6bc34ec03205 tempest-InstanceActionsTestJSON-949444324 tempest-InstanceActionsTestJSON-949444324-project-member] [instance: 2681766c-0468-4170-8c51-03389e2bd1ab] Deallocating network for instance {{(pid=61806) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 700.354373] env[61806]: DEBUG nova.network.neutron [None req-b4447f02-a3c7-4b6e-89ba-6bc34ec03205 tempest-InstanceActionsTestJSON-949444324 tempest-InstanceActionsTestJSON-949444324-project-member] [instance: 2681766c-0468-4170-8c51-03389e2bd1ab] deallocate_for_instance() {{(pid=61806) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 700.371865] env[61806]: DEBUG nova.network.neutron [None req-b4447f02-a3c7-4b6e-89ba-6bc34ec03205 tempest-InstanceActionsTestJSON-949444324 tempest-InstanceActionsTestJSON-949444324-project-member] [instance: 2681766c-0468-4170-8c51-03389e2bd1ab] Instance cache missing network info. {{(pid=61806) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 700.437525] env[61806]: DEBUG oslo_concurrency.lockutils [req-01e199d9-c3a9-437b-b5a2-6ef8a8e5a794 req-778c220a-5391-40ff-921d-df748c9e28a4 service nova] Releasing lock "refresh_cache-294adb58-f23e-4510-a25a-de6b909e3189" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 700.437952] env[61806]: DEBUG nova.compute.manager [req-01e199d9-c3a9-437b-b5a2-6ef8a8e5a794 req-778c220a-5391-40ff-921d-df748c9e28a4 service nova] [instance: 294adb58-f23e-4510-a25a-de6b909e3189] Received event network-vif-deleted-5fb92c74-bcc1-4423-9d6b-394b8c87bb92 {{(pid=61806) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 700.438695] env[61806]: DEBUG oslo_concurrency.lockutils [None req-05ef4399-8310-428c-ba55-af11abf4e445 tempest-SecurityGroupsTestJSON-1213485581 tempest-SecurityGroupsTestJSON-1213485581-project-member] Acquired lock "refresh_cache-294adb58-f23e-4510-a25a-de6b909e3189" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 700.438695] env[61806]: DEBUG nova.network.neutron [None req-05ef4399-8310-428c-ba55-af11abf4e445 tempest-SecurityGroupsTestJSON-1213485581 tempest-SecurityGroupsTestJSON-1213485581-project-member] [instance: 294adb58-f23e-4510-a25a-de6b909e3189] Building network info cache for instance {{(pid=61806) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 700.663929] env[61806]: DEBUG nova.scheduler.client.report [None req-6dff00ec-9995-4616-9033-47ec4f3592c9 tempest-ServerShowV254Test-608938657 tempest-ServerShowV254Test-608938657-project-member] Inventory has not changed for provider a2858be1-fd22-4e08-979e-87ad25293407 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 139, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61806) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 700.874832] env[61806]: DEBUG nova.network.neutron [None req-b4447f02-a3c7-4b6e-89ba-6bc34ec03205 tempest-InstanceActionsTestJSON-949444324 tempest-InstanceActionsTestJSON-949444324-project-member] [instance: 2681766c-0468-4170-8c51-03389e2bd1ab] Updating instance_info_cache with network_info: [] {{(pid=61806) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 700.963307] env[61806]: DEBUG nova.network.neutron [None req-05ef4399-8310-428c-ba55-af11abf4e445 tempest-SecurityGroupsTestJSON-1213485581 tempest-SecurityGroupsTestJSON-1213485581-project-member] [instance: 294adb58-f23e-4510-a25a-de6b909e3189] Instance cache missing network info. {{(pid=61806) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 701.038036] env[61806]: DEBUG nova.network.neutron [None req-05ef4399-8310-428c-ba55-af11abf4e445 tempest-SecurityGroupsTestJSON-1213485581 tempest-SecurityGroupsTestJSON-1213485581-project-member] [instance: 294adb58-f23e-4510-a25a-de6b909e3189] Updating instance_info_cache with network_info: [] {{(pid=61806) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 701.169954] env[61806]: DEBUG oslo_concurrency.lockutils [None req-6dff00ec-9995-4616-9033-47ec4f3592c9 tempest-ServerShowV254Test-608938657 tempest-ServerShowV254Test-608938657-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.964s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 701.172278] env[61806]: DEBUG oslo_concurrency.lockutils [None req-6f137976-e023-4899-ab65-afdf6c94ed87 tempest-ServersAdmin275Test-1945525728 tempest-ServersAdmin275Test-1945525728-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: waited 32.088s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 701.172458] env[61806]: DEBUG nova.objects.instance [None req-6f137976-e023-4899-ab65-afdf6c94ed87 tempest-ServersAdmin275Test-1945525728 tempest-ServersAdmin275Test-1945525728-project-member] [instance: 9db80c3c-e2c3-4034-a79a-d7be924c4ff6] Trying to apply a migration context that does not seem to be set for this instance {{(pid=61806) apply_migration_context /opt/stack/nova/nova/objects/instance.py:1067}} [ 701.188292] env[61806]: INFO nova.scheduler.client.report [None req-6dff00ec-9995-4616-9033-47ec4f3592c9 tempest-ServerShowV254Test-608938657 tempest-ServerShowV254Test-608938657-project-member] Deleted allocations for instance 8016bafc-b287-4df1-bd4b-a00210177a86 [ 701.378093] env[61806]: INFO nova.compute.manager [None req-b4447f02-a3c7-4b6e-89ba-6bc34ec03205 tempest-InstanceActionsTestJSON-949444324 tempest-InstanceActionsTestJSON-949444324-project-member] [instance: 2681766c-0468-4170-8c51-03389e2bd1ab] Took 1.02 seconds to deallocate network for instance. [ 701.544214] env[61806]: DEBUG oslo_concurrency.lockutils [None req-05ef4399-8310-428c-ba55-af11abf4e445 tempest-SecurityGroupsTestJSON-1213485581 tempest-SecurityGroupsTestJSON-1213485581-project-member] Releasing lock "refresh_cache-294adb58-f23e-4510-a25a-de6b909e3189" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 701.544333] env[61806]: DEBUG nova.compute.manager [None req-05ef4399-8310-428c-ba55-af11abf4e445 tempest-SecurityGroupsTestJSON-1213485581 tempest-SecurityGroupsTestJSON-1213485581-project-member] [instance: 294adb58-f23e-4510-a25a-de6b909e3189] Start destroying the instance on the hypervisor. {{(pid=61806) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 701.544520] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-05ef4399-8310-428c-ba55-af11abf4e445 tempest-SecurityGroupsTestJSON-1213485581 tempest-SecurityGroupsTestJSON-1213485581-project-member] [instance: 294adb58-f23e-4510-a25a-de6b909e3189] Destroying instance {{(pid=61806) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 701.544824] env[61806]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-6d54b416-1e28-4759-a1c0-0a59d9dc0f4b {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 701.555025] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-40c526f4-9cc7-475e-bdc2-5966fb06e8a1 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 701.576308] env[61806]: WARNING nova.virt.vmwareapi.vmops [None req-05ef4399-8310-428c-ba55-af11abf4e445 tempest-SecurityGroupsTestJSON-1213485581 tempest-SecurityGroupsTestJSON-1213485581-project-member] [instance: 294adb58-f23e-4510-a25a-de6b909e3189] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 294adb58-f23e-4510-a25a-de6b909e3189 could not be found. [ 701.577146] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-05ef4399-8310-428c-ba55-af11abf4e445 tempest-SecurityGroupsTestJSON-1213485581 tempest-SecurityGroupsTestJSON-1213485581-project-member] [instance: 294adb58-f23e-4510-a25a-de6b909e3189] Instance destroyed {{(pid=61806) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 701.577146] env[61806]: INFO nova.compute.manager [None req-05ef4399-8310-428c-ba55-af11abf4e445 tempest-SecurityGroupsTestJSON-1213485581 tempest-SecurityGroupsTestJSON-1213485581-project-member] [instance: 294adb58-f23e-4510-a25a-de6b909e3189] Took 0.03 seconds to destroy the instance on the hypervisor. [ 701.577146] env[61806]: DEBUG oslo.service.loopingcall [None req-05ef4399-8310-428c-ba55-af11abf4e445 tempest-SecurityGroupsTestJSON-1213485581 tempest-SecurityGroupsTestJSON-1213485581-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61806) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 701.577286] env[61806]: DEBUG nova.compute.manager [-] [instance: 294adb58-f23e-4510-a25a-de6b909e3189] Deallocating network for instance {{(pid=61806) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 701.577286] env[61806]: DEBUG nova.network.neutron [-] [instance: 294adb58-f23e-4510-a25a-de6b909e3189] deallocate_for_instance() {{(pid=61806) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 701.592979] env[61806]: DEBUG nova.network.neutron [-] [instance: 294adb58-f23e-4510-a25a-de6b909e3189] Instance cache missing network info. {{(pid=61806) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 701.695415] env[61806]: DEBUG oslo_concurrency.lockutils [None req-6dff00ec-9995-4616-9033-47ec4f3592c9 tempest-ServerShowV254Test-608938657 tempest-ServerShowV254Test-608938657-project-member] Lock "8016bafc-b287-4df1-bd4b-a00210177a86" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 39.575s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 702.095754] env[61806]: DEBUG nova.network.neutron [-] [instance: 294adb58-f23e-4510-a25a-de6b909e3189] Updating instance_info_cache with network_info: [] {{(pid=61806) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 702.184947] env[61806]: DEBUG oslo_concurrency.lockutils [None req-6f137976-e023-4899-ab65-afdf6c94ed87 tempest-ServersAdmin275Test-1945525728 tempest-ServersAdmin275Test-1945525728-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: held 1.012s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 702.185794] env[61806]: DEBUG oslo_concurrency.lockutils [None req-c9d6a9df-bcc9-4e08-88fa-de36a15b3a76 tempest-ServersAdmin275Test-894191221 tempest-ServersAdmin275Test-894191221-project-admin] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: waited 24.017s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 702.186054] env[61806]: DEBUG nova.objects.instance [None req-c9d6a9df-bcc9-4e08-88fa-de36a15b3a76 tempest-ServersAdmin275Test-894191221 tempest-ServersAdmin275Test-894191221-project-admin] [instance: 9db80c3c-e2c3-4034-a79a-d7be924c4ff6] Trying to apply a migration context that does not seem to be set for this instance {{(pid=61806) apply_migration_context /opt/stack/nova/nova/objects/instance.py:1067}} [ 702.412991] env[61806]: INFO nova.scheduler.client.report [None req-b4447f02-a3c7-4b6e-89ba-6bc34ec03205 tempest-InstanceActionsTestJSON-949444324 tempest-InstanceActionsTestJSON-949444324-project-member] Deleted allocations for instance 2681766c-0468-4170-8c51-03389e2bd1ab [ 702.600250] env[61806]: INFO nova.compute.manager [-] [instance: 294adb58-f23e-4510-a25a-de6b909e3189] Took 1.02 seconds to deallocate network for instance. [ 702.606215] env[61806]: DEBUG nova.compute.claims [None req-05ef4399-8310-428c-ba55-af11abf4e445 tempest-SecurityGroupsTestJSON-1213485581 tempest-SecurityGroupsTestJSON-1213485581-project-member] [instance: 294adb58-f23e-4510-a25a-de6b909e3189] Aborting claim: {{(pid=61806) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 702.606598] env[61806]: DEBUG oslo_concurrency.lockutils [None req-05ef4399-8310-428c-ba55-af11abf4e445 tempest-SecurityGroupsTestJSON-1213485581 tempest-SecurityGroupsTestJSON-1213485581-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 702.922028] env[61806]: DEBUG oslo_concurrency.lockutils [None req-b4447f02-a3c7-4b6e-89ba-6bc34ec03205 tempest-InstanceActionsTestJSON-949444324 tempest-InstanceActionsTestJSON-949444324-project-member] Lock "2681766c-0468-4170-8c51-03389e2bd1ab" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 139.296s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 703.205771] env[61806]: DEBUG oslo_concurrency.lockutils [None req-c9d6a9df-bcc9-4e08-88fa-de36a15b3a76 tempest-ServersAdmin275Test-894191221 tempest-ServersAdmin275Test-894191221-project-admin] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: held 1.020s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 703.206827] env[61806]: DEBUG oslo_concurrency.lockutils [None req-fcbc4cf1-fc11-42b1-9d49-ef556b52782e tempest-ServerRescueNegativeTestJSON-1483258717 tempest-ServerRescueNegativeTestJSON-1483258717-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 22.499s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 703.208282] env[61806]: INFO nova.compute.claims [None req-fcbc4cf1-fc11-42b1-9d49-ef556b52782e tempest-ServerRescueNegativeTestJSON-1483258717 tempest-ServerRescueNegativeTestJSON-1483258717-project-member] [instance: 7ec2497a-7329-4ed4-99ab-dec3d1e8117c] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 703.424534] env[61806]: DEBUG nova.compute.manager [None req-3c3981d2-fa25-4f4b-87d4-a5411e67199e tempest-MultipleCreateTestJSON-1589015462 tempest-MultipleCreateTestJSON-1589015462-project-member] [instance: 5b22c2fc-525c-481b-b84e-e0a7f68f633e] Starting instance... {{(pid=61806) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 703.955178] env[61806]: DEBUG oslo_concurrency.lockutils [None req-3c3981d2-fa25-4f4b-87d4-a5411e67199e tempest-MultipleCreateTestJSON-1589015462 tempest-MultipleCreateTestJSON-1589015462-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 704.610047] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-192d2b5a-252e-4c53-b3e2-3e0f72635a46 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 704.617732] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-01c44b14-f2f9-42e1-b194-96f766e32f5c {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 704.646913] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2dbbf1bb-22c1-4984-811f-3cad2636713e {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 704.653712] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-43517a9a-af92-4e93-a4c3-84cfbcbc914f {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 704.667565] env[61806]: DEBUG nova.compute.provider_tree [None req-fcbc4cf1-fc11-42b1-9d49-ef556b52782e tempest-ServerRescueNegativeTestJSON-1483258717 tempest-ServerRescueNegativeTestJSON-1483258717-project-member] Updating inventory in ProviderTree for provider a2858be1-fd22-4e08-979e-87ad25293407 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 138, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61806) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 705.192794] env[61806]: ERROR nova.scheduler.client.report [None req-fcbc4cf1-fc11-42b1-9d49-ef556b52782e tempest-ServerRescueNegativeTestJSON-1483258717 tempest-ServerRescueNegativeTestJSON-1483258717-project-member] [req-a7c495c1-5bfa-4405-ab8c-52fb14809e6e] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 138, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID a2858be1-fd22-4e08-979e-87ad25293407. Got 409: {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-a7c495c1-5bfa-4405-ab8c-52fb14809e6e"}]} [ 705.211406] env[61806]: DEBUG nova.scheduler.client.report [None req-fcbc4cf1-fc11-42b1-9d49-ef556b52782e tempest-ServerRescueNegativeTestJSON-1483258717 tempest-ServerRescueNegativeTestJSON-1483258717-project-member] Refreshing inventories for resource provider a2858be1-fd22-4e08-979e-87ad25293407 {{(pid=61806) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:818}} [ 705.225711] env[61806]: DEBUG nova.scheduler.client.report [None req-fcbc4cf1-fc11-42b1-9d49-ef556b52782e tempest-ServerRescueNegativeTestJSON-1483258717 tempest-ServerRescueNegativeTestJSON-1483258717-project-member] Updating ProviderTree inventory for provider a2858be1-fd22-4e08-979e-87ad25293407 from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 139, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61806) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:782}} [ 705.225883] env[61806]: DEBUG nova.compute.provider_tree [None req-fcbc4cf1-fc11-42b1-9d49-ef556b52782e tempest-ServerRescueNegativeTestJSON-1483258717 tempest-ServerRescueNegativeTestJSON-1483258717-project-member] Updating inventory in ProviderTree for provider a2858be1-fd22-4e08-979e-87ad25293407 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 139, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61806) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 705.238023] env[61806]: DEBUG nova.scheduler.client.report [None req-fcbc4cf1-fc11-42b1-9d49-ef556b52782e tempest-ServerRescueNegativeTestJSON-1483258717 tempest-ServerRescueNegativeTestJSON-1483258717-project-member] Refreshing aggregate associations for resource provider a2858be1-fd22-4e08-979e-87ad25293407, aggregates: None {{(pid=61806) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:827}} [ 705.265673] env[61806]: DEBUG nova.scheduler.client.report [None req-fcbc4cf1-fc11-42b1-9d49-ef556b52782e tempest-ServerRescueNegativeTestJSON-1483258717 tempest-ServerRescueNegativeTestJSON-1483258717-project-member] Refreshing trait associations for resource provider a2858be1-fd22-4e08-979e-87ad25293407, traits: COMPUTE_IMAGE_TYPE_ISO,COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_SAME_HOST_COLD_MIGRATE,COMPUTE_NET_ATTACH_INTERFACE,COMPUTE_NODE {{(pid=61806) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:839}} [ 705.627581] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-47f297c4-cb3b-4803-90f7-84043b042173 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 705.635341] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e2f5a8ce-cedb-4cdf-89f1-27c955138a88 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 705.664271] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-46640c48-e7b5-49c8-af66-28947d4793b0 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 705.671196] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ed61ff78-4df3-462c-8f91-79ebff400242 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 705.685107] env[61806]: DEBUG nova.compute.provider_tree [None req-fcbc4cf1-fc11-42b1-9d49-ef556b52782e tempest-ServerRescueNegativeTestJSON-1483258717 tempest-ServerRescueNegativeTestJSON-1483258717-project-member] Updating inventory in ProviderTree for provider a2858be1-fd22-4e08-979e-87ad25293407 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 138, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61806) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 706.221661] env[61806]: DEBUG nova.scheduler.client.report [None req-fcbc4cf1-fc11-42b1-9d49-ef556b52782e tempest-ServerRescueNegativeTestJSON-1483258717 tempest-ServerRescueNegativeTestJSON-1483258717-project-member] Updated inventory for provider a2858be1-fd22-4e08-979e-87ad25293407 with generation 69 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 138, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61806) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:971}} [ 706.221661] env[61806]: DEBUG nova.compute.provider_tree [None req-fcbc4cf1-fc11-42b1-9d49-ef556b52782e tempest-ServerRescueNegativeTestJSON-1483258717 tempest-ServerRescueNegativeTestJSON-1483258717-project-member] Updating resource provider a2858be1-fd22-4e08-979e-87ad25293407 generation from 69 to 70 during operation: update_inventory {{(pid=61806) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 706.221661] env[61806]: DEBUG nova.compute.provider_tree [None req-fcbc4cf1-fc11-42b1-9d49-ef556b52782e tempest-ServerRescueNegativeTestJSON-1483258717 tempest-ServerRescueNegativeTestJSON-1483258717-project-member] Updating inventory in ProviderTree for provider a2858be1-fd22-4e08-979e-87ad25293407 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 138, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61806) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 706.725107] env[61806]: DEBUG oslo_concurrency.lockutils [None req-fcbc4cf1-fc11-42b1-9d49-ef556b52782e tempest-ServerRescueNegativeTestJSON-1483258717 tempest-ServerRescueNegativeTestJSON-1483258717-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 3.518s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 706.725632] env[61806]: DEBUG nova.compute.manager [None req-fcbc4cf1-fc11-42b1-9d49-ef556b52782e tempest-ServerRescueNegativeTestJSON-1483258717 tempest-ServerRescueNegativeTestJSON-1483258717-project-member] [instance: 7ec2497a-7329-4ed4-99ab-dec3d1e8117c] Start building networks asynchronously for instance. {{(pid=61806) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 706.729965] env[61806]: DEBUG oslo_concurrency.lockutils [None req-6168822a-f60c-43e0-b0ac-2eb07b4977dc tempest-ServersAdmin275Test-1945525728 tempest-ServersAdmin275Test-1945525728-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 23.768s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 706.730202] env[61806]: DEBUG nova.objects.instance [None req-6168822a-f60c-43e0-b0ac-2eb07b4977dc tempest-ServersAdmin275Test-1945525728 tempest-ServersAdmin275Test-1945525728-project-member] Lazy-loading 'resources' on Instance uuid 9db80c3c-e2c3-4034-a79a-d7be924c4ff6 {{(pid=61806) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 707.235527] env[61806]: DEBUG nova.compute.utils [None req-fcbc4cf1-fc11-42b1-9d49-ef556b52782e tempest-ServerRescueNegativeTestJSON-1483258717 tempest-ServerRescueNegativeTestJSON-1483258717-project-member] Using /dev/sd instead of None {{(pid=61806) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 707.237067] env[61806]: DEBUG nova.compute.manager [None req-fcbc4cf1-fc11-42b1-9d49-ef556b52782e tempest-ServerRescueNegativeTestJSON-1483258717 tempest-ServerRescueNegativeTestJSON-1483258717-project-member] [instance: 7ec2497a-7329-4ed4-99ab-dec3d1e8117c] Allocating IP information in the background. {{(pid=61806) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 707.237244] env[61806]: DEBUG nova.network.neutron [None req-fcbc4cf1-fc11-42b1-9d49-ef556b52782e tempest-ServerRescueNegativeTestJSON-1483258717 tempest-ServerRescueNegativeTestJSON-1483258717-project-member] [instance: 7ec2497a-7329-4ed4-99ab-dec3d1e8117c] allocate_for_instance() {{(pid=61806) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 707.298970] env[61806]: DEBUG nova.policy [None req-fcbc4cf1-fc11-42b1-9d49-ef556b52782e tempest-ServerRescueNegativeTestJSON-1483258717 tempest-ServerRescueNegativeTestJSON-1483258717-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '5925347d6d584c1e8d7de04a84857859', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '596c39abc1c8448882be50c8f7cf5e9a', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61806) authorize /opt/stack/nova/nova/policy.py:201}} [ 707.536929] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-10401dc0-2fb0-43a8-9833-163011a457f0 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 707.548030] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-05562b56-c1fb-44a8-b594-d9605e487885 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 707.578219] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-81ede90a-b24d-47bd-b8a9-777d1660e689 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 707.586144] env[61806]: DEBUG nova.network.neutron [None req-fcbc4cf1-fc11-42b1-9d49-ef556b52782e tempest-ServerRescueNegativeTestJSON-1483258717 tempest-ServerRescueNegativeTestJSON-1483258717-project-member] [instance: 7ec2497a-7329-4ed4-99ab-dec3d1e8117c] Successfully created port: 81f90dee-a680-4d31-9c3d-fbe42fa742a4 {{(pid=61806) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 707.588752] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3e84ff27-92ae-434e-bf03-522a7812ef20 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 707.603083] env[61806]: DEBUG nova.compute.provider_tree [None req-6168822a-f60c-43e0-b0ac-2eb07b4977dc tempest-ServersAdmin275Test-1945525728 tempest-ServersAdmin275Test-1945525728-project-member] Inventory has not changed in ProviderTree for provider: a2858be1-fd22-4e08-979e-87ad25293407 {{(pid=61806) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 707.740342] env[61806]: DEBUG nova.compute.manager [None req-fcbc4cf1-fc11-42b1-9d49-ef556b52782e tempest-ServerRescueNegativeTestJSON-1483258717 tempest-ServerRescueNegativeTestJSON-1483258717-project-member] [instance: 7ec2497a-7329-4ed4-99ab-dec3d1e8117c] Start building block device mappings for instance. {{(pid=61806) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 708.106939] env[61806]: DEBUG nova.scheduler.client.report [None req-6168822a-f60c-43e0-b0ac-2eb07b4977dc tempest-ServersAdmin275Test-1945525728 tempest-ServersAdmin275Test-1945525728-project-member] Inventory has not changed for provider a2858be1-fd22-4e08-979e-87ad25293407 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 138, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61806) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 708.285248] env[61806]: DEBUG nova.compute.manager [req-93e4637b-7f83-43eb-9292-ad83e501b0cb req-8db490bc-82bc-4188-8b6a-be846272ec19 service nova] [instance: 7ec2497a-7329-4ed4-99ab-dec3d1e8117c] Received event network-changed-81f90dee-a680-4d31-9c3d-fbe42fa742a4 {{(pid=61806) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 708.285248] env[61806]: DEBUG nova.compute.manager [req-93e4637b-7f83-43eb-9292-ad83e501b0cb req-8db490bc-82bc-4188-8b6a-be846272ec19 service nova] [instance: 7ec2497a-7329-4ed4-99ab-dec3d1e8117c] Refreshing instance network info cache due to event network-changed-81f90dee-a680-4d31-9c3d-fbe42fa742a4. {{(pid=61806) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 708.285248] env[61806]: DEBUG oslo_concurrency.lockutils [req-93e4637b-7f83-43eb-9292-ad83e501b0cb req-8db490bc-82bc-4188-8b6a-be846272ec19 service nova] Acquiring lock "refresh_cache-7ec2497a-7329-4ed4-99ab-dec3d1e8117c" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 708.285248] env[61806]: DEBUG oslo_concurrency.lockutils [req-93e4637b-7f83-43eb-9292-ad83e501b0cb req-8db490bc-82bc-4188-8b6a-be846272ec19 service nova] Acquired lock "refresh_cache-7ec2497a-7329-4ed4-99ab-dec3d1e8117c" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 708.285248] env[61806]: DEBUG nova.network.neutron [req-93e4637b-7f83-43eb-9292-ad83e501b0cb req-8db490bc-82bc-4188-8b6a-be846272ec19 service nova] [instance: 7ec2497a-7329-4ed4-99ab-dec3d1e8117c] Refreshing network info cache for port 81f90dee-a680-4d31-9c3d-fbe42fa742a4 {{(pid=61806) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 708.470667] env[61806]: ERROR nova.compute.manager [None req-fcbc4cf1-fc11-42b1-9d49-ef556b52782e tempest-ServerRescueNegativeTestJSON-1483258717 tempest-ServerRescueNegativeTestJSON-1483258717-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 81f90dee-a680-4d31-9c3d-fbe42fa742a4, please check neutron logs for more information. [ 708.470667] env[61806]: ERROR nova.compute.manager Traceback (most recent call last): [ 708.470667] env[61806]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 708.470667] env[61806]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 708.470667] env[61806]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 708.470667] env[61806]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 708.470667] env[61806]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 708.470667] env[61806]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 708.470667] env[61806]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 708.470667] env[61806]: ERROR nova.compute.manager self.force_reraise() [ 708.470667] env[61806]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 708.470667] env[61806]: ERROR nova.compute.manager raise self.value [ 708.470667] env[61806]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 708.470667] env[61806]: ERROR nova.compute.manager updated_port = self._update_port( [ 708.470667] env[61806]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 708.470667] env[61806]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 708.471220] env[61806]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 708.471220] env[61806]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 708.471220] env[61806]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 81f90dee-a680-4d31-9c3d-fbe42fa742a4, please check neutron logs for more information. [ 708.471220] env[61806]: ERROR nova.compute.manager [ 708.471220] env[61806]: Traceback (most recent call last): [ 708.471220] env[61806]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 708.471220] env[61806]: listener.cb(fileno) [ 708.471220] env[61806]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 708.471220] env[61806]: result = function(*args, **kwargs) [ 708.471220] env[61806]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 708.471220] env[61806]: return func(*args, **kwargs) [ 708.471220] env[61806]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 708.471220] env[61806]: raise e [ 708.471220] env[61806]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 708.471220] env[61806]: nwinfo = self.network_api.allocate_for_instance( [ 708.471220] env[61806]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 708.471220] env[61806]: created_port_ids = self._update_ports_for_instance( [ 708.471220] env[61806]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 708.471220] env[61806]: with excutils.save_and_reraise_exception(): [ 708.471220] env[61806]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 708.471220] env[61806]: self.force_reraise() [ 708.471220] env[61806]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 708.471220] env[61806]: raise self.value [ 708.471220] env[61806]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 708.471220] env[61806]: updated_port = self._update_port( [ 708.471220] env[61806]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 708.471220] env[61806]: _ensure_no_port_binding_failure(port) [ 708.471220] env[61806]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 708.471220] env[61806]: raise exception.PortBindingFailed(port_id=port['id']) [ 708.472148] env[61806]: nova.exception.PortBindingFailed: Binding failed for port 81f90dee-a680-4d31-9c3d-fbe42fa742a4, please check neutron logs for more information. [ 708.472148] env[61806]: Removing descriptor: 14 [ 708.611816] env[61806]: DEBUG oslo_concurrency.lockutils [None req-6168822a-f60c-43e0-b0ac-2eb07b4977dc tempest-ServersAdmin275Test-1945525728 tempest-ServersAdmin275Test-1945525728-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.882s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 708.614227] env[61806]: DEBUG oslo_concurrency.lockutils [None req-7bb45eee-0cbe-4995-9898-530114f6f104 tempest-ImagesOneServerNegativeTestJSON-606465252 tempest-ImagesOneServerNegativeTestJSON-606465252-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 25.029s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 708.643088] env[61806]: INFO nova.scheduler.client.report [None req-6168822a-f60c-43e0-b0ac-2eb07b4977dc tempest-ServersAdmin275Test-1945525728 tempest-ServersAdmin275Test-1945525728-project-member] Deleted allocations for instance 9db80c3c-e2c3-4034-a79a-d7be924c4ff6 [ 708.752664] env[61806]: DEBUG nova.compute.manager [None req-fcbc4cf1-fc11-42b1-9d49-ef556b52782e tempest-ServerRescueNegativeTestJSON-1483258717 tempest-ServerRescueNegativeTestJSON-1483258717-project-member] [instance: 7ec2497a-7329-4ed4-99ab-dec3d1e8117c] Start spawning the instance on the hypervisor. {{(pid=61806) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 708.776501] env[61806]: DEBUG nova.virt.hardware [None req-fcbc4cf1-fc11-42b1-9d49-ef556b52782e tempest-ServerRescueNegativeTestJSON-1483258717 tempest-ServerRescueNegativeTestJSON-1483258717-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-15T18:05:10Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-15T18:04:52Z,direct_url=,disk_format='vmdk',id=ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='84ba83fce2c349a6988173c5d6fc3b07',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-15T18:04:53Z,virtual_size=,visibility=), allow threads: False {{(pid=61806) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 708.776770] env[61806]: DEBUG nova.virt.hardware [None req-fcbc4cf1-fc11-42b1-9d49-ef556b52782e tempest-ServerRescueNegativeTestJSON-1483258717 tempest-ServerRescueNegativeTestJSON-1483258717-project-member] Flavor limits 0:0:0 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 708.776927] env[61806]: DEBUG nova.virt.hardware [None req-fcbc4cf1-fc11-42b1-9d49-ef556b52782e tempest-ServerRescueNegativeTestJSON-1483258717 tempest-ServerRescueNegativeTestJSON-1483258717-project-member] Image limits 0:0:0 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 708.777127] env[61806]: DEBUG nova.virt.hardware [None req-fcbc4cf1-fc11-42b1-9d49-ef556b52782e tempest-ServerRescueNegativeTestJSON-1483258717 tempest-ServerRescueNegativeTestJSON-1483258717-project-member] Flavor pref 0:0:0 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 708.777287] env[61806]: DEBUG nova.virt.hardware [None req-fcbc4cf1-fc11-42b1-9d49-ef556b52782e tempest-ServerRescueNegativeTestJSON-1483258717 tempest-ServerRescueNegativeTestJSON-1483258717-project-member] Image pref 0:0:0 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 708.777442] env[61806]: DEBUG nova.virt.hardware [None req-fcbc4cf1-fc11-42b1-9d49-ef556b52782e tempest-ServerRescueNegativeTestJSON-1483258717 tempest-ServerRescueNegativeTestJSON-1483258717-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 708.777649] env[61806]: DEBUG nova.virt.hardware [None req-fcbc4cf1-fc11-42b1-9d49-ef556b52782e tempest-ServerRescueNegativeTestJSON-1483258717 tempest-ServerRescueNegativeTestJSON-1483258717-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61806) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 708.777808] env[61806]: DEBUG nova.virt.hardware [None req-fcbc4cf1-fc11-42b1-9d49-ef556b52782e tempest-ServerRescueNegativeTestJSON-1483258717 tempest-ServerRescueNegativeTestJSON-1483258717-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61806) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 708.777975] env[61806]: DEBUG nova.virt.hardware [None req-fcbc4cf1-fc11-42b1-9d49-ef556b52782e tempest-ServerRescueNegativeTestJSON-1483258717 tempest-ServerRescueNegativeTestJSON-1483258717-project-member] Got 1 possible topologies {{(pid=61806) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 708.778288] env[61806]: DEBUG nova.virt.hardware [None req-fcbc4cf1-fc11-42b1-9d49-ef556b52782e tempest-ServerRescueNegativeTestJSON-1483258717 tempest-ServerRescueNegativeTestJSON-1483258717-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61806) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 708.778643] env[61806]: DEBUG nova.virt.hardware [None req-fcbc4cf1-fc11-42b1-9d49-ef556b52782e tempest-ServerRescueNegativeTestJSON-1483258717 tempest-ServerRescueNegativeTestJSON-1483258717-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61806) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 708.779638] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2fda7e82-2476-4b5e-ac48-78c0695338f5 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 708.789739] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3cfbf21c-225d-4e50-ac98-b21f13b8e257 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 708.803948] env[61806]: ERROR nova.compute.manager [None req-fcbc4cf1-fc11-42b1-9d49-ef556b52782e tempest-ServerRescueNegativeTestJSON-1483258717 tempest-ServerRescueNegativeTestJSON-1483258717-project-member] [instance: 7ec2497a-7329-4ed4-99ab-dec3d1e8117c] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 81f90dee-a680-4d31-9c3d-fbe42fa742a4, please check neutron logs for more information. [ 708.803948] env[61806]: ERROR nova.compute.manager [instance: 7ec2497a-7329-4ed4-99ab-dec3d1e8117c] Traceback (most recent call last): [ 708.803948] env[61806]: ERROR nova.compute.manager [instance: 7ec2497a-7329-4ed4-99ab-dec3d1e8117c] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 708.803948] env[61806]: ERROR nova.compute.manager [instance: 7ec2497a-7329-4ed4-99ab-dec3d1e8117c] yield resources [ 708.803948] env[61806]: ERROR nova.compute.manager [instance: 7ec2497a-7329-4ed4-99ab-dec3d1e8117c] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 708.803948] env[61806]: ERROR nova.compute.manager [instance: 7ec2497a-7329-4ed4-99ab-dec3d1e8117c] self.driver.spawn(context, instance, image_meta, [ 708.803948] env[61806]: ERROR nova.compute.manager [instance: 7ec2497a-7329-4ed4-99ab-dec3d1e8117c] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 708.803948] env[61806]: ERROR nova.compute.manager [instance: 7ec2497a-7329-4ed4-99ab-dec3d1e8117c] self._vmops.spawn(context, instance, image_meta, injected_files, [ 708.803948] env[61806]: ERROR nova.compute.manager [instance: 7ec2497a-7329-4ed4-99ab-dec3d1e8117c] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 708.803948] env[61806]: ERROR nova.compute.manager [instance: 7ec2497a-7329-4ed4-99ab-dec3d1e8117c] vm_ref = self.build_virtual_machine(instance, [ 708.803948] env[61806]: ERROR nova.compute.manager [instance: 7ec2497a-7329-4ed4-99ab-dec3d1e8117c] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 708.804383] env[61806]: ERROR nova.compute.manager [instance: 7ec2497a-7329-4ed4-99ab-dec3d1e8117c] vif_infos = vmwarevif.get_vif_info(self._session, [ 708.804383] env[61806]: ERROR nova.compute.manager [instance: 7ec2497a-7329-4ed4-99ab-dec3d1e8117c] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 708.804383] env[61806]: ERROR nova.compute.manager [instance: 7ec2497a-7329-4ed4-99ab-dec3d1e8117c] for vif in network_info: [ 708.804383] env[61806]: ERROR nova.compute.manager [instance: 7ec2497a-7329-4ed4-99ab-dec3d1e8117c] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 708.804383] env[61806]: ERROR nova.compute.manager [instance: 7ec2497a-7329-4ed4-99ab-dec3d1e8117c] return self._sync_wrapper(fn, *args, **kwargs) [ 708.804383] env[61806]: ERROR nova.compute.manager [instance: 7ec2497a-7329-4ed4-99ab-dec3d1e8117c] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 708.804383] env[61806]: ERROR nova.compute.manager [instance: 7ec2497a-7329-4ed4-99ab-dec3d1e8117c] self.wait() [ 708.804383] env[61806]: ERROR nova.compute.manager [instance: 7ec2497a-7329-4ed4-99ab-dec3d1e8117c] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 708.804383] env[61806]: ERROR nova.compute.manager [instance: 7ec2497a-7329-4ed4-99ab-dec3d1e8117c] self[:] = self._gt.wait() [ 708.804383] env[61806]: ERROR nova.compute.manager [instance: 7ec2497a-7329-4ed4-99ab-dec3d1e8117c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 708.804383] env[61806]: ERROR nova.compute.manager [instance: 7ec2497a-7329-4ed4-99ab-dec3d1e8117c] return self._exit_event.wait() [ 708.804383] env[61806]: ERROR nova.compute.manager [instance: 7ec2497a-7329-4ed4-99ab-dec3d1e8117c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 708.804383] env[61806]: ERROR nova.compute.manager [instance: 7ec2497a-7329-4ed4-99ab-dec3d1e8117c] current.throw(*self._exc) [ 708.804783] env[61806]: ERROR nova.compute.manager [instance: 7ec2497a-7329-4ed4-99ab-dec3d1e8117c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 708.804783] env[61806]: ERROR nova.compute.manager [instance: 7ec2497a-7329-4ed4-99ab-dec3d1e8117c] result = function(*args, **kwargs) [ 708.804783] env[61806]: ERROR nova.compute.manager [instance: 7ec2497a-7329-4ed4-99ab-dec3d1e8117c] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 708.804783] env[61806]: ERROR nova.compute.manager [instance: 7ec2497a-7329-4ed4-99ab-dec3d1e8117c] return func(*args, **kwargs) [ 708.804783] env[61806]: ERROR nova.compute.manager [instance: 7ec2497a-7329-4ed4-99ab-dec3d1e8117c] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 708.804783] env[61806]: ERROR nova.compute.manager [instance: 7ec2497a-7329-4ed4-99ab-dec3d1e8117c] raise e [ 708.804783] env[61806]: ERROR nova.compute.manager [instance: 7ec2497a-7329-4ed4-99ab-dec3d1e8117c] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 708.804783] env[61806]: ERROR nova.compute.manager [instance: 7ec2497a-7329-4ed4-99ab-dec3d1e8117c] nwinfo = self.network_api.allocate_for_instance( [ 708.804783] env[61806]: ERROR nova.compute.manager [instance: 7ec2497a-7329-4ed4-99ab-dec3d1e8117c] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 708.804783] env[61806]: ERROR nova.compute.manager [instance: 7ec2497a-7329-4ed4-99ab-dec3d1e8117c] created_port_ids = self._update_ports_for_instance( [ 708.804783] env[61806]: ERROR nova.compute.manager [instance: 7ec2497a-7329-4ed4-99ab-dec3d1e8117c] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 708.804783] env[61806]: ERROR nova.compute.manager [instance: 7ec2497a-7329-4ed4-99ab-dec3d1e8117c] with excutils.save_and_reraise_exception(): [ 708.804783] env[61806]: ERROR nova.compute.manager [instance: 7ec2497a-7329-4ed4-99ab-dec3d1e8117c] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 708.805187] env[61806]: ERROR nova.compute.manager [instance: 7ec2497a-7329-4ed4-99ab-dec3d1e8117c] self.force_reraise() [ 708.805187] env[61806]: ERROR nova.compute.manager [instance: 7ec2497a-7329-4ed4-99ab-dec3d1e8117c] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 708.805187] env[61806]: ERROR nova.compute.manager [instance: 7ec2497a-7329-4ed4-99ab-dec3d1e8117c] raise self.value [ 708.805187] env[61806]: ERROR nova.compute.manager [instance: 7ec2497a-7329-4ed4-99ab-dec3d1e8117c] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 708.805187] env[61806]: ERROR nova.compute.manager [instance: 7ec2497a-7329-4ed4-99ab-dec3d1e8117c] updated_port = self._update_port( [ 708.805187] env[61806]: ERROR nova.compute.manager [instance: 7ec2497a-7329-4ed4-99ab-dec3d1e8117c] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 708.805187] env[61806]: ERROR nova.compute.manager [instance: 7ec2497a-7329-4ed4-99ab-dec3d1e8117c] _ensure_no_port_binding_failure(port) [ 708.805187] env[61806]: ERROR nova.compute.manager [instance: 7ec2497a-7329-4ed4-99ab-dec3d1e8117c] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 708.805187] env[61806]: ERROR nova.compute.manager [instance: 7ec2497a-7329-4ed4-99ab-dec3d1e8117c] raise exception.PortBindingFailed(port_id=port['id']) [ 708.805187] env[61806]: ERROR nova.compute.manager [instance: 7ec2497a-7329-4ed4-99ab-dec3d1e8117c] nova.exception.PortBindingFailed: Binding failed for port 81f90dee-a680-4d31-9c3d-fbe42fa742a4, please check neutron logs for more information. [ 708.805187] env[61806]: ERROR nova.compute.manager [instance: 7ec2497a-7329-4ed4-99ab-dec3d1e8117c] [ 708.805187] env[61806]: INFO nova.compute.manager [None req-fcbc4cf1-fc11-42b1-9d49-ef556b52782e tempest-ServerRescueNegativeTestJSON-1483258717 tempest-ServerRescueNegativeTestJSON-1483258717-project-member] [instance: 7ec2497a-7329-4ed4-99ab-dec3d1e8117c] Terminating instance [ 708.805864] env[61806]: DEBUG oslo_concurrency.lockutils [None req-fcbc4cf1-fc11-42b1-9d49-ef556b52782e tempest-ServerRescueNegativeTestJSON-1483258717 tempest-ServerRescueNegativeTestJSON-1483258717-project-member] Acquiring lock "refresh_cache-7ec2497a-7329-4ed4-99ab-dec3d1e8117c" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 708.806539] env[61806]: DEBUG nova.network.neutron [req-93e4637b-7f83-43eb-9292-ad83e501b0cb req-8db490bc-82bc-4188-8b6a-be846272ec19 service nova] [instance: 7ec2497a-7329-4ed4-99ab-dec3d1e8117c] Instance cache missing network info. {{(pid=61806) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 708.878813] env[61806]: DEBUG nova.network.neutron [req-93e4637b-7f83-43eb-9292-ad83e501b0cb req-8db490bc-82bc-4188-8b6a-be846272ec19 service nova] [instance: 7ec2497a-7329-4ed4-99ab-dec3d1e8117c] Updating instance_info_cache with network_info: [] {{(pid=61806) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 709.156802] env[61806]: DEBUG oslo_concurrency.lockutils [None req-6168822a-f60c-43e0-b0ac-2eb07b4977dc tempest-ServersAdmin275Test-1945525728 tempest-ServersAdmin275Test-1945525728-project-member] Lock "9db80c3c-e2c3-4034-a79a-d7be924c4ff6" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 30.050s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 709.305817] env[61806]: DEBUG oslo_concurrency.lockutils [None req-dbeadda9-7825-4678-af18-5f977e49124f tempest-ServerActionsV293TestJSON-906151111 tempest-ServerActionsV293TestJSON-906151111-project-member] Acquiring lock "0758901a-7093-41d3-b0e2-5c519333abdd" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 709.306605] env[61806]: DEBUG oslo_concurrency.lockutils [None req-dbeadda9-7825-4678-af18-5f977e49124f tempest-ServerActionsV293TestJSON-906151111 tempest-ServerActionsV293TestJSON-906151111-project-member] Lock "0758901a-7093-41d3-b0e2-5c519333abdd" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 709.381120] env[61806]: DEBUG oslo_concurrency.lockutils [req-93e4637b-7f83-43eb-9292-ad83e501b0cb req-8db490bc-82bc-4188-8b6a-be846272ec19 service nova] Releasing lock "refresh_cache-7ec2497a-7329-4ed4-99ab-dec3d1e8117c" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 709.381376] env[61806]: DEBUG oslo_concurrency.lockutils [None req-fcbc4cf1-fc11-42b1-9d49-ef556b52782e tempest-ServerRescueNegativeTestJSON-1483258717 tempest-ServerRescueNegativeTestJSON-1483258717-project-member] Acquired lock "refresh_cache-7ec2497a-7329-4ed4-99ab-dec3d1e8117c" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 709.381566] env[61806]: DEBUG nova.network.neutron [None req-fcbc4cf1-fc11-42b1-9d49-ef556b52782e tempest-ServerRescueNegativeTestJSON-1483258717 tempest-ServerRescueNegativeTestJSON-1483258717-project-member] [instance: 7ec2497a-7329-4ed4-99ab-dec3d1e8117c] Building network info cache for instance {{(pid=61806) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 709.506742] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3b360dae-de73-4fd9-940d-7b9e7b3b3a49 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 709.515482] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ca497233-2bc8-4beb-8abf-a8934eb3c59d {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 709.547602] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5c1aa732-b227-4549-acae-a89720f093a6 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 709.556869] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b428a0bd-4bbd-4f3b-96ef-541a13ab9c46 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 709.570770] env[61806]: DEBUG nova.compute.provider_tree [None req-7bb45eee-0cbe-4995-9898-530114f6f104 tempest-ImagesOneServerNegativeTestJSON-606465252 tempest-ImagesOneServerNegativeTestJSON-606465252-project-member] Inventory has not changed in ProviderTree for provider: a2858be1-fd22-4e08-979e-87ad25293407 {{(pid=61806) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 709.901604] env[61806]: DEBUG nova.network.neutron [None req-fcbc4cf1-fc11-42b1-9d49-ef556b52782e tempest-ServerRescueNegativeTestJSON-1483258717 tempest-ServerRescueNegativeTestJSON-1483258717-project-member] [instance: 7ec2497a-7329-4ed4-99ab-dec3d1e8117c] Instance cache missing network info. {{(pid=61806) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 709.995849] env[61806]: DEBUG nova.network.neutron [None req-fcbc4cf1-fc11-42b1-9d49-ef556b52782e tempest-ServerRescueNegativeTestJSON-1483258717 tempest-ServerRescueNegativeTestJSON-1483258717-project-member] [instance: 7ec2497a-7329-4ed4-99ab-dec3d1e8117c] Updating instance_info_cache with network_info: [] {{(pid=61806) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 710.075503] env[61806]: DEBUG nova.scheduler.client.report [None req-7bb45eee-0cbe-4995-9898-530114f6f104 tempest-ImagesOneServerNegativeTestJSON-606465252 tempest-ImagesOneServerNegativeTestJSON-606465252-project-member] Inventory has not changed for provider a2858be1-fd22-4e08-979e-87ad25293407 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 138, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61806) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 710.311194] env[61806]: DEBUG nova.compute.manager [req-d330b477-b940-41f1-89de-2c4bedddb9d7 req-ed2aca66-429b-452e-a367-fdbe0e6b9a2b service nova] [instance: 7ec2497a-7329-4ed4-99ab-dec3d1e8117c] Received event network-vif-deleted-81f90dee-a680-4d31-9c3d-fbe42fa742a4 {{(pid=61806) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 710.499012] env[61806]: DEBUG oslo_concurrency.lockutils [None req-fcbc4cf1-fc11-42b1-9d49-ef556b52782e tempest-ServerRescueNegativeTestJSON-1483258717 tempest-ServerRescueNegativeTestJSON-1483258717-project-member] Releasing lock "refresh_cache-7ec2497a-7329-4ed4-99ab-dec3d1e8117c" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 710.499395] env[61806]: DEBUG nova.compute.manager [None req-fcbc4cf1-fc11-42b1-9d49-ef556b52782e tempest-ServerRescueNegativeTestJSON-1483258717 tempest-ServerRescueNegativeTestJSON-1483258717-project-member] [instance: 7ec2497a-7329-4ed4-99ab-dec3d1e8117c] Start destroying the instance on the hypervisor. {{(pid=61806) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 710.499616] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-fcbc4cf1-fc11-42b1-9d49-ef556b52782e tempest-ServerRescueNegativeTestJSON-1483258717 tempest-ServerRescueNegativeTestJSON-1483258717-project-member] [instance: 7ec2497a-7329-4ed4-99ab-dec3d1e8117c] Destroying instance {{(pid=61806) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 710.499938] env[61806]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-eb582842-a3dc-40a4-b0fa-0128c8caf9f8 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 710.511688] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e1f031e8-a2e8-43a1-9620-4e6965d9c426 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 710.538020] env[61806]: WARNING nova.virt.vmwareapi.vmops [None req-fcbc4cf1-fc11-42b1-9d49-ef556b52782e tempest-ServerRescueNegativeTestJSON-1483258717 tempest-ServerRescueNegativeTestJSON-1483258717-project-member] [instance: 7ec2497a-7329-4ed4-99ab-dec3d1e8117c] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 7ec2497a-7329-4ed4-99ab-dec3d1e8117c could not be found. [ 710.538020] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-fcbc4cf1-fc11-42b1-9d49-ef556b52782e tempest-ServerRescueNegativeTestJSON-1483258717 tempest-ServerRescueNegativeTestJSON-1483258717-project-member] [instance: 7ec2497a-7329-4ed4-99ab-dec3d1e8117c] Instance destroyed {{(pid=61806) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 710.538020] env[61806]: INFO nova.compute.manager [None req-fcbc4cf1-fc11-42b1-9d49-ef556b52782e tempest-ServerRescueNegativeTestJSON-1483258717 tempest-ServerRescueNegativeTestJSON-1483258717-project-member] [instance: 7ec2497a-7329-4ed4-99ab-dec3d1e8117c] Took 0.04 seconds to destroy the instance on the hypervisor. [ 710.538020] env[61806]: DEBUG oslo.service.loopingcall [None req-fcbc4cf1-fc11-42b1-9d49-ef556b52782e tempest-ServerRescueNegativeTestJSON-1483258717 tempest-ServerRescueNegativeTestJSON-1483258717-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61806) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 710.538384] env[61806]: DEBUG nova.compute.manager [-] [instance: 7ec2497a-7329-4ed4-99ab-dec3d1e8117c] Deallocating network for instance {{(pid=61806) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 710.538384] env[61806]: DEBUG nova.network.neutron [-] [instance: 7ec2497a-7329-4ed4-99ab-dec3d1e8117c] deallocate_for_instance() {{(pid=61806) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 710.557955] env[61806]: DEBUG nova.network.neutron [-] [instance: 7ec2497a-7329-4ed4-99ab-dec3d1e8117c] Instance cache missing network info. {{(pid=61806) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 710.582378] env[61806]: DEBUG oslo_concurrency.lockutils [None req-7bb45eee-0cbe-4995-9898-530114f6f104 tempest-ImagesOneServerNegativeTestJSON-606465252 tempest-ImagesOneServerNegativeTestJSON-606465252-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.967s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 710.582378] env[61806]: ERROR nova.compute.manager [None req-7bb45eee-0cbe-4995-9898-530114f6f104 tempest-ImagesOneServerNegativeTestJSON-606465252 tempest-ImagesOneServerNegativeTestJSON-606465252-project-member] [instance: 8d0cdc81-56ed-4a7a-8221-30bb906e16b2] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 4cd363a6-ef40-4b4b-909f-85a169e5c84e, please check neutron logs for more information. [ 710.582378] env[61806]: ERROR nova.compute.manager [instance: 8d0cdc81-56ed-4a7a-8221-30bb906e16b2] Traceback (most recent call last): [ 710.582378] env[61806]: ERROR nova.compute.manager [instance: 8d0cdc81-56ed-4a7a-8221-30bb906e16b2] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 710.582378] env[61806]: ERROR nova.compute.manager [instance: 8d0cdc81-56ed-4a7a-8221-30bb906e16b2] self.driver.spawn(context, instance, image_meta, [ 710.582378] env[61806]: ERROR nova.compute.manager [instance: 8d0cdc81-56ed-4a7a-8221-30bb906e16b2] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 710.582378] env[61806]: ERROR nova.compute.manager [instance: 8d0cdc81-56ed-4a7a-8221-30bb906e16b2] self._vmops.spawn(context, instance, image_meta, injected_files, [ 710.582378] env[61806]: ERROR nova.compute.manager [instance: 8d0cdc81-56ed-4a7a-8221-30bb906e16b2] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 710.582378] env[61806]: ERROR nova.compute.manager [instance: 8d0cdc81-56ed-4a7a-8221-30bb906e16b2] vm_ref = self.build_virtual_machine(instance, [ 710.582840] env[61806]: ERROR nova.compute.manager [instance: 8d0cdc81-56ed-4a7a-8221-30bb906e16b2] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 710.582840] env[61806]: ERROR nova.compute.manager [instance: 8d0cdc81-56ed-4a7a-8221-30bb906e16b2] vif_infos = vmwarevif.get_vif_info(self._session, [ 710.582840] env[61806]: ERROR nova.compute.manager [instance: 8d0cdc81-56ed-4a7a-8221-30bb906e16b2] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 710.582840] env[61806]: ERROR nova.compute.manager [instance: 8d0cdc81-56ed-4a7a-8221-30bb906e16b2] for vif in network_info: [ 710.582840] env[61806]: ERROR nova.compute.manager [instance: 8d0cdc81-56ed-4a7a-8221-30bb906e16b2] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 710.582840] env[61806]: ERROR nova.compute.manager [instance: 8d0cdc81-56ed-4a7a-8221-30bb906e16b2] return self._sync_wrapper(fn, *args, **kwargs) [ 710.582840] env[61806]: ERROR nova.compute.manager [instance: 8d0cdc81-56ed-4a7a-8221-30bb906e16b2] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 710.582840] env[61806]: ERROR nova.compute.manager [instance: 8d0cdc81-56ed-4a7a-8221-30bb906e16b2] self.wait() [ 710.582840] env[61806]: ERROR nova.compute.manager [instance: 8d0cdc81-56ed-4a7a-8221-30bb906e16b2] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 710.582840] env[61806]: ERROR nova.compute.manager [instance: 8d0cdc81-56ed-4a7a-8221-30bb906e16b2] self[:] = self._gt.wait() [ 710.582840] env[61806]: ERROR nova.compute.manager [instance: 8d0cdc81-56ed-4a7a-8221-30bb906e16b2] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 710.582840] env[61806]: ERROR nova.compute.manager [instance: 8d0cdc81-56ed-4a7a-8221-30bb906e16b2] return self._exit_event.wait() [ 710.582840] env[61806]: ERROR nova.compute.manager [instance: 8d0cdc81-56ed-4a7a-8221-30bb906e16b2] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 710.583489] env[61806]: ERROR nova.compute.manager [instance: 8d0cdc81-56ed-4a7a-8221-30bb906e16b2] current.throw(*self._exc) [ 710.583489] env[61806]: ERROR nova.compute.manager [instance: 8d0cdc81-56ed-4a7a-8221-30bb906e16b2] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 710.583489] env[61806]: ERROR nova.compute.manager [instance: 8d0cdc81-56ed-4a7a-8221-30bb906e16b2] result = function(*args, **kwargs) [ 710.583489] env[61806]: ERROR nova.compute.manager [instance: 8d0cdc81-56ed-4a7a-8221-30bb906e16b2] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 710.583489] env[61806]: ERROR nova.compute.manager [instance: 8d0cdc81-56ed-4a7a-8221-30bb906e16b2] return func(*args, **kwargs) [ 710.583489] env[61806]: ERROR nova.compute.manager [instance: 8d0cdc81-56ed-4a7a-8221-30bb906e16b2] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 710.583489] env[61806]: ERROR nova.compute.manager [instance: 8d0cdc81-56ed-4a7a-8221-30bb906e16b2] raise e [ 710.583489] env[61806]: ERROR nova.compute.manager [instance: 8d0cdc81-56ed-4a7a-8221-30bb906e16b2] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 710.583489] env[61806]: ERROR nova.compute.manager [instance: 8d0cdc81-56ed-4a7a-8221-30bb906e16b2] nwinfo = self.network_api.allocate_for_instance( [ 710.583489] env[61806]: ERROR nova.compute.manager [instance: 8d0cdc81-56ed-4a7a-8221-30bb906e16b2] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 710.583489] env[61806]: ERROR nova.compute.manager [instance: 8d0cdc81-56ed-4a7a-8221-30bb906e16b2] created_port_ids = self._update_ports_for_instance( [ 710.583489] env[61806]: ERROR nova.compute.manager [instance: 8d0cdc81-56ed-4a7a-8221-30bb906e16b2] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 710.583489] env[61806]: ERROR nova.compute.manager [instance: 8d0cdc81-56ed-4a7a-8221-30bb906e16b2] with excutils.save_and_reraise_exception(): [ 710.584042] env[61806]: ERROR nova.compute.manager [instance: 8d0cdc81-56ed-4a7a-8221-30bb906e16b2] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 710.584042] env[61806]: ERROR nova.compute.manager [instance: 8d0cdc81-56ed-4a7a-8221-30bb906e16b2] self.force_reraise() [ 710.584042] env[61806]: ERROR nova.compute.manager [instance: 8d0cdc81-56ed-4a7a-8221-30bb906e16b2] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 710.584042] env[61806]: ERROR nova.compute.manager [instance: 8d0cdc81-56ed-4a7a-8221-30bb906e16b2] raise self.value [ 710.584042] env[61806]: ERROR nova.compute.manager [instance: 8d0cdc81-56ed-4a7a-8221-30bb906e16b2] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 710.584042] env[61806]: ERROR nova.compute.manager [instance: 8d0cdc81-56ed-4a7a-8221-30bb906e16b2] updated_port = self._update_port( [ 710.584042] env[61806]: ERROR nova.compute.manager [instance: 8d0cdc81-56ed-4a7a-8221-30bb906e16b2] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 710.584042] env[61806]: ERROR nova.compute.manager [instance: 8d0cdc81-56ed-4a7a-8221-30bb906e16b2] _ensure_no_port_binding_failure(port) [ 710.584042] env[61806]: ERROR nova.compute.manager [instance: 8d0cdc81-56ed-4a7a-8221-30bb906e16b2] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 710.584042] env[61806]: ERROR nova.compute.manager [instance: 8d0cdc81-56ed-4a7a-8221-30bb906e16b2] raise exception.PortBindingFailed(port_id=port['id']) [ 710.584042] env[61806]: ERROR nova.compute.manager [instance: 8d0cdc81-56ed-4a7a-8221-30bb906e16b2] nova.exception.PortBindingFailed: Binding failed for port 4cd363a6-ef40-4b4b-909f-85a169e5c84e, please check neutron logs for more information. [ 710.584042] env[61806]: ERROR nova.compute.manager [instance: 8d0cdc81-56ed-4a7a-8221-30bb906e16b2] [ 710.584779] env[61806]: DEBUG nova.compute.utils [None req-7bb45eee-0cbe-4995-9898-530114f6f104 tempest-ImagesOneServerNegativeTestJSON-606465252 tempest-ImagesOneServerNegativeTestJSON-606465252-project-member] [instance: 8d0cdc81-56ed-4a7a-8221-30bb906e16b2] Binding failed for port 4cd363a6-ef40-4b4b-909f-85a169e5c84e, please check neutron logs for more information. {{(pid=61806) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 710.584779] env[61806]: DEBUG oslo_concurrency.lockutils [None req-6b55cbbf-00f7-4d4d-87f3-a59cc1c553b0 tempest-ServerRescueNegativeTestJSON-1483258717 tempest-ServerRescueNegativeTestJSON-1483258717-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 25.375s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 710.585710] env[61806]: INFO nova.compute.claims [None req-6b55cbbf-00f7-4d4d-87f3-a59cc1c553b0 tempest-ServerRescueNegativeTestJSON-1483258717 tempest-ServerRescueNegativeTestJSON-1483258717-project-member] [instance: 7ea919b6-12f7-4f4e-a123-413044bbbffc] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 710.589456] env[61806]: DEBUG nova.compute.manager [None req-7bb45eee-0cbe-4995-9898-530114f6f104 tempest-ImagesOneServerNegativeTestJSON-606465252 tempest-ImagesOneServerNegativeTestJSON-606465252-project-member] [instance: 8d0cdc81-56ed-4a7a-8221-30bb906e16b2] Build of instance 8d0cdc81-56ed-4a7a-8221-30bb906e16b2 was re-scheduled: Binding failed for port 4cd363a6-ef40-4b4b-909f-85a169e5c84e, please check neutron logs for more information. {{(pid=61806) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 710.590084] env[61806]: DEBUG nova.compute.manager [None req-7bb45eee-0cbe-4995-9898-530114f6f104 tempest-ImagesOneServerNegativeTestJSON-606465252 tempest-ImagesOneServerNegativeTestJSON-606465252-project-member] [instance: 8d0cdc81-56ed-4a7a-8221-30bb906e16b2] Unplugging VIFs for instance {{(pid=61806) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 710.590176] env[61806]: DEBUG oslo_concurrency.lockutils [None req-7bb45eee-0cbe-4995-9898-530114f6f104 tempest-ImagesOneServerNegativeTestJSON-606465252 tempest-ImagesOneServerNegativeTestJSON-606465252-project-member] Acquiring lock "refresh_cache-8d0cdc81-56ed-4a7a-8221-30bb906e16b2" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 710.590283] env[61806]: DEBUG oslo_concurrency.lockutils [None req-7bb45eee-0cbe-4995-9898-530114f6f104 tempest-ImagesOneServerNegativeTestJSON-606465252 tempest-ImagesOneServerNegativeTestJSON-606465252-project-member] Acquired lock "refresh_cache-8d0cdc81-56ed-4a7a-8221-30bb906e16b2" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 710.590445] env[61806]: DEBUG nova.network.neutron [None req-7bb45eee-0cbe-4995-9898-530114f6f104 tempest-ImagesOneServerNegativeTestJSON-606465252 tempest-ImagesOneServerNegativeTestJSON-606465252-project-member] [instance: 8d0cdc81-56ed-4a7a-8221-30bb906e16b2] Building network info cache for instance {{(pid=61806) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 711.060743] env[61806]: DEBUG nova.network.neutron [-] [instance: 7ec2497a-7329-4ed4-99ab-dec3d1e8117c] Updating instance_info_cache with network_info: [] {{(pid=61806) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 711.110941] env[61806]: DEBUG nova.network.neutron [None req-7bb45eee-0cbe-4995-9898-530114f6f104 tempest-ImagesOneServerNegativeTestJSON-606465252 tempest-ImagesOneServerNegativeTestJSON-606465252-project-member] [instance: 8d0cdc81-56ed-4a7a-8221-30bb906e16b2] Instance cache missing network info. {{(pid=61806) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 711.201242] env[61806]: DEBUG nova.network.neutron [None req-7bb45eee-0cbe-4995-9898-530114f6f104 tempest-ImagesOneServerNegativeTestJSON-606465252 tempest-ImagesOneServerNegativeTestJSON-606465252-project-member] [instance: 8d0cdc81-56ed-4a7a-8221-30bb906e16b2] Updating instance_info_cache with network_info: [] {{(pid=61806) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 711.563624] env[61806]: INFO nova.compute.manager [-] [instance: 7ec2497a-7329-4ed4-99ab-dec3d1e8117c] Took 1.03 seconds to deallocate network for instance. [ 711.566208] env[61806]: DEBUG nova.compute.claims [None req-fcbc4cf1-fc11-42b1-9d49-ef556b52782e tempest-ServerRescueNegativeTestJSON-1483258717 tempest-ServerRescueNegativeTestJSON-1483258717-project-member] [instance: 7ec2497a-7329-4ed4-99ab-dec3d1e8117c] Aborting claim: {{(pid=61806) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 711.566327] env[61806]: DEBUG oslo_concurrency.lockutils [None req-fcbc4cf1-fc11-42b1-9d49-ef556b52782e tempest-ServerRescueNegativeTestJSON-1483258717 tempest-ServerRescueNegativeTestJSON-1483258717-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 711.703358] env[61806]: DEBUG oslo_concurrency.lockutils [None req-7bb45eee-0cbe-4995-9898-530114f6f104 tempest-ImagesOneServerNegativeTestJSON-606465252 tempest-ImagesOneServerNegativeTestJSON-606465252-project-member] Releasing lock "refresh_cache-8d0cdc81-56ed-4a7a-8221-30bb906e16b2" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 711.703665] env[61806]: DEBUG nova.compute.manager [None req-7bb45eee-0cbe-4995-9898-530114f6f104 tempest-ImagesOneServerNegativeTestJSON-606465252 tempest-ImagesOneServerNegativeTestJSON-606465252-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61806) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 711.703807] env[61806]: DEBUG nova.compute.manager [None req-7bb45eee-0cbe-4995-9898-530114f6f104 tempest-ImagesOneServerNegativeTestJSON-606465252 tempest-ImagesOneServerNegativeTestJSON-606465252-project-member] [instance: 8d0cdc81-56ed-4a7a-8221-30bb906e16b2] Deallocating network for instance {{(pid=61806) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 711.706960] env[61806]: DEBUG nova.network.neutron [None req-7bb45eee-0cbe-4995-9898-530114f6f104 tempest-ImagesOneServerNegativeTestJSON-606465252 tempest-ImagesOneServerNegativeTestJSON-606465252-project-member] [instance: 8d0cdc81-56ed-4a7a-8221-30bb906e16b2] deallocate_for_instance() {{(pid=61806) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 711.720377] env[61806]: DEBUG nova.network.neutron [None req-7bb45eee-0cbe-4995-9898-530114f6f104 tempest-ImagesOneServerNegativeTestJSON-606465252 tempest-ImagesOneServerNegativeTestJSON-606465252-project-member] [instance: 8d0cdc81-56ed-4a7a-8221-30bb906e16b2] Instance cache missing network info. {{(pid=61806) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 711.955855] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-79d5511f-36ac-4010-8ab3-2397a3c4b7d6 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 711.964943] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d13a58d2-6bb8-4ce4-b3e9-6fb855c04b35 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 711.998606] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-28ac2f9b-c877-44f9-8fe6-060517bfb282 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 712.007217] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-74aca732-a019-4151-9730-e2ec244cc1aa {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 712.023311] env[61806]: DEBUG nova.compute.provider_tree [None req-6b55cbbf-00f7-4d4d-87f3-a59cc1c553b0 tempest-ServerRescueNegativeTestJSON-1483258717 tempest-ServerRescueNegativeTestJSON-1483258717-project-member] Inventory has not changed in ProviderTree for provider: a2858be1-fd22-4e08-979e-87ad25293407 {{(pid=61806) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 712.224437] env[61806]: DEBUG nova.network.neutron [None req-7bb45eee-0cbe-4995-9898-530114f6f104 tempest-ImagesOneServerNegativeTestJSON-606465252 tempest-ImagesOneServerNegativeTestJSON-606465252-project-member] [instance: 8d0cdc81-56ed-4a7a-8221-30bb906e16b2] Updating instance_info_cache with network_info: [] {{(pid=61806) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 712.526448] env[61806]: DEBUG nova.scheduler.client.report [None req-6b55cbbf-00f7-4d4d-87f3-a59cc1c553b0 tempest-ServerRescueNegativeTestJSON-1483258717 tempest-ServerRescueNegativeTestJSON-1483258717-project-member] Inventory has not changed for provider a2858be1-fd22-4e08-979e-87ad25293407 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 138, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61806) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 712.726853] env[61806]: INFO nova.compute.manager [None req-7bb45eee-0cbe-4995-9898-530114f6f104 tempest-ImagesOneServerNegativeTestJSON-606465252 tempest-ImagesOneServerNegativeTestJSON-606465252-project-member] [instance: 8d0cdc81-56ed-4a7a-8221-30bb906e16b2] Took 1.02 seconds to deallocate network for instance. [ 713.032215] env[61806]: DEBUG oslo_concurrency.lockutils [None req-6b55cbbf-00f7-4d4d-87f3-a59cc1c553b0 tempest-ServerRescueNegativeTestJSON-1483258717 tempest-ServerRescueNegativeTestJSON-1483258717-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.448s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 713.032759] env[61806]: DEBUG nova.compute.manager [None req-6b55cbbf-00f7-4d4d-87f3-a59cc1c553b0 tempest-ServerRescueNegativeTestJSON-1483258717 tempest-ServerRescueNegativeTestJSON-1483258717-project-member] [instance: 7ea919b6-12f7-4f4e-a123-413044bbbffc] Start building networks asynchronously for instance. {{(pid=61806) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 713.035935] env[61806]: DEBUG oslo_concurrency.lockutils [None req-865353ee-538a-40fc-bfba-502aba0b6a20 tempest-VolumesAdminNegativeTest-1387177911 tempest-VolumesAdminNegativeTest-1387177911-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 24.923s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 713.537523] env[61806]: DEBUG nova.compute.utils [None req-6b55cbbf-00f7-4d4d-87f3-a59cc1c553b0 tempest-ServerRescueNegativeTestJSON-1483258717 tempest-ServerRescueNegativeTestJSON-1483258717-project-member] Using /dev/sd instead of None {{(pid=61806) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 713.538915] env[61806]: DEBUG nova.compute.manager [None req-6b55cbbf-00f7-4d4d-87f3-a59cc1c553b0 tempest-ServerRescueNegativeTestJSON-1483258717 tempest-ServerRescueNegativeTestJSON-1483258717-project-member] [instance: 7ea919b6-12f7-4f4e-a123-413044bbbffc] Allocating IP information in the background. {{(pid=61806) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 713.539163] env[61806]: DEBUG nova.network.neutron [None req-6b55cbbf-00f7-4d4d-87f3-a59cc1c553b0 tempest-ServerRescueNegativeTestJSON-1483258717 tempest-ServerRescueNegativeTestJSON-1483258717-project-member] [instance: 7ea919b6-12f7-4f4e-a123-413044bbbffc] allocate_for_instance() {{(pid=61806) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 713.599847] env[61806]: DEBUG nova.policy [None req-6b55cbbf-00f7-4d4d-87f3-a59cc1c553b0 tempest-ServerRescueNegativeTestJSON-1483258717 tempest-ServerRescueNegativeTestJSON-1483258717-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '5925347d6d584c1e8d7de04a84857859', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '596c39abc1c8448882be50c8f7cf5e9a', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61806) authorize /opt/stack/nova/nova/policy.py:201}} [ 713.762869] env[61806]: INFO nova.scheduler.client.report [None req-7bb45eee-0cbe-4995-9898-530114f6f104 tempest-ImagesOneServerNegativeTestJSON-606465252 tempest-ImagesOneServerNegativeTestJSON-606465252-project-member] Deleted allocations for instance 8d0cdc81-56ed-4a7a-8221-30bb906e16b2 [ 713.913272] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bf218025-2f3d-4728-a358-e558038fe548 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 713.923451] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-395101bb-bc2c-44f6-b37d-f9bcd156e4c2 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 713.955457] env[61806]: DEBUG nova.network.neutron [None req-6b55cbbf-00f7-4d4d-87f3-a59cc1c553b0 tempest-ServerRescueNegativeTestJSON-1483258717 tempest-ServerRescueNegativeTestJSON-1483258717-project-member] [instance: 7ea919b6-12f7-4f4e-a123-413044bbbffc] Successfully created port: 61c8084d-6ad0-442a-967a-46f14ffa6671 {{(pid=61806) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 713.957863] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-075e4456-8083-4f91-9b06-d674172fbd6e {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 713.965647] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7eb25187-f87b-45ed-81c4-bb6d70b9e0d7 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 713.978968] env[61806]: DEBUG nova.compute.provider_tree [None req-865353ee-538a-40fc-bfba-502aba0b6a20 tempest-VolumesAdminNegativeTest-1387177911 tempest-VolumesAdminNegativeTest-1387177911-project-member] Inventory has not changed in ProviderTree for provider: a2858be1-fd22-4e08-979e-87ad25293407 {{(pid=61806) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 714.044505] env[61806]: DEBUG nova.compute.manager [None req-6b55cbbf-00f7-4d4d-87f3-a59cc1c553b0 tempest-ServerRescueNegativeTestJSON-1483258717 tempest-ServerRescueNegativeTestJSON-1483258717-project-member] [instance: 7ea919b6-12f7-4f4e-a123-413044bbbffc] Start building block device mappings for instance. {{(pid=61806) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 714.276963] env[61806]: DEBUG oslo_concurrency.lockutils [None req-7bb45eee-0cbe-4995-9898-530114f6f104 tempest-ImagesOneServerNegativeTestJSON-606465252 tempest-ImagesOneServerNegativeTestJSON-606465252-project-member] Lock "8d0cdc81-56ed-4a7a-8221-30bb906e16b2" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 148.671s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 714.481589] env[61806]: DEBUG nova.scheduler.client.report [None req-865353ee-538a-40fc-bfba-502aba0b6a20 tempest-VolumesAdminNegativeTest-1387177911 tempest-VolumesAdminNegativeTest-1387177911-project-member] Inventory has not changed for provider a2858be1-fd22-4e08-979e-87ad25293407 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 138, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61806) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 714.778289] env[61806]: DEBUG nova.compute.manager [None req-d17e7876-7c69-44af-acb2-32ed225ce945 tempest-ServerActionsTestOtherA-1637843849 tempest-ServerActionsTestOtherA-1637843849-project-member] [instance: 15476ac0-289a-4e04-aa9d-4244c658e962] Starting instance... {{(pid=61806) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 714.948402] env[61806]: DEBUG nova.compute.manager [req-3cf3cacd-49a7-4251-ac79-d9aab7571468 req-5368049e-68cc-4a12-b839-072eb04de969 service nova] [instance: 7ea919b6-12f7-4f4e-a123-413044bbbffc] Received event network-changed-61c8084d-6ad0-442a-967a-46f14ffa6671 {{(pid=61806) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 714.948605] env[61806]: DEBUG nova.compute.manager [req-3cf3cacd-49a7-4251-ac79-d9aab7571468 req-5368049e-68cc-4a12-b839-072eb04de969 service nova] [instance: 7ea919b6-12f7-4f4e-a123-413044bbbffc] Refreshing instance network info cache due to event network-changed-61c8084d-6ad0-442a-967a-46f14ffa6671. {{(pid=61806) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 714.948815] env[61806]: DEBUG oslo_concurrency.lockutils [req-3cf3cacd-49a7-4251-ac79-d9aab7571468 req-5368049e-68cc-4a12-b839-072eb04de969 service nova] Acquiring lock "refresh_cache-7ea919b6-12f7-4f4e-a123-413044bbbffc" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 714.948960] env[61806]: DEBUG oslo_concurrency.lockutils [req-3cf3cacd-49a7-4251-ac79-d9aab7571468 req-5368049e-68cc-4a12-b839-072eb04de969 service nova] Acquired lock "refresh_cache-7ea919b6-12f7-4f4e-a123-413044bbbffc" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 714.949182] env[61806]: DEBUG nova.network.neutron [req-3cf3cacd-49a7-4251-ac79-d9aab7571468 req-5368049e-68cc-4a12-b839-072eb04de969 service nova] [instance: 7ea919b6-12f7-4f4e-a123-413044bbbffc] Refreshing network info cache for port 61c8084d-6ad0-442a-967a-46f14ffa6671 {{(pid=61806) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 714.986698] env[61806]: DEBUG oslo_concurrency.lockutils [None req-865353ee-538a-40fc-bfba-502aba0b6a20 tempest-VolumesAdminNegativeTest-1387177911 tempest-VolumesAdminNegativeTest-1387177911-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.951s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 714.987114] env[61806]: ERROR nova.compute.manager [None req-865353ee-538a-40fc-bfba-502aba0b6a20 tempest-VolumesAdminNegativeTest-1387177911 tempest-VolumesAdminNegativeTest-1387177911-project-member] [instance: b7d55e87-95db-4c45-9417-bbece4fbb98d] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 03736e5c-7351-4a9e-ab04-f54729f3099f, please check neutron logs for more information. [ 714.987114] env[61806]: ERROR nova.compute.manager [instance: b7d55e87-95db-4c45-9417-bbece4fbb98d] Traceback (most recent call last): [ 714.987114] env[61806]: ERROR nova.compute.manager [instance: b7d55e87-95db-4c45-9417-bbece4fbb98d] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 714.987114] env[61806]: ERROR nova.compute.manager [instance: b7d55e87-95db-4c45-9417-bbece4fbb98d] self.driver.spawn(context, instance, image_meta, [ 714.987114] env[61806]: ERROR nova.compute.manager [instance: b7d55e87-95db-4c45-9417-bbece4fbb98d] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 714.987114] env[61806]: ERROR nova.compute.manager [instance: b7d55e87-95db-4c45-9417-bbece4fbb98d] self._vmops.spawn(context, instance, image_meta, injected_files, [ 714.987114] env[61806]: ERROR nova.compute.manager [instance: b7d55e87-95db-4c45-9417-bbece4fbb98d] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 714.987114] env[61806]: ERROR nova.compute.manager [instance: b7d55e87-95db-4c45-9417-bbece4fbb98d] vm_ref = self.build_virtual_machine(instance, [ 714.987114] env[61806]: ERROR nova.compute.manager [instance: b7d55e87-95db-4c45-9417-bbece4fbb98d] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 714.987114] env[61806]: ERROR nova.compute.manager [instance: b7d55e87-95db-4c45-9417-bbece4fbb98d] vif_infos = vmwarevif.get_vif_info(self._session, [ 714.987114] env[61806]: ERROR nova.compute.manager [instance: b7d55e87-95db-4c45-9417-bbece4fbb98d] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 714.987508] env[61806]: ERROR nova.compute.manager [instance: b7d55e87-95db-4c45-9417-bbece4fbb98d] for vif in network_info: [ 714.987508] env[61806]: ERROR nova.compute.manager [instance: b7d55e87-95db-4c45-9417-bbece4fbb98d] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 714.987508] env[61806]: ERROR nova.compute.manager [instance: b7d55e87-95db-4c45-9417-bbece4fbb98d] return self._sync_wrapper(fn, *args, **kwargs) [ 714.987508] env[61806]: ERROR nova.compute.manager [instance: b7d55e87-95db-4c45-9417-bbece4fbb98d] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 714.987508] env[61806]: ERROR nova.compute.manager [instance: b7d55e87-95db-4c45-9417-bbece4fbb98d] self.wait() [ 714.987508] env[61806]: ERROR nova.compute.manager [instance: b7d55e87-95db-4c45-9417-bbece4fbb98d] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 714.987508] env[61806]: ERROR nova.compute.manager [instance: b7d55e87-95db-4c45-9417-bbece4fbb98d] self[:] = self._gt.wait() [ 714.987508] env[61806]: ERROR nova.compute.manager [instance: b7d55e87-95db-4c45-9417-bbece4fbb98d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 714.987508] env[61806]: ERROR nova.compute.manager [instance: b7d55e87-95db-4c45-9417-bbece4fbb98d] return self._exit_event.wait() [ 714.987508] env[61806]: ERROR nova.compute.manager [instance: b7d55e87-95db-4c45-9417-bbece4fbb98d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 714.987508] env[61806]: ERROR nova.compute.manager [instance: b7d55e87-95db-4c45-9417-bbece4fbb98d] current.throw(*self._exc) [ 714.987508] env[61806]: ERROR nova.compute.manager [instance: b7d55e87-95db-4c45-9417-bbece4fbb98d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 714.987508] env[61806]: ERROR nova.compute.manager [instance: b7d55e87-95db-4c45-9417-bbece4fbb98d] result = function(*args, **kwargs) [ 714.987882] env[61806]: ERROR nova.compute.manager [instance: b7d55e87-95db-4c45-9417-bbece4fbb98d] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 714.987882] env[61806]: ERROR nova.compute.manager [instance: b7d55e87-95db-4c45-9417-bbece4fbb98d] return func(*args, **kwargs) [ 714.987882] env[61806]: ERROR nova.compute.manager [instance: b7d55e87-95db-4c45-9417-bbece4fbb98d] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 714.987882] env[61806]: ERROR nova.compute.manager [instance: b7d55e87-95db-4c45-9417-bbece4fbb98d] raise e [ 714.987882] env[61806]: ERROR nova.compute.manager [instance: b7d55e87-95db-4c45-9417-bbece4fbb98d] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 714.987882] env[61806]: ERROR nova.compute.manager [instance: b7d55e87-95db-4c45-9417-bbece4fbb98d] nwinfo = self.network_api.allocate_for_instance( [ 714.987882] env[61806]: ERROR nova.compute.manager [instance: b7d55e87-95db-4c45-9417-bbece4fbb98d] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 714.987882] env[61806]: ERROR nova.compute.manager [instance: b7d55e87-95db-4c45-9417-bbece4fbb98d] created_port_ids = self._update_ports_for_instance( [ 714.987882] env[61806]: ERROR nova.compute.manager [instance: b7d55e87-95db-4c45-9417-bbece4fbb98d] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 714.987882] env[61806]: ERROR nova.compute.manager [instance: b7d55e87-95db-4c45-9417-bbece4fbb98d] with excutils.save_and_reraise_exception(): [ 714.987882] env[61806]: ERROR nova.compute.manager [instance: b7d55e87-95db-4c45-9417-bbece4fbb98d] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 714.987882] env[61806]: ERROR nova.compute.manager [instance: b7d55e87-95db-4c45-9417-bbece4fbb98d] self.force_reraise() [ 714.987882] env[61806]: ERROR nova.compute.manager [instance: b7d55e87-95db-4c45-9417-bbece4fbb98d] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 714.988309] env[61806]: ERROR nova.compute.manager [instance: b7d55e87-95db-4c45-9417-bbece4fbb98d] raise self.value [ 714.988309] env[61806]: ERROR nova.compute.manager [instance: b7d55e87-95db-4c45-9417-bbece4fbb98d] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 714.988309] env[61806]: ERROR nova.compute.manager [instance: b7d55e87-95db-4c45-9417-bbece4fbb98d] updated_port = self._update_port( [ 714.988309] env[61806]: ERROR nova.compute.manager [instance: b7d55e87-95db-4c45-9417-bbece4fbb98d] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 714.988309] env[61806]: ERROR nova.compute.manager [instance: b7d55e87-95db-4c45-9417-bbece4fbb98d] _ensure_no_port_binding_failure(port) [ 714.988309] env[61806]: ERROR nova.compute.manager [instance: b7d55e87-95db-4c45-9417-bbece4fbb98d] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 714.988309] env[61806]: ERROR nova.compute.manager [instance: b7d55e87-95db-4c45-9417-bbece4fbb98d] raise exception.PortBindingFailed(port_id=port['id']) [ 714.988309] env[61806]: ERROR nova.compute.manager [instance: b7d55e87-95db-4c45-9417-bbece4fbb98d] nova.exception.PortBindingFailed: Binding failed for port 03736e5c-7351-4a9e-ab04-f54729f3099f, please check neutron logs for more information. [ 714.988309] env[61806]: ERROR nova.compute.manager [instance: b7d55e87-95db-4c45-9417-bbece4fbb98d] [ 714.988309] env[61806]: DEBUG nova.compute.utils [None req-865353ee-538a-40fc-bfba-502aba0b6a20 tempest-VolumesAdminNegativeTest-1387177911 tempest-VolumesAdminNegativeTest-1387177911-project-member] [instance: b7d55e87-95db-4c45-9417-bbece4fbb98d] Binding failed for port 03736e5c-7351-4a9e-ab04-f54729f3099f, please check neutron logs for more information. {{(pid=61806) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 714.988965] env[61806]: DEBUG oslo_concurrency.lockutils [None req-81a55da1-45c4-45f8-af56-f8f7d424c8aa tempest-ServersTestJSON-430716070 tempest-ServersTestJSON-430716070-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 25.189s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 714.994928] env[61806]: DEBUG nova.compute.manager [None req-865353ee-538a-40fc-bfba-502aba0b6a20 tempest-VolumesAdminNegativeTest-1387177911 tempest-VolumesAdminNegativeTest-1387177911-project-member] [instance: b7d55e87-95db-4c45-9417-bbece4fbb98d] Build of instance b7d55e87-95db-4c45-9417-bbece4fbb98d was re-scheduled: Binding failed for port 03736e5c-7351-4a9e-ab04-f54729f3099f, please check neutron logs for more information. {{(pid=61806) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 714.996595] env[61806]: DEBUG nova.compute.manager [None req-865353ee-538a-40fc-bfba-502aba0b6a20 tempest-VolumesAdminNegativeTest-1387177911 tempest-VolumesAdminNegativeTest-1387177911-project-member] [instance: b7d55e87-95db-4c45-9417-bbece4fbb98d] Unplugging VIFs for instance {{(pid=61806) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 714.996879] env[61806]: DEBUG oslo_concurrency.lockutils [None req-865353ee-538a-40fc-bfba-502aba0b6a20 tempest-VolumesAdminNegativeTest-1387177911 tempest-VolumesAdminNegativeTest-1387177911-project-member] Acquiring lock "refresh_cache-b7d55e87-95db-4c45-9417-bbece4fbb98d" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 714.996959] env[61806]: DEBUG oslo_concurrency.lockutils [None req-865353ee-538a-40fc-bfba-502aba0b6a20 tempest-VolumesAdminNegativeTest-1387177911 tempest-VolumesAdminNegativeTest-1387177911-project-member] Acquired lock "refresh_cache-b7d55e87-95db-4c45-9417-bbece4fbb98d" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 714.997134] env[61806]: DEBUG nova.network.neutron [None req-865353ee-538a-40fc-bfba-502aba0b6a20 tempest-VolumesAdminNegativeTest-1387177911 tempest-VolumesAdminNegativeTest-1387177911-project-member] [instance: b7d55e87-95db-4c45-9417-bbece4fbb98d] Building network info cache for instance {{(pid=61806) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 715.053054] env[61806]: DEBUG nova.compute.manager [None req-6b55cbbf-00f7-4d4d-87f3-a59cc1c553b0 tempest-ServerRescueNegativeTestJSON-1483258717 tempest-ServerRescueNegativeTestJSON-1483258717-project-member] [instance: 7ea919b6-12f7-4f4e-a123-413044bbbffc] Start spawning the instance on the hypervisor. {{(pid=61806) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 715.085021] env[61806]: DEBUG nova.virt.hardware [None req-6b55cbbf-00f7-4d4d-87f3-a59cc1c553b0 tempest-ServerRescueNegativeTestJSON-1483258717 tempest-ServerRescueNegativeTestJSON-1483258717-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-15T18:05:10Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-15T18:04:52Z,direct_url=,disk_format='vmdk',id=ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='84ba83fce2c349a6988173c5d6fc3b07',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-15T18:04:53Z,virtual_size=,visibility=), allow threads: False {{(pid=61806) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 715.085324] env[61806]: DEBUG nova.virt.hardware [None req-6b55cbbf-00f7-4d4d-87f3-a59cc1c553b0 tempest-ServerRescueNegativeTestJSON-1483258717 tempest-ServerRescueNegativeTestJSON-1483258717-project-member] Flavor limits 0:0:0 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 715.085438] env[61806]: DEBUG nova.virt.hardware [None req-6b55cbbf-00f7-4d4d-87f3-a59cc1c553b0 tempest-ServerRescueNegativeTestJSON-1483258717 tempest-ServerRescueNegativeTestJSON-1483258717-project-member] Image limits 0:0:0 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 715.085618] env[61806]: DEBUG nova.virt.hardware [None req-6b55cbbf-00f7-4d4d-87f3-a59cc1c553b0 tempest-ServerRescueNegativeTestJSON-1483258717 tempest-ServerRescueNegativeTestJSON-1483258717-project-member] Flavor pref 0:0:0 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 715.085765] env[61806]: DEBUG nova.virt.hardware [None req-6b55cbbf-00f7-4d4d-87f3-a59cc1c553b0 tempest-ServerRescueNegativeTestJSON-1483258717 tempest-ServerRescueNegativeTestJSON-1483258717-project-member] Image pref 0:0:0 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 715.085912] env[61806]: DEBUG nova.virt.hardware [None req-6b55cbbf-00f7-4d4d-87f3-a59cc1c553b0 tempest-ServerRescueNegativeTestJSON-1483258717 tempest-ServerRescueNegativeTestJSON-1483258717-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 715.086305] env[61806]: DEBUG nova.virt.hardware [None req-6b55cbbf-00f7-4d4d-87f3-a59cc1c553b0 tempest-ServerRescueNegativeTestJSON-1483258717 tempest-ServerRescueNegativeTestJSON-1483258717-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61806) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 715.086516] env[61806]: DEBUG nova.virt.hardware [None req-6b55cbbf-00f7-4d4d-87f3-a59cc1c553b0 tempest-ServerRescueNegativeTestJSON-1483258717 tempest-ServerRescueNegativeTestJSON-1483258717-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61806) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 715.086693] env[61806]: DEBUG nova.virt.hardware [None req-6b55cbbf-00f7-4d4d-87f3-a59cc1c553b0 tempest-ServerRescueNegativeTestJSON-1483258717 tempest-ServerRescueNegativeTestJSON-1483258717-project-member] Got 1 possible topologies {{(pid=61806) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 715.086859] env[61806]: DEBUG nova.virt.hardware [None req-6b55cbbf-00f7-4d4d-87f3-a59cc1c553b0 tempest-ServerRescueNegativeTestJSON-1483258717 tempest-ServerRescueNegativeTestJSON-1483258717-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61806) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 715.087542] env[61806]: DEBUG nova.virt.hardware [None req-6b55cbbf-00f7-4d4d-87f3-a59cc1c553b0 tempest-ServerRescueNegativeTestJSON-1483258717 tempest-ServerRescueNegativeTestJSON-1483258717-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61806) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 715.088187] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-77697b40-b3d4-4a45-a5f8-6f74b12827f6 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 715.096100] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ef3cfbb6-9426-420e-9bca-5db56e6cae8b {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 715.100906] env[61806]: ERROR nova.compute.manager [None req-6b55cbbf-00f7-4d4d-87f3-a59cc1c553b0 tempest-ServerRescueNegativeTestJSON-1483258717 tempest-ServerRescueNegativeTestJSON-1483258717-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 61c8084d-6ad0-442a-967a-46f14ffa6671, please check neutron logs for more information. [ 715.100906] env[61806]: ERROR nova.compute.manager Traceback (most recent call last): [ 715.100906] env[61806]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 715.100906] env[61806]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 715.100906] env[61806]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 715.100906] env[61806]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 715.100906] env[61806]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 715.100906] env[61806]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 715.100906] env[61806]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 715.100906] env[61806]: ERROR nova.compute.manager self.force_reraise() [ 715.100906] env[61806]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 715.100906] env[61806]: ERROR nova.compute.manager raise self.value [ 715.100906] env[61806]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 715.100906] env[61806]: ERROR nova.compute.manager updated_port = self._update_port( [ 715.100906] env[61806]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 715.100906] env[61806]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 715.101433] env[61806]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 715.101433] env[61806]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 715.101433] env[61806]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 61c8084d-6ad0-442a-967a-46f14ffa6671, please check neutron logs for more information. [ 715.101433] env[61806]: ERROR nova.compute.manager [ 715.101433] env[61806]: Traceback (most recent call last): [ 715.101433] env[61806]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 715.101433] env[61806]: listener.cb(fileno) [ 715.101433] env[61806]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 715.101433] env[61806]: result = function(*args, **kwargs) [ 715.101433] env[61806]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 715.101433] env[61806]: return func(*args, **kwargs) [ 715.101433] env[61806]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 715.101433] env[61806]: raise e [ 715.101433] env[61806]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 715.101433] env[61806]: nwinfo = self.network_api.allocate_for_instance( [ 715.101433] env[61806]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 715.101433] env[61806]: created_port_ids = self._update_ports_for_instance( [ 715.101433] env[61806]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 715.101433] env[61806]: with excutils.save_and_reraise_exception(): [ 715.101433] env[61806]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 715.101433] env[61806]: self.force_reraise() [ 715.101433] env[61806]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 715.101433] env[61806]: raise self.value [ 715.101433] env[61806]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 715.101433] env[61806]: updated_port = self._update_port( [ 715.101433] env[61806]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 715.101433] env[61806]: _ensure_no_port_binding_failure(port) [ 715.101433] env[61806]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 715.101433] env[61806]: raise exception.PortBindingFailed(port_id=port['id']) [ 715.103421] env[61806]: nova.exception.PortBindingFailed: Binding failed for port 61c8084d-6ad0-442a-967a-46f14ffa6671, please check neutron logs for more information. [ 715.103421] env[61806]: Removing descriptor: 14 [ 715.111567] env[61806]: ERROR nova.compute.manager [None req-6b55cbbf-00f7-4d4d-87f3-a59cc1c553b0 tempest-ServerRescueNegativeTestJSON-1483258717 tempest-ServerRescueNegativeTestJSON-1483258717-project-member] [instance: 7ea919b6-12f7-4f4e-a123-413044bbbffc] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 61c8084d-6ad0-442a-967a-46f14ffa6671, please check neutron logs for more information. [ 715.111567] env[61806]: ERROR nova.compute.manager [instance: 7ea919b6-12f7-4f4e-a123-413044bbbffc] Traceback (most recent call last): [ 715.111567] env[61806]: ERROR nova.compute.manager [instance: 7ea919b6-12f7-4f4e-a123-413044bbbffc] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 715.111567] env[61806]: ERROR nova.compute.manager [instance: 7ea919b6-12f7-4f4e-a123-413044bbbffc] yield resources [ 715.111567] env[61806]: ERROR nova.compute.manager [instance: 7ea919b6-12f7-4f4e-a123-413044bbbffc] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 715.111567] env[61806]: ERROR nova.compute.manager [instance: 7ea919b6-12f7-4f4e-a123-413044bbbffc] self.driver.spawn(context, instance, image_meta, [ 715.111567] env[61806]: ERROR nova.compute.manager [instance: 7ea919b6-12f7-4f4e-a123-413044bbbffc] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 715.111567] env[61806]: ERROR nova.compute.manager [instance: 7ea919b6-12f7-4f4e-a123-413044bbbffc] self._vmops.spawn(context, instance, image_meta, injected_files, [ 715.111567] env[61806]: ERROR nova.compute.manager [instance: 7ea919b6-12f7-4f4e-a123-413044bbbffc] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 715.111567] env[61806]: ERROR nova.compute.manager [instance: 7ea919b6-12f7-4f4e-a123-413044bbbffc] vm_ref = self.build_virtual_machine(instance, [ 715.111567] env[61806]: ERROR nova.compute.manager [instance: 7ea919b6-12f7-4f4e-a123-413044bbbffc] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 715.111960] env[61806]: ERROR nova.compute.manager [instance: 7ea919b6-12f7-4f4e-a123-413044bbbffc] vif_infos = vmwarevif.get_vif_info(self._session, [ 715.111960] env[61806]: ERROR nova.compute.manager [instance: 7ea919b6-12f7-4f4e-a123-413044bbbffc] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 715.111960] env[61806]: ERROR nova.compute.manager [instance: 7ea919b6-12f7-4f4e-a123-413044bbbffc] for vif in network_info: [ 715.111960] env[61806]: ERROR nova.compute.manager [instance: 7ea919b6-12f7-4f4e-a123-413044bbbffc] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 715.111960] env[61806]: ERROR nova.compute.manager [instance: 7ea919b6-12f7-4f4e-a123-413044bbbffc] return self._sync_wrapper(fn, *args, **kwargs) [ 715.111960] env[61806]: ERROR nova.compute.manager [instance: 7ea919b6-12f7-4f4e-a123-413044bbbffc] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 715.111960] env[61806]: ERROR nova.compute.manager [instance: 7ea919b6-12f7-4f4e-a123-413044bbbffc] self.wait() [ 715.111960] env[61806]: ERROR nova.compute.manager [instance: 7ea919b6-12f7-4f4e-a123-413044bbbffc] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 715.111960] env[61806]: ERROR nova.compute.manager [instance: 7ea919b6-12f7-4f4e-a123-413044bbbffc] self[:] = self._gt.wait() [ 715.111960] env[61806]: ERROR nova.compute.manager [instance: 7ea919b6-12f7-4f4e-a123-413044bbbffc] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 715.111960] env[61806]: ERROR nova.compute.manager [instance: 7ea919b6-12f7-4f4e-a123-413044bbbffc] return self._exit_event.wait() [ 715.111960] env[61806]: ERROR nova.compute.manager [instance: 7ea919b6-12f7-4f4e-a123-413044bbbffc] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 715.111960] env[61806]: ERROR nova.compute.manager [instance: 7ea919b6-12f7-4f4e-a123-413044bbbffc] current.throw(*self._exc) [ 715.112453] env[61806]: ERROR nova.compute.manager [instance: 7ea919b6-12f7-4f4e-a123-413044bbbffc] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 715.112453] env[61806]: ERROR nova.compute.manager [instance: 7ea919b6-12f7-4f4e-a123-413044bbbffc] result = function(*args, **kwargs) [ 715.112453] env[61806]: ERROR nova.compute.manager [instance: 7ea919b6-12f7-4f4e-a123-413044bbbffc] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 715.112453] env[61806]: ERROR nova.compute.manager [instance: 7ea919b6-12f7-4f4e-a123-413044bbbffc] return func(*args, **kwargs) [ 715.112453] env[61806]: ERROR nova.compute.manager [instance: 7ea919b6-12f7-4f4e-a123-413044bbbffc] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 715.112453] env[61806]: ERROR nova.compute.manager [instance: 7ea919b6-12f7-4f4e-a123-413044bbbffc] raise e [ 715.112453] env[61806]: ERROR nova.compute.manager [instance: 7ea919b6-12f7-4f4e-a123-413044bbbffc] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 715.112453] env[61806]: ERROR nova.compute.manager [instance: 7ea919b6-12f7-4f4e-a123-413044bbbffc] nwinfo = self.network_api.allocate_for_instance( [ 715.112453] env[61806]: ERROR nova.compute.manager [instance: 7ea919b6-12f7-4f4e-a123-413044bbbffc] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 715.112453] env[61806]: ERROR nova.compute.manager [instance: 7ea919b6-12f7-4f4e-a123-413044bbbffc] created_port_ids = self._update_ports_for_instance( [ 715.112453] env[61806]: ERROR nova.compute.manager [instance: 7ea919b6-12f7-4f4e-a123-413044bbbffc] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 715.112453] env[61806]: ERROR nova.compute.manager [instance: 7ea919b6-12f7-4f4e-a123-413044bbbffc] with excutils.save_and_reraise_exception(): [ 715.112453] env[61806]: ERROR nova.compute.manager [instance: 7ea919b6-12f7-4f4e-a123-413044bbbffc] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 715.112855] env[61806]: ERROR nova.compute.manager [instance: 7ea919b6-12f7-4f4e-a123-413044bbbffc] self.force_reraise() [ 715.112855] env[61806]: ERROR nova.compute.manager [instance: 7ea919b6-12f7-4f4e-a123-413044bbbffc] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 715.112855] env[61806]: ERROR nova.compute.manager [instance: 7ea919b6-12f7-4f4e-a123-413044bbbffc] raise self.value [ 715.112855] env[61806]: ERROR nova.compute.manager [instance: 7ea919b6-12f7-4f4e-a123-413044bbbffc] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 715.112855] env[61806]: ERROR nova.compute.manager [instance: 7ea919b6-12f7-4f4e-a123-413044bbbffc] updated_port = self._update_port( [ 715.112855] env[61806]: ERROR nova.compute.manager [instance: 7ea919b6-12f7-4f4e-a123-413044bbbffc] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 715.112855] env[61806]: ERROR nova.compute.manager [instance: 7ea919b6-12f7-4f4e-a123-413044bbbffc] _ensure_no_port_binding_failure(port) [ 715.112855] env[61806]: ERROR nova.compute.manager [instance: 7ea919b6-12f7-4f4e-a123-413044bbbffc] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 715.112855] env[61806]: ERROR nova.compute.manager [instance: 7ea919b6-12f7-4f4e-a123-413044bbbffc] raise exception.PortBindingFailed(port_id=port['id']) [ 715.112855] env[61806]: ERROR nova.compute.manager [instance: 7ea919b6-12f7-4f4e-a123-413044bbbffc] nova.exception.PortBindingFailed: Binding failed for port 61c8084d-6ad0-442a-967a-46f14ffa6671, please check neutron logs for more information. [ 715.112855] env[61806]: ERROR nova.compute.manager [instance: 7ea919b6-12f7-4f4e-a123-413044bbbffc] [ 715.112855] env[61806]: INFO nova.compute.manager [None req-6b55cbbf-00f7-4d4d-87f3-a59cc1c553b0 tempest-ServerRescueNegativeTestJSON-1483258717 tempest-ServerRescueNegativeTestJSON-1483258717-project-member] [instance: 7ea919b6-12f7-4f4e-a123-413044bbbffc] Terminating instance [ 715.114081] env[61806]: DEBUG oslo_concurrency.lockutils [None req-6b55cbbf-00f7-4d4d-87f3-a59cc1c553b0 tempest-ServerRescueNegativeTestJSON-1483258717 tempest-ServerRescueNegativeTestJSON-1483258717-project-member] Acquiring lock "refresh_cache-7ea919b6-12f7-4f4e-a123-413044bbbffc" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 715.300273] env[61806]: DEBUG oslo_concurrency.lockutils [None req-d17e7876-7c69-44af-acb2-32ed225ce945 tempest-ServerActionsTestOtherA-1637843849 tempest-ServerActionsTestOtherA-1637843849-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 715.472834] env[61806]: DEBUG nova.network.neutron [req-3cf3cacd-49a7-4251-ac79-d9aab7571468 req-5368049e-68cc-4a12-b839-072eb04de969 service nova] [instance: 7ea919b6-12f7-4f4e-a123-413044bbbffc] Instance cache missing network info. {{(pid=61806) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 715.517253] env[61806]: DEBUG nova.network.neutron [None req-865353ee-538a-40fc-bfba-502aba0b6a20 tempest-VolumesAdminNegativeTest-1387177911 tempest-VolumesAdminNegativeTest-1387177911-project-member] [instance: b7d55e87-95db-4c45-9417-bbece4fbb98d] Instance cache missing network info. {{(pid=61806) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 715.597681] env[61806]: DEBUG nova.network.neutron [None req-865353ee-538a-40fc-bfba-502aba0b6a20 tempest-VolumesAdminNegativeTest-1387177911 tempest-VolumesAdminNegativeTest-1387177911-project-member] [instance: b7d55e87-95db-4c45-9417-bbece4fbb98d] Updating instance_info_cache with network_info: [] {{(pid=61806) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 715.624692] env[61806]: DEBUG nova.network.neutron [req-3cf3cacd-49a7-4251-ac79-d9aab7571468 req-5368049e-68cc-4a12-b839-072eb04de969 service nova] [instance: 7ea919b6-12f7-4f4e-a123-413044bbbffc] Updating instance_info_cache with network_info: [] {{(pid=61806) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 715.952825] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-10025ae0-f6ec-4543-b1d4-59f983779fc6 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 715.961735] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-25d91ac5-7986-40e5-86dd-45de5751a95c {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 715.991785] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0f19ef88-d9b3-4350-a266-3458e48ce47f {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 715.999215] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9a6f5fb8-25fa-4e39-8c6b-e1419d75b62e {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 716.012381] env[61806]: DEBUG nova.compute.provider_tree [None req-81a55da1-45c4-45f8-af56-f8f7d424c8aa tempest-ServersTestJSON-430716070 tempest-ServersTestJSON-430716070-project-member] Inventory has not changed in ProviderTree for provider: a2858be1-fd22-4e08-979e-87ad25293407 {{(pid=61806) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 716.101261] env[61806]: DEBUG oslo_concurrency.lockutils [None req-865353ee-538a-40fc-bfba-502aba0b6a20 tempest-VolumesAdminNegativeTest-1387177911 tempest-VolumesAdminNegativeTest-1387177911-project-member] Releasing lock "refresh_cache-b7d55e87-95db-4c45-9417-bbece4fbb98d" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 716.101537] env[61806]: DEBUG nova.compute.manager [None req-865353ee-538a-40fc-bfba-502aba0b6a20 tempest-VolumesAdminNegativeTest-1387177911 tempest-VolumesAdminNegativeTest-1387177911-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61806) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 716.101691] env[61806]: DEBUG nova.compute.manager [None req-865353ee-538a-40fc-bfba-502aba0b6a20 tempest-VolumesAdminNegativeTest-1387177911 tempest-VolumesAdminNegativeTest-1387177911-project-member] [instance: b7d55e87-95db-4c45-9417-bbece4fbb98d] Deallocating network for instance {{(pid=61806) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 716.101862] env[61806]: DEBUG nova.network.neutron [None req-865353ee-538a-40fc-bfba-502aba0b6a20 tempest-VolumesAdminNegativeTest-1387177911 tempest-VolumesAdminNegativeTest-1387177911-project-member] [instance: b7d55e87-95db-4c45-9417-bbece4fbb98d] deallocate_for_instance() {{(pid=61806) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 716.121197] env[61806]: DEBUG nova.network.neutron [None req-865353ee-538a-40fc-bfba-502aba0b6a20 tempest-VolumesAdminNegativeTest-1387177911 tempest-VolumesAdminNegativeTest-1387177911-project-member] [instance: b7d55e87-95db-4c45-9417-bbece4fbb98d] Instance cache missing network info. {{(pid=61806) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 716.127269] env[61806]: DEBUG oslo_concurrency.lockutils [req-3cf3cacd-49a7-4251-ac79-d9aab7571468 req-5368049e-68cc-4a12-b839-072eb04de969 service nova] Releasing lock "refresh_cache-7ea919b6-12f7-4f4e-a123-413044bbbffc" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 716.127657] env[61806]: DEBUG oslo_concurrency.lockutils [None req-6b55cbbf-00f7-4d4d-87f3-a59cc1c553b0 tempest-ServerRescueNegativeTestJSON-1483258717 tempest-ServerRescueNegativeTestJSON-1483258717-project-member] Acquired lock "refresh_cache-7ea919b6-12f7-4f4e-a123-413044bbbffc" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 716.127837] env[61806]: DEBUG nova.network.neutron [None req-6b55cbbf-00f7-4d4d-87f3-a59cc1c553b0 tempest-ServerRescueNegativeTestJSON-1483258717 tempest-ServerRescueNegativeTestJSON-1483258717-project-member] [instance: 7ea919b6-12f7-4f4e-a123-413044bbbffc] Building network info cache for instance {{(pid=61806) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 716.520627] env[61806]: DEBUG nova.scheduler.client.report [None req-81a55da1-45c4-45f8-af56-f8f7d424c8aa tempest-ServersTestJSON-430716070 tempest-ServersTestJSON-430716070-project-member] Inventory has not changed for provider a2858be1-fd22-4e08-979e-87ad25293407 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 138, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61806) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 716.623469] env[61806]: DEBUG nova.network.neutron [None req-865353ee-538a-40fc-bfba-502aba0b6a20 tempest-VolumesAdminNegativeTest-1387177911 tempest-VolumesAdminNegativeTest-1387177911-project-member] [instance: b7d55e87-95db-4c45-9417-bbece4fbb98d] Updating instance_info_cache with network_info: [] {{(pid=61806) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 716.810926] env[61806]: DEBUG nova.network.neutron [None req-6b55cbbf-00f7-4d4d-87f3-a59cc1c553b0 tempest-ServerRescueNegativeTestJSON-1483258717 tempest-ServerRescueNegativeTestJSON-1483258717-project-member] [instance: 7ea919b6-12f7-4f4e-a123-413044bbbffc] Instance cache missing network info. {{(pid=61806) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 716.966593] env[61806]: DEBUG nova.network.neutron [None req-6b55cbbf-00f7-4d4d-87f3-a59cc1c553b0 tempest-ServerRescueNegativeTestJSON-1483258717 tempest-ServerRescueNegativeTestJSON-1483258717-project-member] [instance: 7ea919b6-12f7-4f4e-a123-413044bbbffc] Updating instance_info_cache with network_info: [] {{(pid=61806) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 716.980364] env[61806]: DEBUG nova.compute.manager [req-eeb2d69c-fd8f-4793-9e55-93d158d32614 req-d0c99196-94d6-40d6-82fc-cd5b1ec9b8ff service nova] [instance: 7ea919b6-12f7-4f4e-a123-413044bbbffc] Received event network-vif-deleted-61c8084d-6ad0-442a-967a-46f14ffa6671 {{(pid=61806) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 717.026437] env[61806]: DEBUG oslo_concurrency.lockutils [None req-81a55da1-45c4-45f8-af56-f8f7d424c8aa tempest-ServersTestJSON-430716070 tempest-ServersTestJSON-430716070-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.037s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 717.027323] env[61806]: ERROR nova.compute.manager [None req-81a55da1-45c4-45f8-af56-f8f7d424c8aa tempest-ServersTestJSON-430716070 tempest-ServersTestJSON-430716070-project-member] [instance: eca2e3da-b38c-482e-b9b6-4ad75d83caa1] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 007a0ffc-be2b-4f45-b7a2-5df4203fb2da, please check neutron logs for more information. [ 717.027323] env[61806]: ERROR nova.compute.manager [instance: eca2e3da-b38c-482e-b9b6-4ad75d83caa1] Traceback (most recent call last): [ 717.027323] env[61806]: ERROR nova.compute.manager [instance: eca2e3da-b38c-482e-b9b6-4ad75d83caa1] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 717.027323] env[61806]: ERROR nova.compute.manager [instance: eca2e3da-b38c-482e-b9b6-4ad75d83caa1] self.driver.spawn(context, instance, image_meta, [ 717.027323] env[61806]: ERROR nova.compute.manager [instance: eca2e3da-b38c-482e-b9b6-4ad75d83caa1] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 717.027323] env[61806]: ERROR nova.compute.manager [instance: eca2e3da-b38c-482e-b9b6-4ad75d83caa1] self._vmops.spawn(context, instance, image_meta, injected_files, [ 717.027323] env[61806]: ERROR nova.compute.manager [instance: eca2e3da-b38c-482e-b9b6-4ad75d83caa1] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 717.027323] env[61806]: ERROR nova.compute.manager [instance: eca2e3da-b38c-482e-b9b6-4ad75d83caa1] vm_ref = self.build_virtual_machine(instance, [ 717.027323] env[61806]: ERROR nova.compute.manager [instance: eca2e3da-b38c-482e-b9b6-4ad75d83caa1] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 717.027323] env[61806]: ERROR nova.compute.manager [instance: eca2e3da-b38c-482e-b9b6-4ad75d83caa1] vif_infos = vmwarevif.get_vif_info(self._session, [ 717.027323] env[61806]: ERROR nova.compute.manager [instance: eca2e3da-b38c-482e-b9b6-4ad75d83caa1] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 717.027660] env[61806]: ERROR nova.compute.manager [instance: eca2e3da-b38c-482e-b9b6-4ad75d83caa1] for vif in network_info: [ 717.027660] env[61806]: ERROR nova.compute.manager [instance: eca2e3da-b38c-482e-b9b6-4ad75d83caa1] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 717.027660] env[61806]: ERROR nova.compute.manager [instance: eca2e3da-b38c-482e-b9b6-4ad75d83caa1] return self._sync_wrapper(fn, *args, **kwargs) [ 717.027660] env[61806]: ERROR nova.compute.manager [instance: eca2e3da-b38c-482e-b9b6-4ad75d83caa1] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 717.027660] env[61806]: ERROR nova.compute.manager [instance: eca2e3da-b38c-482e-b9b6-4ad75d83caa1] self.wait() [ 717.027660] env[61806]: ERROR nova.compute.manager [instance: eca2e3da-b38c-482e-b9b6-4ad75d83caa1] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 717.027660] env[61806]: ERROR nova.compute.manager [instance: eca2e3da-b38c-482e-b9b6-4ad75d83caa1] self[:] = self._gt.wait() [ 717.027660] env[61806]: ERROR nova.compute.manager [instance: eca2e3da-b38c-482e-b9b6-4ad75d83caa1] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 717.027660] env[61806]: ERROR nova.compute.manager [instance: eca2e3da-b38c-482e-b9b6-4ad75d83caa1] return self._exit_event.wait() [ 717.027660] env[61806]: ERROR nova.compute.manager [instance: eca2e3da-b38c-482e-b9b6-4ad75d83caa1] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 717.027660] env[61806]: ERROR nova.compute.manager [instance: eca2e3da-b38c-482e-b9b6-4ad75d83caa1] current.throw(*self._exc) [ 717.027660] env[61806]: ERROR nova.compute.manager [instance: eca2e3da-b38c-482e-b9b6-4ad75d83caa1] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 717.027660] env[61806]: ERROR nova.compute.manager [instance: eca2e3da-b38c-482e-b9b6-4ad75d83caa1] result = function(*args, **kwargs) [ 717.028913] env[61806]: ERROR nova.compute.manager [instance: eca2e3da-b38c-482e-b9b6-4ad75d83caa1] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 717.028913] env[61806]: ERROR nova.compute.manager [instance: eca2e3da-b38c-482e-b9b6-4ad75d83caa1] return func(*args, **kwargs) [ 717.028913] env[61806]: ERROR nova.compute.manager [instance: eca2e3da-b38c-482e-b9b6-4ad75d83caa1] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 717.028913] env[61806]: ERROR nova.compute.manager [instance: eca2e3da-b38c-482e-b9b6-4ad75d83caa1] raise e [ 717.028913] env[61806]: ERROR nova.compute.manager [instance: eca2e3da-b38c-482e-b9b6-4ad75d83caa1] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 717.028913] env[61806]: ERROR nova.compute.manager [instance: eca2e3da-b38c-482e-b9b6-4ad75d83caa1] nwinfo = self.network_api.allocate_for_instance( [ 717.028913] env[61806]: ERROR nova.compute.manager [instance: eca2e3da-b38c-482e-b9b6-4ad75d83caa1] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 717.028913] env[61806]: ERROR nova.compute.manager [instance: eca2e3da-b38c-482e-b9b6-4ad75d83caa1] created_port_ids = self._update_ports_for_instance( [ 717.028913] env[61806]: ERROR nova.compute.manager [instance: eca2e3da-b38c-482e-b9b6-4ad75d83caa1] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 717.028913] env[61806]: ERROR nova.compute.manager [instance: eca2e3da-b38c-482e-b9b6-4ad75d83caa1] with excutils.save_and_reraise_exception(): [ 717.028913] env[61806]: ERROR nova.compute.manager [instance: eca2e3da-b38c-482e-b9b6-4ad75d83caa1] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 717.028913] env[61806]: ERROR nova.compute.manager [instance: eca2e3da-b38c-482e-b9b6-4ad75d83caa1] self.force_reraise() [ 717.028913] env[61806]: ERROR nova.compute.manager [instance: eca2e3da-b38c-482e-b9b6-4ad75d83caa1] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 717.029430] env[61806]: ERROR nova.compute.manager [instance: eca2e3da-b38c-482e-b9b6-4ad75d83caa1] raise self.value [ 717.029430] env[61806]: ERROR nova.compute.manager [instance: eca2e3da-b38c-482e-b9b6-4ad75d83caa1] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 717.029430] env[61806]: ERROR nova.compute.manager [instance: eca2e3da-b38c-482e-b9b6-4ad75d83caa1] updated_port = self._update_port( [ 717.029430] env[61806]: ERROR nova.compute.manager [instance: eca2e3da-b38c-482e-b9b6-4ad75d83caa1] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 717.029430] env[61806]: ERROR nova.compute.manager [instance: eca2e3da-b38c-482e-b9b6-4ad75d83caa1] _ensure_no_port_binding_failure(port) [ 717.029430] env[61806]: ERROR nova.compute.manager [instance: eca2e3da-b38c-482e-b9b6-4ad75d83caa1] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 717.029430] env[61806]: ERROR nova.compute.manager [instance: eca2e3da-b38c-482e-b9b6-4ad75d83caa1] raise exception.PortBindingFailed(port_id=port['id']) [ 717.029430] env[61806]: ERROR nova.compute.manager [instance: eca2e3da-b38c-482e-b9b6-4ad75d83caa1] nova.exception.PortBindingFailed: Binding failed for port 007a0ffc-be2b-4f45-b7a2-5df4203fb2da, please check neutron logs for more information. [ 717.029430] env[61806]: ERROR nova.compute.manager [instance: eca2e3da-b38c-482e-b9b6-4ad75d83caa1] [ 717.029430] env[61806]: DEBUG nova.compute.utils [None req-81a55da1-45c4-45f8-af56-f8f7d424c8aa tempest-ServersTestJSON-430716070 tempest-ServersTestJSON-430716070-project-member] [instance: eca2e3da-b38c-482e-b9b6-4ad75d83caa1] Binding failed for port 007a0ffc-be2b-4f45-b7a2-5df4203fb2da, please check neutron logs for more information. {{(pid=61806) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 717.030091] env[61806]: DEBUG oslo_concurrency.lockutils [None req-3c3981d2-fa25-4f4b-87d4-a5411e67199e tempest-MultipleCreateTestJSON-1589015462 tempest-MultipleCreateTestJSON-1589015462-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 24.798s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 717.032104] env[61806]: INFO nova.compute.claims [None req-3c3981d2-fa25-4f4b-87d4-a5411e67199e tempest-MultipleCreateTestJSON-1589015462 tempest-MultipleCreateTestJSON-1589015462-project-member] [instance: 7211d568-cc68-4e50-80ad-b78878d1deab] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 717.035035] env[61806]: DEBUG nova.compute.manager [None req-81a55da1-45c4-45f8-af56-f8f7d424c8aa tempest-ServersTestJSON-430716070 tempest-ServersTestJSON-430716070-project-member] [instance: eca2e3da-b38c-482e-b9b6-4ad75d83caa1] Build of instance eca2e3da-b38c-482e-b9b6-4ad75d83caa1 was re-scheduled: Binding failed for port 007a0ffc-be2b-4f45-b7a2-5df4203fb2da, please check neutron logs for more information. {{(pid=61806) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 717.035573] env[61806]: DEBUG nova.compute.manager [None req-81a55da1-45c4-45f8-af56-f8f7d424c8aa tempest-ServersTestJSON-430716070 tempest-ServersTestJSON-430716070-project-member] [instance: eca2e3da-b38c-482e-b9b6-4ad75d83caa1] Unplugging VIFs for instance {{(pid=61806) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 717.035804] env[61806]: DEBUG oslo_concurrency.lockutils [None req-81a55da1-45c4-45f8-af56-f8f7d424c8aa tempest-ServersTestJSON-430716070 tempest-ServersTestJSON-430716070-project-member] Acquiring lock "refresh_cache-eca2e3da-b38c-482e-b9b6-4ad75d83caa1" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 717.035950] env[61806]: DEBUG oslo_concurrency.lockutils [None req-81a55da1-45c4-45f8-af56-f8f7d424c8aa tempest-ServersTestJSON-430716070 tempest-ServersTestJSON-430716070-project-member] Acquired lock "refresh_cache-eca2e3da-b38c-482e-b9b6-4ad75d83caa1" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 717.036122] env[61806]: DEBUG nova.network.neutron [None req-81a55da1-45c4-45f8-af56-f8f7d424c8aa tempest-ServersTestJSON-430716070 tempest-ServersTestJSON-430716070-project-member] [instance: eca2e3da-b38c-482e-b9b6-4ad75d83caa1] Building network info cache for instance {{(pid=61806) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 717.126760] env[61806]: INFO nova.compute.manager [None req-865353ee-538a-40fc-bfba-502aba0b6a20 tempest-VolumesAdminNegativeTest-1387177911 tempest-VolumesAdminNegativeTest-1387177911-project-member] [instance: b7d55e87-95db-4c45-9417-bbece4fbb98d] Took 1.02 seconds to deallocate network for instance. [ 717.469814] env[61806]: DEBUG oslo_concurrency.lockutils [None req-6b55cbbf-00f7-4d4d-87f3-a59cc1c553b0 tempest-ServerRescueNegativeTestJSON-1483258717 tempest-ServerRescueNegativeTestJSON-1483258717-project-member] Releasing lock "refresh_cache-7ea919b6-12f7-4f4e-a123-413044bbbffc" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 717.470285] env[61806]: DEBUG nova.compute.manager [None req-6b55cbbf-00f7-4d4d-87f3-a59cc1c553b0 tempest-ServerRescueNegativeTestJSON-1483258717 tempest-ServerRescueNegativeTestJSON-1483258717-project-member] [instance: 7ea919b6-12f7-4f4e-a123-413044bbbffc] Start destroying the instance on the hypervisor. {{(pid=61806) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 717.470486] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-6b55cbbf-00f7-4d4d-87f3-a59cc1c553b0 tempest-ServerRescueNegativeTestJSON-1483258717 tempest-ServerRescueNegativeTestJSON-1483258717-project-member] [instance: 7ea919b6-12f7-4f4e-a123-413044bbbffc] Destroying instance {{(pid=61806) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 717.470794] env[61806]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-8263ab89-8ac9-43a4-ae06-1b01777da67a {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 717.483009] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9788f493-3534-435f-8b48-500843fbc402 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 717.511912] env[61806]: WARNING nova.virt.vmwareapi.vmops [None req-6b55cbbf-00f7-4d4d-87f3-a59cc1c553b0 tempest-ServerRescueNegativeTestJSON-1483258717 tempest-ServerRescueNegativeTestJSON-1483258717-project-member] [instance: 7ea919b6-12f7-4f4e-a123-413044bbbffc] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 7ea919b6-12f7-4f4e-a123-413044bbbffc could not be found. [ 717.512162] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-6b55cbbf-00f7-4d4d-87f3-a59cc1c553b0 tempest-ServerRescueNegativeTestJSON-1483258717 tempest-ServerRescueNegativeTestJSON-1483258717-project-member] [instance: 7ea919b6-12f7-4f4e-a123-413044bbbffc] Instance destroyed {{(pid=61806) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 717.512346] env[61806]: INFO nova.compute.manager [None req-6b55cbbf-00f7-4d4d-87f3-a59cc1c553b0 tempest-ServerRescueNegativeTestJSON-1483258717 tempest-ServerRescueNegativeTestJSON-1483258717-project-member] [instance: 7ea919b6-12f7-4f4e-a123-413044bbbffc] Took 0.04 seconds to destroy the instance on the hypervisor. [ 717.512596] env[61806]: DEBUG oslo.service.loopingcall [None req-6b55cbbf-00f7-4d4d-87f3-a59cc1c553b0 tempest-ServerRescueNegativeTestJSON-1483258717 tempest-ServerRescueNegativeTestJSON-1483258717-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61806) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 717.512812] env[61806]: DEBUG nova.compute.manager [-] [instance: 7ea919b6-12f7-4f4e-a123-413044bbbffc] Deallocating network for instance {{(pid=61806) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 717.512976] env[61806]: DEBUG nova.network.neutron [-] [instance: 7ea919b6-12f7-4f4e-a123-413044bbbffc] deallocate_for_instance() {{(pid=61806) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 717.530881] env[61806]: DEBUG nova.network.neutron [-] [instance: 7ea919b6-12f7-4f4e-a123-413044bbbffc] Instance cache missing network info. {{(pid=61806) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 717.566507] env[61806]: DEBUG nova.network.neutron [None req-81a55da1-45c4-45f8-af56-f8f7d424c8aa tempest-ServersTestJSON-430716070 tempest-ServersTestJSON-430716070-project-member] [instance: eca2e3da-b38c-482e-b9b6-4ad75d83caa1] Instance cache missing network info. {{(pid=61806) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 717.703880] env[61806]: DEBUG nova.network.neutron [None req-81a55da1-45c4-45f8-af56-f8f7d424c8aa tempest-ServersTestJSON-430716070 tempest-ServersTestJSON-430716070-project-member] [instance: eca2e3da-b38c-482e-b9b6-4ad75d83caa1] Updating instance_info_cache with network_info: [] {{(pid=61806) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 718.037794] env[61806]: DEBUG nova.network.neutron [-] [instance: 7ea919b6-12f7-4f4e-a123-413044bbbffc] Updating instance_info_cache with network_info: [] {{(pid=61806) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 718.169409] env[61806]: INFO nova.scheduler.client.report [None req-865353ee-538a-40fc-bfba-502aba0b6a20 tempest-VolumesAdminNegativeTest-1387177911 tempest-VolumesAdminNegativeTest-1387177911-project-member] Deleted allocations for instance b7d55e87-95db-4c45-9417-bbece4fbb98d [ 718.208880] env[61806]: DEBUG oslo_concurrency.lockutils [None req-81a55da1-45c4-45f8-af56-f8f7d424c8aa tempest-ServersTestJSON-430716070 tempest-ServersTestJSON-430716070-project-member] Releasing lock "refresh_cache-eca2e3da-b38c-482e-b9b6-4ad75d83caa1" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 718.209179] env[61806]: DEBUG nova.compute.manager [None req-81a55da1-45c4-45f8-af56-f8f7d424c8aa tempest-ServersTestJSON-430716070 tempest-ServersTestJSON-430716070-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61806) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 718.209398] env[61806]: DEBUG nova.compute.manager [None req-81a55da1-45c4-45f8-af56-f8f7d424c8aa tempest-ServersTestJSON-430716070 tempest-ServersTestJSON-430716070-project-member] [instance: eca2e3da-b38c-482e-b9b6-4ad75d83caa1] Deallocating network for instance {{(pid=61806) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 718.209651] env[61806]: DEBUG nova.network.neutron [None req-81a55da1-45c4-45f8-af56-f8f7d424c8aa tempest-ServersTestJSON-430716070 tempest-ServersTestJSON-430716070-project-member] [instance: eca2e3da-b38c-482e-b9b6-4ad75d83caa1] deallocate_for_instance() {{(pid=61806) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 718.240921] env[61806]: DEBUG nova.network.neutron [None req-81a55da1-45c4-45f8-af56-f8f7d424c8aa tempest-ServersTestJSON-430716070 tempest-ServersTestJSON-430716070-project-member] [instance: eca2e3da-b38c-482e-b9b6-4ad75d83caa1] Instance cache missing network info. {{(pid=61806) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 718.468535] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4cb8a885-f5c3-4368-b86a-3369c2040e4f {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 718.477242] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a7f21ff3-5c79-4f8a-a5da-56860c67f9ea {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 718.513924] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1d3a5c42-af89-4e58-a610-cf0801d6a16a {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 718.522269] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5f95c4e2-21e0-4739-a356-a1e323991767 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 718.536157] env[61806]: DEBUG nova.compute.provider_tree [None req-3c3981d2-fa25-4f4b-87d4-a5411e67199e tempest-MultipleCreateTestJSON-1589015462 tempest-MultipleCreateTestJSON-1589015462-project-member] Updating inventory in ProviderTree for provider a2858be1-fd22-4e08-979e-87ad25293407 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 139, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61806) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 718.544470] env[61806]: INFO nova.compute.manager [-] [instance: 7ea919b6-12f7-4f4e-a123-413044bbbffc] Took 1.03 seconds to deallocate network for instance. [ 718.546839] env[61806]: DEBUG nova.compute.claims [None req-6b55cbbf-00f7-4d4d-87f3-a59cc1c553b0 tempest-ServerRescueNegativeTestJSON-1483258717 tempest-ServerRescueNegativeTestJSON-1483258717-project-member] [instance: 7ea919b6-12f7-4f4e-a123-413044bbbffc] Aborting claim: {{(pid=61806) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 718.547042] env[61806]: DEBUG oslo_concurrency.lockutils [None req-6b55cbbf-00f7-4d4d-87f3-a59cc1c553b0 tempest-ServerRescueNegativeTestJSON-1483258717 tempest-ServerRescueNegativeTestJSON-1483258717-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 718.683337] env[61806]: DEBUG oslo_concurrency.lockutils [None req-865353ee-538a-40fc-bfba-502aba0b6a20 tempest-VolumesAdminNegativeTest-1387177911 tempest-VolumesAdminNegativeTest-1387177911-project-member] Lock "b7d55e87-95db-4c45-9417-bbece4fbb98d" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 152.054s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 718.744313] env[61806]: DEBUG nova.network.neutron [None req-81a55da1-45c4-45f8-af56-f8f7d424c8aa tempest-ServersTestJSON-430716070 tempest-ServersTestJSON-430716070-project-member] [instance: eca2e3da-b38c-482e-b9b6-4ad75d83caa1] Updating instance_info_cache with network_info: [] {{(pid=61806) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 719.061843] env[61806]: ERROR nova.scheduler.client.report [None req-3c3981d2-fa25-4f4b-87d4-a5411e67199e tempest-MultipleCreateTestJSON-1589015462 tempest-MultipleCreateTestJSON-1589015462-project-member] [req-d8f76542-29cb-4fa4-9fd0-61aed16bd0cf] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 139, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID a2858be1-fd22-4e08-979e-87ad25293407. Got 409: {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-d8f76542-29cb-4fa4-9fd0-61aed16bd0cf"}]} [ 719.080026] env[61806]: DEBUG nova.scheduler.client.report [None req-3c3981d2-fa25-4f4b-87d4-a5411e67199e tempest-MultipleCreateTestJSON-1589015462 tempest-MultipleCreateTestJSON-1589015462-project-member] Refreshing inventories for resource provider a2858be1-fd22-4e08-979e-87ad25293407 {{(pid=61806) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:818}} [ 719.104912] env[61806]: DEBUG nova.scheduler.client.report [None req-3c3981d2-fa25-4f4b-87d4-a5411e67199e tempest-MultipleCreateTestJSON-1589015462 tempest-MultipleCreateTestJSON-1589015462-project-member] Updating ProviderTree inventory for provider a2858be1-fd22-4e08-979e-87ad25293407 from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 138, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61806) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:782}} [ 719.105164] env[61806]: DEBUG nova.compute.provider_tree [None req-3c3981d2-fa25-4f4b-87d4-a5411e67199e tempest-MultipleCreateTestJSON-1589015462 tempest-MultipleCreateTestJSON-1589015462-project-member] Updating inventory in ProviderTree for provider a2858be1-fd22-4e08-979e-87ad25293407 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 138, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61806) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 719.118406] env[61806]: DEBUG nova.scheduler.client.report [None req-3c3981d2-fa25-4f4b-87d4-a5411e67199e tempest-MultipleCreateTestJSON-1589015462 tempest-MultipleCreateTestJSON-1589015462-project-member] Refreshing aggregate associations for resource provider a2858be1-fd22-4e08-979e-87ad25293407, aggregates: None {{(pid=61806) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:827}} [ 719.146022] env[61806]: DEBUG nova.scheduler.client.report [None req-3c3981d2-fa25-4f4b-87d4-a5411e67199e tempest-MultipleCreateTestJSON-1589015462 tempest-MultipleCreateTestJSON-1589015462-project-member] Refreshing trait associations for resource provider a2858be1-fd22-4e08-979e-87ad25293407, traits: COMPUTE_IMAGE_TYPE_ISO,COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_SAME_HOST_COLD_MIGRATE,COMPUTE_NET_ATTACH_INTERFACE,COMPUTE_NODE {{(pid=61806) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:839}} [ 719.187762] env[61806]: DEBUG nova.compute.manager [None req-1382c347-5a79-4289-8b32-002f296b7ae3 tempest-TenantUsagesTestJSON-255073140 tempest-TenantUsagesTestJSON-255073140-project-member] [instance: 949c1050-5003-4519-b24c-8904de21a676] Starting instance... {{(pid=61806) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 719.250398] env[61806]: INFO nova.compute.manager [None req-81a55da1-45c4-45f8-af56-f8f7d424c8aa tempest-ServersTestJSON-430716070 tempest-ServersTestJSON-430716070-project-member] [instance: eca2e3da-b38c-482e-b9b6-4ad75d83caa1] Took 1.04 seconds to deallocate network for instance. [ 719.577030] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ab1bbde0-9bfe-40a4-89cc-172801b5fba0 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 719.585696] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0c484816-9890-4e5c-95a5-3ac71f2ea0d8 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 719.619218] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-77ad6750-2c3b-4d9d-80b8-9da338b75798 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 719.628666] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c48d164a-3e4d-42fc-8eb7-c365719f1587 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 719.641892] env[61806]: DEBUG nova.compute.provider_tree [None req-3c3981d2-fa25-4f4b-87d4-a5411e67199e tempest-MultipleCreateTestJSON-1589015462 tempest-MultipleCreateTestJSON-1589015462-project-member] Updating inventory in ProviderTree for provider a2858be1-fd22-4e08-979e-87ad25293407 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 139, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61806) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 719.716400] env[61806]: DEBUG oslo_concurrency.lockutils [None req-1382c347-5a79-4289-8b32-002f296b7ae3 tempest-TenantUsagesTestJSON-255073140 tempest-TenantUsagesTestJSON-255073140-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 719.841707] env[61806]: DEBUG oslo_concurrency.lockutils [None req-5af76806-1cdf-4002-962d-34b8992ee9dc tempest-VolumesAdminNegativeTest-1387177911 tempest-VolumesAdminNegativeTest-1387177911-project-member] Acquiring lock "4373b735-31cf-4b53-b655-38555cf212a5" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 719.842258] env[61806]: DEBUG oslo_concurrency.lockutils [None req-5af76806-1cdf-4002-962d-34b8992ee9dc tempest-VolumesAdminNegativeTest-1387177911 tempest-VolumesAdminNegativeTest-1387177911-project-member] Lock "4373b735-31cf-4b53-b655-38555cf212a5" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 719.981729] env[61806]: DEBUG oslo_concurrency.lockutils [None req-61b80e06-de2f-412b-be27-b9318eeb5c25 tempest-ServerAddressesTestJSON-661730579 tempest-ServerAddressesTestJSON-661730579-project-member] Acquiring lock "a842df44-d8a9-4376-b9fc-5ca19a68a4b7" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 719.981950] env[61806]: DEBUG oslo_concurrency.lockutils [None req-61b80e06-de2f-412b-be27-b9318eeb5c25 tempest-ServerAddressesTestJSON-661730579 tempest-ServerAddressesTestJSON-661730579-project-member] Lock "a842df44-d8a9-4376-b9fc-5ca19a68a4b7" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 720.164576] env[61806]: ERROR nova.scheduler.client.report [None req-3c3981d2-fa25-4f4b-87d4-a5411e67199e tempest-MultipleCreateTestJSON-1589015462 tempest-MultipleCreateTestJSON-1589015462-project-member] [req-f188bb91-0615-4426-9ee7-25664bad2ddb] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 139, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID a2858be1-fd22-4e08-979e-87ad25293407. Got 409: {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-f188bb91-0615-4426-9ee7-25664bad2ddb"}]} [ 720.181081] env[61806]: DEBUG nova.scheduler.client.report [None req-3c3981d2-fa25-4f4b-87d4-a5411e67199e tempest-MultipleCreateTestJSON-1589015462 tempest-MultipleCreateTestJSON-1589015462-project-member] Refreshing inventories for resource provider a2858be1-fd22-4e08-979e-87ad25293407 {{(pid=61806) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:818}} [ 720.194702] env[61806]: DEBUG nova.scheduler.client.report [None req-3c3981d2-fa25-4f4b-87d4-a5411e67199e tempest-MultipleCreateTestJSON-1589015462 tempest-MultipleCreateTestJSON-1589015462-project-member] Updating ProviderTree inventory for provider a2858be1-fd22-4e08-979e-87ad25293407 from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 138, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61806) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:782}} [ 720.194941] env[61806]: DEBUG nova.compute.provider_tree [None req-3c3981d2-fa25-4f4b-87d4-a5411e67199e tempest-MultipleCreateTestJSON-1589015462 tempest-MultipleCreateTestJSON-1589015462-project-member] Updating inventory in ProviderTree for provider a2858be1-fd22-4e08-979e-87ad25293407 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 138, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61806) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 720.206642] env[61806]: DEBUG nova.scheduler.client.report [None req-3c3981d2-fa25-4f4b-87d4-a5411e67199e tempest-MultipleCreateTestJSON-1589015462 tempest-MultipleCreateTestJSON-1589015462-project-member] Refreshing aggregate associations for resource provider a2858be1-fd22-4e08-979e-87ad25293407, aggregates: None {{(pid=61806) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:827}} [ 720.223840] env[61806]: DEBUG nova.scheduler.client.report [None req-3c3981d2-fa25-4f4b-87d4-a5411e67199e tempest-MultipleCreateTestJSON-1589015462 tempest-MultipleCreateTestJSON-1589015462-project-member] Refreshing trait associations for resource provider a2858be1-fd22-4e08-979e-87ad25293407, traits: COMPUTE_IMAGE_TYPE_ISO,COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_SAME_HOST_COLD_MIGRATE,COMPUTE_NET_ATTACH_INTERFACE,COMPUTE_NODE {{(pid=61806) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:839}} [ 720.279283] env[61806]: INFO nova.scheduler.client.report [None req-81a55da1-45c4-45f8-af56-f8f7d424c8aa tempest-ServersTestJSON-430716070 tempest-ServersTestJSON-430716070-project-member] Deleted allocations for instance eca2e3da-b38c-482e-b9b6-4ad75d83caa1 [ 720.683024] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d5739e8a-b709-45af-ab5d-a0b2daf84be6 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 720.690718] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f2239b8f-1ea6-4fc0-8a92-992b6f7202ef {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 720.722455] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-07b20eab-80fc-42aa-aa09-e818b604545b {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 720.730973] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a9d5c8df-37ff-42d8-9e2f-6eea9f785b3d {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 720.746331] env[61806]: DEBUG nova.compute.provider_tree [None req-3c3981d2-fa25-4f4b-87d4-a5411e67199e tempest-MultipleCreateTestJSON-1589015462 tempest-MultipleCreateTestJSON-1589015462-project-member] Updating inventory in ProviderTree for provider a2858be1-fd22-4e08-979e-87ad25293407 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 139, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61806) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 720.789638] env[61806]: DEBUG oslo_concurrency.lockutils [None req-81a55da1-45c4-45f8-af56-f8f7d424c8aa tempest-ServersTestJSON-430716070 tempest-ServersTestJSON-430716070-project-member] Lock "eca2e3da-b38c-482e-b9b6-4ad75d83caa1" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 151.299s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 721.284415] env[61806]: DEBUG nova.scheduler.client.report [None req-3c3981d2-fa25-4f4b-87d4-a5411e67199e tempest-MultipleCreateTestJSON-1589015462 tempest-MultipleCreateTestJSON-1589015462-project-member] Updated inventory for provider a2858be1-fd22-4e08-979e-87ad25293407 with generation 73 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 139, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61806) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:971}} [ 721.284415] env[61806]: DEBUG nova.compute.provider_tree [None req-3c3981d2-fa25-4f4b-87d4-a5411e67199e tempest-MultipleCreateTestJSON-1589015462 tempest-MultipleCreateTestJSON-1589015462-project-member] Updating resource provider a2858be1-fd22-4e08-979e-87ad25293407 generation from 73 to 74 during operation: update_inventory {{(pid=61806) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 721.284415] env[61806]: DEBUG nova.compute.provider_tree [None req-3c3981d2-fa25-4f4b-87d4-a5411e67199e tempest-MultipleCreateTestJSON-1589015462 tempest-MultipleCreateTestJSON-1589015462-project-member] Updating inventory in ProviderTree for provider a2858be1-fd22-4e08-979e-87ad25293407 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 139, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61806) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 721.294704] env[61806]: DEBUG nova.compute.manager [None req-3bf756e0-c56e-4902-9c2b-b224a1d09a7c tempest-ListServersNegativeTestJSON-272187588 tempest-ListServersNegativeTestJSON-272187588-project-member] [instance: f2446d04-15b8-4b3d-8932-c37df67b0f8f] Starting instance... {{(pid=61806) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 721.790352] env[61806]: DEBUG oslo_concurrency.lockutils [None req-3c3981d2-fa25-4f4b-87d4-a5411e67199e tempest-MultipleCreateTestJSON-1589015462 tempest-MultipleCreateTestJSON-1589015462-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 4.760s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 721.790877] env[61806]: DEBUG nova.compute.manager [None req-3c3981d2-fa25-4f4b-87d4-a5411e67199e tempest-MultipleCreateTestJSON-1589015462 tempest-MultipleCreateTestJSON-1589015462-project-member] [instance: 7211d568-cc68-4e50-80ad-b78878d1deab] Start building networks asynchronously for instance. {{(pid=61806) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 721.796227] env[61806]: DEBUG oslo_concurrency.lockutils [None req-2bf81f39-4b8d-4851-a793-760af008ffe2 tempest-ImagesOneServerTestJSON-1003762139 tempest-ImagesOneServerTestJSON-1003762139-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 25.277s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 721.837379] env[61806]: DEBUG oslo_concurrency.lockutils [None req-3bf756e0-c56e-4902-9c2b-b224a1d09a7c tempest-ListServersNegativeTestJSON-272187588 tempest-ListServersNegativeTestJSON-272187588-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 722.298326] env[61806]: DEBUG nova.compute.utils [None req-3c3981d2-fa25-4f4b-87d4-a5411e67199e tempest-MultipleCreateTestJSON-1589015462 tempest-MultipleCreateTestJSON-1589015462-project-member] Using /dev/sd instead of None {{(pid=61806) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 722.302699] env[61806]: DEBUG nova.compute.manager [None req-3c3981d2-fa25-4f4b-87d4-a5411e67199e tempest-MultipleCreateTestJSON-1589015462 tempest-MultipleCreateTestJSON-1589015462-project-member] [instance: 7211d568-cc68-4e50-80ad-b78878d1deab] Allocating IP information in the background. {{(pid=61806) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 722.302699] env[61806]: DEBUG nova.network.neutron [None req-3c3981d2-fa25-4f4b-87d4-a5411e67199e tempest-MultipleCreateTestJSON-1589015462 tempest-MultipleCreateTestJSON-1589015462-project-member] [instance: 7211d568-cc68-4e50-80ad-b78878d1deab] allocate_for_instance() {{(pid=61806) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 722.376527] env[61806]: DEBUG nova.policy [None req-3c3981d2-fa25-4f4b-87d4-a5411e67199e tempest-MultipleCreateTestJSON-1589015462 tempest-MultipleCreateTestJSON-1589015462-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'b5a92ad3d119419085f56d7dea0975f5', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'dc40805c39f34d07be9d07861ad5841d', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61806) authorize /opt/stack/nova/nova/policy.py:201}} [ 722.732696] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dcb058c9-5853-4c40-ad7c-2add258f9a88 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 722.742136] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-74013798-b3b0-4943-a964-16cc039b1c37 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 722.777156] env[61806]: DEBUG nova.network.neutron [None req-3c3981d2-fa25-4f4b-87d4-a5411e67199e tempest-MultipleCreateTestJSON-1589015462 tempest-MultipleCreateTestJSON-1589015462-project-member] [instance: 7211d568-cc68-4e50-80ad-b78878d1deab] Successfully created port: 2aa127d2-ff55-470b-822d-60c2d2d409d6 {{(pid=61806) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 722.780403] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7e9f645a-5178-4754-991b-38b9eb1d904c {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 722.787994] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-db74e589-f344-40e1-b1e8-e4253df691a8 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 722.801378] env[61806]: DEBUG nova.compute.provider_tree [None req-2bf81f39-4b8d-4851-a793-760af008ffe2 tempest-ImagesOneServerTestJSON-1003762139 tempest-ImagesOneServerTestJSON-1003762139-project-member] Inventory has not changed in ProviderTree for provider: a2858be1-fd22-4e08-979e-87ad25293407 {{(pid=61806) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 722.803914] env[61806]: DEBUG nova.compute.manager [None req-3c3981d2-fa25-4f4b-87d4-a5411e67199e tempest-MultipleCreateTestJSON-1589015462 tempest-MultipleCreateTestJSON-1589015462-project-member] [instance: 7211d568-cc68-4e50-80ad-b78878d1deab] Start building block device mappings for instance. {{(pid=61806) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 723.305400] env[61806]: DEBUG nova.scheduler.client.report [None req-2bf81f39-4b8d-4851-a793-760af008ffe2 tempest-ImagesOneServerTestJSON-1003762139 tempest-ImagesOneServerTestJSON-1003762139-project-member] Inventory has not changed for provider a2858be1-fd22-4e08-979e-87ad25293407 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 139, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61806) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 723.813842] env[61806]: DEBUG oslo_concurrency.lockutils [None req-2bf81f39-4b8d-4851-a793-760af008ffe2 tempest-ImagesOneServerTestJSON-1003762139 tempest-ImagesOneServerTestJSON-1003762139-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.018s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 723.814503] env[61806]: ERROR nova.compute.manager [None req-2bf81f39-4b8d-4851-a793-760af008ffe2 tempest-ImagesOneServerTestJSON-1003762139 tempest-ImagesOneServerTestJSON-1003762139-project-member] [instance: 294f0a37-7589-405d-a324-13099c80f458] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port b9eb0672-5644-46ba-9b38-f6a90aef16ee, please check neutron logs for more information. [ 723.814503] env[61806]: ERROR nova.compute.manager [instance: 294f0a37-7589-405d-a324-13099c80f458] Traceback (most recent call last): [ 723.814503] env[61806]: ERROR nova.compute.manager [instance: 294f0a37-7589-405d-a324-13099c80f458] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 723.814503] env[61806]: ERROR nova.compute.manager [instance: 294f0a37-7589-405d-a324-13099c80f458] self.driver.spawn(context, instance, image_meta, [ 723.814503] env[61806]: ERROR nova.compute.manager [instance: 294f0a37-7589-405d-a324-13099c80f458] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 723.814503] env[61806]: ERROR nova.compute.manager [instance: 294f0a37-7589-405d-a324-13099c80f458] self._vmops.spawn(context, instance, image_meta, injected_files, [ 723.814503] env[61806]: ERROR nova.compute.manager [instance: 294f0a37-7589-405d-a324-13099c80f458] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 723.814503] env[61806]: ERROR nova.compute.manager [instance: 294f0a37-7589-405d-a324-13099c80f458] vm_ref = self.build_virtual_machine(instance, [ 723.814503] env[61806]: ERROR nova.compute.manager [instance: 294f0a37-7589-405d-a324-13099c80f458] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 723.814503] env[61806]: ERROR nova.compute.manager [instance: 294f0a37-7589-405d-a324-13099c80f458] vif_infos = vmwarevif.get_vif_info(self._session, [ 723.814503] env[61806]: ERROR nova.compute.manager [instance: 294f0a37-7589-405d-a324-13099c80f458] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 723.814911] env[61806]: ERROR nova.compute.manager [instance: 294f0a37-7589-405d-a324-13099c80f458] for vif in network_info: [ 723.814911] env[61806]: ERROR nova.compute.manager [instance: 294f0a37-7589-405d-a324-13099c80f458] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 723.814911] env[61806]: ERROR nova.compute.manager [instance: 294f0a37-7589-405d-a324-13099c80f458] return self._sync_wrapper(fn, *args, **kwargs) [ 723.814911] env[61806]: ERROR nova.compute.manager [instance: 294f0a37-7589-405d-a324-13099c80f458] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 723.814911] env[61806]: ERROR nova.compute.manager [instance: 294f0a37-7589-405d-a324-13099c80f458] self.wait() [ 723.814911] env[61806]: ERROR nova.compute.manager [instance: 294f0a37-7589-405d-a324-13099c80f458] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 723.814911] env[61806]: ERROR nova.compute.manager [instance: 294f0a37-7589-405d-a324-13099c80f458] self[:] = self._gt.wait() [ 723.814911] env[61806]: ERROR nova.compute.manager [instance: 294f0a37-7589-405d-a324-13099c80f458] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 723.814911] env[61806]: ERROR nova.compute.manager [instance: 294f0a37-7589-405d-a324-13099c80f458] return self._exit_event.wait() [ 723.814911] env[61806]: ERROR nova.compute.manager [instance: 294f0a37-7589-405d-a324-13099c80f458] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 723.814911] env[61806]: ERROR nova.compute.manager [instance: 294f0a37-7589-405d-a324-13099c80f458] result = hub.switch() [ 723.814911] env[61806]: ERROR nova.compute.manager [instance: 294f0a37-7589-405d-a324-13099c80f458] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 723.814911] env[61806]: ERROR nova.compute.manager [instance: 294f0a37-7589-405d-a324-13099c80f458] return self.greenlet.switch() [ 723.815352] env[61806]: ERROR nova.compute.manager [instance: 294f0a37-7589-405d-a324-13099c80f458] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 723.815352] env[61806]: ERROR nova.compute.manager [instance: 294f0a37-7589-405d-a324-13099c80f458] result = function(*args, **kwargs) [ 723.815352] env[61806]: ERROR nova.compute.manager [instance: 294f0a37-7589-405d-a324-13099c80f458] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 723.815352] env[61806]: ERROR nova.compute.manager [instance: 294f0a37-7589-405d-a324-13099c80f458] return func(*args, **kwargs) [ 723.815352] env[61806]: ERROR nova.compute.manager [instance: 294f0a37-7589-405d-a324-13099c80f458] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 723.815352] env[61806]: ERROR nova.compute.manager [instance: 294f0a37-7589-405d-a324-13099c80f458] raise e [ 723.815352] env[61806]: ERROR nova.compute.manager [instance: 294f0a37-7589-405d-a324-13099c80f458] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 723.815352] env[61806]: ERROR nova.compute.manager [instance: 294f0a37-7589-405d-a324-13099c80f458] nwinfo = self.network_api.allocate_for_instance( [ 723.815352] env[61806]: ERROR nova.compute.manager [instance: 294f0a37-7589-405d-a324-13099c80f458] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 723.815352] env[61806]: ERROR nova.compute.manager [instance: 294f0a37-7589-405d-a324-13099c80f458] created_port_ids = self._update_ports_for_instance( [ 723.815352] env[61806]: ERROR nova.compute.manager [instance: 294f0a37-7589-405d-a324-13099c80f458] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 723.815352] env[61806]: ERROR nova.compute.manager [instance: 294f0a37-7589-405d-a324-13099c80f458] with excutils.save_and_reraise_exception(): [ 723.815352] env[61806]: ERROR nova.compute.manager [instance: 294f0a37-7589-405d-a324-13099c80f458] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 723.815783] env[61806]: ERROR nova.compute.manager [instance: 294f0a37-7589-405d-a324-13099c80f458] self.force_reraise() [ 723.815783] env[61806]: ERROR nova.compute.manager [instance: 294f0a37-7589-405d-a324-13099c80f458] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 723.815783] env[61806]: ERROR nova.compute.manager [instance: 294f0a37-7589-405d-a324-13099c80f458] raise self.value [ 723.815783] env[61806]: ERROR nova.compute.manager [instance: 294f0a37-7589-405d-a324-13099c80f458] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 723.815783] env[61806]: ERROR nova.compute.manager [instance: 294f0a37-7589-405d-a324-13099c80f458] updated_port = self._update_port( [ 723.815783] env[61806]: ERROR nova.compute.manager [instance: 294f0a37-7589-405d-a324-13099c80f458] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 723.815783] env[61806]: ERROR nova.compute.manager [instance: 294f0a37-7589-405d-a324-13099c80f458] _ensure_no_port_binding_failure(port) [ 723.815783] env[61806]: ERROR nova.compute.manager [instance: 294f0a37-7589-405d-a324-13099c80f458] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 723.815783] env[61806]: ERROR nova.compute.manager [instance: 294f0a37-7589-405d-a324-13099c80f458] raise exception.PortBindingFailed(port_id=port['id']) [ 723.815783] env[61806]: ERROR nova.compute.manager [instance: 294f0a37-7589-405d-a324-13099c80f458] nova.exception.PortBindingFailed: Binding failed for port b9eb0672-5644-46ba-9b38-f6a90aef16ee, please check neutron logs for more information. [ 723.815783] env[61806]: ERROR nova.compute.manager [instance: 294f0a37-7589-405d-a324-13099c80f458] [ 723.816618] env[61806]: DEBUG nova.compute.utils [None req-2bf81f39-4b8d-4851-a793-760af008ffe2 tempest-ImagesOneServerTestJSON-1003762139 tempest-ImagesOneServerTestJSON-1003762139-project-member] [instance: 294f0a37-7589-405d-a324-13099c80f458] Binding failed for port b9eb0672-5644-46ba-9b38-f6a90aef16ee, please check neutron logs for more information. {{(pid=61806) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 723.817044] env[61806]: DEBUG nova.compute.manager [None req-3c3981d2-fa25-4f4b-87d4-a5411e67199e tempest-MultipleCreateTestJSON-1589015462 tempest-MultipleCreateTestJSON-1589015462-project-member] [instance: 7211d568-cc68-4e50-80ad-b78878d1deab] Start spawning the instance on the hypervisor. {{(pid=61806) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 723.822053] env[61806]: DEBUG nova.compute.manager [None req-2bf81f39-4b8d-4851-a793-760af008ffe2 tempest-ImagesOneServerTestJSON-1003762139 tempest-ImagesOneServerTestJSON-1003762139-project-member] [instance: 294f0a37-7589-405d-a324-13099c80f458] Build of instance 294f0a37-7589-405d-a324-13099c80f458 was re-scheduled: Binding failed for port b9eb0672-5644-46ba-9b38-f6a90aef16ee, please check neutron logs for more information. {{(pid=61806) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 723.822053] env[61806]: DEBUG nova.compute.manager [None req-2bf81f39-4b8d-4851-a793-760af008ffe2 tempest-ImagesOneServerTestJSON-1003762139 tempest-ImagesOneServerTestJSON-1003762139-project-member] [instance: 294f0a37-7589-405d-a324-13099c80f458] Unplugging VIFs for instance {{(pid=61806) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 723.822053] env[61806]: DEBUG oslo_concurrency.lockutils [None req-2bf81f39-4b8d-4851-a793-760af008ffe2 tempest-ImagesOneServerTestJSON-1003762139 tempest-ImagesOneServerTestJSON-1003762139-project-member] Acquiring lock "refresh_cache-294f0a37-7589-405d-a324-13099c80f458" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 723.822053] env[61806]: DEBUG oslo_concurrency.lockutils [None req-2bf81f39-4b8d-4851-a793-760af008ffe2 tempest-ImagesOneServerTestJSON-1003762139 tempest-ImagesOneServerTestJSON-1003762139-project-member] Acquired lock "refresh_cache-294f0a37-7589-405d-a324-13099c80f458" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 723.822339] env[61806]: DEBUG nova.network.neutron [None req-2bf81f39-4b8d-4851-a793-760af008ffe2 tempest-ImagesOneServerTestJSON-1003762139 tempest-ImagesOneServerTestJSON-1003762139-project-member] [instance: 294f0a37-7589-405d-a324-13099c80f458] Building network info cache for instance {{(pid=61806) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 723.823025] env[61806]: DEBUG oslo_concurrency.lockutils [None req-0f0a69c6-2310-4272-abc1-520940b4bcf0 tempest-ServersTestFqdnHostnames-356624376 tempest-ServersTestFqdnHostnames-356624376-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 24.789s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 723.873876] env[61806]: DEBUG nova.virt.hardware [None req-3c3981d2-fa25-4f4b-87d4-a5411e67199e tempest-MultipleCreateTestJSON-1589015462 tempest-MultipleCreateTestJSON-1589015462-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-15T18:05:10Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-15T18:04:52Z,direct_url=,disk_format='vmdk',id=ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='84ba83fce2c349a6988173c5d6fc3b07',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-15T18:04:53Z,virtual_size=,visibility=), allow threads: False {{(pid=61806) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 723.873998] env[61806]: DEBUG nova.virt.hardware [None req-3c3981d2-fa25-4f4b-87d4-a5411e67199e tempest-MultipleCreateTestJSON-1589015462 tempest-MultipleCreateTestJSON-1589015462-project-member] Flavor limits 0:0:0 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 723.874352] env[61806]: DEBUG nova.virt.hardware [None req-3c3981d2-fa25-4f4b-87d4-a5411e67199e tempest-MultipleCreateTestJSON-1589015462 tempest-MultipleCreateTestJSON-1589015462-project-member] Image limits 0:0:0 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 723.874352] env[61806]: DEBUG nova.virt.hardware [None req-3c3981d2-fa25-4f4b-87d4-a5411e67199e tempest-MultipleCreateTestJSON-1589015462 tempest-MultipleCreateTestJSON-1589015462-project-member] Flavor pref 0:0:0 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 723.874465] env[61806]: DEBUG nova.virt.hardware [None req-3c3981d2-fa25-4f4b-87d4-a5411e67199e tempest-MultipleCreateTestJSON-1589015462 tempest-MultipleCreateTestJSON-1589015462-project-member] Image pref 0:0:0 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 723.875063] env[61806]: DEBUG nova.virt.hardware [None req-3c3981d2-fa25-4f4b-87d4-a5411e67199e tempest-MultipleCreateTestJSON-1589015462 tempest-MultipleCreateTestJSON-1589015462-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 723.875063] env[61806]: DEBUG nova.virt.hardware [None req-3c3981d2-fa25-4f4b-87d4-a5411e67199e tempest-MultipleCreateTestJSON-1589015462 tempest-MultipleCreateTestJSON-1589015462-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61806) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 723.875063] env[61806]: DEBUG nova.virt.hardware [None req-3c3981d2-fa25-4f4b-87d4-a5411e67199e tempest-MultipleCreateTestJSON-1589015462 tempest-MultipleCreateTestJSON-1589015462-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61806) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 723.875223] env[61806]: DEBUG nova.virt.hardware [None req-3c3981d2-fa25-4f4b-87d4-a5411e67199e tempest-MultipleCreateTestJSON-1589015462 tempest-MultipleCreateTestJSON-1589015462-project-member] Got 1 possible topologies {{(pid=61806) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 723.875532] env[61806]: DEBUG nova.virt.hardware [None req-3c3981d2-fa25-4f4b-87d4-a5411e67199e tempest-MultipleCreateTestJSON-1589015462 tempest-MultipleCreateTestJSON-1589015462-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61806) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 723.875532] env[61806]: DEBUG nova.virt.hardware [None req-3c3981d2-fa25-4f4b-87d4-a5411e67199e tempest-MultipleCreateTestJSON-1589015462 tempest-MultipleCreateTestJSON-1589015462-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61806) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 723.876390] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e87470c2-2bf2-45f2-99f2-136997194f69 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 723.887915] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e08d53c0-f326-49e6-8a03-e0fc4729eb0d {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 724.008198] env[61806]: DEBUG nova.compute.manager [req-1d6ced2b-3cfe-4a70-b9e6-09e376f2421e req-2e2b7aa9-1bf8-4d2d-8a3d-85dad9991060 service nova] [instance: 7211d568-cc68-4e50-80ad-b78878d1deab] Received event network-changed-2aa127d2-ff55-470b-822d-60c2d2d409d6 {{(pid=61806) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 724.008198] env[61806]: DEBUG nova.compute.manager [req-1d6ced2b-3cfe-4a70-b9e6-09e376f2421e req-2e2b7aa9-1bf8-4d2d-8a3d-85dad9991060 service nova] [instance: 7211d568-cc68-4e50-80ad-b78878d1deab] Refreshing instance network info cache due to event network-changed-2aa127d2-ff55-470b-822d-60c2d2d409d6. {{(pid=61806) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 724.008198] env[61806]: DEBUG oslo_concurrency.lockutils [req-1d6ced2b-3cfe-4a70-b9e6-09e376f2421e req-2e2b7aa9-1bf8-4d2d-8a3d-85dad9991060 service nova] Acquiring lock "refresh_cache-7211d568-cc68-4e50-80ad-b78878d1deab" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 724.008198] env[61806]: DEBUG oslo_concurrency.lockutils [req-1d6ced2b-3cfe-4a70-b9e6-09e376f2421e req-2e2b7aa9-1bf8-4d2d-8a3d-85dad9991060 service nova] Acquired lock "refresh_cache-7211d568-cc68-4e50-80ad-b78878d1deab" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 724.008198] env[61806]: DEBUG nova.network.neutron [req-1d6ced2b-3cfe-4a70-b9e6-09e376f2421e req-2e2b7aa9-1bf8-4d2d-8a3d-85dad9991060 service nova] [instance: 7211d568-cc68-4e50-80ad-b78878d1deab] Refreshing network info cache for port 2aa127d2-ff55-470b-822d-60c2d2d409d6 {{(pid=61806) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 724.278512] env[61806]: ERROR nova.compute.manager [None req-3c3981d2-fa25-4f4b-87d4-a5411e67199e tempest-MultipleCreateTestJSON-1589015462 tempest-MultipleCreateTestJSON-1589015462-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 2aa127d2-ff55-470b-822d-60c2d2d409d6, please check neutron logs for more information. [ 724.278512] env[61806]: ERROR nova.compute.manager Traceback (most recent call last): [ 724.278512] env[61806]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 724.278512] env[61806]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 724.278512] env[61806]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 724.278512] env[61806]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 724.278512] env[61806]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 724.278512] env[61806]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 724.278512] env[61806]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 724.278512] env[61806]: ERROR nova.compute.manager self.force_reraise() [ 724.278512] env[61806]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 724.278512] env[61806]: ERROR nova.compute.manager raise self.value [ 724.278512] env[61806]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 724.278512] env[61806]: ERROR nova.compute.manager updated_port = self._update_port( [ 724.278512] env[61806]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 724.278512] env[61806]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 724.279149] env[61806]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 724.279149] env[61806]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 724.279149] env[61806]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 2aa127d2-ff55-470b-822d-60c2d2d409d6, please check neutron logs for more information. [ 724.279149] env[61806]: ERROR nova.compute.manager [ 724.279149] env[61806]: Traceback (most recent call last): [ 724.279149] env[61806]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 724.279149] env[61806]: listener.cb(fileno) [ 724.279149] env[61806]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 724.279149] env[61806]: result = function(*args, **kwargs) [ 724.279149] env[61806]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 724.279149] env[61806]: return func(*args, **kwargs) [ 724.279149] env[61806]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 724.279149] env[61806]: raise e [ 724.279149] env[61806]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 724.279149] env[61806]: nwinfo = self.network_api.allocate_for_instance( [ 724.279149] env[61806]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 724.279149] env[61806]: created_port_ids = self._update_ports_for_instance( [ 724.279149] env[61806]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 724.279149] env[61806]: with excutils.save_and_reraise_exception(): [ 724.279149] env[61806]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 724.279149] env[61806]: self.force_reraise() [ 724.279149] env[61806]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 724.279149] env[61806]: raise self.value [ 724.279149] env[61806]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 724.279149] env[61806]: updated_port = self._update_port( [ 724.279149] env[61806]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 724.279149] env[61806]: _ensure_no_port_binding_failure(port) [ 724.279149] env[61806]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 724.279149] env[61806]: raise exception.PortBindingFailed(port_id=port['id']) [ 724.280054] env[61806]: nova.exception.PortBindingFailed: Binding failed for port 2aa127d2-ff55-470b-822d-60c2d2d409d6, please check neutron logs for more information. [ 724.280054] env[61806]: Removing descriptor: 14 [ 724.280054] env[61806]: ERROR nova.compute.manager [None req-3c3981d2-fa25-4f4b-87d4-a5411e67199e tempest-MultipleCreateTestJSON-1589015462 tempest-MultipleCreateTestJSON-1589015462-project-member] [instance: 7211d568-cc68-4e50-80ad-b78878d1deab] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 2aa127d2-ff55-470b-822d-60c2d2d409d6, please check neutron logs for more information. [ 724.280054] env[61806]: ERROR nova.compute.manager [instance: 7211d568-cc68-4e50-80ad-b78878d1deab] Traceback (most recent call last): [ 724.280054] env[61806]: ERROR nova.compute.manager [instance: 7211d568-cc68-4e50-80ad-b78878d1deab] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 724.280054] env[61806]: ERROR nova.compute.manager [instance: 7211d568-cc68-4e50-80ad-b78878d1deab] yield resources [ 724.280054] env[61806]: ERROR nova.compute.manager [instance: 7211d568-cc68-4e50-80ad-b78878d1deab] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 724.280054] env[61806]: ERROR nova.compute.manager [instance: 7211d568-cc68-4e50-80ad-b78878d1deab] self.driver.spawn(context, instance, image_meta, [ 724.280054] env[61806]: ERROR nova.compute.manager [instance: 7211d568-cc68-4e50-80ad-b78878d1deab] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 724.280054] env[61806]: ERROR nova.compute.manager [instance: 7211d568-cc68-4e50-80ad-b78878d1deab] self._vmops.spawn(context, instance, image_meta, injected_files, [ 724.280054] env[61806]: ERROR nova.compute.manager [instance: 7211d568-cc68-4e50-80ad-b78878d1deab] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 724.280054] env[61806]: ERROR nova.compute.manager [instance: 7211d568-cc68-4e50-80ad-b78878d1deab] vm_ref = self.build_virtual_machine(instance, [ 724.280472] env[61806]: ERROR nova.compute.manager [instance: 7211d568-cc68-4e50-80ad-b78878d1deab] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 724.280472] env[61806]: ERROR nova.compute.manager [instance: 7211d568-cc68-4e50-80ad-b78878d1deab] vif_infos = vmwarevif.get_vif_info(self._session, [ 724.280472] env[61806]: ERROR nova.compute.manager [instance: 7211d568-cc68-4e50-80ad-b78878d1deab] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 724.280472] env[61806]: ERROR nova.compute.manager [instance: 7211d568-cc68-4e50-80ad-b78878d1deab] for vif in network_info: [ 724.280472] env[61806]: ERROR nova.compute.manager [instance: 7211d568-cc68-4e50-80ad-b78878d1deab] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 724.280472] env[61806]: ERROR nova.compute.manager [instance: 7211d568-cc68-4e50-80ad-b78878d1deab] return self._sync_wrapper(fn, *args, **kwargs) [ 724.280472] env[61806]: ERROR nova.compute.manager [instance: 7211d568-cc68-4e50-80ad-b78878d1deab] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 724.280472] env[61806]: ERROR nova.compute.manager [instance: 7211d568-cc68-4e50-80ad-b78878d1deab] self.wait() [ 724.280472] env[61806]: ERROR nova.compute.manager [instance: 7211d568-cc68-4e50-80ad-b78878d1deab] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 724.280472] env[61806]: ERROR nova.compute.manager [instance: 7211d568-cc68-4e50-80ad-b78878d1deab] self[:] = self._gt.wait() [ 724.280472] env[61806]: ERROR nova.compute.manager [instance: 7211d568-cc68-4e50-80ad-b78878d1deab] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 724.280472] env[61806]: ERROR nova.compute.manager [instance: 7211d568-cc68-4e50-80ad-b78878d1deab] return self._exit_event.wait() [ 724.280472] env[61806]: ERROR nova.compute.manager [instance: 7211d568-cc68-4e50-80ad-b78878d1deab] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 724.280900] env[61806]: ERROR nova.compute.manager [instance: 7211d568-cc68-4e50-80ad-b78878d1deab] result = hub.switch() [ 724.280900] env[61806]: ERROR nova.compute.manager [instance: 7211d568-cc68-4e50-80ad-b78878d1deab] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 724.280900] env[61806]: ERROR nova.compute.manager [instance: 7211d568-cc68-4e50-80ad-b78878d1deab] return self.greenlet.switch() [ 724.280900] env[61806]: ERROR nova.compute.manager [instance: 7211d568-cc68-4e50-80ad-b78878d1deab] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 724.280900] env[61806]: ERROR nova.compute.manager [instance: 7211d568-cc68-4e50-80ad-b78878d1deab] result = function(*args, **kwargs) [ 724.280900] env[61806]: ERROR nova.compute.manager [instance: 7211d568-cc68-4e50-80ad-b78878d1deab] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 724.280900] env[61806]: ERROR nova.compute.manager [instance: 7211d568-cc68-4e50-80ad-b78878d1deab] return func(*args, **kwargs) [ 724.280900] env[61806]: ERROR nova.compute.manager [instance: 7211d568-cc68-4e50-80ad-b78878d1deab] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 724.280900] env[61806]: ERROR nova.compute.manager [instance: 7211d568-cc68-4e50-80ad-b78878d1deab] raise e [ 724.280900] env[61806]: ERROR nova.compute.manager [instance: 7211d568-cc68-4e50-80ad-b78878d1deab] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 724.280900] env[61806]: ERROR nova.compute.manager [instance: 7211d568-cc68-4e50-80ad-b78878d1deab] nwinfo = self.network_api.allocate_for_instance( [ 724.280900] env[61806]: ERROR nova.compute.manager [instance: 7211d568-cc68-4e50-80ad-b78878d1deab] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 724.280900] env[61806]: ERROR nova.compute.manager [instance: 7211d568-cc68-4e50-80ad-b78878d1deab] created_port_ids = self._update_ports_for_instance( [ 724.282694] env[61806]: ERROR nova.compute.manager [instance: 7211d568-cc68-4e50-80ad-b78878d1deab] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 724.282694] env[61806]: ERROR nova.compute.manager [instance: 7211d568-cc68-4e50-80ad-b78878d1deab] with excutils.save_and_reraise_exception(): [ 724.282694] env[61806]: ERROR nova.compute.manager [instance: 7211d568-cc68-4e50-80ad-b78878d1deab] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 724.282694] env[61806]: ERROR nova.compute.manager [instance: 7211d568-cc68-4e50-80ad-b78878d1deab] self.force_reraise() [ 724.282694] env[61806]: ERROR nova.compute.manager [instance: 7211d568-cc68-4e50-80ad-b78878d1deab] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 724.282694] env[61806]: ERROR nova.compute.manager [instance: 7211d568-cc68-4e50-80ad-b78878d1deab] raise self.value [ 724.282694] env[61806]: ERROR nova.compute.manager [instance: 7211d568-cc68-4e50-80ad-b78878d1deab] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 724.282694] env[61806]: ERROR nova.compute.manager [instance: 7211d568-cc68-4e50-80ad-b78878d1deab] updated_port = self._update_port( [ 724.282694] env[61806]: ERROR nova.compute.manager [instance: 7211d568-cc68-4e50-80ad-b78878d1deab] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 724.282694] env[61806]: ERROR nova.compute.manager [instance: 7211d568-cc68-4e50-80ad-b78878d1deab] _ensure_no_port_binding_failure(port) [ 724.282694] env[61806]: ERROR nova.compute.manager [instance: 7211d568-cc68-4e50-80ad-b78878d1deab] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 724.282694] env[61806]: ERROR nova.compute.manager [instance: 7211d568-cc68-4e50-80ad-b78878d1deab] raise exception.PortBindingFailed(port_id=port['id']) [ 724.283065] env[61806]: ERROR nova.compute.manager [instance: 7211d568-cc68-4e50-80ad-b78878d1deab] nova.exception.PortBindingFailed: Binding failed for port 2aa127d2-ff55-470b-822d-60c2d2d409d6, please check neutron logs for more information. [ 724.283065] env[61806]: ERROR nova.compute.manager [instance: 7211d568-cc68-4e50-80ad-b78878d1deab] [ 724.283065] env[61806]: INFO nova.compute.manager [None req-3c3981d2-fa25-4f4b-87d4-a5411e67199e tempest-MultipleCreateTestJSON-1589015462 tempest-MultipleCreateTestJSON-1589015462-project-member] [instance: 7211d568-cc68-4e50-80ad-b78878d1deab] Terminating instance [ 724.283260] env[61806]: DEBUG oslo_concurrency.lockutils [None req-3c3981d2-fa25-4f4b-87d4-a5411e67199e tempest-MultipleCreateTestJSON-1589015462 tempest-MultipleCreateTestJSON-1589015462-project-member] Acquiring lock "refresh_cache-7211d568-cc68-4e50-80ad-b78878d1deab" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 724.349285] env[61806]: DEBUG nova.network.neutron [None req-2bf81f39-4b8d-4851-a793-760af008ffe2 tempest-ImagesOneServerTestJSON-1003762139 tempest-ImagesOneServerTestJSON-1003762139-project-member] [instance: 294f0a37-7589-405d-a324-13099c80f458] Instance cache missing network info. {{(pid=61806) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 724.443515] env[61806]: DEBUG nova.network.neutron [None req-2bf81f39-4b8d-4851-a793-760af008ffe2 tempest-ImagesOneServerTestJSON-1003762139 tempest-ImagesOneServerTestJSON-1003762139-project-member] [instance: 294f0a37-7589-405d-a324-13099c80f458] Updating instance_info_cache with network_info: [] {{(pid=61806) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 724.530790] env[61806]: DEBUG nova.network.neutron [req-1d6ced2b-3cfe-4a70-b9e6-09e376f2421e req-2e2b7aa9-1bf8-4d2d-8a3d-85dad9991060 service nova] [instance: 7211d568-cc68-4e50-80ad-b78878d1deab] Instance cache missing network info. {{(pid=61806) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 724.606779] env[61806]: DEBUG nova.network.neutron [req-1d6ced2b-3cfe-4a70-b9e6-09e376f2421e req-2e2b7aa9-1bf8-4d2d-8a3d-85dad9991060 service nova] [instance: 7211d568-cc68-4e50-80ad-b78878d1deab] Updating instance_info_cache with network_info: [] {{(pid=61806) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 724.737776] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4d869077-ec6d-4073-ab8d-7e3de058c0f0 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 724.745433] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b23fe255-081a-4195-b18a-190320f7c3e9 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 724.775214] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8da1e364-9370-45c6-8066-7ddf5d7b14c3 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 724.782388] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d0e611e9-26be-4180-a0dc-a60ed4dafdee {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 724.795758] env[61806]: DEBUG nova.compute.provider_tree [None req-0f0a69c6-2310-4272-abc1-520940b4bcf0 tempest-ServersTestFqdnHostnames-356624376 tempest-ServersTestFqdnHostnames-356624376-project-member] Inventory has not changed in ProviderTree for provider: a2858be1-fd22-4e08-979e-87ad25293407 {{(pid=61806) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 724.950240] env[61806]: DEBUG oslo_concurrency.lockutils [None req-2bf81f39-4b8d-4851-a793-760af008ffe2 tempest-ImagesOneServerTestJSON-1003762139 tempest-ImagesOneServerTestJSON-1003762139-project-member] Releasing lock "refresh_cache-294f0a37-7589-405d-a324-13099c80f458" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 724.950500] env[61806]: DEBUG nova.compute.manager [None req-2bf81f39-4b8d-4851-a793-760af008ffe2 tempest-ImagesOneServerTestJSON-1003762139 tempest-ImagesOneServerTestJSON-1003762139-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61806) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 724.950692] env[61806]: DEBUG nova.compute.manager [None req-2bf81f39-4b8d-4851-a793-760af008ffe2 tempest-ImagesOneServerTestJSON-1003762139 tempest-ImagesOneServerTestJSON-1003762139-project-member] [instance: 294f0a37-7589-405d-a324-13099c80f458] Deallocating network for instance {{(pid=61806) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 724.950993] env[61806]: DEBUG nova.network.neutron [None req-2bf81f39-4b8d-4851-a793-760af008ffe2 tempest-ImagesOneServerTestJSON-1003762139 tempest-ImagesOneServerTestJSON-1003762139-project-member] [instance: 294f0a37-7589-405d-a324-13099c80f458] deallocate_for_instance() {{(pid=61806) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 724.978905] env[61806]: DEBUG nova.network.neutron [None req-2bf81f39-4b8d-4851-a793-760af008ffe2 tempest-ImagesOneServerTestJSON-1003762139 tempest-ImagesOneServerTestJSON-1003762139-project-member] [instance: 294f0a37-7589-405d-a324-13099c80f458] Instance cache missing network info. {{(pid=61806) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 725.111757] env[61806]: DEBUG oslo_concurrency.lockutils [req-1d6ced2b-3cfe-4a70-b9e6-09e376f2421e req-2e2b7aa9-1bf8-4d2d-8a3d-85dad9991060 service nova] Releasing lock "refresh_cache-7211d568-cc68-4e50-80ad-b78878d1deab" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 725.112273] env[61806]: DEBUG oslo_concurrency.lockutils [None req-3c3981d2-fa25-4f4b-87d4-a5411e67199e tempest-MultipleCreateTestJSON-1589015462 tempest-MultipleCreateTestJSON-1589015462-project-member] Acquired lock "refresh_cache-7211d568-cc68-4e50-80ad-b78878d1deab" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 725.112524] env[61806]: DEBUG nova.network.neutron [None req-3c3981d2-fa25-4f4b-87d4-a5411e67199e tempest-MultipleCreateTestJSON-1589015462 tempest-MultipleCreateTestJSON-1589015462-project-member] [instance: 7211d568-cc68-4e50-80ad-b78878d1deab] Building network info cache for instance {{(pid=61806) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 725.299168] env[61806]: DEBUG nova.scheduler.client.report [None req-0f0a69c6-2310-4272-abc1-520940b4bcf0 tempest-ServersTestFqdnHostnames-356624376 tempest-ServersTestFqdnHostnames-356624376-project-member] Inventory has not changed for provider a2858be1-fd22-4e08-979e-87ad25293407 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 139, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61806) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 725.485942] env[61806]: DEBUG nova.network.neutron [None req-2bf81f39-4b8d-4851-a793-760af008ffe2 tempest-ImagesOneServerTestJSON-1003762139 tempest-ImagesOneServerTestJSON-1003762139-project-member] [instance: 294f0a37-7589-405d-a324-13099c80f458] Updating instance_info_cache with network_info: [] {{(pid=61806) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 725.629749] env[61806]: DEBUG nova.network.neutron [None req-3c3981d2-fa25-4f4b-87d4-a5411e67199e tempest-MultipleCreateTestJSON-1589015462 tempest-MultipleCreateTestJSON-1589015462-project-member] [instance: 7211d568-cc68-4e50-80ad-b78878d1deab] Instance cache missing network info. {{(pid=61806) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 725.718287] env[61806]: DEBUG nova.network.neutron [None req-3c3981d2-fa25-4f4b-87d4-a5411e67199e tempest-MultipleCreateTestJSON-1589015462 tempest-MultipleCreateTestJSON-1589015462-project-member] [instance: 7211d568-cc68-4e50-80ad-b78878d1deab] Updating instance_info_cache with network_info: [] {{(pid=61806) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 725.807759] env[61806]: DEBUG oslo_concurrency.lockutils [None req-0f0a69c6-2310-4272-abc1-520940b4bcf0 tempest-ServersTestFqdnHostnames-356624376 tempest-ServersTestFqdnHostnames-356624376-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.985s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 725.808453] env[61806]: ERROR nova.compute.manager [None req-0f0a69c6-2310-4272-abc1-520940b4bcf0 tempest-ServersTestFqdnHostnames-356624376 tempest-ServersTestFqdnHostnames-356624376-project-member] [instance: 46c34ed0-9bac-4803-8d5c-3a3346a64117] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port b68f927d-fcde-4af6-ba22-2354a0b31a3c, please check neutron logs for more information. [ 725.808453] env[61806]: ERROR nova.compute.manager [instance: 46c34ed0-9bac-4803-8d5c-3a3346a64117] Traceback (most recent call last): [ 725.808453] env[61806]: ERROR nova.compute.manager [instance: 46c34ed0-9bac-4803-8d5c-3a3346a64117] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 725.808453] env[61806]: ERROR nova.compute.manager [instance: 46c34ed0-9bac-4803-8d5c-3a3346a64117] self.driver.spawn(context, instance, image_meta, [ 725.808453] env[61806]: ERROR nova.compute.manager [instance: 46c34ed0-9bac-4803-8d5c-3a3346a64117] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 725.808453] env[61806]: ERROR nova.compute.manager [instance: 46c34ed0-9bac-4803-8d5c-3a3346a64117] self._vmops.spawn(context, instance, image_meta, injected_files, [ 725.808453] env[61806]: ERROR nova.compute.manager [instance: 46c34ed0-9bac-4803-8d5c-3a3346a64117] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 725.808453] env[61806]: ERROR nova.compute.manager [instance: 46c34ed0-9bac-4803-8d5c-3a3346a64117] vm_ref = self.build_virtual_machine(instance, [ 725.808453] env[61806]: ERROR nova.compute.manager [instance: 46c34ed0-9bac-4803-8d5c-3a3346a64117] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 725.808453] env[61806]: ERROR nova.compute.manager [instance: 46c34ed0-9bac-4803-8d5c-3a3346a64117] vif_infos = vmwarevif.get_vif_info(self._session, [ 725.808453] env[61806]: ERROR nova.compute.manager [instance: 46c34ed0-9bac-4803-8d5c-3a3346a64117] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 725.808856] env[61806]: ERROR nova.compute.manager [instance: 46c34ed0-9bac-4803-8d5c-3a3346a64117] for vif in network_info: [ 725.808856] env[61806]: ERROR nova.compute.manager [instance: 46c34ed0-9bac-4803-8d5c-3a3346a64117] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 725.808856] env[61806]: ERROR nova.compute.manager [instance: 46c34ed0-9bac-4803-8d5c-3a3346a64117] return self._sync_wrapper(fn, *args, **kwargs) [ 725.808856] env[61806]: ERROR nova.compute.manager [instance: 46c34ed0-9bac-4803-8d5c-3a3346a64117] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 725.808856] env[61806]: ERROR nova.compute.manager [instance: 46c34ed0-9bac-4803-8d5c-3a3346a64117] self.wait() [ 725.808856] env[61806]: ERROR nova.compute.manager [instance: 46c34ed0-9bac-4803-8d5c-3a3346a64117] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 725.808856] env[61806]: ERROR nova.compute.manager [instance: 46c34ed0-9bac-4803-8d5c-3a3346a64117] self[:] = self._gt.wait() [ 725.808856] env[61806]: ERROR nova.compute.manager [instance: 46c34ed0-9bac-4803-8d5c-3a3346a64117] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 725.808856] env[61806]: ERROR nova.compute.manager [instance: 46c34ed0-9bac-4803-8d5c-3a3346a64117] return self._exit_event.wait() [ 725.808856] env[61806]: ERROR nova.compute.manager [instance: 46c34ed0-9bac-4803-8d5c-3a3346a64117] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 725.808856] env[61806]: ERROR nova.compute.manager [instance: 46c34ed0-9bac-4803-8d5c-3a3346a64117] current.throw(*self._exc) [ 725.808856] env[61806]: ERROR nova.compute.manager [instance: 46c34ed0-9bac-4803-8d5c-3a3346a64117] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 725.808856] env[61806]: ERROR nova.compute.manager [instance: 46c34ed0-9bac-4803-8d5c-3a3346a64117] result = function(*args, **kwargs) [ 725.809293] env[61806]: ERROR nova.compute.manager [instance: 46c34ed0-9bac-4803-8d5c-3a3346a64117] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 725.809293] env[61806]: ERROR nova.compute.manager [instance: 46c34ed0-9bac-4803-8d5c-3a3346a64117] return func(*args, **kwargs) [ 725.809293] env[61806]: ERROR nova.compute.manager [instance: 46c34ed0-9bac-4803-8d5c-3a3346a64117] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 725.809293] env[61806]: ERROR nova.compute.manager [instance: 46c34ed0-9bac-4803-8d5c-3a3346a64117] raise e [ 725.809293] env[61806]: ERROR nova.compute.manager [instance: 46c34ed0-9bac-4803-8d5c-3a3346a64117] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 725.809293] env[61806]: ERROR nova.compute.manager [instance: 46c34ed0-9bac-4803-8d5c-3a3346a64117] nwinfo = self.network_api.allocate_for_instance( [ 725.809293] env[61806]: ERROR nova.compute.manager [instance: 46c34ed0-9bac-4803-8d5c-3a3346a64117] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 725.809293] env[61806]: ERROR nova.compute.manager [instance: 46c34ed0-9bac-4803-8d5c-3a3346a64117] created_port_ids = self._update_ports_for_instance( [ 725.809293] env[61806]: ERROR nova.compute.manager [instance: 46c34ed0-9bac-4803-8d5c-3a3346a64117] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 725.809293] env[61806]: ERROR nova.compute.manager [instance: 46c34ed0-9bac-4803-8d5c-3a3346a64117] with excutils.save_and_reraise_exception(): [ 725.809293] env[61806]: ERROR nova.compute.manager [instance: 46c34ed0-9bac-4803-8d5c-3a3346a64117] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 725.809293] env[61806]: ERROR nova.compute.manager [instance: 46c34ed0-9bac-4803-8d5c-3a3346a64117] self.force_reraise() [ 725.809293] env[61806]: ERROR nova.compute.manager [instance: 46c34ed0-9bac-4803-8d5c-3a3346a64117] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 725.809780] env[61806]: ERROR nova.compute.manager [instance: 46c34ed0-9bac-4803-8d5c-3a3346a64117] raise self.value [ 725.809780] env[61806]: ERROR nova.compute.manager [instance: 46c34ed0-9bac-4803-8d5c-3a3346a64117] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 725.809780] env[61806]: ERROR nova.compute.manager [instance: 46c34ed0-9bac-4803-8d5c-3a3346a64117] updated_port = self._update_port( [ 725.809780] env[61806]: ERROR nova.compute.manager [instance: 46c34ed0-9bac-4803-8d5c-3a3346a64117] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 725.809780] env[61806]: ERROR nova.compute.manager [instance: 46c34ed0-9bac-4803-8d5c-3a3346a64117] _ensure_no_port_binding_failure(port) [ 725.809780] env[61806]: ERROR nova.compute.manager [instance: 46c34ed0-9bac-4803-8d5c-3a3346a64117] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 725.809780] env[61806]: ERROR nova.compute.manager [instance: 46c34ed0-9bac-4803-8d5c-3a3346a64117] raise exception.PortBindingFailed(port_id=port['id']) [ 725.809780] env[61806]: ERROR nova.compute.manager [instance: 46c34ed0-9bac-4803-8d5c-3a3346a64117] nova.exception.PortBindingFailed: Binding failed for port b68f927d-fcde-4af6-ba22-2354a0b31a3c, please check neutron logs for more information. [ 725.809780] env[61806]: ERROR nova.compute.manager [instance: 46c34ed0-9bac-4803-8d5c-3a3346a64117] [ 725.809780] env[61806]: DEBUG nova.compute.utils [None req-0f0a69c6-2310-4272-abc1-520940b4bcf0 tempest-ServersTestFqdnHostnames-356624376 tempest-ServersTestFqdnHostnames-356624376-project-member] [instance: 46c34ed0-9bac-4803-8d5c-3a3346a64117] Binding failed for port b68f927d-fcde-4af6-ba22-2354a0b31a3c, please check neutron logs for more information. {{(pid=61806) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 725.810674] env[61806]: DEBUG oslo_concurrency.lockutils [None req-05ef4399-8310-428c-ba55-af11abf4e445 tempest-SecurityGroupsTestJSON-1213485581 tempest-SecurityGroupsTestJSON-1213485581-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 23.204s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 725.813908] env[61806]: DEBUG nova.compute.manager [None req-0f0a69c6-2310-4272-abc1-520940b4bcf0 tempest-ServersTestFqdnHostnames-356624376 tempest-ServersTestFqdnHostnames-356624376-project-member] [instance: 46c34ed0-9bac-4803-8d5c-3a3346a64117] Build of instance 46c34ed0-9bac-4803-8d5c-3a3346a64117 was re-scheduled: Binding failed for port b68f927d-fcde-4af6-ba22-2354a0b31a3c, please check neutron logs for more information. {{(pid=61806) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 725.814370] env[61806]: DEBUG nova.compute.manager [None req-0f0a69c6-2310-4272-abc1-520940b4bcf0 tempest-ServersTestFqdnHostnames-356624376 tempest-ServersTestFqdnHostnames-356624376-project-member] [instance: 46c34ed0-9bac-4803-8d5c-3a3346a64117] Unplugging VIFs for instance {{(pid=61806) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 725.814604] env[61806]: DEBUG oslo_concurrency.lockutils [None req-0f0a69c6-2310-4272-abc1-520940b4bcf0 tempest-ServersTestFqdnHostnames-356624376 tempest-ServersTestFqdnHostnames-356624376-project-member] Acquiring lock "refresh_cache-46c34ed0-9bac-4803-8d5c-3a3346a64117" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 725.814751] env[61806]: DEBUG oslo_concurrency.lockutils [None req-0f0a69c6-2310-4272-abc1-520940b4bcf0 tempest-ServersTestFqdnHostnames-356624376 tempest-ServersTestFqdnHostnames-356624376-project-member] Acquired lock "refresh_cache-46c34ed0-9bac-4803-8d5c-3a3346a64117" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 725.814964] env[61806]: DEBUG nova.network.neutron [None req-0f0a69c6-2310-4272-abc1-520940b4bcf0 tempest-ServersTestFqdnHostnames-356624376 tempest-ServersTestFqdnHostnames-356624376-project-member] [instance: 46c34ed0-9bac-4803-8d5c-3a3346a64117] Building network info cache for instance {{(pid=61806) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 725.988230] env[61806]: INFO nova.compute.manager [None req-2bf81f39-4b8d-4851-a793-760af008ffe2 tempest-ImagesOneServerTestJSON-1003762139 tempest-ImagesOneServerTestJSON-1003762139-project-member] [instance: 294f0a37-7589-405d-a324-13099c80f458] Took 1.04 seconds to deallocate network for instance. [ 726.059132] env[61806]: DEBUG nova.compute.manager [req-3d58f377-70ee-4a97-a537-4cc30c523fd0 req-d4612db5-5ab9-4426-b3bb-57d66e26dd55 service nova] [instance: 7211d568-cc68-4e50-80ad-b78878d1deab] Received event network-vif-deleted-2aa127d2-ff55-470b-822d-60c2d2d409d6 {{(pid=61806) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 726.222949] env[61806]: DEBUG oslo_concurrency.lockutils [None req-3c3981d2-fa25-4f4b-87d4-a5411e67199e tempest-MultipleCreateTestJSON-1589015462 tempest-MultipleCreateTestJSON-1589015462-project-member] Releasing lock "refresh_cache-7211d568-cc68-4e50-80ad-b78878d1deab" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 726.223420] env[61806]: DEBUG nova.compute.manager [None req-3c3981d2-fa25-4f4b-87d4-a5411e67199e tempest-MultipleCreateTestJSON-1589015462 tempest-MultipleCreateTestJSON-1589015462-project-member] [instance: 7211d568-cc68-4e50-80ad-b78878d1deab] Start destroying the instance on the hypervisor. {{(pid=61806) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 726.223609] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-3c3981d2-fa25-4f4b-87d4-a5411e67199e tempest-MultipleCreateTestJSON-1589015462 tempest-MultipleCreateTestJSON-1589015462-project-member] [instance: 7211d568-cc68-4e50-80ad-b78878d1deab] Destroying instance {{(pid=61806) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 726.223912] env[61806]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-d171d583-d8e7-4d0f-ba51-33c681bf0c81 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 726.234415] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7b2f652b-72ca-417e-a023-419d94aa2ebc {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 726.258380] env[61806]: WARNING nova.virt.vmwareapi.vmops [None req-3c3981d2-fa25-4f4b-87d4-a5411e67199e tempest-MultipleCreateTestJSON-1589015462 tempest-MultipleCreateTestJSON-1589015462-project-member] [instance: 7211d568-cc68-4e50-80ad-b78878d1deab] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 7211d568-cc68-4e50-80ad-b78878d1deab could not be found. [ 726.258651] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-3c3981d2-fa25-4f4b-87d4-a5411e67199e tempest-MultipleCreateTestJSON-1589015462 tempest-MultipleCreateTestJSON-1589015462-project-member] [instance: 7211d568-cc68-4e50-80ad-b78878d1deab] Instance destroyed {{(pid=61806) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 726.258816] env[61806]: INFO nova.compute.manager [None req-3c3981d2-fa25-4f4b-87d4-a5411e67199e tempest-MultipleCreateTestJSON-1589015462 tempest-MultipleCreateTestJSON-1589015462-project-member] [instance: 7211d568-cc68-4e50-80ad-b78878d1deab] Took 0.04 seconds to destroy the instance on the hypervisor. [ 726.260151] env[61806]: DEBUG oslo.service.loopingcall [None req-3c3981d2-fa25-4f4b-87d4-a5411e67199e tempest-MultipleCreateTestJSON-1589015462 tempest-MultipleCreateTestJSON-1589015462-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61806) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 726.260151] env[61806]: DEBUG nova.compute.manager [-] [instance: 7211d568-cc68-4e50-80ad-b78878d1deab] Deallocating network for instance {{(pid=61806) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 726.260151] env[61806]: DEBUG nova.network.neutron [-] [instance: 7211d568-cc68-4e50-80ad-b78878d1deab] deallocate_for_instance() {{(pid=61806) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 726.278288] env[61806]: DEBUG nova.network.neutron [-] [instance: 7211d568-cc68-4e50-80ad-b78878d1deab] Instance cache missing network info. {{(pid=61806) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 726.336671] env[61806]: DEBUG nova.network.neutron [None req-0f0a69c6-2310-4272-abc1-520940b4bcf0 tempest-ServersTestFqdnHostnames-356624376 tempest-ServersTestFqdnHostnames-356624376-project-member] [instance: 46c34ed0-9bac-4803-8d5c-3a3346a64117] Instance cache missing network info. {{(pid=61806) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 726.441178] env[61806]: DEBUG nova.network.neutron [None req-0f0a69c6-2310-4272-abc1-520940b4bcf0 tempest-ServersTestFqdnHostnames-356624376 tempest-ServersTestFqdnHostnames-356624376-project-member] [instance: 46c34ed0-9bac-4803-8d5c-3a3346a64117] Updating instance_info_cache with network_info: [] {{(pid=61806) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 726.680318] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-10f54fac-d61a-4a43-a0b9-b09a68e6e3eb {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 726.688026] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e8e20ad0-0920-4c7b-b9f4-8d7563c2d1ab {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 726.719469] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-63ecb211-31b0-418c-87ec-4bc71897ac74 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 726.728017] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e7483e3d-5487-4a82-bf74-3d1145c122b3 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 726.740438] env[61806]: DEBUG nova.compute.provider_tree [None req-05ef4399-8310-428c-ba55-af11abf4e445 tempest-SecurityGroupsTestJSON-1213485581 tempest-SecurityGroupsTestJSON-1213485581-project-member] Inventory has not changed in ProviderTree for provider: a2858be1-fd22-4e08-979e-87ad25293407 {{(pid=61806) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 726.780647] env[61806]: DEBUG nova.network.neutron [-] [instance: 7211d568-cc68-4e50-80ad-b78878d1deab] Updating instance_info_cache with network_info: [] {{(pid=61806) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 726.948891] env[61806]: DEBUG oslo_concurrency.lockutils [None req-0f0a69c6-2310-4272-abc1-520940b4bcf0 tempest-ServersTestFqdnHostnames-356624376 tempest-ServersTestFqdnHostnames-356624376-project-member] Releasing lock "refresh_cache-46c34ed0-9bac-4803-8d5c-3a3346a64117" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 726.949154] env[61806]: DEBUG nova.compute.manager [None req-0f0a69c6-2310-4272-abc1-520940b4bcf0 tempest-ServersTestFqdnHostnames-356624376 tempest-ServersTestFqdnHostnames-356624376-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61806) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 726.949776] env[61806]: DEBUG nova.compute.manager [None req-0f0a69c6-2310-4272-abc1-520940b4bcf0 tempest-ServersTestFqdnHostnames-356624376 tempest-ServersTestFqdnHostnames-356624376-project-member] [instance: 46c34ed0-9bac-4803-8d5c-3a3346a64117] Deallocating network for instance {{(pid=61806) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 726.949776] env[61806]: DEBUG nova.network.neutron [None req-0f0a69c6-2310-4272-abc1-520940b4bcf0 tempest-ServersTestFqdnHostnames-356624376 tempest-ServersTestFqdnHostnames-356624376-project-member] [instance: 46c34ed0-9bac-4803-8d5c-3a3346a64117] deallocate_for_instance() {{(pid=61806) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 726.968546] env[61806]: DEBUG nova.network.neutron [None req-0f0a69c6-2310-4272-abc1-520940b4bcf0 tempest-ServersTestFqdnHostnames-356624376 tempest-ServersTestFqdnHostnames-356624376-project-member] [instance: 46c34ed0-9bac-4803-8d5c-3a3346a64117] Instance cache missing network info. {{(pid=61806) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 727.026900] env[61806]: INFO nova.scheduler.client.report [None req-2bf81f39-4b8d-4851-a793-760af008ffe2 tempest-ImagesOneServerTestJSON-1003762139 tempest-ImagesOneServerTestJSON-1003762139-project-member] Deleted allocations for instance 294f0a37-7589-405d-a324-13099c80f458 [ 727.243938] env[61806]: DEBUG nova.scheduler.client.report [None req-05ef4399-8310-428c-ba55-af11abf4e445 tempest-SecurityGroupsTestJSON-1213485581 tempest-SecurityGroupsTestJSON-1213485581-project-member] Inventory has not changed for provider a2858be1-fd22-4e08-979e-87ad25293407 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 139, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61806) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 727.283814] env[61806]: INFO nova.compute.manager [-] [instance: 7211d568-cc68-4e50-80ad-b78878d1deab] Took 1.02 seconds to deallocate network for instance. [ 727.286473] env[61806]: DEBUG nova.compute.claims [None req-3c3981d2-fa25-4f4b-87d4-a5411e67199e tempest-MultipleCreateTestJSON-1589015462 tempest-MultipleCreateTestJSON-1589015462-project-member] [instance: 7211d568-cc68-4e50-80ad-b78878d1deab] Aborting claim: {{(pid=61806) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 727.286682] env[61806]: DEBUG oslo_concurrency.lockutils [None req-3c3981d2-fa25-4f4b-87d4-a5411e67199e tempest-MultipleCreateTestJSON-1589015462 tempest-MultipleCreateTestJSON-1589015462-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 727.474069] env[61806]: DEBUG nova.network.neutron [None req-0f0a69c6-2310-4272-abc1-520940b4bcf0 tempest-ServersTestFqdnHostnames-356624376 tempest-ServersTestFqdnHostnames-356624376-project-member] [instance: 46c34ed0-9bac-4803-8d5c-3a3346a64117] Updating instance_info_cache with network_info: [] {{(pid=61806) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 727.540671] env[61806]: DEBUG oslo_concurrency.lockutils [None req-2bf81f39-4b8d-4851-a793-760af008ffe2 tempest-ImagesOneServerTestJSON-1003762139 tempest-ImagesOneServerTestJSON-1003762139-project-member] Lock "294f0a37-7589-405d-a324-13099c80f458" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 156.130s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 727.750482] env[61806]: DEBUG oslo_concurrency.lockutils [None req-05ef4399-8310-428c-ba55-af11abf4e445 tempest-SecurityGroupsTestJSON-1213485581 tempest-SecurityGroupsTestJSON-1213485581-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.940s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 727.751201] env[61806]: ERROR nova.compute.manager [None req-05ef4399-8310-428c-ba55-af11abf4e445 tempest-SecurityGroupsTestJSON-1213485581 tempest-SecurityGroupsTestJSON-1213485581-project-member] [instance: 294adb58-f23e-4510-a25a-de6b909e3189] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 5fb92c74-bcc1-4423-9d6b-394b8c87bb92, please check neutron logs for more information. [ 727.751201] env[61806]: ERROR nova.compute.manager [instance: 294adb58-f23e-4510-a25a-de6b909e3189] Traceback (most recent call last): [ 727.751201] env[61806]: ERROR nova.compute.manager [instance: 294adb58-f23e-4510-a25a-de6b909e3189] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 727.751201] env[61806]: ERROR nova.compute.manager [instance: 294adb58-f23e-4510-a25a-de6b909e3189] self.driver.spawn(context, instance, image_meta, [ 727.751201] env[61806]: ERROR nova.compute.manager [instance: 294adb58-f23e-4510-a25a-de6b909e3189] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 727.751201] env[61806]: ERROR nova.compute.manager [instance: 294adb58-f23e-4510-a25a-de6b909e3189] self._vmops.spawn(context, instance, image_meta, injected_files, [ 727.751201] env[61806]: ERROR nova.compute.manager [instance: 294adb58-f23e-4510-a25a-de6b909e3189] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 727.751201] env[61806]: ERROR nova.compute.manager [instance: 294adb58-f23e-4510-a25a-de6b909e3189] vm_ref = self.build_virtual_machine(instance, [ 727.751201] env[61806]: ERROR nova.compute.manager [instance: 294adb58-f23e-4510-a25a-de6b909e3189] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 727.751201] env[61806]: ERROR nova.compute.manager [instance: 294adb58-f23e-4510-a25a-de6b909e3189] vif_infos = vmwarevif.get_vif_info(self._session, [ 727.751201] env[61806]: ERROR nova.compute.manager [instance: 294adb58-f23e-4510-a25a-de6b909e3189] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 727.751636] env[61806]: ERROR nova.compute.manager [instance: 294adb58-f23e-4510-a25a-de6b909e3189] for vif in network_info: [ 727.751636] env[61806]: ERROR nova.compute.manager [instance: 294adb58-f23e-4510-a25a-de6b909e3189] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 727.751636] env[61806]: ERROR nova.compute.manager [instance: 294adb58-f23e-4510-a25a-de6b909e3189] return self._sync_wrapper(fn, *args, **kwargs) [ 727.751636] env[61806]: ERROR nova.compute.manager [instance: 294adb58-f23e-4510-a25a-de6b909e3189] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 727.751636] env[61806]: ERROR nova.compute.manager [instance: 294adb58-f23e-4510-a25a-de6b909e3189] self.wait() [ 727.751636] env[61806]: ERROR nova.compute.manager [instance: 294adb58-f23e-4510-a25a-de6b909e3189] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 727.751636] env[61806]: ERROR nova.compute.manager [instance: 294adb58-f23e-4510-a25a-de6b909e3189] self[:] = self._gt.wait() [ 727.751636] env[61806]: ERROR nova.compute.manager [instance: 294adb58-f23e-4510-a25a-de6b909e3189] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 727.751636] env[61806]: ERROR nova.compute.manager [instance: 294adb58-f23e-4510-a25a-de6b909e3189] return self._exit_event.wait() [ 727.751636] env[61806]: ERROR nova.compute.manager [instance: 294adb58-f23e-4510-a25a-de6b909e3189] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 727.751636] env[61806]: ERROR nova.compute.manager [instance: 294adb58-f23e-4510-a25a-de6b909e3189] current.throw(*self._exc) [ 727.751636] env[61806]: ERROR nova.compute.manager [instance: 294adb58-f23e-4510-a25a-de6b909e3189] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 727.751636] env[61806]: ERROR nova.compute.manager [instance: 294adb58-f23e-4510-a25a-de6b909e3189] result = function(*args, **kwargs) [ 727.752057] env[61806]: ERROR nova.compute.manager [instance: 294adb58-f23e-4510-a25a-de6b909e3189] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 727.752057] env[61806]: ERROR nova.compute.manager [instance: 294adb58-f23e-4510-a25a-de6b909e3189] return func(*args, **kwargs) [ 727.752057] env[61806]: ERROR nova.compute.manager [instance: 294adb58-f23e-4510-a25a-de6b909e3189] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 727.752057] env[61806]: ERROR nova.compute.manager [instance: 294adb58-f23e-4510-a25a-de6b909e3189] raise e [ 727.752057] env[61806]: ERROR nova.compute.manager [instance: 294adb58-f23e-4510-a25a-de6b909e3189] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 727.752057] env[61806]: ERROR nova.compute.manager [instance: 294adb58-f23e-4510-a25a-de6b909e3189] nwinfo = self.network_api.allocate_for_instance( [ 727.752057] env[61806]: ERROR nova.compute.manager [instance: 294adb58-f23e-4510-a25a-de6b909e3189] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 727.752057] env[61806]: ERROR nova.compute.manager [instance: 294adb58-f23e-4510-a25a-de6b909e3189] created_port_ids = self._update_ports_for_instance( [ 727.752057] env[61806]: ERROR nova.compute.manager [instance: 294adb58-f23e-4510-a25a-de6b909e3189] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 727.752057] env[61806]: ERROR nova.compute.manager [instance: 294adb58-f23e-4510-a25a-de6b909e3189] with excutils.save_and_reraise_exception(): [ 727.752057] env[61806]: ERROR nova.compute.manager [instance: 294adb58-f23e-4510-a25a-de6b909e3189] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 727.752057] env[61806]: ERROR nova.compute.manager [instance: 294adb58-f23e-4510-a25a-de6b909e3189] self.force_reraise() [ 727.752057] env[61806]: ERROR nova.compute.manager [instance: 294adb58-f23e-4510-a25a-de6b909e3189] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 727.752515] env[61806]: ERROR nova.compute.manager [instance: 294adb58-f23e-4510-a25a-de6b909e3189] raise self.value [ 727.752515] env[61806]: ERROR nova.compute.manager [instance: 294adb58-f23e-4510-a25a-de6b909e3189] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 727.752515] env[61806]: ERROR nova.compute.manager [instance: 294adb58-f23e-4510-a25a-de6b909e3189] updated_port = self._update_port( [ 727.752515] env[61806]: ERROR nova.compute.manager [instance: 294adb58-f23e-4510-a25a-de6b909e3189] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 727.752515] env[61806]: ERROR nova.compute.manager [instance: 294adb58-f23e-4510-a25a-de6b909e3189] _ensure_no_port_binding_failure(port) [ 727.752515] env[61806]: ERROR nova.compute.manager [instance: 294adb58-f23e-4510-a25a-de6b909e3189] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 727.752515] env[61806]: ERROR nova.compute.manager [instance: 294adb58-f23e-4510-a25a-de6b909e3189] raise exception.PortBindingFailed(port_id=port['id']) [ 727.752515] env[61806]: ERROR nova.compute.manager [instance: 294adb58-f23e-4510-a25a-de6b909e3189] nova.exception.PortBindingFailed: Binding failed for port 5fb92c74-bcc1-4423-9d6b-394b8c87bb92, please check neutron logs for more information. [ 727.752515] env[61806]: ERROR nova.compute.manager [instance: 294adb58-f23e-4510-a25a-de6b909e3189] [ 727.752515] env[61806]: DEBUG nova.compute.utils [None req-05ef4399-8310-428c-ba55-af11abf4e445 tempest-SecurityGroupsTestJSON-1213485581 tempest-SecurityGroupsTestJSON-1213485581-project-member] [instance: 294adb58-f23e-4510-a25a-de6b909e3189] Binding failed for port 5fb92c74-bcc1-4423-9d6b-394b8c87bb92, please check neutron logs for more information. {{(pid=61806) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 727.754081] env[61806]: DEBUG oslo_concurrency.lockutils [None req-3c3981d2-fa25-4f4b-87d4-a5411e67199e tempest-MultipleCreateTestJSON-1589015462 tempest-MultipleCreateTestJSON-1589015462-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 23.799s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 727.755913] env[61806]: INFO nova.compute.claims [None req-3c3981d2-fa25-4f4b-87d4-a5411e67199e tempest-MultipleCreateTestJSON-1589015462 tempest-MultipleCreateTestJSON-1589015462-project-member] [instance: 5b22c2fc-525c-481b-b84e-e0a7f68f633e] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 727.761858] env[61806]: DEBUG nova.compute.manager [None req-05ef4399-8310-428c-ba55-af11abf4e445 tempest-SecurityGroupsTestJSON-1213485581 tempest-SecurityGroupsTestJSON-1213485581-project-member] [instance: 294adb58-f23e-4510-a25a-de6b909e3189] Build of instance 294adb58-f23e-4510-a25a-de6b909e3189 was re-scheduled: Binding failed for port 5fb92c74-bcc1-4423-9d6b-394b8c87bb92, please check neutron logs for more information. {{(pid=61806) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 727.762013] env[61806]: DEBUG nova.compute.manager [None req-05ef4399-8310-428c-ba55-af11abf4e445 tempest-SecurityGroupsTestJSON-1213485581 tempest-SecurityGroupsTestJSON-1213485581-project-member] [instance: 294adb58-f23e-4510-a25a-de6b909e3189] Unplugging VIFs for instance {{(pid=61806) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 727.762259] env[61806]: DEBUG oslo_concurrency.lockutils [None req-05ef4399-8310-428c-ba55-af11abf4e445 tempest-SecurityGroupsTestJSON-1213485581 tempest-SecurityGroupsTestJSON-1213485581-project-member] Acquiring lock "refresh_cache-294adb58-f23e-4510-a25a-de6b909e3189" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 727.762478] env[61806]: DEBUG oslo_concurrency.lockutils [None req-05ef4399-8310-428c-ba55-af11abf4e445 tempest-SecurityGroupsTestJSON-1213485581 tempest-SecurityGroupsTestJSON-1213485581-project-member] Acquired lock "refresh_cache-294adb58-f23e-4510-a25a-de6b909e3189" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 727.762660] env[61806]: DEBUG nova.network.neutron [None req-05ef4399-8310-428c-ba55-af11abf4e445 tempest-SecurityGroupsTestJSON-1213485581 tempest-SecurityGroupsTestJSON-1213485581-project-member] [instance: 294adb58-f23e-4510-a25a-de6b909e3189] Building network info cache for instance {{(pid=61806) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 727.977364] env[61806]: INFO nova.compute.manager [None req-0f0a69c6-2310-4272-abc1-520940b4bcf0 tempest-ServersTestFqdnHostnames-356624376 tempest-ServersTestFqdnHostnames-356624376-project-member] [instance: 46c34ed0-9bac-4803-8d5c-3a3346a64117] Took 1.03 seconds to deallocate network for instance. [ 728.046807] env[61806]: DEBUG nova.compute.manager [None req-3bf756e0-c56e-4902-9c2b-b224a1d09a7c tempest-ListServersNegativeTestJSON-272187588 tempest-ListServersNegativeTestJSON-272187588-project-member] [instance: ea8d39a1-066d-4eb4-888d-776b0e45f684] Starting instance... {{(pid=61806) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 728.269075] env[61806]: DEBUG oslo_concurrency.lockutils [None req-8aa43bc8-f134-4682-8a54-ce9aa25e4ee5 tempest-AttachVolumeNegativeTest-1683033704 tempest-AttachVolumeNegativeTest-1683033704-project-member] Acquiring lock "43dcce0d-4fdb-4ee1-8309-92e92d2331a9" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 728.269623] env[61806]: DEBUG oslo_concurrency.lockutils [None req-8aa43bc8-f134-4682-8a54-ce9aa25e4ee5 tempest-AttachVolumeNegativeTest-1683033704 tempest-AttachVolumeNegativeTest-1683033704-project-member] Lock "43dcce0d-4fdb-4ee1-8309-92e92d2331a9" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 728.293832] env[61806]: DEBUG nova.network.neutron [None req-05ef4399-8310-428c-ba55-af11abf4e445 tempest-SecurityGroupsTestJSON-1213485581 tempest-SecurityGroupsTestJSON-1213485581-project-member] [instance: 294adb58-f23e-4510-a25a-de6b909e3189] Instance cache missing network info. {{(pid=61806) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 728.375109] env[61806]: DEBUG nova.network.neutron [None req-05ef4399-8310-428c-ba55-af11abf4e445 tempest-SecurityGroupsTestJSON-1213485581 tempest-SecurityGroupsTestJSON-1213485581-project-member] [instance: 294adb58-f23e-4510-a25a-de6b909e3189] Updating instance_info_cache with network_info: [] {{(pid=61806) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 728.568680] env[61806]: DEBUG oslo_concurrency.lockutils [None req-3bf756e0-c56e-4902-9c2b-b224a1d09a7c tempest-ListServersNegativeTestJSON-272187588 tempest-ListServersNegativeTestJSON-272187588-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 728.877758] env[61806]: DEBUG oslo_concurrency.lockutils [None req-05ef4399-8310-428c-ba55-af11abf4e445 tempest-SecurityGroupsTestJSON-1213485581 tempest-SecurityGroupsTestJSON-1213485581-project-member] Releasing lock "refresh_cache-294adb58-f23e-4510-a25a-de6b909e3189" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 728.878038] env[61806]: DEBUG nova.compute.manager [None req-05ef4399-8310-428c-ba55-af11abf4e445 tempest-SecurityGroupsTestJSON-1213485581 tempest-SecurityGroupsTestJSON-1213485581-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61806) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 728.878174] env[61806]: DEBUG nova.compute.manager [None req-05ef4399-8310-428c-ba55-af11abf4e445 tempest-SecurityGroupsTestJSON-1213485581 tempest-SecurityGroupsTestJSON-1213485581-project-member] [instance: 294adb58-f23e-4510-a25a-de6b909e3189] Deallocating network for instance {{(pid=61806) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 728.878337] env[61806]: DEBUG nova.network.neutron [None req-05ef4399-8310-428c-ba55-af11abf4e445 tempest-SecurityGroupsTestJSON-1213485581 tempest-SecurityGroupsTestJSON-1213485581-project-member] [instance: 294adb58-f23e-4510-a25a-de6b909e3189] deallocate_for_instance() {{(pid=61806) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 728.901716] env[61806]: DEBUG nova.network.neutron [None req-05ef4399-8310-428c-ba55-af11abf4e445 tempest-SecurityGroupsTestJSON-1213485581 tempest-SecurityGroupsTestJSON-1213485581-project-member] [instance: 294adb58-f23e-4510-a25a-de6b909e3189] Instance cache missing network info. {{(pid=61806) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 729.017132] env[61806]: INFO nova.scheduler.client.report [None req-0f0a69c6-2310-4272-abc1-520940b4bcf0 tempest-ServersTestFqdnHostnames-356624376 tempest-ServersTestFqdnHostnames-356624376-project-member] Deleted allocations for instance 46c34ed0-9bac-4803-8d5c-3a3346a64117 [ 729.148629] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ef10df32-b594-408a-b483-47a638f99d13 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 729.158547] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b66a7158-57f4-4044-a602-a2668657e5c4 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 729.190894] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-adf652a6-1d55-4dac-887f-4f32fde10b67 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 729.198759] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b69b7d60-7484-4dc8-9a2c-74781fe8ab70 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 729.212859] env[61806]: DEBUG nova.compute.provider_tree [None req-3c3981d2-fa25-4f4b-87d4-a5411e67199e tempest-MultipleCreateTestJSON-1589015462 tempest-MultipleCreateTestJSON-1589015462-project-member] Inventory has not changed in ProviderTree for provider: a2858be1-fd22-4e08-979e-87ad25293407 {{(pid=61806) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 729.405807] env[61806]: DEBUG nova.network.neutron [None req-05ef4399-8310-428c-ba55-af11abf4e445 tempest-SecurityGroupsTestJSON-1213485581 tempest-SecurityGroupsTestJSON-1213485581-project-member] [instance: 294adb58-f23e-4510-a25a-de6b909e3189] Updating instance_info_cache with network_info: [] {{(pid=61806) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 729.532375] env[61806]: DEBUG oslo_concurrency.lockutils [None req-0f0a69c6-2310-4272-abc1-520940b4bcf0 tempest-ServersTestFqdnHostnames-356624376 tempest-ServersTestFqdnHostnames-356624376-project-member] Lock "46c34ed0-9bac-4803-8d5c-3a3346a64117" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 157.228s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 729.716222] env[61806]: DEBUG nova.scheduler.client.report [None req-3c3981d2-fa25-4f4b-87d4-a5411e67199e tempest-MultipleCreateTestJSON-1589015462 tempest-MultipleCreateTestJSON-1589015462-project-member] Inventory has not changed for provider a2858be1-fd22-4e08-979e-87ad25293407 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 139, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61806) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 729.909587] env[61806]: INFO nova.compute.manager [None req-05ef4399-8310-428c-ba55-af11abf4e445 tempest-SecurityGroupsTestJSON-1213485581 tempest-SecurityGroupsTestJSON-1213485581-project-member] [instance: 294adb58-f23e-4510-a25a-de6b909e3189] Took 1.03 seconds to deallocate network for instance. [ 730.034718] env[61806]: DEBUG nova.compute.manager [None req-3bf756e0-c56e-4902-9c2b-b224a1d09a7c tempest-ListServersNegativeTestJSON-272187588 tempest-ListServersNegativeTestJSON-272187588-project-member] [instance: 198dc1a4-050d-47b2-8bc4-30af5534a789] Starting instance... {{(pid=61806) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 730.221164] env[61806]: DEBUG oslo_concurrency.lockutils [None req-3c3981d2-fa25-4f4b-87d4-a5411e67199e tempest-MultipleCreateTestJSON-1589015462 tempest-MultipleCreateTestJSON-1589015462-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.467s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 730.221679] env[61806]: DEBUG nova.compute.manager [None req-3c3981d2-fa25-4f4b-87d4-a5411e67199e tempest-MultipleCreateTestJSON-1589015462 tempest-MultipleCreateTestJSON-1589015462-project-member] [instance: 5b22c2fc-525c-481b-b84e-e0a7f68f633e] Start building networks asynchronously for instance. {{(pid=61806) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 730.224212] env[61806]: DEBUG oslo_concurrency.lockutils [None req-fcbc4cf1-fc11-42b1-9d49-ef556b52782e tempest-ServerRescueNegativeTestJSON-1483258717 tempest-ServerRescueNegativeTestJSON-1483258717-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 18.658s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 730.562150] env[61806]: DEBUG oslo_concurrency.lockutils [None req-3bf756e0-c56e-4902-9c2b-b224a1d09a7c tempest-ListServersNegativeTestJSON-272187588 tempest-ListServersNegativeTestJSON-272187588-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 730.728750] env[61806]: DEBUG nova.compute.utils [None req-3c3981d2-fa25-4f4b-87d4-a5411e67199e tempest-MultipleCreateTestJSON-1589015462 tempest-MultipleCreateTestJSON-1589015462-project-member] Using /dev/sd instead of None {{(pid=61806) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 730.733229] env[61806]: DEBUG nova.compute.manager [None req-3c3981d2-fa25-4f4b-87d4-a5411e67199e tempest-MultipleCreateTestJSON-1589015462 tempest-MultipleCreateTestJSON-1589015462-project-member] [instance: 5b22c2fc-525c-481b-b84e-e0a7f68f633e] Allocating IP information in the background. {{(pid=61806) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 730.733421] env[61806]: DEBUG nova.network.neutron [None req-3c3981d2-fa25-4f4b-87d4-a5411e67199e tempest-MultipleCreateTestJSON-1589015462 tempest-MultipleCreateTestJSON-1589015462-project-member] [instance: 5b22c2fc-525c-481b-b84e-e0a7f68f633e] allocate_for_instance() {{(pid=61806) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 730.806622] env[61806]: DEBUG nova.policy [None req-3c3981d2-fa25-4f4b-87d4-a5411e67199e tempest-MultipleCreateTestJSON-1589015462 tempest-MultipleCreateTestJSON-1589015462-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'b5a92ad3d119419085f56d7dea0975f5', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'dc40805c39f34d07be9d07861ad5841d', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61806) authorize /opt/stack/nova/nova/policy.py:201}} [ 730.949530] env[61806]: INFO nova.scheduler.client.report [None req-05ef4399-8310-428c-ba55-af11abf4e445 tempest-SecurityGroupsTestJSON-1213485581 tempest-SecurityGroupsTestJSON-1213485581-project-member] Deleted allocations for instance 294adb58-f23e-4510-a25a-de6b909e3189 [ 731.143888] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fbb32dfd-be83-40f9-a5d2-33d78ffdf17b {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 731.151923] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7d671bc5-2562-4472-8ec0-53c381b9e945 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 731.183588] env[61806]: DEBUG nova.network.neutron [None req-3c3981d2-fa25-4f4b-87d4-a5411e67199e tempest-MultipleCreateTestJSON-1589015462 tempest-MultipleCreateTestJSON-1589015462-project-member] [instance: 5b22c2fc-525c-481b-b84e-e0a7f68f633e] Successfully created port: 6cacc512-7619-41e7-8fcb-b88df2c7cf8b {{(pid=61806) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 731.186946] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0841432a-ba8a-4af9-9a7b-51bfd60784fe {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 731.193805] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4c2261fd-e86d-4cc7-87ca-eb49c90f57d2 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 731.207023] env[61806]: DEBUG nova.compute.provider_tree [None req-fcbc4cf1-fc11-42b1-9d49-ef556b52782e tempest-ServerRescueNegativeTestJSON-1483258717 tempest-ServerRescueNegativeTestJSON-1483258717-project-member] Inventory has not changed in ProviderTree for provider: a2858be1-fd22-4e08-979e-87ad25293407 {{(pid=61806) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 731.233706] env[61806]: DEBUG nova.compute.manager [None req-3c3981d2-fa25-4f4b-87d4-a5411e67199e tempest-MultipleCreateTestJSON-1589015462 tempest-MultipleCreateTestJSON-1589015462-project-member] [instance: 5b22c2fc-525c-481b-b84e-e0a7f68f633e] Start building block device mappings for instance. {{(pid=61806) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 731.461926] env[61806]: DEBUG oslo_concurrency.lockutils [None req-05ef4399-8310-428c-ba55-af11abf4e445 tempest-SecurityGroupsTestJSON-1213485581 tempest-SecurityGroupsTestJSON-1213485581-project-member] Lock "294adb58-f23e-4510-a25a-de6b909e3189" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 155.690s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 731.710965] env[61806]: DEBUG nova.scheduler.client.report [None req-fcbc4cf1-fc11-42b1-9d49-ef556b52782e tempest-ServerRescueNegativeTestJSON-1483258717 tempest-ServerRescueNegativeTestJSON-1483258717-project-member] Inventory has not changed for provider a2858be1-fd22-4e08-979e-87ad25293407 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 139, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61806) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 731.966803] env[61806]: DEBUG nova.compute.manager [None req-f0377b51-dab9-4f68-8592-ca0f31b4ba51 tempest-FloatingIPsAssociationNegativeTestJSON-30494360 tempest-FloatingIPsAssociationNegativeTestJSON-30494360-project-member] [instance: 37c07b8d-e2a0-4aa6-b7ba-194feed837e5] Starting instance... {{(pid=61806) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 732.217066] env[61806]: DEBUG oslo_concurrency.lockutils [None req-fcbc4cf1-fc11-42b1-9d49-ef556b52782e tempest-ServerRescueNegativeTestJSON-1483258717 tempest-ServerRescueNegativeTestJSON-1483258717-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.993s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 732.218128] env[61806]: ERROR nova.compute.manager [None req-fcbc4cf1-fc11-42b1-9d49-ef556b52782e tempest-ServerRescueNegativeTestJSON-1483258717 tempest-ServerRescueNegativeTestJSON-1483258717-project-member] [instance: 7ec2497a-7329-4ed4-99ab-dec3d1e8117c] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 81f90dee-a680-4d31-9c3d-fbe42fa742a4, please check neutron logs for more information. [ 732.218128] env[61806]: ERROR nova.compute.manager [instance: 7ec2497a-7329-4ed4-99ab-dec3d1e8117c] Traceback (most recent call last): [ 732.218128] env[61806]: ERROR nova.compute.manager [instance: 7ec2497a-7329-4ed4-99ab-dec3d1e8117c] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 732.218128] env[61806]: ERROR nova.compute.manager [instance: 7ec2497a-7329-4ed4-99ab-dec3d1e8117c] self.driver.spawn(context, instance, image_meta, [ 732.218128] env[61806]: ERROR nova.compute.manager [instance: 7ec2497a-7329-4ed4-99ab-dec3d1e8117c] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 732.218128] env[61806]: ERROR nova.compute.manager [instance: 7ec2497a-7329-4ed4-99ab-dec3d1e8117c] self._vmops.spawn(context, instance, image_meta, injected_files, [ 732.218128] env[61806]: ERROR nova.compute.manager [instance: 7ec2497a-7329-4ed4-99ab-dec3d1e8117c] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 732.218128] env[61806]: ERROR nova.compute.manager [instance: 7ec2497a-7329-4ed4-99ab-dec3d1e8117c] vm_ref = self.build_virtual_machine(instance, [ 732.218128] env[61806]: ERROR nova.compute.manager [instance: 7ec2497a-7329-4ed4-99ab-dec3d1e8117c] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 732.218128] env[61806]: ERROR nova.compute.manager [instance: 7ec2497a-7329-4ed4-99ab-dec3d1e8117c] vif_infos = vmwarevif.get_vif_info(self._session, [ 732.218128] env[61806]: ERROR nova.compute.manager [instance: 7ec2497a-7329-4ed4-99ab-dec3d1e8117c] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 732.218482] env[61806]: ERROR nova.compute.manager [instance: 7ec2497a-7329-4ed4-99ab-dec3d1e8117c] for vif in network_info: [ 732.218482] env[61806]: ERROR nova.compute.manager [instance: 7ec2497a-7329-4ed4-99ab-dec3d1e8117c] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 732.218482] env[61806]: ERROR nova.compute.manager [instance: 7ec2497a-7329-4ed4-99ab-dec3d1e8117c] return self._sync_wrapper(fn, *args, **kwargs) [ 732.218482] env[61806]: ERROR nova.compute.manager [instance: 7ec2497a-7329-4ed4-99ab-dec3d1e8117c] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 732.218482] env[61806]: ERROR nova.compute.manager [instance: 7ec2497a-7329-4ed4-99ab-dec3d1e8117c] self.wait() [ 732.218482] env[61806]: ERROR nova.compute.manager [instance: 7ec2497a-7329-4ed4-99ab-dec3d1e8117c] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 732.218482] env[61806]: ERROR nova.compute.manager [instance: 7ec2497a-7329-4ed4-99ab-dec3d1e8117c] self[:] = self._gt.wait() [ 732.218482] env[61806]: ERROR nova.compute.manager [instance: 7ec2497a-7329-4ed4-99ab-dec3d1e8117c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 732.218482] env[61806]: ERROR nova.compute.manager [instance: 7ec2497a-7329-4ed4-99ab-dec3d1e8117c] return self._exit_event.wait() [ 732.218482] env[61806]: ERROR nova.compute.manager [instance: 7ec2497a-7329-4ed4-99ab-dec3d1e8117c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 732.218482] env[61806]: ERROR nova.compute.manager [instance: 7ec2497a-7329-4ed4-99ab-dec3d1e8117c] current.throw(*self._exc) [ 732.218482] env[61806]: ERROR nova.compute.manager [instance: 7ec2497a-7329-4ed4-99ab-dec3d1e8117c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 732.218482] env[61806]: ERROR nova.compute.manager [instance: 7ec2497a-7329-4ed4-99ab-dec3d1e8117c] result = function(*args, **kwargs) [ 732.218845] env[61806]: ERROR nova.compute.manager [instance: 7ec2497a-7329-4ed4-99ab-dec3d1e8117c] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 732.218845] env[61806]: ERROR nova.compute.manager [instance: 7ec2497a-7329-4ed4-99ab-dec3d1e8117c] return func(*args, **kwargs) [ 732.218845] env[61806]: ERROR nova.compute.manager [instance: 7ec2497a-7329-4ed4-99ab-dec3d1e8117c] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 732.218845] env[61806]: ERROR nova.compute.manager [instance: 7ec2497a-7329-4ed4-99ab-dec3d1e8117c] raise e [ 732.218845] env[61806]: ERROR nova.compute.manager [instance: 7ec2497a-7329-4ed4-99ab-dec3d1e8117c] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 732.218845] env[61806]: ERROR nova.compute.manager [instance: 7ec2497a-7329-4ed4-99ab-dec3d1e8117c] nwinfo = self.network_api.allocate_for_instance( [ 732.218845] env[61806]: ERROR nova.compute.manager [instance: 7ec2497a-7329-4ed4-99ab-dec3d1e8117c] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 732.218845] env[61806]: ERROR nova.compute.manager [instance: 7ec2497a-7329-4ed4-99ab-dec3d1e8117c] created_port_ids = self._update_ports_for_instance( [ 732.218845] env[61806]: ERROR nova.compute.manager [instance: 7ec2497a-7329-4ed4-99ab-dec3d1e8117c] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 732.218845] env[61806]: ERROR nova.compute.manager [instance: 7ec2497a-7329-4ed4-99ab-dec3d1e8117c] with excutils.save_and_reraise_exception(): [ 732.218845] env[61806]: ERROR nova.compute.manager [instance: 7ec2497a-7329-4ed4-99ab-dec3d1e8117c] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 732.218845] env[61806]: ERROR nova.compute.manager [instance: 7ec2497a-7329-4ed4-99ab-dec3d1e8117c] self.force_reraise() [ 732.218845] env[61806]: ERROR nova.compute.manager [instance: 7ec2497a-7329-4ed4-99ab-dec3d1e8117c] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 732.219217] env[61806]: ERROR nova.compute.manager [instance: 7ec2497a-7329-4ed4-99ab-dec3d1e8117c] raise self.value [ 732.219217] env[61806]: ERROR nova.compute.manager [instance: 7ec2497a-7329-4ed4-99ab-dec3d1e8117c] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 732.219217] env[61806]: ERROR nova.compute.manager [instance: 7ec2497a-7329-4ed4-99ab-dec3d1e8117c] updated_port = self._update_port( [ 732.219217] env[61806]: ERROR nova.compute.manager [instance: 7ec2497a-7329-4ed4-99ab-dec3d1e8117c] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 732.219217] env[61806]: ERROR nova.compute.manager [instance: 7ec2497a-7329-4ed4-99ab-dec3d1e8117c] _ensure_no_port_binding_failure(port) [ 732.219217] env[61806]: ERROR nova.compute.manager [instance: 7ec2497a-7329-4ed4-99ab-dec3d1e8117c] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 732.219217] env[61806]: ERROR nova.compute.manager [instance: 7ec2497a-7329-4ed4-99ab-dec3d1e8117c] raise exception.PortBindingFailed(port_id=port['id']) [ 732.219217] env[61806]: ERROR nova.compute.manager [instance: 7ec2497a-7329-4ed4-99ab-dec3d1e8117c] nova.exception.PortBindingFailed: Binding failed for port 81f90dee-a680-4d31-9c3d-fbe42fa742a4, please check neutron logs for more information. [ 732.219217] env[61806]: ERROR nova.compute.manager [instance: 7ec2497a-7329-4ed4-99ab-dec3d1e8117c] [ 732.219493] env[61806]: DEBUG nova.compute.utils [None req-fcbc4cf1-fc11-42b1-9d49-ef556b52782e tempest-ServerRescueNegativeTestJSON-1483258717 tempest-ServerRescueNegativeTestJSON-1483258717-project-member] [instance: 7ec2497a-7329-4ed4-99ab-dec3d1e8117c] Binding failed for port 81f90dee-a680-4d31-9c3d-fbe42fa742a4, please check neutron logs for more information. {{(pid=61806) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 732.222839] env[61806]: DEBUG oslo_concurrency.lockutils [None req-d17e7876-7c69-44af-acb2-32ed225ce945 tempest-ServerActionsTestOtherA-1637843849 tempest-ServerActionsTestOtherA-1637843849-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 16.923s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 732.229702] env[61806]: INFO nova.compute.claims [None req-d17e7876-7c69-44af-acb2-32ed225ce945 tempest-ServerActionsTestOtherA-1637843849 tempest-ServerActionsTestOtherA-1637843849-project-member] [instance: 15476ac0-289a-4e04-aa9d-4244c658e962] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 732.237224] env[61806]: DEBUG nova.compute.manager [None req-fcbc4cf1-fc11-42b1-9d49-ef556b52782e tempest-ServerRescueNegativeTestJSON-1483258717 tempest-ServerRescueNegativeTestJSON-1483258717-project-member] [instance: 7ec2497a-7329-4ed4-99ab-dec3d1e8117c] Build of instance 7ec2497a-7329-4ed4-99ab-dec3d1e8117c was re-scheduled: Binding failed for port 81f90dee-a680-4d31-9c3d-fbe42fa742a4, please check neutron logs for more information. {{(pid=61806) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 732.237313] env[61806]: DEBUG nova.compute.manager [None req-fcbc4cf1-fc11-42b1-9d49-ef556b52782e tempest-ServerRescueNegativeTestJSON-1483258717 tempest-ServerRescueNegativeTestJSON-1483258717-project-member] [instance: 7ec2497a-7329-4ed4-99ab-dec3d1e8117c] Unplugging VIFs for instance {{(pid=61806) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 732.237541] env[61806]: DEBUG oslo_concurrency.lockutils [None req-fcbc4cf1-fc11-42b1-9d49-ef556b52782e tempest-ServerRescueNegativeTestJSON-1483258717 tempest-ServerRescueNegativeTestJSON-1483258717-project-member] Acquiring lock "refresh_cache-7ec2497a-7329-4ed4-99ab-dec3d1e8117c" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 732.237683] env[61806]: DEBUG oslo_concurrency.lockutils [None req-fcbc4cf1-fc11-42b1-9d49-ef556b52782e tempest-ServerRescueNegativeTestJSON-1483258717 tempest-ServerRescueNegativeTestJSON-1483258717-project-member] Acquired lock "refresh_cache-7ec2497a-7329-4ed4-99ab-dec3d1e8117c" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 732.237857] env[61806]: DEBUG nova.network.neutron [None req-fcbc4cf1-fc11-42b1-9d49-ef556b52782e tempest-ServerRescueNegativeTestJSON-1483258717 tempest-ServerRescueNegativeTestJSON-1483258717-project-member] [instance: 7ec2497a-7329-4ed4-99ab-dec3d1e8117c] Building network info cache for instance {{(pid=61806) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 732.248078] env[61806]: DEBUG nova.compute.manager [None req-3c3981d2-fa25-4f4b-87d4-a5411e67199e tempest-MultipleCreateTestJSON-1589015462 tempest-MultipleCreateTestJSON-1589015462-project-member] [instance: 5b22c2fc-525c-481b-b84e-e0a7f68f633e] Start spawning the instance on the hypervisor. {{(pid=61806) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 732.279050] env[61806]: DEBUG nova.virt.hardware [None req-3c3981d2-fa25-4f4b-87d4-a5411e67199e tempest-MultipleCreateTestJSON-1589015462 tempest-MultipleCreateTestJSON-1589015462-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-15T18:05:10Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-15T18:04:52Z,direct_url=,disk_format='vmdk',id=ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='84ba83fce2c349a6988173c5d6fc3b07',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-15T18:04:53Z,virtual_size=,visibility=), allow threads: False {{(pid=61806) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 732.279349] env[61806]: DEBUG nova.virt.hardware [None req-3c3981d2-fa25-4f4b-87d4-a5411e67199e tempest-MultipleCreateTestJSON-1589015462 tempest-MultipleCreateTestJSON-1589015462-project-member] Flavor limits 0:0:0 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 732.279718] env[61806]: DEBUG nova.virt.hardware [None req-3c3981d2-fa25-4f4b-87d4-a5411e67199e tempest-MultipleCreateTestJSON-1589015462 tempest-MultipleCreateTestJSON-1589015462-project-member] Image limits 0:0:0 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 732.279922] env[61806]: DEBUG nova.virt.hardware [None req-3c3981d2-fa25-4f4b-87d4-a5411e67199e tempest-MultipleCreateTestJSON-1589015462 tempest-MultipleCreateTestJSON-1589015462-project-member] Flavor pref 0:0:0 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 732.280291] env[61806]: DEBUG nova.virt.hardware [None req-3c3981d2-fa25-4f4b-87d4-a5411e67199e tempest-MultipleCreateTestJSON-1589015462 tempest-MultipleCreateTestJSON-1589015462-project-member] Image pref 0:0:0 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 732.280550] env[61806]: DEBUG nova.virt.hardware [None req-3c3981d2-fa25-4f4b-87d4-a5411e67199e tempest-MultipleCreateTestJSON-1589015462 tempest-MultipleCreateTestJSON-1589015462-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 732.281085] env[61806]: DEBUG nova.virt.hardware [None req-3c3981d2-fa25-4f4b-87d4-a5411e67199e tempest-MultipleCreateTestJSON-1589015462 tempest-MultipleCreateTestJSON-1589015462-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61806) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 732.282164] env[61806]: DEBUG nova.virt.hardware [None req-3c3981d2-fa25-4f4b-87d4-a5411e67199e tempest-MultipleCreateTestJSON-1589015462 tempest-MultipleCreateTestJSON-1589015462-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61806) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 732.282412] env[61806]: DEBUG nova.virt.hardware [None req-3c3981d2-fa25-4f4b-87d4-a5411e67199e tempest-MultipleCreateTestJSON-1589015462 tempest-MultipleCreateTestJSON-1589015462-project-member] Got 1 possible topologies {{(pid=61806) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 732.282612] env[61806]: DEBUG nova.virt.hardware [None req-3c3981d2-fa25-4f4b-87d4-a5411e67199e tempest-MultipleCreateTestJSON-1589015462 tempest-MultipleCreateTestJSON-1589015462-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61806) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 732.282789] env[61806]: DEBUG nova.virt.hardware [None req-3c3981d2-fa25-4f4b-87d4-a5411e67199e tempest-MultipleCreateTestJSON-1589015462 tempest-MultipleCreateTestJSON-1589015462-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61806) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 732.283690] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f93639d6-dcac-475f-9757-446f5f878407 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 732.295419] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bb2910e1-f329-4974-b398-6f5632c6fbc5 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 732.394525] env[61806]: DEBUG nova.compute.manager [req-a758ca59-8c70-40f4-976e-081256593e0a req-a005b7ae-2da4-459c-8da6-7b05b0b9b6c4 service nova] [instance: 5b22c2fc-525c-481b-b84e-e0a7f68f633e] Received event network-changed-6cacc512-7619-41e7-8fcb-b88df2c7cf8b {{(pid=61806) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 732.394663] env[61806]: DEBUG nova.compute.manager [req-a758ca59-8c70-40f4-976e-081256593e0a req-a005b7ae-2da4-459c-8da6-7b05b0b9b6c4 service nova] [instance: 5b22c2fc-525c-481b-b84e-e0a7f68f633e] Refreshing instance network info cache due to event network-changed-6cacc512-7619-41e7-8fcb-b88df2c7cf8b. {{(pid=61806) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 732.394899] env[61806]: DEBUG oslo_concurrency.lockutils [req-a758ca59-8c70-40f4-976e-081256593e0a req-a005b7ae-2da4-459c-8da6-7b05b0b9b6c4 service nova] Acquiring lock "refresh_cache-5b22c2fc-525c-481b-b84e-e0a7f68f633e" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 732.395199] env[61806]: DEBUG oslo_concurrency.lockutils [req-a758ca59-8c70-40f4-976e-081256593e0a req-a005b7ae-2da4-459c-8da6-7b05b0b9b6c4 service nova] Acquired lock "refresh_cache-5b22c2fc-525c-481b-b84e-e0a7f68f633e" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 732.395405] env[61806]: DEBUG nova.network.neutron [req-a758ca59-8c70-40f4-976e-081256593e0a req-a005b7ae-2da4-459c-8da6-7b05b0b9b6c4 service nova] [instance: 5b22c2fc-525c-481b-b84e-e0a7f68f633e] Refreshing network info cache for port 6cacc512-7619-41e7-8fcb-b88df2c7cf8b {{(pid=61806) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 732.494399] env[61806]: DEBUG oslo_concurrency.lockutils [None req-f0377b51-dab9-4f68-8592-ca0f31b4ba51 tempest-FloatingIPsAssociationNegativeTestJSON-30494360 tempest-FloatingIPsAssociationNegativeTestJSON-30494360-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 732.668394] env[61806]: ERROR nova.compute.manager [None req-3c3981d2-fa25-4f4b-87d4-a5411e67199e tempest-MultipleCreateTestJSON-1589015462 tempest-MultipleCreateTestJSON-1589015462-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 6cacc512-7619-41e7-8fcb-b88df2c7cf8b, please check neutron logs for more information. [ 732.668394] env[61806]: ERROR nova.compute.manager Traceback (most recent call last): [ 732.668394] env[61806]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 732.668394] env[61806]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 732.668394] env[61806]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 732.668394] env[61806]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 732.668394] env[61806]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 732.668394] env[61806]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 732.668394] env[61806]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 732.668394] env[61806]: ERROR nova.compute.manager self.force_reraise() [ 732.668394] env[61806]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 732.668394] env[61806]: ERROR nova.compute.manager raise self.value [ 732.668394] env[61806]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 732.668394] env[61806]: ERROR nova.compute.manager updated_port = self._update_port( [ 732.668394] env[61806]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 732.668394] env[61806]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 732.669013] env[61806]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 732.669013] env[61806]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 732.669013] env[61806]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 6cacc512-7619-41e7-8fcb-b88df2c7cf8b, please check neutron logs for more information. [ 732.669013] env[61806]: ERROR nova.compute.manager [ 732.669013] env[61806]: Traceback (most recent call last): [ 732.669013] env[61806]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 732.669013] env[61806]: listener.cb(fileno) [ 732.669013] env[61806]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 732.669013] env[61806]: result = function(*args, **kwargs) [ 732.669013] env[61806]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 732.669013] env[61806]: return func(*args, **kwargs) [ 732.669013] env[61806]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 732.669013] env[61806]: raise e [ 732.669013] env[61806]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 732.669013] env[61806]: nwinfo = self.network_api.allocate_for_instance( [ 732.669013] env[61806]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 732.669013] env[61806]: created_port_ids = self._update_ports_for_instance( [ 732.669013] env[61806]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 732.669013] env[61806]: with excutils.save_and_reraise_exception(): [ 732.669013] env[61806]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 732.669013] env[61806]: self.force_reraise() [ 732.669013] env[61806]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 732.669013] env[61806]: raise self.value [ 732.669013] env[61806]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 732.669013] env[61806]: updated_port = self._update_port( [ 732.669013] env[61806]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 732.669013] env[61806]: _ensure_no_port_binding_failure(port) [ 732.669013] env[61806]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 732.669013] env[61806]: raise exception.PortBindingFailed(port_id=port['id']) [ 732.670041] env[61806]: nova.exception.PortBindingFailed: Binding failed for port 6cacc512-7619-41e7-8fcb-b88df2c7cf8b, please check neutron logs for more information. [ 732.670041] env[61806]: Removing descriptor: 14 [ 732.670041] env[61806]: ERROR nova.compute.manager [None req-3c3981d2-fa25-4f4b-87d4-a5411e67199e tempest-MultipleCreateTestJSON-1589015462 tempest-MultipleCreateTestJSON-1589015462-project-member] [instance: 5b22c2fc-525c-481b-b84e-e0a7f68f633e] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 6cacc512-7619-41e7-8fcb-b88df2c7cf8b, please check neutron logs for more information. [ 732.670041] env[61806]: ERROR nova.compute.manager [instance: 5b22c2fc-525c-481b-b84e-e0a7f68f633e] Traceback (most recent call last): [ 732.670041] env[61806]: ERROR nova.compute.manager [instance: 5b22c2fc-525c-481b-b84e-e0a7f68f633e] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 732.670041] env[61806]: ERROR nova.compute.manager [instance: 5b22c2fc-525c-481b-b84e-e0a7f68f633e] yield resources [ 732.670041] env[61806]: ERROR nova.compute.manager [instance: 5b22c2fc-525c-481b-b84e-e0a7f68f633e] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 732.670041] env[61806]: ERROR nova.compute.manager [instance: 5b22c2fc-525c-481b-b84e-e0a7f68f633e] self.driver.spawn(context, instance, image_meta, [ 732.670041] env[61806]: ERROR nova.compute.manager [instance: 5b22c2fc-525c-481b-b84e-e0a7f68f633e] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 732.670041] env[61806]: ERROR nova.compute.manager [instance: 5b22c2fc-525c-481b-b84e-e0a7f68f633e] self._vmops.spawn(context, instance, image_meta, injected_files, [ 732.670041] env[61806]: ERROR nova.compute.manager [instance: 5b22c2fc-525c-481b-b84e-e0a7f68f633e] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 732.670041] env[61806]: ERROR nova.compute.manager [instance: 5b22c2fc-525c-481b-b84e-e0a7f68f633e] vm_ref = self.build_virtual_machine(instance, [ 732.670479] env[61806]: ERROR nova.compute.manager [instance: 5b22c2fc-525c-481b-b84e-e0a7f68f633e] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 732.670479] env[61806]: ERROR nova.compute.manager [instance: 5b22c2fc-525c-481b-b84e-e0a7f68f633e] vif_infos = vmwarevif.get_vif_info(self._session, [ 732.670479] env[61806]: ERROR nova.compute.manager [instance: 5b22c2fc-525c-481b-b84e-e0a7f68f633e] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 732.670479] env[61806]: ERROR nova.compute.manager [instance: 5b22c2fc-525c-481b-b84e-e0a7f68f633e] for vif in network_info: [ 732.670479] env[61806]: ERROR nova.compute.manager [instance: 5b22c2fc-525c-481b-b84e-e0a7f68f633e] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 732.670479] env[61806]: ERROR nova.compute.manager [instance: 5b22c2fc-525c-481b-b84e-e0a7f68f633e] return self._sync_wrapper(fn, *args, **kwargs) [ 732.670479] env[61806]: ERROR nova.compute.manager [instance: 5b22c2fc-525c-481b-b84e-e0a7f68f633e] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 732.670479] env[61806]: ERROR nova.compute.manager [instance: 5b22c2fc-525c-481b-b84e-e0a7f68f633e] self.wait() [ 732.670479] env[61806]: ERROR nova.compute.manager [instance: 5b22c2fc-525c-481b-b84e-e0a7f68f633e] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 732.670479] env[61806]: ERROR nova.compute.manager [instance: 5b22c2fc-525c-481b-b84e-e0a7f68f633e] self[:] = self._gt.wait() [ 732.670479] env[61806]: ERROR nova.compute.manager [instance: 5b22c2fc-525c-481b-b84e-e0a7f68f633e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 732.670479] env[61806]: ERROR nova.compute.manager [instance: 5b22c2fc-525c-481b-b84e-e0a7f68f633e] return self._exit_event.wait() [ 732.670479] env[61806]: ERROR nova.compute.manager [instance: 5b22c2fc-525c-481b-b84e-e0a7f68f633e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 732.670914] env[61806]: ERROR nova.compute.manager [instance: 5b22c2fc-525c-481b-b84e-e0a7f68f633e] result = hub.switch() [ 732.670914] env[61806]: ERROR nova.compute.manager [instance: 5b22c2fc-525c-481b-b84e-e0a7f68f633e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 732.670914] env[61806]: ERROR nova.compute.manager [instance: 5b22c2fc-525c-481b-b84e-e0a7f68f633e] return self.greenlet.switch() [ 732.670914] env[61806]: ERROR nova.compute.manager [instance: 5b22c2fc-525c-481b-b84e-e0a7f68f633e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 732.670914] env[61806]: ERROR nova.compute.manager [instance: 5b22c2fc-525c-481b-b84e-e0a7f68f633e] result = function(*args, **kwargs) [ 732.670914] env[61806]: ERROR nova.compute.manager [instance: 5b22c2fc-525c-481b-b84e-e0a7f68f633e] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 732.670914] env[61806]: ERROR nova.compute.manager [instance: 5b22c2fc-525c-481b-b84e-e0a7f68f633e] return func(*args, **kwargs) [ 732.670914] env[61806]: ERROR nova.compute.manager [instance: 5b22c2fc-525c-481b-b84e-e0a7f68f633e] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 732.670914] env[61806]: ERROR nova.compute.manager [instance: 5b22c2fc-525c-481b-b84e-e0a7f68f633e] raise e [ 732.670914] env[61806]: ERROR nova.compute.manager [instance: 5b22c2fc-525c-481b-b84e-e0a7f68f633e] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 732.670914] env[61806]: ERROR nova.compute.manager [instance: 5b22c2fc-525c-481b-b84e-e0a7f68f633e] nwinfo = self.network_api.allocate_for_instance( [ 732.670914] env[61806]: ERROR nova.compute.manager [instance: 5b22c2fc-525c-481b-b84e-e0a7f68f633e] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 732.670914] env[61806]: ERROR nova.compute.manager [instance: 5b22c2fc-525c-481b-b84e-e0a7f68f633e] created_port_ids = self._update_ports_for_instance( [ 732.671372] env[61806]: ERROR nova.compute.manager [instance: 5b22c2fc-525c-481b-b84e-e0a7f68f633e] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 732.671372] env[61806]: ERROR nova.compute.manager [instance: 5b22c2fc-525c-481b-b84e-e0a7f68f633e] with excutils.save_and_reraise_exception(): [ 732.671372] env[61806]: ERROR nova.compute.manager [instance: 5b22c2fc-525c-481b-b84e-e0a7f68f633e] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 732.671372] env[61806]: ERROR nova.compute.manager [instance: 5b22c2fc-525c-481b-b84e-e0a7f68f633e] self.force_reraise() [ 732.671372] env[61806]: ERROR nova.compute.manager [instance: 5b22c2fc-525c-481b-b84e-e0a7f68f633e] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 732.671372] env[61806]: ERROR nova.compute.manager [instance: 5b22c2fc-525c-481b-b84e-e0a7f68f633e] raise self.value [ 732.671372] env[61806]: ERROR nova.compute.manager [instance: 5b22c2fc-525c-481b-b84e-e0a7f68f633e] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 732.671372] env[61806]: ERROR nova.compute.manager [instance: 5b22c2fc-525c-481b-b84e-e0a7f68f633e] updated_port = self._update_port( [ 732.671372] env[61806]: ERROR nova.compute.manager [instance: 5b22c2fc-525c-481b-b84e-e0a7f68f633e] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 732.671372] env[61806]: ERROR nova.compute.manager [instance: 5b22c2fc-525c-481b-b84e-e0a7f68f633e] _ensure_no_port_binding_failure(port) [ 732.671372] env[61806]: ERROR nova.compute.manager [instance: 5b22c2fc-525c-481b-b84e-e0a7f68f633e] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 732.671372] env[61806]: ERROR nova.compute.manager [instance: 5b22c2fc-525c-481b-b84e-e0a7f68f633e] raise exception.PortBindingFailed(port_id=port['id']) [ 732.671820] env[61806]: ERROR nova.compute.manager [instance: 5b22c2fc-525c-481b-b84e-e0a7f68f633e] nova.exception.PortBindingFailed: Binding failed for port 6cacc512-7619-41e7-8fcb-b88df2c7cf8b, please check neutron logs for more information. [ 732.671820] env[61806]: ERROR nova.compute.manager [instance: 5b22c2fc-525c-481b-b84e-e0a7f68f633e] [ 732.671820] env[61806]: INFO nova.compute.manager [None req-3c3981d2-fa25-4f4b-87d4-a5411e67199e tempest-MultipleCreateTestJSON-1589015462 tempest-MultipleCreateTestJSON-1589015462-project-member] [instance: 5b22c2fc-525c-481b-b84e-e0a7f68f633e] Terminating instance [ 732.672135] env[61806]: DEBUG oslo_concurrency.lockutils [None req-3c3981d2-fa25-4f4b-87d4-a5411e67199e tempest-MultipleCreateTestJSON-1589015462 tempest-MultipleCreateTestJSON-1589015462-project-member] Acquiring lock "refresh_cache-5b22c2fc-525c-481b-b84e-e0a7f68f633e" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 732.766952] env[61806]: DEBUG nova.network.neutron [None req-fcbc4cf1-fc11-42b1-9d49-ef556b52782e tempest-ServerRescueNegativeTestJSON-1483258717 tempest-ServerRescueNegativeTestJSON-1483258717-project-member] [instance: 7ec2497a-7329-4ed4-99ab-dec3d1e8117c] Instance cache missing network info. {{(pid=61806) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 732.862134] env[61806]: DEBUG nova.network.neutron [None req-fcbc4cf1-fc11-42b1-9d49-ef556b52782e tempest-ServerRescueNegativeTestJSON-1483258717 tempest-ServerRescueNegativeTestJSON-1483258717-project-member] [instance: 7ec2497a-7329-4ed4-99ab-dec3d1e8117c] Updating instance_info_cache with network_info: [] {{(pid=61806) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 732.918685] env[61806]: DEBUG nova.network.neutron [req-a758ca59-8c70-40f4-976e-081256593e0a req-a005b7ae-2da4-459c-8da6-7b05b0b9b6c4 service nova] [instance: 5b22c2fc-525c-481b-b84e-e0a7f68f633e] Instance cache missing network info. {{(pid=61806) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 733.021553] env[61806]: DEBUG nova.network.neutron [req-a758ca59-8c70-40f4-976e-081256593e0a req-a005b7ae-2da4-459c-8da6-7b05b0b9b6c4 service nova] [instance: 5b22c2fc-525c-481b-b84e-e0a7f68f633e] Updating instance_info_cache with network_info: [] {{(pid=61806) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 733.364841] env[61806]: DEBUG oslo_concurrency.lockutils [None req-fcbc4cf1-fc11-42b1-9d49-ef556b52782e tempest-ServerRescueNegativeTestJSON-1483258717 tempest-ServerRescueNegativeTestJSON-1483258717-project-member] Releasing lock "refresh_cache-7ec2497a-7329-4ed4-99ab-dec3d1e8117c" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 733.365283] env[61806]: DEBUG nova.compute.manager [None req-fcbc4cf1-fc11-42b1-9d49-ef556b52782e tempest-ServerRescueNegativeTestJSON-1483258717 tempest-ServerRescueNegativeTestJSON-1483258717-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61806) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 733.365283] env[61806]: DEBUG nova.compute.manager [None req-fcbc4cf1-fc11-42b1-9d49-ef556b52782e tempest-ServerRescueNegativeTestJSON-1483258717 tempest-ServerRescueNegativeTestJSON-1483258717-project-member] [instance: 7ec2497a-7329-4ed4-99ab-dec3d1e8117c] Deallocating network for instance {{(pid=61806) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 733.365512] env[61806]: DEBUG nova.network.neutron [None req-fcbc4cf1-fc11-42b1-9d49-ef556b52782e tempest-ServerRescueNegativeTestJSON-1483258717 tempest-ServerRescueNegativeTestJSON-1483258717-project-member] [instance: 7ec2497a-7329-4ed4-99ab-dec3d1e8117c] deallocate_for_instance() {{(pid=61806) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 733.386919] env[61806]: DEBUG nova.network.neutron [None req-fcbc4cf1-fc11-42b1-9d49-ef556b52782e tempest-ServerRescueNegativeTestJSON-1483258717 tempest-ServerRescueNegativeTestJSON-1483258717-project-member] [instance: 7ec2497a-7329-4ed4-99ab-dec3d1e8117c] Instance cache missing network info. {{(pid=61806) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 733.525274] env[61806]: DEBUG oslo_concurrency.lockutils [req-a758ca59-8c70-40f4-976e-081256593e0a req-a005b7ae-2da4-459c-8da6-7b05b0b9b6c4 service nova] Releasing lock "refresh_cache-5b22c2fc-525c-481b-b84e-e0a7f68f633e" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 733.525274] env[61806]: DEBUG oslo_concurrency.lockutils [None req-3c3981d2-fa25-4f4b-87d4-a5411e67199e tempest-MultipleCreateTestJSON-1589015462 tempest-MultipleCreateTestJSON-1589015462-project-member] Acquired lock "refresh_cache-5b22c2fc-525c-481b-b84e-e0a7f68f633e" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 733.525428] env[61806]: DEBUG nova.network.neutron [None req-3c3981d2-fa25-4f4b-87d4-a5411e67199e tempest-MultipleCreateTestJSON-1589015462 tempest-MultipleCreateTestJSON-1589015462-project-member] [instance: 5b22c2fc-525c-481b-b84e-e0a7f68f633e] Building network info cache for instance {{(pid=61806) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 733.590854] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-91059f45-9bf6-472e-bf2f-98283ba0745b {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 733.598852] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-89d827ac-80a5-4b08-81b7-031856ce046c {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 733.636454] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-59f7ea72-49c9-4762-812c-d2d723d0053a {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 733.644574] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cdf1cf72-8317-45de-b72b-2d16ed1104d2 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 733.658417] env[61806]: DEBUG nova.compute.provider_tree [None req-d17e7876-7c69-44af-acb2-32ed225ce945 tempest-ServerActionsTestOtherA-1637843849 tempest-ServerActionsTestOtherA-1637843849-project-member] Inventory has not changed in ProviderTree for provider: a2858be1-fd22-4e08-979e-87ad25293407 {{(pid=61806) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 733.890938] env[61806]: DEBUG nova.network.neutron [None req-fcbc4cf1-fc11-42b1-9d49-ef556b52782e tempest-ServerRescueNegativeTestJSON-1483258717 tempest-ServerRescueNegativeTestJSON-1483258717-project-member] [instance: 7ec2497a-7329-4ed4-99ab-dec3d1e8117c] Updating instance_info_cache with network_info: [] {{(pid=61806) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 734.050255] env[61806]: DEBUG nova.network.neutron [None req-3c3981d2-fa25-4f4b-87d4-a5411e67199e tempest-MultipleCreateTestJSON-1589015462 tempest-MultipleCreateTestJSON-1589015462-project-member] [instance: 5b22c2fc-525c-481b-b84e-e0a7f68f633e] Instance cache missing network info. {{(pid=61806) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 734.126544] env[61806]: DEBUG oslo_service.periodic_task [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=61806) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 734.126730] env[61806]: DEBUG oslo_service.periodic_task [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Running periodic task ComputeManager._sync_scheduler_instance_info {{(pid=61806) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 734.164148] env[61806]: DEBUG nova.scheduler.client.report [None req-d17e7876-7c69-44af-acb2-32ed225ce945 tempest-ServerActionsTestOtherA-1637843849 tempest-ServerActionsTestOtherA-1637843849-project-member] Inventory has not changed for provider a2858be1-fd22-4e08-979e-87ad25293407 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 139, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61806) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 734.201731] env[61806]: DEBUG nova.network.neutron [None req-3c3981d2-fa25-4f4b-87d4-a5411e67199e tempest-MultipleCreateTestJSON-1589015462 tempest-MultipleCreateTestJSON-1589015462-project-member] [instance: 5b22c2fc-525c-481b-b84e-e0a7f68f633e] Updating instance_info_cache with network_info: [] {{(pid=61806) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 734.393455] env[61806]: INFO nova.compute.manager [None req-fcbc4cf1-fc11-42b1-9d49-ef556b52782e tempest-ServerRescueNegativeTestJSON-1483258717 tempest-ServerRescueNegativeTestJSON-1483258717-project-member] [instance: 7ec2497a-7329-4ed4-99ab-dec3d1e8117c] Took 1.03 seconds to deallocate network for instance. [ 734.427322] env[61806]: DEBUG nova.compute.manager [req-7e519233-512a-40ed-8ab5-746873f19b4d req-1a84b591-42f0-4843-ada0-da41d0f29012 service nova] [instance: 5b22c2fc-525c-481b-b84e-e0a7f68f633e] Received event network-vif-deleted-6cacc512-7619-41e7-8fcb-b88df2c7cf8b {{(pid=61806) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 734.638657] env[61806]: DEBUG oslo_service.periodic_task [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=61806) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 734.639109] env[61806]: DEBUG nova.compute.manager [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Starting heal instance info cache {{(pid=61806) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9926}} [ 734.639109] env[61806]: DEBUG nova.compute.manager [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Rebuilding the list of instances to heal {{(pid=61806) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9930}} [ 734.669057] env[61806]: DEBUG oslo_concurrency.lockutils [None req-d17e7876-7c69-44af-acb2-32ed225ce945 tempest-ServerActionsTestOtherA-1637843849 tempest-ServerActionsTestOtherA-1637843849-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.446s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 734.669848] env[61806]: DEBUG nova.compute.manager [None req-d17e7876-7c69-44af-acb2-32ed225ce945 tempest-ServerActionsTestOtherA-1637843849 tempest-ServerActionsTestOtherA-1637843849-project-member] [instance: 15476ac0-289a-4e04-aa9d-4244c658e962] Start building networks asynchronously for instance. {{(pid=61806) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 734.673969] env[61806]: DEBUG oslo_concurrency.lockutils [None req-6b55cbbf-00f7-4d4d-87f3-a59cc1c553b0 tempest-ServerRescueNegativeTestJSON-1483258717 tempest-ServerRescueNegativeTestJSON-1483258717-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 16.127s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 734.704577] env[61806]: DEBUG oslo_concurrency.lockutils [None req-3c3981d2-fa25-4f4b-87d4-a5411e67199e tempest-MultipleCreateTestJSON-1589015462 tempest-MultipleCreateTestJSON-1589015462-project-member] Releasing lock "refresh_cache-5b22c2fc-525c-481b-b84e-e0a7f68f633e" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 734.704944] env[61806]: DEBUG nova.compute.manager [None req-3c3981d2-fa25-4f4b-87d4-a5411e67199e tempest-MultipleCreateTestJSON-1589015462 tempest-MultipleCreateTestJSON-1589015462-project-member] [instance: 5b22c2fc-525c-481b-b84e-e0a7f68f633e] Start destroying the instance on the hypervisor. {{(pid=61806) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 734.705450] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-3c3981d2-fa25-4f4b-87d4-a5411e67199e tempest-MultipleCreateTestJSON-1589015462 tempest-MultipleCreateTestJSON-1589015462-project-member] [instance: 5b22c2fc-525c-481b-b84e-e0a7f68f633e] Destroying instance {{(pid=61806) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 734.706094] env[61806]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-0491db34-41c6-4f13-863c-74cb646d60f1 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 734.715698] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ee8b370e-17fc-44aa-be29-51dfc764fa9a {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 734.739883] env[61806]: WARNING nova.virt.vmwareapi.vmops [None req-3c3981d2-fa25-4f4b-87d4-a5411e67199e tempest-MultipleCreateTestJSON-1589015462 tempest-MultipleCreateTestJSON-1589015462-project-member] [instance: 5b22c2fc-525c-481b-b84e-e0a7f68f633e] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 5b22c2fc-525c-481b-b84e-e0a7f68f633e could not be found. [ 734.739883] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-3c3981d2-fa25-4f4b-87d4-a5411e67199e tempest-MultipleCreateTestJSON-1589015462 tempest-MultipleCreateTestJSON-1589015462-project-member] [instance: 5b22c2fc-525c-481b-b84e-e0a7f68f633e] Instance destroyed {{(pid=61806) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 734.739883] env[61806]: INFO nova.compute.manager [None req-3c3981d2-fa25-4f4b-87d4-a5411e67199e tempest-MultipleCreateTestJSON-1589015462 tempest-MultipleCreateTestJSON-1589015462-project-member] [instance: 5b22c2fc-525c-481b-b84e-e0a7f68f633e] Took 0.03 seconds to destroy the instance on the hypervisor. [ 734.739883] env[61806]: DEBUG oslo.service.loopingcall [None req-3c3981d2-fa25-4f4b-87d4-a5411e67199e tempest-MultipleCreateTestJSON-1589015462 tempest-MultipleCreateTestJSON-1589015462-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61806) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 734.739883] env[61806]: DEBUG nova.compute.manager [-] [instance: 5b22c2fc-525c-481b-b84e-e0a7f68f633e] Deallocating network for instance {{(pid=61806) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 734.739883] env[61806]: DEBUG nova.network.neutron [-] [instance: 5b22c2fc-525c-481b-b84e-e0a7f68f633e] deallocate_for_instance() {{(pid=61806) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 734.756818] env[61806]: DEBUG nova.network.neutron [-] [instance: 5b22c2fc-525c-481b-b84e-e0a7f68f633e] Instance cache missing network info. {{(pid=61806) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 734.792772] env[61806]: DEBUG oslo_concurrency.lockutils [None req-9c52c9e9-816e-4638-b13a-a63a9756140c tempest-SecurityGroupsTestJSON-1213485581 tempest-SecurityGroupsTestJSON-1213485581-project-member] Acquiring lock "b41531f2-f28c-4d82-9682-0b557bbaa491" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 734.793007] env[61806]: DEBUG oslo_concurrency.lockutils [None req-9c52c9e9-816e-4638-b13a-a63a9756140c tempest-SecurityGroupsTestJSON-1213485581 tempest-SecurityGroupsTestJSON-1213485581-project-member] Lock "b41531f2-f28c-4d82-9682-0b557bbaa491" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 735.144304] env[61806]: DEBUG nova.compute.manager [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] [instance: 7ea919b6-12f7-4f4e-a123-413044bbbffc] Skipping network cache update for instance because it is Building. {{(pid=61806) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9939}} [ 735.144594] env[61806]: DEBUG nova.compute.manager [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] [instance: 7211d568-cc68-4e50-80ad-b78878d1deab] Skipping network cache update for instance because it is Building. {{(pid=61806) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9939}} [ 735.144646] env[61806]: DEBUG nova.compute.manager [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] [instance: 5b22c2fc-525c-481b-b84e-e0a7f68f633e] Skipping network cache update for instance because it is Building. {{(pid=61806) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9939}} [ 735.144754] env[61806]: DEBUG nova.compute.manager [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] [instance: 15476ac0-289a-4e04-aa9d-4244c658e962] Skipping network cache update for instance because it is Building. {{(pid=61806) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9939}} [ 735.144882] env[61806]: DEBUG nova.compute.manager [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Didn't find any instances for network info cache update. {{(pid=61806) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10012}} [ 735.145141] env[61806]: DEBUG oslo_service.periodic_task [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=61806) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 735.145246] env[61806]: DEBUG oslo_service.periodic_task [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=61806) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 735.145392] env[61806]: DEBUG oslo_service.periodic_task [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=61806) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 735.145541] env[61806]: DEBUG oslo_service.periodic_task [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=61806) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 735.145694] env[61806]: DEBUG oslo_service.periodic_task [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=61806) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 735.145839] env[61806]: DEBUG oslo_service.periodic_task [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=61806) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 735.145965] env[61806]: DEBUG nova.compute.manager [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=61806) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10545}} [ 735.146119] env[61806]: DEBUG oslo_service.periodic_task [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Running periodic task ComputeManager.update_available_resource {{(pid=61806) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 735.183201] env[61806]: DEBUG nova.compute.utils [None req-d17e7876-7c69-44af-acb2-32ed225ce945 tempest-ServerActionsTestOtherA-1637843849 tempest-ServerActionsTestOtherA-1637843849-project-member] Using /dev/sd instead of None {{(pid=61806) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 735.185665] env[61806]: DEBUG nova.compute.manager [None req-d17e7876-7c69-44af-acb2-32ed225ce945 tempest-ServerActionsTestOtherA-1637843849 tempest-ServerActionsTestOtherA-1637843849-project-member] [instance: 15476ac0-289a-4e04-aa9d-4244c658e962] Allocating IP information in the background. {{(pid=61806) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 735.185854] env[61806]: DEBUG nova.network.neutron [None req-d17e7876-7c69-44af-acb2-32ed225ce945 tempest-ServerActionsTestOtherA-1637843849 tempest-ServerActionsTestOtherA-1637843849-project-member] [instance: 15476ac0-289a-4e04-aa9d-4244c658e962] allocate_for_instance() {{(pid=61806) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 735.261261] env[61806]: DEBUG nova.network.neutron [-] [instance: 5b22c2fc-525c-481b-b84e-e0a7f68f633e] Updating instance_info_cache with network_info: [] {{(pid=61806) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 735.271124] env[61806]: DEBUG nova.policy [None req-d17e7876-7c69-44af-acb2-32ed225ce945 tempest-ServerActionsTestOtherA-1637843849 tempest-ServerActionsTestOtherA-1637843849-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '76e14242bd514e92a8f639a40b31e0cd', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '2dee81e76857407f94be1a77a95641be', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61806) authorize /opt/stack/nova/nova/policy.py:201}} [ 735.424956] env[61806]: INFO nova.scheduler.client.report [None req-fcbc4cf1-fc11-42b1-9d49-ef556b52782e tempest-ServerRescueNegativeTestJSON-1483258717 tempest-ServerRescueNegativeTestJSON-1483258717-project-member] Deleted allocations for instance 7ec2497a-7329-4ed4-99ab-dec3d1e8117c [ 735.543950] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8d2f8803-a765-43d5-a932-5a972597d86d {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 735.551946] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-83cde443-7a1f-44a4-bc8d-d5a9b0893629 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 735.589115] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-59a55660-8d8d-434c-80ca-239eb52ef481 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 735.596819] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-04360783-2b47-4eb9-9611-d63d95770bc3 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 735.611752] env[61806]: DEBUG nova.compute.provider_tree [None req-6b55cbbf-00f7-4d4d-87f3-a59cc1c553b0 tempest-ServerRescueNegativeTestJSON-1483258717 tempest-ServerRescueNegativeTestJSON-1483258717-project-member] Inventory has not changed in ProviderTree for provider: a2858be1-fd22-4e08-979e-87ad25293407 {{(pid=61806) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 735.649053] env[61806]: DEBUG oslo_concurrency.lockutils [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 735.692872] env[61806]: DEBUG nova.compute.manager [None req-d17e7876-7c69-44af-acb2-32ed225ce945 tempest-ServerActionsTestOtherA-1637843849 tempest-ServerActionsTestOtherA-1637843849-project-member] [instance: 15476ac0-289a-4e04-aa9d-4244c658e962] Start building block device mappings for instance. {{(pid=61806) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 735.765150] env[61806]: INFO nova.compute.manager [-] [instance: 5b22c2fc-525c-481b-b84e-e0a7f68f633e] Took 1.02 seconds to deallocate network for instance. [ 735.765949] env[61806]: DEBUG nova.compute.claims [None req-3c3981d2-fa25-4f4b-87d4-a5411e67199e tempest-MultipleCreateTestJSON-1589015462 tempest-MultipleCreateTestJSON-1589015462-project-member] [instance: 5b22c2fc-525c-481b-b84e-e0a7f68f633e] Aborting claim: {{(pid=61806) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 735.766274] env[61806]: DEBUG oslo_concurrency.lockutils [None req-3c3981d2-fa25-4f4b-87d4-a5411e67199e tempest-MultipleCreateTestJSON-1589015462 tempest-MultipleCreateTestJSON-1589015462-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 735.814652] env[61806]: DEBUG nova.network.neutron [None req-d17e7876-7c69-44af-acb2-32ed225ce945 tempest-ServerActionsTestOtherA-1637843849 tempest-ServerActionsTestOtherA-1637843849-project-member] [instance: 15476ac0-289a-4e04-aa9d-4244c658e962] Successfully created port: 76b9d660-f9cc-4d3f-ae94-28e2b3d3b46d {{(pid=61806) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 735.936380] env[61806]: DEBUG oslo_concurrency.lockutils [None req-fcbc4cf1-fc11-42b1-9d49-ef556b52782e tempest-ServerRescueNegativeTestJSON-1483258717 tempest-ServerRescueNegativeTestJSON-1483258717-project-member] Lock "7ec2497a-7329-4ed4-99ab-dec3d1e8117c" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 156.296s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 736.114534] env[61806]: DEBUG nova.scheduler.client.report [None req-6b55cbbf-00f7-4d4d-87f3-a59cc1c553b0 tempest-ServerRescueNegativeTestJSON-1483258717 tempest-ServerRescueNegativeTestJSON-1483258717-project-member] Inventory has not changed for provider a2858be1-fd22-4e08-979e-87ad25293407 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 139, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61806) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 736.442147] env[61806]: DEBUG nova.compute.manager [None req-da1d2f3c-f6ea-47f5-8c64-65c86bf2d676 tempest-ServerRescueTestJSON-811948373 tempest-ServerRescueTestJSON-811948373-project-member] [instance: c61df5d3-7f43-48d4-a2c0-a8972372c0ba] Starting instance... {{(pid=61806) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 736.619574] env[61806]: DEBUG oslo_concurrency.lockutils [None req-6b55cbbf-00f7-4d4d-87f3-a59cc1c553b0 tempest-ServerRescueNegativeTestJSON-1483258717 tempest-ServerRescueNegativeTestJSON-1483258717-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.946s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 736.620235] env[61806]: ERROR nova.compute.manager [None req-6b55cbbf-00f7-4d4d-87f3-a59cc1c553b0 tempest-ServerRescueNegativeTestJSON-1483258717 tempest-ServerRescueNegativeTestJSON-1483258717-project-member] [instance: 7ea919b6-12f7-4f4e-a123-413044bbbffc] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 61c8084d-6ad0-442a-967a-46f14ffa6671, please check neutron logs for more information. [ 736.620235] env[61806]: ERROR nova.compute.manager [instance: 7ea919b6-12f7-4f4e-a123-413044bbbffc] Traceback (most recent call last): [ 736.620235] env[61806]: ERROR nova.compute.manager [instance: 7ea919b6-12f7-4f4e-a123-413044bbbffc] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 736.620235] env[61806]: ERROR nova.compute.manager [instance: 7ea919b6-12f7-4f4e-a123-413044bbbffc] self.driver.spawn(context, instance, image_meta, [ 736.620235] env[61806]: ERROR nova.compute.manager [instance: 7ea919b6-12f7-4f4e-a123-413044bbbffc] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 736.620235] env[61806]: ERROR nova.compute.manager [instance: 7ea919b6-12f7-4f4e-a123-413044bbbffc] self._vmops.spawn(context, instance, image_meta, injected_files, [ 736.620235] env[61806]: ERROR nova.compute.manager [instance: 7ea919b6-12f7-4f4e-a123-413044bbbffc] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 736.620235] env[61806]: ERROR nova.compute.manager [instance: 7ea919b6-12f7-4f4e-a123-413044bbbffc] vm_ref = self.build_virtual_machine(instance, [ 736.620235] env[61806]: ERROR nova.compute.manager [instance: 7ea919b6-12f7-4f4e-a123-413044bbbffc] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 736.620235] env[61806]: ERROR nova.compute.manager [instance: 7ea919b6-12f7-4f4e-a123-413044bbbffc] vif_infos = vmwarevif.get_vif_info(self._session, [ 736.620235] env[61806]: ERROR nova.compute.manager [instance: 7ea919b6-12f7-4f4e-a123-413044bbbffc] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 736.620733] env[61806]: ERROR nova.compute.manager [instance: 7ea919b6-12f7-4f4e-a123-413044bbbffc] for vif in network_info: [ 736.620733] env[61806]: ERROR nova.compute.manager [instance: 7ea919b6-12f7-4f4e-a123-413044bbbffc] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 736.620733] env[61806]: ERROR nova.compute.manager [instance: 7ea919b6-12f7-4f4e-a123-413044bbbffc] return self._sync_wrapper(fn, *args, **kwargs) [ 736.620733] env[61806]: ERROR nova.compute.manager [instance: 7ea919b6-12f7-4f4e-a123-413044bbbffc] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 736.620733] env[61806]: ERROR nova.compute.manager [instance: 7ea919b6-12f7-4f4e-a123-413044bbbffc] self.wait() [ 736.620733] env[61806]: ERROR nova.compute.manager [instance: 7ea919b6-12f7-4f4e-a123-413044bbbffc] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 736.620733] env[61806]: ERROR nova.compute.manager [instance: 7ea919b6-12f7-4f4e-a123-413044bbbffc] self[:] = self._gt.wait() [ 736.620733] env[61806]: ERROR nova.compute.manager [instance: 7ea919b6-12f7-4f4e-a123-413044bbbffc] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 736.620733] env[61806]: ERROR nova.compute.manager [instance: 7ea919b6-12f7-4f4e-a123-413044bbbffc] return self._exit_event.wait() [ 736.620733] env[61806]: ERROR nova.compute.manager [instance: 7ea919b6-12f7-4f4e-a123-413044bbbffc] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 736.620733] env[61806]: ERROR nova.compute.manager [instance: 7ea919b6-12f7-4f4e-a123-413044bbbffc] current.throw(*self._exc) [ 736.620733] env[61806]: ERROR nova.compute.manager [instance: 7ea919b6-12f7-4f4e-a123-413044bbbffc] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 736.620733] env[61806]: ERROR nova.compute.manager [instance: 7ea919b6-12f7-4f4e-a123-413044bbbffc] result = function(*args, **kwargs) [ 736.621230] env[61806]: ERROR nova.compute.manager [instance: 7ea919b6-12f7-4f4e-a123-413044bbbffc] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 736.621230] env[61806]: ERROR nova.compute.manager [instance: 7ea919b6-12f7-4f4e-a123-413044bbbffc] return func(*args, **kwargs) [ 736.621230] env[61806]: ERROR nova.compute.manager [instance: 7ea919b6-12f7-4f4e-a123-413044bbbffc] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 736.621230] env[61806]: ERROR nova.compute.manager [instance: 7ea919b6-12f7-4f4e-a123-413044bbbffc] raise e [ 736.621230] env[61806]: ERROR nova.compute.manager [instance: 7ea919b6-12f7-4f4e-a123-413044bbbffc] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 736.621230] env[61806]: ERROR nova.compute.manager [instance: 7ea919b6-12f7-4f4e-a123-413044bbbffc] nwinfo = self.network_api.allocate_for_instance( [ 736.621230] env[61806]: ERROR nova.compute.manager [instance: 7ea919b6-12f7-4f4e-a123-413044bbbffc] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 736.621230] env[61806]: ERROR nova.compute.manager [instance: 7ea919b6-12f7-4f4e-a123-413044bbbffc] created_port_ids = self._update_ports_for_instance( [ 736.621230] env[61806]: ERROR nova.compute.manager [instance: 7ea919b6-12f7-4f4e-a123-413044bbbffc] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 736.621230] env[61806]: ERROR nova.compute.manager [instance: 7ea919b6-12f7-4f4e-a123-413044bbbffc] with excutils.save_and_reraise_exception(): [ 736.621230] env[61806]: ERROR nova.compute.manager [instance: 7ea919b6-12f7-4f4e-a123-413044bbbffc] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 736.621230] env[61806]: ERROR nova.compute.manager [instance: 7ea919b6-12f7-4f4e-a123-413044bbbffc] self.force_reraise() [ 736.621230] env[61806]: ERROR nova.compute.manager [instance: 7ea919b6-12f7-4f4e-a123-413044bbbffc] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 736.621733] env[61806]: ERROR nova.compute.manager [instance: 7ea919b6-12f7-4f4e-a123-413044bbbffc] raise self.value [ 736.621733] env[61806]: ERROR nova.compute.manager [instance: 7ea919b6-12f7-4f4e-a123-413044bbbffc] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 736.621733] env[61806]: ERROR nova.compute.manager [instance: 7ea919b6-12f7-4f4e-a123-413044bbbffc] updated_port = self._update_port( [ 736.621733] env[61806]: ERROR nova.compute.manager [instance: 7ea919b6-12f7-4f4e-a123-413044bbbffc] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 736.621733] env[61806]: ERROR nova.compute.manager [instance: 7ea919b6-12f7-4f4e-a123-413044bbbffc] _ensure_no_port_binding_failure(port) [ 736.621733] env[61806]: ERROR nova.compute.manager [instance: 7ea919b6-12f7-4f4e-a123-413044bbbffc] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 736.621733] env[61806]: ERROR nova.compute.manager [instance: 7ea919b6-12f7-4f4e-a123-413044bbbffc] raise exception.PortBindingFailed(port_id=port['id']) [ 736.621733] env[61806]: ERROR nova.compute.manager [instance: 7ea919b6-12f7-4f4e-a123-413044bbbffc] nova.exception.PortBindingFailed: Binding failed for port 61c8084d-6ad0-442a-967a-46f14ffa6671, please check neutron logs for more information. [ 736.621733] env[61806]: ERROR nova.compute.manager [instance: 7ea919b6-12f7-4f4e-a123-413044bbbffc] [ 736.621733] env[61806]: DEBUG nova.compute.utils [None req-6b55cbbf-00f7-4d4d-87f3-a59cc1c553b0 tempest-ServerRescueNegativeTestJSON-1483258717 tempest-ServerRescueNegativeTestJSON-1483258717-project-member] [instance: 7ea919b6-12f7-4f4e-a123-413044bbbffc] Binding failed for port 61c8084d-6ad0-442a-967a-46f14ffa6671, please check neutron logs for more information. {{(pid=61806) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 736.625694] env[61806]: DEBUG oslo_concurrency.lockutils [None req-1382c347-5a79-4289-8b32-002f296b7ae3 tempest-TenantUsagesTestJSON-255073140 tempest-TenantUsagesTestJSON-255073140-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 16.909s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 736.627195] env[61806]: INFO nova.compute.claims [None req-1382c347-5a79-4289-8b32-002f296b7ae3 tempest-TenantUsagesTestJSON-255073140 tempest-TenantUsagesTestJSON-255073140-project-member] [instance: 949c1050-5003-4519-b24c-8904de21a676] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 736.630753] env[61806]: DEBUG nova.compute.manager [None req-6b55cbbf-00f7-4d4d-87f3-a59cc1c553b0 tempest-ServerRescueNegativeTestJSON-1483258717 tempest-ServerRescueNegativeTestJSON-1483258717-project-member] [instance: 7ea919b6-12f7-4f4e-a123-413044bbbffc] Build of instance 7ea919b6-12f7-4f4e-a123-413044bbbffc was re-scheduled: Binding failed for port 61c8084d-6ad0-442a-967a-46f14ffa6671, please check neutron logs for more information. {{(pid=61806) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 736.631237] env[61806]: DEBUG nova.compute.manager [None req-6b55cbbf-00f7-4d4d-87f3-a59cc1c553b0 tempest-ServerRescueNegativeTestJSON-1483258717 tempest-ServerRescueNegativeTestJSON-1483258717-project-member] [instance: 7ea919b6-12f7-4f4e-a123-413044bbbffc] Unplugging VIFs for instance {{(pid=61806) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 736.631488] env[61806]: DEBUG oslo_concurrency.lockutils [None req-6b55cbbf-00f7-4d4d-87f3-a59cc1c553b0 tempest-ServerRescueNegativeTestJSON-1483258717 tempest-ServerRescueNegativeTestJSON-1483258717-project-member] Acquiring lock "refresh_cache-7ea919b6-12f7-4f4e-a123-413044bbbffc" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 736.631647] env[61806]: DEBUG oslo_concurrency.lockutils [None req-6b55cbbf-00f7-4d4d-87f3-a59cc1c553b0 tempest-ServerRescueNegativeTestJSON-1483258717 tempest-ServerRescueNegativeTestJSON-1483258717-project-member] Acquired lock "refresh_cache-7ea919b6-12f7-4f4e-a123-413044bbbffc" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 736.632111] env[61806]: DEBUG nova.network.neutron [None req-6b55cbbf-00f7-4d4d-87f3-a59cc1c553b0 tempest-ServerRescueNegativeTestJSON-1483258717 tempest-ServerRescueNegativeTestJSON-1483258717-project-member] [instance: 7ea919b6-12f7-4f4e-a123-413044bbbffc] Building network info cache for instance {{(pid=61806) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 736.706421] env[61806]: DEBUG nova.compute.manager [None req-d17e7876-7c69-44af-acb2-32ed225ce945 tempest-ServerActionsTestOtherA-1637843849 tempest-ServerActionsTestOtherA-1637843849-project-member] [instance: 15476ac0-289a-4e04-aa9d-4244c658e962] Start spawning the instance on the hypervisor. {{(pid=61806) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 736.736814] env[61806]: DEBUG nova.virt.hardware [None req-d17e7876-7c69-44af-acb2-32ed225ce945 tempest-ServerActionsTestOtherA-1637843849 tempest-ServerActionsTestOtherA-1637843849-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-15T18:05:10Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-15T18:04:52Z,direct_url=,disk_format='vmdk',id=ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='84ba83fce2c349a6988173c5d6fc3b07',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-15T18:04:53Z,virtual_size=,visibility=), allow threads: False {{(pid=61806) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 736.737091] env[61806]: DEBUG nova.virt.hardware [None req-d17e7876-7c69-44af-acb2-32ed225ce945 tempest-ServerActionsTestOtherA-1637843849 tempest-ServerActionsTestOtherA-1637843849-project-member] Flavor limits 0:0:0 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 736.737255] env[61806]: DEBUG nova.virt.hardware [None req-d17e7876-7c69-44af-acb2-32ed225ce945 tempest-ServerActionsTestOtherA-1637843849 tempest-ServerActionsTestOtherA-1637843849-project-member] Image limits 0:0:0 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 736.737443] env[61806]: DEBUG nova.virt.hardware [None req-d17e7876-7c69-44af-acb2-32ed225ce945 tempest-ServerActionsTestOtherA-1637843849 tempest-ServerActionsTestOtherA-1637843849-project-member] Flavor pref 0:0:0 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 736.737591] env[61806]: DEBUG nova.virt.hardware [None req-d17e7876-7c69-44af-acb2-32ed225ce945 tempest-ServerActionsTestOtherA-1637843849 tempest-ServerActionsTestOtherA-1637843849-project-member] Image pref 0:0:0 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 736.737736] env[61806]: DEBUG nova.virt.hardware [None req-d17e7876-7c69-44af-acb2-32ed225ce945 tempest-ServerActionsTestOtherA-1637843849 tempest-ServerActionsTestOtherA-1637843849-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 736.737945] env[61806]: DEBUG nova.virt.hardware [None req-d17e7876-7c69-44af-acb2-32ed225ce945 tempest-ServerActionsTestOtherA-1637843849 tempest-ServerActionsTestOtherA-1637843849-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61806) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 736.739479] env[61806]: DEBUG nova.virt.hardware [None req-d17e7876-7c69-44af-acb2-32ed225ce945 tempest-ServerActionsTestOtherA-1637843849 tempest-ServerActionsTestOtherA-1637843849-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61806) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 736.739724] env[61806]: DEBUG nova.virt.hardware [None req-d17e7876-7c69-44af-acb2-32ed225ce945 tempest-ServerActionsTestOtherA-1637843849 tempest-ServerActionsTestOtherA-1637843849-project-member] Got 1 possible topologies {{(pid=61806) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 736.739909] env[61806]: DEBUG nova.virt.hardware [None req-d17e7876-7c69-44af-acb2-32ed225ce945 tempest-ServerActionsTestOtherA-1637843849 tempest-ServerActionsTestOtherA-1637843849-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61806) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 736.740109] env[61806]: DEBUG nova.virt.hardware [None req-d17e7876-7c69-44af-acb2-32ed225ce945 tempest-ServerActionsTestOtherA-1637843849 tempest-ServerActionsTestOtherA-1637843849-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61806) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 736.740988] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f8a7f4b7-4163-44eb-8b4f-ff4d782a2573 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 736.749995] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e9e1038f-a8a4-4056-ab52-d1bddff14b47 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 736.964714] env[61806]: DEBUG oslo_concurrency.lockutils [None req-da1d2f3c-f6ea-47f5-8c64-65c86bf2d676 tempest-ServerRescueTestJSON-811948373 tempest-ServerRescueTestJSON-811948373-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 737.160141] env[61806]: DEBUG nova.network.neutron [None req-6b55cbbf-00f7-4d4d-87f3-a59cc1c553b0 tempest-ServerRescueNegativeTestJSON-1483258717 tempest-ServerRescueNegativeTestJSON-1483258717-project-member] [instance: 7ea919b6-12f7-4f4e-a123-413044bbbffc] Instance cache missing network info. {{(pid=61806) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 737.178505] env[61806]: DEBUG nova.compute.manager [req-16517620-a93e-4261-a6c2-3d99ff14639c req-81495b78-72e7-4b82-89a8-6e7aefa5bfd7 service nova] [instance: 15476ac0-289a-4e04-aa9d-4244c658e962] Received event network-changed-76b9d660-f9cc-4d3f-ae94-28e2b3d3b46d {{(pid=61806) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 737.178505] env[61806]: DEBUG nova.compute.manager [req-16517620-a93e-4261-a6c2-3d99ff14639c req-81495b78-72e7-4b82-89a8-6e7aefa5bfd7 service nova] [instance: 15476ac0-289a-4e04-aa9d-4244c658e962] Refreshing instance network info cache due to event network-changed-76b9d660-f9cc-4d3f-ae94-28e2b3d3b46d. {{(pid=61806) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 737.178596] env[61806]: DEBUG oslo_concurrency.lockutils [req-16517620-a93e-4261-a6c2-3d99ff14639c req-81495b78-72e7-4b82-89a8-6e7aefa5bfd7 service nova] Acquiring lock "refresh_cache-15476ac0-289a-4e04-aa9d-4244c658e962" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 737.178711] env[61806]: DEBUG oslo_concurrency.lockutils [req-16517620-a93e-4261-a6c2-3d99ff14639c req-81495b78-72e7-4b82-89a8-6e7aefa5bfd7 service nova] Acquired lock "refresh_cache-15476ac0-289a-4e04-aa9d-4244c658e962" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 737.178875] env[61806]: DEBUG nova.network.neutron [req-16517620-a93e-4261-a6c2-3d99ff14639c req-81495b78-72e7-4b82-89a8-6e7aefa5bfd7 service nova] [instance: 15476ac0-289a-4e04-aa9d-4244c658e962] Refreshing network info cache for port 76b9d660-f9cc-4d3f-ae94-28e2b3d3b46d {{(pid=61806) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 737.262477] env[61806]: DEBUG nova.network.neutron [None req-6b55cbbf-00f7-4d4d-87f3-a59cc1c553b0 tempest-ServerRescueNegativeTestJSON-1483258717 tempest-ServerRescueNegativeTestJSON-1483258717-project-member] [instance: 7ea919b6-12f7-4f4e-a123-413044bbbffc] Updating instance_info_cache with network_info: [] {{(pid=61806) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 737.365979] env[61806]: ERROR nova.compute.manager [None req-d17e7876-7c69-44af-acb2-32ed225ce945 tempest-ServerActionsTestOtherA-1637843849 tempest-ServerActionsTestOtherA-1637843849-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 76b9d660-f9cc-4d3f-ae94-28e2b3d3b46d, please check neutron logs for more information. [ 737.365979] env[61806]: ERROR nova.compute.manager Traceback (most recent call last): [ 737.365979] env[61806]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 737.365979] env[61806]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 737.365979] env[61806]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 737.365979] env[61806]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 737.365979] env[61806]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 737.365979] env[61806]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 737.365979] env[61806]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 737.365979] env[61806]: ERROR nova.compute.manager self.force_reraise() [ 737.365979] env[61806]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 737.365979] env[61806]: ERROR nova.compute.manager raise self.value [ 737.365979] env[61806]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 737.365979] env[61806]: ERROR nova.compute.manager updated_port = self._update_port( [ 737.365979] env[61806]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 737.365979] env[61806]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 737.366671] env[61806]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 737.366671] env[61806]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 737.366671] env[61806]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 76b9d660-f9cc-4d3f-ae94-28e2b3d3b46d, please check neutron logs for more information. [ 737.366671] env[61806]: ERROR nova.compute.manager [ 737.366671] env[61806]: Traceback (most recent call last): [ 737.366671] env[61806]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 737.366671] env[61806]: listener.cb(fileno) [ 737.366671] env[61806]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 737.366671] env[61806]: result = function(*args, **kwargs) [ 737.366671] env[61806]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 737.366671] env[61806]: return func(*args, **kwargs) [ 737.366671] env[61806]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 737.366671] env[61806]: raise e [ 737.366671] env[61806]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 737.366671] env[61806]: nwinfo = self.network_api.allocate_for_instance( [ 737.366671] env[61806]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 737.366671] env[61806]: created_port_ids = self._update_ports_for_instance( [ 737.366671] env[61806]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 737.366671] env[61806]: with excutils.save_and_reraise_exception(): [ 737.366671] env[61806]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 737.366671] env[61806]: self.force_reraise() [ 737.366671] env[61806]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 737.366671] env[61806]: raise self.value [ 737.366671] env[61806]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 737.366671] env[61806]: updated_port = self._update_port( [ 737.366671] env[61806]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 737.366671] env[61806]: _ensure_no_port_binding_failure(port) [ 737.366671] env[61806]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 737.366671] env[61806]: raise exception.PortBindingFailed(port_id=port['id']) [ 737.367616] env[61806]: nova.exception.PortBindingFailed: Binding failed for port 76b9d660-f9cc-4d3f-ae94-28e2b3d3b46d, please check neutron logs for more information. [ 737.367616] env[61806]: Removing descriptor: 14 [ 737.367616] env[61806]: ERROR nova.compute.manager [None req-d17e7876-7c69-44af-acb2-32ed225ce945 tempest-ServerActionsTestOtherA-1637843849 tempest-ServerActionsTestOtherA-1637843849-project-member] [instance: 15476ac0-289a-4e04-aa9d-4244c658e962] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 76b9d660-f9cc-4d3f-ae94-28e2b3d3b46d, please check neutron logs for more information. [ 737.367616] env[61806]: ERROR nova.compute.manager [instance: 15476ac0-289a-4e04-aa9d-4244c658e962] Traceback (most recent call last): [ 737.367616] env[61806]: ERROR nova.compute.manager [instance: 15476ac0-289a-4e04-aa9d-4244c658e962] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 737.367616] env[61806]: ERROR nova.compute.manager [instance: 15476ac0-289a-4e04-aa9d-4244c658e962] yield resources [ 737.367616] env[61806]: ERROR nova.compute.manager [instance: 15476ac0-289a-4e04-aa9d-4244c658e962] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 737.367616] env[61806]: ERROR nova.compute.manager [instance: 15476ac0-289a-4e04-aa9d-4244c658e962] self.driver.spawn(context, instance, image_meta, [ 737.367616] env[61806]: ERROR nova.compute.manager [instance: 15476ac0-289a-4e04-aa9d-4244c658e962] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 737.367616] env[61806]: ERROR nova.compute.manager [instance: 15476ac0-289a-4e04-aa9d-4244c658e962] self._vmops.spawn(context, instance, image_meta, injected_files, [ 737.367616] env[61806]: ERROR nova.compute.manager [instance: 15476ac0-289a-4e04-aa9d-4244c658e962] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 737.367616] env[61806]: ERROR nova.compute.manager [instance: 15476ac0-289a-4e04-aa9d-4244c658e962] vm_ref = self.build_virtual_machine(instance, [ 737.367992] env[61806]: ERROR nova.compute.manager [instance: 15476ac0-289a-4e04-aa9d-4244c658e962] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 737.367992] env[61806]: ERROR nova.compute.manager [instance: 15476ac0-289a-4e04-aa9d-4244c658e962] vif_infos = vmwarevif.get_vif_info(self._session, [ 737.367992] env[61806]: ERROR nova.compute.manager [instance: 15476ac0-289a-4e04-aa9d-4244c658e962] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 737.367992] env[61806]: ERROR nova.compute.manager [instance: 15476ac0-289a-4e04-aa9d-4244c658e962] for vif in network_info: [ 737.367992] env[61806]: ERROR nova.compute.manager [instance: 15476ac0-289a-4e04-aa9d-4244c658e962] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 737.367992] env[61806]: ERROR nova.compute.manager [instance: 15476ac0-289a-4e04-aa9d-4244c658e962] return self._sync_wrapper(fn, *args, **kwargs) [ 737.367992] env[61806]: ERROR nova.compute.manager [instance: 15476ac0-289a-4e04-aa9d-4244c658e962] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 737.367992] env[61806]: ERROR nova.compute.manager [instance: 15476ac0-289a-4e04-aa9d-4244c658e962] self.wait() [ 737.367992] env[61806]: ERROR nova.compute.manager [instance: 15476ac0-289a-4e04-aa9d-4244c658e962] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 737.367992] env[61806]: ERROR nova.compute.manager [instance: 15476ac0-289a-4e04-aa9d-4244c658e962] self[:] = self._gt.wait() [ 737.367992] env[61806]: ERROR nova.compute.manager [instance: 15476ac0-289a-4e04-aa9d-4244c658e962] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 737.367992] env[61806]: ERROR nova.compute.manager [instance: 15476ac0-289a-4e04-aa9d-4244c658e962] return self._exit_event.wait() [ 737.367992] env[61806]: ERROR nova.compute.manager [instance: 15476ac0-289a-4e04-aa9d-4244c658e962] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 737.368391] env[61806]: ERROR nova.compute.manager [instance: 15476ac0-289a-4e04-aa9d-4244c658e962] result = hub.switch() [ 737.368391] env[61806]: ERROR nova.compute.manager [instance: 15476ac0-289a-4e04-aa9d-4244c658e962] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 737.368391] env[61806]: ERROR nova.compute.manager [instance: 15476ac0-289a-4e04-aa9d-4244c658e962] return self.greenlet.switch() [ 737.368391] env[61806]: ERROR nova.compute.manager [instance: 15476ac0-289a-4e04-aa9d-4244c658e962] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 737.368391] env[61806]: ERROR nova.compute.manager [instance: 15476ac0-289a-4e04-aa9d-4244c658e962] result = function(*args, **kwargs) [ 737.368391] env[61806]: ERROR nova.compute.manager [instance: 15476ac0-289a-4e04-aa9d-4244c658e962] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 737.368391] env[61806]: ERROR nova.compute.manager [instance: 15476ac0-289a-4e04-aa9d-4244c658e962] return func(*args, **kwargs) [ 737.368391] env[61806]: ERROR nova.compute.manager [instance: 15476ac0-289a-4e04-aa9d-4244c658e962] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 737.368391] env[61806]: ERROR nova.compute.manager [instance: 15476ac0-289a-4e04-aa9d-4244c658e962] raise e [ 737.368391] env[61806]: ERROR nova.compute.manager [instance: 15476ac0-289a-4e04-aa9d-4244c658e962] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 737.368391] env[61806]: ERROR nova.compute.manager [instance: 15476ac0-289a-4e04-aa9d-4244c658e962] nwinfo = self.network_api.allocate_for_instance( [ 737.368391] env[61806]: ERROR nova.compute.manager [instance: 15476ac0-289a-4e04-aa9d-4244c658e962] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 737.368391] env[61806]: ERROR nova.compute.manager [instance: 15476ac0-289a-4e04-aa9d-4244c658e962] created_port_ids = self._update_ports_for_instance( [ 737.368789] env[61806]: ERROR nova.compute.manager [instance: 15476ac0-289a-4e04-aa9d-4244c658e962] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 737.368789] env[61806]: ERROR nova.compute.manager [instance: 15476ac0-289a-4e04-aa9d-4244c658e962] with excutils.save_and_reraise_exception(): [ 737.368789] env[61806]: ERROR nova.compute.manager [instance: 15476ac0-289a-4e04-aa9d-4244c658e962] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 737.368789] env[61806]: ERROR nova.compute.manager [instance: 15476ac0-289a-4e04-aa9d-4244c658e962] self.force_reraise() [ 737.368789] env[61806]: ERROR nova.compute.manager [instance: 15476ac0-289a-4e04-aa9d-4244c658e962] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 737.368789] env[61806]: ERROR nova.compute.manager [instance: 15476ac0-289a-4e04-aa9d-4244c658e962] raise self.value [ 737.368789] env[61806]: ERROR nova.compute.manager [instance: 15476ac0-289a-4e04-aa9d-4244c658e962] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 737.368789] env[61806]: ERROR nova.compute.manager [instance: 15476ac0-289a-4e04-aa9d-4244c658e962] updated_port = self._update_port( [ 737.368789] env[61806]: ERROR nova.compute.manager [instance: 15476ac0-289a-4e04-aa9d-4244c658e962] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 737.368789] env[61806]: ERROR nova.compute.manager [instance: 15476ac0-289a-4e04-aa9d-4244c658e962] _ensure_no_port_binding_failure(port) [ 737.368789] env[61806]: ERROR nova.compute.manager [instance: 15476ac0-289a-4e04-aa9d-4244c658e962] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 737.368789] env[61806]: ERROR nova.compute.manager [instance: 15476ac0-289a-4e04-aa9d-4244c658e962] raise exception.PortBindingFailed(port_id=port['id']) [ 737.369212] env[61806]: ERROR nova.compute.manager [instance: 15476ac0-289a-4e04-aa9d-4244c658e962] nova.exception.PortBindingFailed: Binding failed for port 76b9d660-f9cc-4d3f-ae94-28e2b3d3b46d, please check neutron logs for more information. [ 737.369212] env[61806]: ERROR nova.compute.manager [instance: 15476ac0-289a-4e04-aa9d-4244c658e962] [ 737.369212] env[61806]: INFO nova.compute.manager [None req-d17e7876-7c69-44af-acb2-32ed225ce945 tempest-ServerActionsTestOtherA-1637843849 tempest-ServerActionsTestOtherA-1637843849-project-member] [instance: 15476ac0-289a-4e04-aa9d-4244c658e962] Terminating instance [ 737.369932] env[61806]: DEBUG oslo_concurrency.lockutils [None req-d17e7876-7c69-44af-acb2-32ed225ce945 tempest-ServerActionsTestOtherA-1637843849 tempest-ServerActionsTestOtherA-1637843849-project-member] Acquiring lock "refresh_cache-15476ac0-289a-4e04-aa9d-4244c658e962" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 737.700507] env[61806]: DEBUG nova.network.neutron [req-16517620-a93e-4261-a6c2-3d99ff14639c req-81495b78-72e7-4b82-89a8-6e7aefa5bfd7 service nova] [instance: 15476ac0-289a-4e04-aa9d-4244c658e962] Instance cache missing network info. {{(pid=61806) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 737.765568] env[61806]: DEBUG oslo_concurrency.lockutils [None req-6b55cbbf-00f7-4d4d-87f3-a59cc1c553b0 tempest-ServerRescueNegativeTestJSON-1483258717 tempest-ServerRescueNegativeTestJSON-1483258717-project-member] Releasing lock "refresh_cache-7ea919b6-12f7-4f4e-a123-413044bbbffc" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 737.765816] env[61806]: DEBUG nova.compute.manager [None req-6b55cbbf-00f7-4d4d-87f3-a59cc1c553b0 tempest-ServerRescueNegativeTestJSON-1483258717 tempest-ServerRescueNegativeTestJSON-1483258717-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61806) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 737.766010] env[61806]: DEBUG nova.compute.manager [None req-6b55cbbf-00f7-4d4d-87f3-a59cc1c553b0 tempest-ServerRescueNegativeTestJSON-1483258717 tempest-ServerRescueNegativeTestJSON-1483258717-project-member] [instance: 7ea919b6-12f7-4f4e-a123-413044bbbffc] Deallocating network for instance {{(pid=61806) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 737.766187] env[61806]: DEBUG nova.network.neutron [None req-6b55cbbf-00f7-4d4d-87f3-a59cc1c553b0 tempest-ServerRescueNegativeTestJSON-1483258717 tempest-ServerRescueNegativeTestJSON-1483258717-project-member] [instance: 7ea919b6-12f7-4f4e-a123-413044bbbffc] deallocate_for_instance() {{(pid=61806) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 737.787521] env[61806]: DEBUG nova.network.neutron [None req-6b55cbbf-00f7-4d4d-87f3-a59cc1c553b0 tempest-ServerRescueNegativeTestJSON-1483258717 tempest-ServerRescueNegativeTestJSON-1483258717-project-member] [instance: 7ea919b6-12f7-4f4e-a123-413044bbbffc] Instance cache missing network info. {{(pid=61806) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 737.821130] env[61806]: DEBUG nova.network.neutron [req-16517620-a93e-4261-a6c2-3d99ff14639c req-81495b78-72e7-4b82-89a8-6e7aefa5bfd7 service nova] [instance: 15476ac0-289a-4e04-aa9d-4244c658e962] Updating instance_info_cache with network_info: [] {{(pid=61806) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 737.973122] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-591f4de9-d00c-4da1-9d05-81bd1a17f5d7 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 737.980641] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a71ca11e-41e5-41c7-b84b-a96e8be02def {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 738.009459] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3d97e732-5049-4996-81d9-540c5928553a {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 738.016839] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-45608ffb-10ee-4c9d-8f8a-736368bfdbc1 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 738.029855] env[61806]: DEBUG nova.compute.provider_tree [None req-1382c347-5a79-4289-8b32-002f296b7ae3 tempest-TenantUsagesTestJSON-255073140 tempest-TenantUsagesTestJSON-255073140-project-member] Inventory has not changed in ProviderTree for provider: a2858be1-fd22-4e08-979e-87ad25293407 {{(pid=61806) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 738.167603] env[61806]: DEBUG oslo_concurrency.lockutils [None req-7fa92f02-ad2b-4004-a9a7-cb08f83ec0f1 tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Acquiring lock "67d1bdd8-cf8f-4fb5-bd3c-42008e4d60e3" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 738.168188] env[61806]: DEBUG oslo_concurrency.lockutils [None req-7fa92f02-ad2b-4004-a9a7-cb08f83ec0f1 tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Lock "67d1bdd8-cf8f-4fb5-bd3c-42008e4d60e3" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 738.290681] env[61806]: DEBUG nova.network.neutron [None req-6b55cbbf-00f7-4d4d-87f3-a59cc1c553b0 tempest-ServerRescueNegativeTestJSON-1483258717 tempest-ServerRescueNegativeTestJSON-1483258717-project-member] [instance: 7ea919b6-12f7-4f4e-a123-413044bbbffc] Updating instance_info_cache with network_info: [] {{(pid=61806) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 738.323367] env[61806]: DEBUG oslo_concurrency.lockutils [req-16517620-a93e-4261-a6c2-3d99ff14639c req-81495b78-72e7-4b82-89a8-6e7aefa5bfd7 service nova] Releasing lock "refresh_cache-15476ac0-289a-4e04-aa9d-4244c658e962" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 738.323810] env[61806]: DEBUG oslo_concurrency.lockutils [None req-d17e7876-7c69-44af-acb2-32ed225ce945 tempest-ServerActionsTestOtherA-1637843849 tempest-ServerActionsTestOtherA-1637843849-project-member] Acquired lock "refresh_cache-15476ac0-289a-4e04-aa9d-4244c658e962" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 738.324164] env[61806]: DEBUG nova.network.neutron [None req-d17e7876-7c69-44af-acb2-32ed225ce945 tempest-ServerActionsTestOtherA-1637843849 tempest-ServerActionsTestOtherA-1637843849-project-member] [instance: 15476ac0-289a-4e04-aa9d-4244c658e962] Building network info cache for instance {{(pid=61806) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 738.533376] env[61806]: DEBUG nova.scheduler.client.report [None req-1382c347-5a79-4289-8b32-002f296b7ae3 tempest-TenantUsagesTestJSON-255073140 tempest-TenantUsagesTestJSON-255073140-project-member] Inventory has not changed for provider a2858be1-fd22-4e08-979e-87ad25293407 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 139, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61806) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 738.793564] env[61806]: INFO nova.compute.manager [None req-6b55cbbf-00f7-4d4d-87f3-a59cc1c553b0 tempest-ServerRescueNegativeTestJSON-1483258717 tempest-ServerRescueNegativeTestJSON-1483258717-project-member] [instance: 7ea919b6-12f7-4f4e-a123-413044bbbffc] Took 1.03 seconds to deallocate network for instance. [ 738.849710] env[61806]: DEBUG nova.network.neutron [None req-d17e7876-7c69-44af-acb2-32ed225ce945 tempest-ServerActionsTestOtherA-1637843849 tempest-ServerActionsTestOtherA-1637843849-project-member] [instance: 15476ac0-289a-4e04-aa9d-4244c658e962] Instance cache missing network info. {{(pid=61806) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 738.936132] env[61806]: DEBUG nova.network.neutron [None req-d17e7876-7c69-44af-acb2-32ed225ce945 tempest-ServerActionsTestOtherA-1637843849 tempest-ServerActionsTestOtherA-1637843849-project-member] [instance: 15476ac0-289a-4e04-aa9d-4244c658e962] Updating instance_info_cache with network_info: [] {{(pid=61806) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 739.038614] env[61806]: DEBUG oslo_concurrency.lockutils [None req-1382c347-5a79-4289-8b32-002f296b7ae3 tempest-TenantUsagesTestJSON-255073140 tempest-TenantUsagesTestJSON-255073140-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.413s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 739.039239] env[61806]: DEBUG nova.compute.manager [None req-1382c347-5a79-4289-8b32-002f296b7ae3 tempest-TenantUsagesTestJSON-255073140 tempest-TenantUsagesTestJSON-255073140-project-member] [instance: 949c1050-5003-4519-b24c-8904de21a676] Start building networks asynchronously for instance. {{(pid=61806) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 739.041950] env[61806]: DEBUG oslo_concurrency.lockutils [None req-3bf756e0-c56e-4902-9c2b-b224a1d09a7c tempest-ListServersNegativeTestJSON-272187588 tempest-ListServersNegativeTestJSON-272187588-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 17.205s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 739.043296] env[61806]: INFO nova.compute.claims [None req-3bf756e0-c56e-4902-9c2b-b224a1d09a7c tempest-ListServersNegativeTestJSON-272187588 tempest-ListServersNegativeTestJSON-272187588-project-member] [instance: f2446d04-15b8-4b3d-8932-c37df67b0f8f] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 739.202868] env[61806]: DEBUG nova.compute.manager [req-5da90610-f71e-4d82-ad22-92e928745a09 req-0f755aec-ddb2-4627-8b9b-726728181ec0 service nova] [instance: 15476ac0-289a-4e04-aa9d-4244c658e962] Received event network-vif-deleted-76b9d660-f9cc-4d3f-ae94-28e2b3d3b46d {{(pid=61806) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 739.438930] env[61806]: DEBUG oslo_concurrency.lockutils [None req-d17e7876-7c69-44af-acb2-32ed225ce945 tempest-ServerActionsTestOtherA-1637843849 tempest-ServerActionsTestOtherA-1637843849-project-member] Releasing lock "refresh_cache-15476ac0-289a-4e04-aa9d-4244c658e962" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 739.439618] env[61806]: DEBUG nova.compute.manager [None req-d17e7876-7c69-44af-acb2-32ed225ce945 tempest-ServerActionsTestOtherA-1637843849 tempest-ServerActionsTestOtherA-1637843849-project-member] [instance: 15476ac0-289a-4e04-aa9d-4244c658e962] Start destroying the instance on the hypervisor. {{(pid=61806) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 739.439849] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-d17e7876-7c69-44af-acb2-32ed225ce945 tempest-ServerActionsTestOtherA-1637843849 tempest-ServerActionsTestOtherA-1637843849-project-member] [instance: 15476ac0-289a-4e04-aa9d-4244c658e962] Destroying instance {{(pid=61806) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 739.440197] env[61806]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-d8e06068-136b-431f-8aca-27288fc505df {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 739.449168] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6520c274-99e3-4669-9013-bde4e55fbc2e {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 739.471094] env[61806]: WARNING nova.virt.vmwareapi.vmops [None req-d17e7876-7c69-44af-acb2-32ed225ce945 tempest-ServerActionsTestOtherA-1637843849 tempest-ServerActionsTestOtherA-1637843849-project-member] [instance: 15476ac0-289a-4e04-aa9d-4244c658e962] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 15476ac0-289a-4e04-aa9d-4244c658e962 could not be found. [ 739.471400] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-d17e7876-7c69-44af-acb2-32ed225ce945 tempest-ServerActionsTestOtherA-1637843849 tempest-ServerActionsTestOtherA-1637843849-project-member] [instance: 15476ac0-289a-4e04-aa9d-4244c658e962] Instance destroyed {{(pid=61806) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 739.471653] env[61806]: INFO nova.compute.manager [None req-d17e7876-7c69-44af-acb2-32ed225ce945 tempest-ServerActionsTestOtherA-1637843849 tempest-ServerActionsTestOtherA-1637843849-project-member] [instance: 15476ac0-289a-4e04-aa9d-4244c658e962] Took 0.03 seconds to destroy the instance on the hypervisor. [ 739.471975] env[61806]: DEBUG oslo.service.loopingcall [None req-d17e7876-7c69-44af-acb2-32ed225ce945 tempest-ServerActionsTestOtherA-1637843849 tempest-ServerActionsTestOtherA-1637843849-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61806) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 739.472255] env[61806]: DEBUG nova.compute.manager [-] [instance: 15476ac0-289a-4e04-aa9d-4244c658e962] Deallocating network for instance {{(pid=61806) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 739.472380] env[61806]: DEBUG nova.network.neutron [-] [instance: 15476ac0-289a-4e04-aa9d-4244c658e962] deallocate_for_instance() {{(pid=61806) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 739.486918] env[61806]: DEBUG nova.network.neutron [-] [instance: 15476ac0-289a-4e04-aa9d-4244c658e962] Instance cache missing network info. {{(pid=61806) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 739.550308] env[61806]: DEBUG nova.compute.utils [None req-1382c347-5a79-4289-8b32-002f296b7ae3 tempest-TenantUsagesTestJSON-255073140 tempest-TenantUsagesTestJSON-255073140-project-member] Using /dev/sd instead of None {{(pid=61806) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 739.551831] env[61806]: DEBUG nova.compute.manager [None req-1382c347-5a79-4289-8b32-002f296b7ae3 tempest-TenantUsagesTestJSON-255073140 tempest-TenantUsagesTestJSON-255073140-project-member] [instance: 949c1050-5003-4519-b24c-8904de21a676] Allocating IP information in the background. {{(pid=61806) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 739.551970] env[61806]: DEBUG nova.network.neutron [None req-1382c347-5a79-4289-8b32-002f296b7ae3 tempest-TenantUsagesTestJSON-255073140 tempest-TenantUsagesTestJSON-255073140-project-member] [instance: 949c1050-5003-4519-b24c-8904de21a676] allocate_for_instance() {{(pid=61806) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 739.602216] env[61806]: DEBUG nova.policy [None req-1382c347-5a79-4289-8b32-002f296b7ae3 tempest-TenantUsagesTestJSON-255073140 tempest-TenantUsagesTestJSON-255073140-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'f50fd694dbe240818c024467cc4872bf', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'c6163a69c7734653ab24b136123d2557', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61806) authorize /opt/stack/nova/nova/policy.py:201}} [ 739.829049] env[61806]: INFO nova.scheduler.client.report [None req-6b55cbbf-00f7-4d4d-87f3-a59cc1c553b0 tempest-ServerRescueNegativeTestJSON-1483258717 tempest-ServerRescueNegativeTestJSON-1483258717-project-member] Deleted allocations for instance 7ea919b6-12f7-4f4e-a123-413044bbbffc [ 739.988143] env[61806]: DEBUG nova.network.neutron [None req-1382c347-5a79-4289-8b32-002f296b7ae3 tempest-TenantUsagesTestJSON-255073140 tempest-TenantUsagesTestJSON-255073140-project-member] [instance: 949c1050-5003-4519-b24c-8904de21a676] Successfully created port: 1add6960-077d-421a-b55c-d82b1d6d40eb {{(pid=61806) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 739.990890] env[61806]: DEBUG nova.network.neutron [-] [instance: 15476ac0-289a-4e04-aa9d-4244c658e962] Updating instance_info_cache with network_info: [] {{(pid=61806) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 740.055346] env[61806]: DEBUG nova.compute.manager [None req-1382c347-5a79-4289-8b32-002f296b7ae3 tempest-TenantUsagesTestJSON-255073140 tempest-TenantUsagesTestJSON-255073140-project-member] [instance: 949c1050-5003-4519-b24c-8904de21a676] Start building block device mappings for instance. {{(pid=61806) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 740.335592] env[61806]: DEBUG oslo_concurrency.lockutils [None req-6b55cbbf-00f7-4d4d-87f3-a59cc1c553b0 tempest-ServerRescueNegativeTestJSON-1483258717 tempest-ServerRescueNegativeTestJSON-1483258717-project-member] Lock "7ea919b6-12f7-4f4e-a123-413044bbbffc" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 159.842s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 740.405369] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7d6cec04-af7d-4925-bfb0-076567973b87 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 740.415576] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7112416a-d0e9-4d1c-a6dc-c7e8e453a138 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 740.449334] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9e78b599-02a3-4a8d-ba51-057347ea5b9e {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 740.457349] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-11d73696-fc0e-4a28-97a9-13350daee6cb {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 740.472912] env[61806]: DEBUG nova.compute.provider_tree [None req-3bf756e0-c56e-4902-9c2b-b224a1d09a7c tempest-ListServersNegativeTestJSON-272187588 tempest-ListServersNegativeTestJSON-272187588-project-member] Inventory has not changed in ProviderTree for provider: a2858be1-fd22-4e08-979e-87ad25293407 {{(pid=61806) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 740.492542] env[61806]: INFO nova.compute.manager [-] [instance: 15476ac0-289a-4e04-aa9d-4244c658e962] Took 1.02 seconds to deallocate network for instance. [ 740.494630] env[61806]: DEBUG nova.compute.claims [None req-d17e7876-7c69-44af-acb2-32ed225ce945 tempest-ServerActionsTestOtherA-1637843849 tempest-ServerActionsTestOtherA-1637843849-project-member] [instance: 15476ac0-289a-4e04-aa9d-4244c658e962] Aborting claim: {{(pid=61806) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 740.494807] env[61806]: DEBUG oslo_concurrency.lockutils [None req-d17e7876-7c69-44af-acb2-32ed225ce945 tempest-ServerActionsTestOtherA-1637843849 tempest-ServerActionsTestOtherA-1637843849-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 740.840193] env[61806]: DEBUG nova.compute.manager [None req-b5a6ece2-b279-4bbd-a3d1-16582c4d8894 tempest-ServersTestMultiNic-1480543082 tempest-ServersTestMultiNic-1480543082-project-member] [instance: 24c83131-1451-429f-84fe-dcd070898d65] Starting instance... {{(pid=61806) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 740.935931] env[61806]: ERROR nova.compute.manager [None req-1382c347-5a79-4289-8b32-002f296b7ae3 tempest-TenantUsagesTestJSON-255073140 tempest-TenantUsagesTestJSON-255073140-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 1add6960-077d-421a-b55c-d82b1d6d40eb, please check neutron logs for more information. [ 740.935931] env[61806]: ERROR nova.compute.manager Traceback (most recent call last): [ 740.935931] env[61806]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 740.935931] env[61806]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 740.935931] env[61806]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 740.935931] env[61806]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 740.935931] env[61806]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 740.935931] env[61806]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 740.935931] env[61806]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 740.935931] env[61806]: ERROR nova.compute.manager self.force_reraise() [ 740.935931] env[61806]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 740.935931] env[61806]: ERROR nova.compute.manager raise self.value [ 740.935931] env[61806]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 740.935931] env[61806]: ERROR nova.compute.manager updated_port = self._update_port( [ 740.935931] env[61806]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 740.935931] env[61806]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 740.936898] env[61806]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 740.936898] env[61806]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 740.936898] env[61806]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 1add6960-077d-421a-b55c-d82b1d6d40eb, please check neutron logs for more information. [ 740.936898] env[61806]: ERROR nova.compute.manager [ 740.936898] env[61806]: Traceback (most recent call last): [ 740.936898] env[61806]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 740.936898] env[61806]: listener.cb(fileno) [ 740.936898] env[61806]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 740.936898] env[61806]: result = function(*args, **kwargs) [ 740.936898] env[61806]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 740.936898] env[61806]: return func(*args, **kwargs) [ 740.936898] env[61806]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 740.936898] env[61806]: raise e [ 740.936898] env[61806]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 740.936898] env[61806]: nwinfo = self.network_api.allocate_for_instance( [ 740.936898] env[61806]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 740.936898] env[61806]: created_port_ids = self._update_ports_for_instance( [ 740.936898] env[61806]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 740.936898] env[61806]: with excutils.save_and_reraise_exception(): [ 740.936898] env[61806]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 740.936898] env[61806]: self.force_reraise() [ 740.936898] env[61806]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 740.936898] env[61806]: raise self.value [ 740.936898] env[61806]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 740.936898] env[61806]: updated_port = self._update_port( [ 740.936898] env[61806]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 740.936898] env[61806]: _ensure_no_port_binding_failure(port) [ 740.936898] env[61806]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 740.936898] env[61806]: raise exception.PortBindingFailed(port_id=port['id']) [ 740.938682] env[61806]: nova.exception.PortBindingFailed: Binding failed for port 1add6960-077d-421a-b55c-d82b1d6d40eb, please check neutron logs for more information. [ 740.938682] env[61806]: Removing descriptor: 14 [ 740.978022] env[61806]: DEBUG nova.scheduler.client.report [None req-3bf756e0-c56e-4902-9c2b-b224a1d09a7c tempest-ListServersNegativeTestJSON-272187588 tempest-ListServersNegativeTestJSON-272187588-project-member] Inventory has not changed for provider a2858be1-fd22-4e08-979e-87ad25293407 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 139, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61806) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 741.065980] env[61806]: DEBUG nova.compute.manager [None req-1382c347-5a79-4289-8b32-002f296b7ae3 tempest-TenantUsagesTestJSON-255073140 tempest-TenantUsagesTestJSON-255073140-project-member] [instance: 949c1050-5003-4519-b24c-8904de21a676] Start spawning the instance on the hypervisor. {{(pid=61806) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 741.092981] env[61806]: DEBUG nova.virt.hardware [None req-1382c347-5a79-4289-8b32-002f296b7ae3 tempest-TenantUsagesTestJSON-255073140 tempest-TenantUsagesTestJSON-255073140-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-15T18:05:10Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-15T18:04:52Z,direct_url=,disk_format='vmdk',id=ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='84ba83fce2c349a6988173c5d6fc3b07',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-15T18:04:53Z,virtual_size=,visibility=), allow threads: False {{(pid=61806) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 741.093237] env[61806]: DEBUG nova.virt.hardware [None req-1382c347-5a79-4289-8b32-002f296b7ae3 tempest-TenantUsagesTestJSON-255073140 tempest-TenantUsagesTestJSON-255073140-project-member] Flavor limits 0:0:0 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 741.093395] env[61806]: DEBUG nova.virt.hardware [None req-1382c347-5a79-4289-8b32-002f296b7ae3 tempest-TenantUsagesTestJSON-255073140 tempest-TenantUsagesTestJSON-255073140-project-member] Image limits 0:0:0 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 741.093578] env[61806]: DEBUG nova.virt.hardware [None req-1382c347-5a79-4289-8b32-002f296b7ae3 tempest-TenantUsagesTestJSON-255073140 tempest-TenantUsagesTestJSON-255073140-project-member] Flavor pref 0:0:0 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 741.093725] env[61806]: DEBUG nova.virt.hardware [None req-1382c347-5a79-4289-8b32-002f296b7ae3 tempest-TenantUsagesTestJSON-255073140 tempest-TenantUsagesTestJSON-255073140-project-member] Image pref 0:0:0 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 741.093868] env[61806]: DEBUG nova.virt.hardware [None req-1382c347-5a79-4289-8b32-002f296b7ae3 tempest-TenantUsagesTestJSON-255073140 tempest-TenantUsagesTestJSON-255073140-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 741.094081] env[61806]: DEBUG nova.virt.hardware [None req-1382c347-5a79-4289-8b32-002f296b7ae3 tempest-TenantUsagesTestJSON-255073140 tempest-TenantUsagesTestJSON-255073140-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61806) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 741.094240] env[61806]: DEBUG nova.virt.hardware [None req-1382c347-5a79-4289-8b32-002f296b7ae3 tempest-TenantUsagesTestJSON-255073140 tempest-TenantUsagesTestJSON-255073140-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61806) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 741.094402] env[61806]: DEBUG nova.virt.hardware [None req-1382c347-5a79-4289-8b32-002f296b7ae3 tempest-TenantUsagesTestJSON-255073140 tempest-TenantUsagesTestJSON-255073140-project-member] Got 1 possible topologies {{(pid=61806) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 741.094611] env[61806]: DEBUG nova.virt.hardware [None req-1382c347-5a79-4289-8b32-002f296b7ae3 tempest-TenantUsagesTestJSON-255073140 tempest-TenantUsagesTestJSON-255073140-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61806) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 741.094812] env[61806]: DEBUG nova.virt.hardware [None req-1382c347-5a79-4289-8b32-002f296b7ae3 tempest-TenantUsagesTestJSON-255073140 tempest-TenantUsagesTestJSON-255073140-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61806) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 741.095664] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-af5abb29-3e77-48ae-bad2-304050b33772 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 741.103551] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-75043a2a-caee-447b-aff9-687cbf638001 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 741.122744] env[61806]: ERROR nova.compute.manager [None req-1382c347-5a79-4289-8b32-002f296b7ae3 tempest-TenantUsagesTestJSON-255073140 tempest-TenantUsagesTestJSON-255073140-project-member] [instance: 949c1050-5003-4519-b24c-8904de21a676] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 1add6960-077d-421a-b55c-d82b1d6d40eb, please check neutron logs for more information. [ 741.122744] env[61806]: ERROR nova.compute.manager [instance: 949c1050-5003-4519-b24c-8904de21a676] Traceback (most recent call last): [ 741.122744] env[61806]: ERROR nova.compute.manager [instance: 949c1050-5003-4519-b24c-8904de21a676] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 741.122744] env[61806]: ERROR nova.compute.manager [instance: 949c1050-5003-4519-b24c-8904de21a676] yield resources [ 741.122744] env[61806]: ERROR nova.compute.manager [instance: 949c1050-5003-4519-b24c-8904de21a676] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 741.122744] env[61806]: ERROR nova.compute.manager [instance: 949c1050-5003-4519-b24c-8904de21a676] self.driver.spawn(context, instance, image_meta, [ 741.122744] env[61806]: ERROR nova.compute.manager [instance: 949c1050-5003-4519-b24c-8904de21a676] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 741.122744] env[61806]: ERROR nova.compute.manager [instance: 949c1050-5003-4519-b24c-8904de21a676] self._vmops.spawn(context, instance, image_meta, injected_files, [ 741.122744] env[61806]: ERROR nova.compute.manager [instance: 949c1050-5003-4519-b24c-8904de21a676] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 741.122744] env[61806]: ERROR nova.compute.manager [instance: 949c1050-5003-4519-b24c-8904de21a676] vm_ref = self.build_virtual_machine(instance, [ 741.122744] env[61806]: ERROR nova.compute.manager [instance: 949c1050-5003-4519-b24c-8904de21a676] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 741.123447] env[61806]: ERROR nova.compute.manager [instance: 949c1050-5003-4519-b24c-8904de21a676] vif_infos = vmwarevif.get_vif_info(self._session, [ 741.123447] env[61806]: ERROR nova.compute.manager [instance: 949c1050-5003-4519-b24c-8904de21a676] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 741.123447] env[61806]: ERROR nova.compute.manager [instance: 949c1050-5003-4519-b24c-8904de21a676] for vif in network_info: [ 741.123447] env[61806]: ERROR nova.compute.manager [instance: 949c1050-5003-4519-b24c-8904de21a676] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 741.123447] env[61806]: ERROR nova.compute.manager [instance: 949c1050-5003-4519-b24c-8904de21a676] return self._sync_wrapper(fn, *args, **kwargs) [ 741.123447] env[61806]: ERROR nova.compute.manager [instance: 949c1050-5003-4519-b24c-8904de21a676] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 741.123447] env[61806]: ERROR nova.compute.manager [instance: 949c1050-5003-4519-b24c-8904de21a676] self.wait() [ 741.123447] env[61806]: ERROR nova.compute.manager [instance: 949c1050-5003-4519-b24c-8904de21a676] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 741.123447] env[61806]: ERROR nova.compute.manager [instance: 949c1050-5003-4519-b24c-8904de21a676] self[:] = self._gt.wait() [ 741.123447] env[61806]: ERROR nova.compute.manager [instance: 949c1050-5003-4519-b24c-8904de21a676] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 741.123447] env[61806]: ERROR nova.compute.manager [instance: 949c1050-5003-4519-b24c-8904de21a676] return self._exit_event.wait() [ 741.123447] env[61806]: ERROR nova.compute.manager [instance: 949c1050-5003-4519-b24c-8904de21a676] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 741.123447] env[61806]: ERROR nova.compute.manager [instance: 949c1050-5003-4519-b24c-8904de21a676] current.throw(*self._exc) [ 741.124164] env[61806]: ERROR nova.compute.manager [instance: 949c1050-5003-4519-b24c-8904de21a676] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 741.124164] env[61806]: ERROR nova.compute.manager [instance: 949c1050-5003-4519-b24c-8904de21a676] result = function(*args, **kwargs) [ 741.124164] env[61806]: ERROR nova.compute.manager [instance: 949c1050-5003-4519-b24c-8904de21a676] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 741.124164] env[61806]: ERROR nova.compute.manager [instance: 949c1050-5003-4519-b24c-8904de21a676] return func(*args, **kwargs) [ 741.124164] env[61806]: ERROR nova.compute.manager [instance: 949c1050-5003-4519-b24c-8904de21a676] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 741.124164] env[61806]: ERROR nova.compute.manager [instance: 949c1050-5003-4519-b24c-8904de21a676] raise e [ 741.124164] env[61806]: ERROR nova.compute.manager [instance: 949c1050-5003-4519-b24c-8904de21a676] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 741.124164] env[61806]: ERROR nova.compute.manager [instance: 949c1050-5003-4519-b24c-8904de21a676] nwinfo = self.network_api.allocate_for_instance( [ 741.124164] env[61806]: ERROR nova.compute.manager [instance: 949c1050-5003-4519-b24c-8904de21a676] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 741.124164] env[61806]: ERROR nova.compute.manager [instance: 949c1050-5003-4519-b24c-8904de21a676] created_port_ids = self._update_ports_for_instance( [ 741.124164] env[61806]: ERROR nova.compute.manager [instance: 949c1050-5003-4519-b24c-8904de21a676] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 741.124164] env[61806]: ERROR nova.compute.manager [instance: 949c1050-5003-4519-b24c-8904de21a676] with excutils.save_and_reraise_exception(): [ 741.124164] env[61806]: ERROR nova.compute.manager [instance: 949c1050-5003-4519-b24c-8904de21a676] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 741.124643] env[61806]: ERROR nova.compute.manager [instance: 949c1050-5003-4519-b24c-8904de21a676] self.force_reraise() [ 741.124643] env[61806]: ERROR nova.compute.manager [instance: 949c1050-5003-4519-b24c-8904de21a676] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 741.124643] env[61806]: ERROR nova.compute.manager [instance: 949c1050-5003-4519-b24c-8904de21a676] raise self.value [ 741.124643] env[61806]: ERROR nova.compute.manager [instance: 949c1050-5003-4519-b24c-8904de21a676] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 741.124643] env[61806]: ERROR nova.compute.manager [instance: 949c1050-5003-4519-b24c-8904de21a676] updated_port = self._update_port( [ 741.124643] env[61806]: ERROR nova.compute.manager [instance: 949c1050-5003-4519-b24c-8904de21a676] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 741.124643] env[61806]: ERROR nova.compute.manager [instance: 949c1050-5003-4519-b24c-8904de21a676] _ensure_no_port_binding_failure(port) [ 741.124643] env[61806]: ERROR nova.compute.manager [instance: 949c1050-5003-4519-b24c-8904de21a676] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 741.124643] env[61806]: ERROR nova.compute.manager [instance: 949c1050-5003-4519-b24c-8904de21a676] raise exception.PortBindingFailed(port_id=port['id']) [ 741.124643] env[61806]: ERROR nova.compute.manager [instance: 949c1050-5003-4519-b24c-8904de21a676] nova.exception.PortBindingFailed: Binding failed for port 1add6960-077d-421a-b55c-d82b1d6d40eb, please check neutron logs for more information. [ 741.124643] env[61806]: ERROR nova.compute.manager [instance: 949c1050-5003-4519-b24c-8904de21a676] [ 741.124643] env[61806]: INFO nova.compute.manager [None req-1382c347-5a79-4289-8b32-002f296b7ae3 tempest-TenantUsagesTestJSON-255073140 tempest-TenantUsagesTestJSON-255073140-project-member] [instance: 949c1050-5003-4519-b24c-8904de21a676] Terminating instance [ 741.125297] env[61806]: DEBUG oslo_concurrency.lockutils [None req-1382c347-5a79-4289-8b32-002f296b7ae3 tempest-TenantUsagesTestJSON-255073140 tempest-TenantUsagesTestJSON-255073140-project-member] Acquiring lock "refresh_cache-949c1050-5003-4519-b24c-8904de21a676" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 741.125297] env[61806]: DEBUG oslo_concurrency.lockutils [None req-1382c347-5a79-4289-8b32-002f296b7ae3 tempest-TenantUsagesTestJSON-255073140 tempest-TenantUsagesTestJSON-255073140-project-member] Acquired lock "refresh_cache-949c1050-5003-4519-b24c-8904de21a676" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 741.125297] env[61806]: DEBUG nova.network.neutron [None req-1382c347-5a79-4289-8b32-002f296b7ae3 tempest-TenantUsagesTestJSON-255073140 tempest-TenantUsagesTestJSON-255073140-project-member] [instance: 949c1050-5003-4519-b24c-8904de21a676] Building network info cache for instance {{(pid=61806) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 741.243589] env[61806]: DEBUG nova.compute.manager [req-b0a84cb1-7baa-4935-9157-d2ce19735c79 req-e2babf4f-a06f-490d-bfc1-69e98abdfc97 service nova] [instance: 949c1050-5003-4519-b24c-8904de21a676] Received event network-changed-1add6960-077d-421a-b55c-d82b1d6d40eb {{(pid=61806) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 741.243786] env[61806]: DEBUG nova.compute.manager [req-b0a84cb1-7baa-4935-9157-d2ce19735c79 req-e2babf4f-a06f-490d-bfc1-69e98abdfc97 service nova] [instance: 949c1050-5003-4519-b24c-8904de21a676] Refreshing instance network info cache due to event network-changed-1add6960-077d-421a-b55c-d82b1d6d40eb. {{(pid=61806) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 741.243976] env[61806]: DEBUG oslo_concurrency.lockutils [req-b0a84cb1-7baa-4935-9157-d2ce19735c79 req-e2babf4f-a06f-490d-bfc1-69e98abdfc97 service nova] Acquiring lock "refresh_cache-949c1050-5003-4519-b24c-8904de21a676" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 741.365763] env[61806]: DEBUG oslo_concurrency.lockutils [None req-b5a6ece2-b279-4bbd-a3d1-16582c4d8894 tempest-ServersTestMultiNic-1480543082 tempest-ServersTestMultiNic-1480543082-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 741.486978] env[61806]: DEBUG oslo_concurrency.lockutils [None req-3bf756e0-c56e-4902-9c2b-b224a1d09a7c tempest-ListServersNegativeTestJSON-272187588 tempest-ListServersNegativeTestJSON-272187588-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.445s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 741.487543] env[61806]: DEBUG nova.compute.manager [None req-3bf756e0-c56e-4902-9c2b-b224a1d09a7c tempest-ListServersNegativeTestJSON-272187588 tempest-ListServersNegativeTestJSON-272187588-project-member] [instance: f2446d04-15b8-4b3d-8932-c37df67b0f8f] Start building networks asynchronously for instance. {{(pid=61806) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 741.494832] env[61806]: DEBUG oslo_concurrency.lockutils [None req-3c3981d2-fa25-4f4b-87d4-a5411e67199e tempest-MultipleCreateTestJSON-1589015462 tempest-MultipleCreateTestJSON-1589015462-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 14.204s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 741.647361] env[61806]: DEBUG nova.network.neutron [None req-1382c347-5a79-4289-8b32-002f296b7ae3 tempest-TenantUsagesTestJSON-255073140 tempest-TenantUsagesTestJSON-255073140-project-member] [instance: 949c1050-5003-4519-b24c-8904de21a676] Instance cache missing network info. {{(pid=61806) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 741.711524] env[61806]: DEBUG nova.network.neutron [None req-1382c347-5a79-4289-8b32-002f296b7ae3 tempest-TenantUsagesTestJSON-255073140 tempest-TenantUsagesTestJSON-255073140-project-member] [instance: 949c1050-5003-4519-b24c-8904de21a676] Updating instance_info_cache with network_info: [] {{(pid=61806) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 741.999718] env[61806]: DEBUG nova.compute.utils [None req-3bf756e0-c56e-4902-9c2b-b224a1d09a7c tempest-ListServersNegativeTestJSON-272187588 tempest-ListServersNegativeTestJSON-272187588-project-member] Using /dev/sd instead of None {{(pid=61806) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 742.003724] env[61806]: DEBUG nova.compute.manager [None req-3bf756e0-c56e-4902-9c2b-b224a1d09a7c tempest-ListServersNegativeTestJSON-272187588 tempest-ListServersNegativeTestJSON-272187588-project-member] [instance: f2446d04-15b8-4b3d-8932-c37df67b0f8f] Allocating IP information in the background. {{(pid=61806) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 742.003894] env[61806]: DEBUG nova.network.neutron [None req-3bf756e0-c56e-4902-9c2b-b224a1d09a7c tempest-ListServersNegativeTestJSON-272187588 tempest-ListServersNegativeTestJSON-272187588-project-member] [instance: f2446d04-15b8-4b3d-8932-c37df67b0f8f] allocate_for_instance() {{(pid=61806) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 742.065065] env[61806]: DEBUG nova.policy [None req-3bf756e0-c56e-4902-9c2b-b224a1d09a7c tempest-ListServersNegativeTestJSON-272187588 tempest-ListServersNegativeTestJSON-272187588-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '6228fec9ded14ef9974634c9a872374e', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'fb3d07cc8f764d15bc6fab0f299da5e8', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61806) authorize /opt/stack/nova/nova/policy.py:201}} [ 742.218104] env[61806]: DEBUG oslo_concurrency.lockutils [None req-1382c347-5a79-4289-8b32-002f296b7ae3 tempest-TenantUsagesTestJSON-255073140 tempest-TenantUsagesTestJSON-255073140-project-member] Releasing lock "refresh_cache-949c1050-5003-4519-b24c-8904de21a676" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 742.218104] env[61806]: DEBUG nova.compute.manager [None req-1382c347-5a79-4289-8b32-002f296b7ae3 tempest-TenantUsagesTestJSON-255073140 tempest-TenantUsagesTestJSON-255073140-project-member] [instance: 949c1050-5003-4519-b24c-8904de21a676] Start destroying the instance on the hypervisor. {{(pid=61806) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 742.218104] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-1382c347-5a79-4289-8b32-002f296b7ae3 tempest-TenantUsagesTestJSON-255073140 tempest-TenantUsagesTestJSON-255073140-project-member] [instance: 949c1050-5003-4519-b24c-8904de21a676] Destroying instance {{(pid=61806) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 742.218289] env[61806]: DEBUG oslo_concurrency.lockutils [req-b0a84cb1-7baa-4935-9157-d2ce19735c79 req-e2babf4f-a06f-490d-bfc1-69e98abdfc97 service nova] Acquired lock "refresh_cache-949c1050-5003-4519-b24c-8904de21a676" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 742.218487] env[61806]: DEBUG nova.network.neutron [req-b0a84cb1-7baa-4935-9157-d2ce19735c79 req-e2babf4f-a06f-490d-bfc1-69e98abdfc97 service nova] [instance: 949c1050-5003-4519-b24c-8904de21a676] Refreshing network info cache for port 1add6960-077d-421a-b55c-d82b1d6d40eb {{(pid=61806) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 742.219538] env[61806]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-7828db98-a3fc-4099-9a62-d0f538617c94 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 742.230944] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-21af665d-6e07-4118-86b0-b13ceda0ae6f {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 742.265134] env[61806]: WARNING nova.virt.vmwareapi.vmops [None req-1382c347-5a79-4289-8b32-002f296b7ae3 tempest-TenantUsagesTestJSON-255073140 tempest-TenantUsagesTestJSON-255073140-project-member] [instance: 949c1050-5003-4519-b24c-8904de21a676] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 949c1050-5003-4519-b24c-8904de21a676 could not be found. [ 742.265390] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-1382c347-5a79-4289-8b32-002f296b7ae3 tempest-TenantUsagesTestJSON-255073140 tempest-TenantUsagesTestJSON-255073140-project-member] [instance: 949c1050-5003-4519-b24c-8904de21a676] Instance destroyed {{(pid=61806) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 742.265576] env[61806]: INFO nova.compute.manager [None req-1382c347-5a79-4289-8b32-002f296b7ae3 tempest-TenantUsagesTestJSON-255073140 tempest-TenantUsagesTestJSON-255073140-project-member] [instance: 949c1050-5003-4519-b24c-8904de21a676] Took 0.05 seconds to destroy the instance on the hypervisor. [ 742.265819] env[61806]: DEBUG oslo.service.loopingcall [None req-1382c347-5a79-4289-8b32-002f296b7ae3 tempest-TenantUsagesTestJSON-255073140 tempest-TenantUsagesTestJSON-255073140-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61806) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 742.266514] env[61806]: DEBUG nova.compute.manager [-] [instance: 949c1050-5003-4519-b24c-8904de21a676] Deallocating network for instance {{(pid=61806) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 742.266615] env[61806]: DEBUG nova.network.neutron [-] [instance: 949c1050-5003-4519-b24c-8904de21a676] deallocate_for_instance() {{(pid=61806) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 742.292934] env[61806]: DEBUG nova.network.neutron [-] [instance: 949c1050-5003-4519-b24c-8904de21a676] Instance cache missing network info. {{(pid=61806) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 742.340771] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b5ace4c7-507e-477a-a860-15d128bcdfaa {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 742.352242] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c46431bf-7203-4abe-a1d5-95e97dfb959d {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 742.389517] env[61806]: DEBUG nova.network.neutron [None req-3bf756e0-c56e-4902-9c2b-b224a1d09a7c tempest-ListServersNegativeTestJSON-272187588 tempest-ListServersNegativeTestJSON-272187588-project-member] [instance: f2446d04-15b8-4b3d-8932-c37df67b0f8f] Successfully created port: ad624cd6-204b-4189-a855-10dca931bd79 {{(pid=61806) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 742.392124] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-594f7efd-441d-4068-a330-dacec1bec963 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 742.400724] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dde94b82-b287-411f-99d6-11d3866ac93a {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 742.414121] env[61806]: DEBUG nova.compute.provider_tree [None req-3c3981d2-fa25-4f4b-87d4-a5411e67199e tempest-MultipleCreateTestJSON-1589015462 tempest-MultipleCreateTestJSON-1589015462-project-member] Inventory has not changed in ProviderTree for provider: a2858be1-fd22-4e08-979e-87ad25293407 {{(pid=61806) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 742.507106] env[61806]: DEBUG nova.compute.manager [None req-3bf756e0-c56e-4902-9c2b-b224a1d09a7c tempest-ListServersNegativeTestJSON-272187588 tempest-ListServersNegativeTestJSON-272187588-project-member] [instance: f2446d04-15b8-4b3d-8932-c37df67b0f8f] Start building block device mappings for instance. {{(pid=61806) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 742.743793] env[61806]: DEBUG nova.network.neutron [req-b0a84cb1-7baa-4935-9157-d2ce19735c79 req-e2babf4f-a06f-490d-bfc1-69e98abdfc97 service nova] [instance: 949c1050-5003-4519-b24c-8904de21a676] Instance cache missing network info. {{(pid=61806) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 742.797393] env[61806]: DEBUG nova.network.neutron [-] [instance: 949c1050-5003-4519-b24c-8904de21a676] Updating instance_info_cache with network_info: [] {{(pid=61806) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 742.815056] env[61806]: DEBUG nova.network.neutron [req-b0a84cb1-7baa-4935-9157-d2ce19735c79 req-e2babf4f-a06f-490d-bfc1-69e98abdfc97 service nova] [instance: 949c1050-5003-4519-b24c-8904de21a676] Updating instance_info_cache with network_info: [] {{(pid=61806) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 742.918754] env[61806]: DEBUG nova.scheduler.client.report [None req-3c3981d2-fa25-4f4b-87d4-a5411e67199e tempest-MultipleCreateTestJSON-1589015462 tempest-MultipleCreateTestJSON-1589015462-project-member] Inventory has not changed for provider a2858be1-fd22-4e08-979e-87ad25293407 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 139, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61806) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 743.264259] env[61806]: ERROR nova.compute.manager [None req-3bf756e0-c56e-4902-9c2b-b224a1d09a7c tempest-ListServersNegativeTestJSON-272187588 tempest-ListServersNegativeTestJSON-272187588-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port ad624cd6-204b-4189-a855-10dca931bd79, please check neutron logs for more information. [ 743.264259] env[61806]: ERROR nova.compute.manager Traceback (most recent call last): [ 743.264259] env[61806]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 743.264259] env[61806]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 743.264259] env[61806]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 743.264259] env[61806]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 743.264259] env[61806]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 743.264259] env[61806]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 743.264259] env[61806]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 743.264259] env[61806]: ERROR nova.compute.manager self.force_reraise() [ 743.264259] env[61806]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 743.264259] env[61806]: ERROR nova.compute.manager raise self.value [ 743.264259] env[61806]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 743.264259] env[61806]: ERROR nova.compute.manager updated_port = self._update_port( [ 743.264259] env[61806]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 743.264259] env[61806]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 743.265097] env[61806]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 743.265097] env[61806]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 743.265097] env[61806]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port ad624cd6-204b-4189-a855-10dca931bd79, please check neutron logs for more information. [ 743.265097] env[61806]: ERROR nova.compute.manager [ 743.265097] env[61806]: Traceback (most recent call last): [ 743.265097] env[61806]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 743.265097] env[61806]: listener.cb(fileno) [ 743.265097] env[61806]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 743.265097] env[61806]: result = function(*args, **kwargs) [ 743.265097] env[61806]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 743.265097] env[61806]: return func(*args, **kwargs) [ 743.265097] env[61806]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 743.265097] env[61806]: raise e [ 743.265097] env[61806]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 743.265097] env[61806]: nwinfo = self.network_api.allocate_for_instance( [ 743.265097] env[61806]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 743.265097] env[61806]: created_port_ids = self._update_ports_for_instance( [ 743.265097] env[61806]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 743.265097] env[61806]: with excutils.save_and_reraise_exception(): [ 743.265097] env[61806]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 743.265097] env[61806]: self.force_reraise() [ 743.265097] env[61806]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 743.265097] env[61806]: raise self.value [ 743.265097] env[61806]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 743.265097] env[61806]: updated_port = self._update_port( [ 743.265097] env[61806]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 743.265097] env[61806]: _ensure_no_port_binding_failure(port) [ 743.265097] env[61806]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 743.265097] env[61806]: raise exception.PortBindingFailed(port_id=port['id']) [ 743.266072] env[61806]: nova.exception.PortBindingFailed: Binding failed for port ad624cd6-204b-4189-a855-10dca931bd79, please check neutron logs for more information. [ 743.266072] env[61806]: Removing descriptor: 14 [ 743.275809] env[61806]: DEBUG nova.compute.manager [req-471db471-fc87-4586-83cc-b32bd7c6934a req-5d30ec46-1a46-4bc6-97ca-7483698ad66d service nova] [instance: f2446d04-15b8-4b3d-8932-c37df67b0f8f] Received event network-changed-ad624cd6-204b-4189-a855-10dca931bd79 {{(pid=61806) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 743.275938] env[61806]: DEBUG nova.compute.manager [req-471db471-fc87-4586-83cc-b32bd7c6934a req-5d30ec46-1a46-4bc6-97ca-7483698ad66d service nova] [instance: f2446d04-15b8-4b3d-8932-c37df67b0f8f] Refreshing instance network info cache due to event network-changed-ad624cd6-204b-4189-a855-10dca931bd79. {{(pid=61806) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 743.276196] env[61806]: DEBUG oslo_concurrency.lockutils [req-471db471-fc87-4586-83cc-b32bd7c6934a req-5d30ec46-1a46-4bc6-97ca-7483698ad66d service nova] Acquiring lock "refresh_cache-f2446d04-15b8-4b3d-8932-c37df67b0f8f" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 743.276259] env[61806]: DEBUG oslo_concurrency.lockutils [req-471db471-fc87-4586-83cc-b32bd7c6934a req-5d30ec46-1a46-4bc6-97ca-7483698ad66d service nova] Acquired lock "refresh_cache-f2446d04-15b8-4b3d-8932-c37df67b0f8f" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 743.276415] env[61806]: DEBUG nova.network.neutron [req-471db471-fc87-4586-83cc-b32bd7c6934a req-5d30ec46-1a46-4bc6-97ca-7483698ad66d service nova] [instance: f2446d04-15b8-4b3d-8932-c37df67b0f8f] Refreshing network info cache for port ad624cd6-204b-4189-a855-10dca931bd79 {{(pid=61806) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 743.300045] env[61806]: INFO nova.compute.manager [-] [instance: 949c1050-5003-4519-b24c-8904de21a676] Took 1.03 seconds to deallocate network for instance. [ 743.302412] env[61806]: DEBUG nova.compute.claims [None req-1382c347-5a79-4289-8b32-002f296b7ae3 tempest-TenantUsagesTestJSON-255073140 tempest-TenantUsagesTestJSON-255073140-project-member] [instance: 949c1050-5003-4519-b24c-8904de21a676] Aborting claim: {{(pid=61806) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 743.303065] env[61806]: DEBUG oslo_concurrency.lockutils [None req-1382c347-5a79-4289-8b32-002f296b7ae3 tempest-TenantUsagesTestJSON-255073140 tempest-TenantUsagesTestJSON-255073140-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 743.318060] env[61806]: DEBUG oslo_concurrency.lockutils [req-b0a84cb1-7baa-4935-9157-d2ce19735c79 req-e2babf4f-a06f-490d-bfc1-69e98abdfc97 service nova] Releasing lock "refresh_cache-949c1050-5003-4519-b24c-8904de21a676" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 743.318368] env[61806]: DEBUG nova.compute.manager [req-b0a84cb1-7baa-4935-9157-d2ce19735c79 req-e2babf4f-a06f-490d-bfc1-69e98abdfc97 service nova] [instance: 949c1050-5003-4519-b24c-8904de21a676] Received event network-vif-deleted-1add6960-077d-421a-b55c-d82b1d6d40eb {{(pid=61806) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 743.425326] env[61806]: DEBUG oslo_concurrency.lockutils [None req-3c3981d2-fa25-4f4b-87d4-a5411e67199e tempest-MultipleCreateTestJSON-1589015462 tempest-MultipleCreateTestJSON-1589015462-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.935s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 743.425838] env[61806]: ERROR nova.compute.manager [None req-3c3981d2-fa25-4f4b-87d4-a5411e67199e tempest-MultipleCreateTestJSON-1589015462 tempest-MultipleCreateTestJSON-1589015462-project-member] [instance: 7211d568-cc68-4e50-80ad-b78878d1deab] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 2aa127d2-ff55-470b-822d-60c2d2d409d6, please check neutron logs for more information. [ 743.425838] env[61806]: ERROR nova.compute.manager [instance: 7211d568-cc68-4e50-80ad-b78878d1deab] Traceback (most recent call last): [ 743.425838] env[61806]: ERROR nova.compute.manager [instance: 7211d568-cc68-4e50-80ad-b78878d1deab] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 743.425838] env[61806]: ERROR nova.compute.manager [instance: 7211d568-cc68-4e50-80ad-b78878d1deab] self.driver.spawn(context, instance, image_meta, [ 743.425838] env[61806]: ERROR nova.compute.manager [instance: 7211d568-cc68-4e50-80ad-b78878d1deab] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 743.425838] env[61806]: ERROR nova.compute.manager [instance: 7211d568-cc68-4e50-80ad-b78878d1deab] self._vmops.spawn(context, instance, image_meta, injected_files, [ 743.425838] env[61806]: ERROR nova.compute.manager [instance: 7211d568-cc68-4e50-80ad-b78878d1deab] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 743.425838] env[61806]: ERROR nova.compute.manager [instance: 7211d568-cc68-4e50-80ad-b78878d1deab] vm_ref = self.build_virtual_machine(instance, [ 743.425838] env[61806]: ERROR nova.compute.manager [instance: 7211d568-cc68-4e50-80ad-b78878d1deab] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 743.425838] env[61806]: ERROR nova.compute.manager [instance: 7211d568-cc68-4e50-80ad-b78878d1deab] vif_infos = vmwarevif.get_vif_info(self._session, [ 743.425838] env[61806]: ERROR nova.compute.manager [instance: 7211d568-cc68-4e50-80ad-b78878d1deab] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 743.426202] env[61806]: ERROR nova.compute.manager [instance: 7211d568-cc68-4e50-80ad-b78878d1deab] for vif in network_info: [ 743.426202] env[61806]: ERROR nova.compute.manager [instance: 7211d568-cc68-4e50-80ad-b78878d1deab] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 743.426202] env[61806]: ERROR nova.compute.manager [instance: 7211d568-cc68-4e50-80ad-b78878d1deab] return self._sync_wrapper(fn, *args, **kwargs) [ 743.426202] env[61806]: ERROR nova.compute.manager [instance: 7211d568-cc68-4e50-80ad-b78878d1deab] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 743.426202] env[61806]: ERROR nova.compute.manager [instance: 7211d568-cc68-4e50-80ad-b78878d1deab] self.wait() [ 743.426202] env[61806]: ERROR nova.compute.manager [instance: 7211d568-cc68-4e50-80ad-b78878d1deab] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 743.426202] env[61806]: ERROR nova.compute.manager [instance: 7211d568-cc68-4e50-80ad-b78878d1deab] self[:] = self._gt.wait() [ 743.426202] env[61806]: ERROR nova.compute.manager [instance: 7211d568-cc68-4e50-80ad-b78878d1deab] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 743.426202] env[61806]: ERROR nova.compute.manager [instance: 7211d568-cc68-4e50-80ad-b78878d1deab] return self._exit_event.wait() [ 743.426202] env[61806]: ERROR nova.compute.manager [instance: 7211d568-cc68-4e50-80ad-b78878d1deab] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 743.426202] env[61806]: ERROR nova.compute.manager [instance: 7211d568-cc68-4e50-80ad-b78878d1deab] result = hub.switch() [ 743.426202] env[61806]: ERROR nova.compute.manager [instance: 7211d568-cc68-4e50-80ad-b78878d1deab] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 743.426202] env[61806]: ERROR nova.compute.manager [instance: 7211d568-cc68-4e50-80ad-b78878d1deab] return self.greenlet.switch() [ 743.426637] env[61806]: ERROR nova.compute.manager [instance: 7211d568-cc68-4e50-80ad-b78878d1deab] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 743.426637] env[61806]: ERROR nova.compute.manager [instance: 7211d568-cc68-4e50-80ad-b78878d1deab] result = function(*args, **kwargs) [ 743.426637] env[61806]: ERROR nova.compute.manager [instance: 7211d568-cc68-4e50-80ad-b78878d1deab] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 743.426637] env[61806]: ERROR nova.compute.manager [instance: 7211d568-cc68-4e50-80ad-b78878d1deab] return func(*args, **kwargs) [ 743.426637] env[61806]: ERROR nova.compute.manager [instance: 7211d568-cc68-4e50-80ad-b78878d1deab] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 743.426637] env[61806]: ERROR nova.compute.manager [instance: 7211d568-cc68-4e50-80ad-b78878d1deab] raise e [ 743.426637] env[61806]: ERROR nova.compute.manager [instance: 7211d568-cc68-4e50-80ad-b78878d1deab] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 743.426637] env[61806]: ERROR nova.compute.manager [instance: 7211d568-cc68-4e50-80ad-b78878d1deab] nwinfo = self.network_api.allocate_for_instance( [ 743.426637] env[61806]: ERROR nova.compute.manager [instance: 7211d568-cc68-4e50-80ad-b78878d1deab] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 743.426637] env[61806]: ERROR nova.compute.manager [instance: 7211d568-cc68-4e50-80ad-b78878d1deab] created_port_ids = self._update_ports_for_instance( [ 743.426637] env[61806]: ERROR nova.compute.manager [instance: 7211d568-cc68-4e50-80ad-b78878d1deab] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 743.426637] env[61806]: ERROR nova.compute.manager [instance: 7211d568-cc68-4e50-80ad-b78878d1deab] with excutils.save_and_reraise_exception(): [ 743.426637] env[61806]: ERROR nova.compute.manager [instance: 7211d568-cc68-4e50-80ad-b78878d1deab] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 743.427069] env[61806]: ERROR nova.compute.manager [instance: 7211d568-cc68-4e50-80ad-b78878d1deab] self.force_reraise() [ 743.427069] env[61806]: ERROR nova.compute.manager [instance: 7211d568-cc68-4e50-80ad-b78878d1deab] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 743.427069] env[61806]: ERROR nova.compute.manager [instance: 7211d568-cc68-4e50-80ad-b78878d1deab] raise self.value [ 743.427069] env[61806]: ERROR nova.compute.manager [instance: 7211d568-cc68-4e50-80ad-b78878d1deab] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 743.427069] env[61806]: ERROR nova.compute.manager [instance: 7211d568-cc68-4e50-80ad-b78878d1deab] updated_port = self._update_port( [ 743.427069] env[61806]: ERROR nova.compute.manager [instance: 7211d568-cc68-4e50-80ad-b78878d1deab] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 743.427069] env[61806]: ERROR nova.compute.manager [instance: 7211d568-cc68-4e50-80ad-b78878d1deab] _ensure_no_port_binding_failure(port) [ 743.427069] env[61806]: ERROR nova.compute.manager [instance: 7211d568-cc68-4e50-80ad-b78878d1deab] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 743.427069] env[61806]: ERROR nova.compute.manager [instance: 7211d568-cc68-4e50-80ad-b78878d1deab] raise exception.PortBindingFailed(port_id=port['id']) [ 743.427069] env[61806]: ERROR nova.compute.manager [instance: 7211d568-cc68-4e50-80ad-b78878d1deab] nova.exception.PortBindingFailed: Binding failed for port 2aa127d2-ff55-470b-822d-60c2d2d409d6, please check neutron logs for more information. [ 743.427069] env[61806]: ERROR nova.compute.manager [instance: 7211d568-cc68-4e50-80ad-b78878d1deab] [ 743.427445] env[61806]: DEBUG nova.compute.utils [None req-3c3981d2-fa25-4f4b-87d4-a5411e67199e tempest-MultipleCreateTestJSON-1589015462 tempest-MultipleCreateTestJSON-1589015462-project-member] [instance: 7211d568-cc68-4e50-80ad-b78878d1deab] Binding failed for port 2aa127d2-ff55-470b-822d-60c2d2d409d6, please check neutron logs for more information. {{(pid=61806) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 743.428224] env[61806]: DEBUG oslo_concurrency.lockutils [None req-3bf756e0-c56e-4902-9c2b-b224a1d09a7c tempest-ListServersNegativeTestJSON-272187588 tempest-ListServersNegativeTestJSON-272187588-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 14.860s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 743.429716] env[61806]: INFO nova.compute.claims [None req-3bf756e0-c56e-4902-9c2b-b224a1d09a7c tempest-ListServersNegativeTestJSON-272187588 tempest-ListServersNegativeTestJSON-272187588-project-member] [instance: ea8d39a1-066d-4eb4-888d-776b0e45f684] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 743.433287] env[61806]: DEBUG nova.compute.manager [None req-3c3981d2-fa25-4f4b-87d4-a5411e67199e tempest-MultipleCreateTestJSON-1589015462 tempest-MultipleCreateTestJSON-1589015462-project-member] [instance: 7211d568-cc68-4e50-80ad-b78878d1deab] Build of instance 7211d568-cc68-4e50-80ad-b78878d1deab was re-scheduled: Binding failed for port 2aa127d2-ff55-470b-822d-60c2d2d409d6, please check neutron logs for more information. {{(pid=61806) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 743.433744] env[61806]: DEBUG nova.compute.manager [None req-3c3981d2-fa25-4f4b-87d4-a5411e67199e tempest-MultipleCreateTestJSON-1589015462 tempest-MultipleCreateTestJSON-1589015462-project-member] [instance: 7211d568-cc68-4e50-80ad-b78878d1deab] Unplugging VIFs for instance {{(pid=61806) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 743.433978] env[61806]: DEBUG oslo_concurrency.lockutils [None req-3c3981d2-fa25-4f4b-87d4-a5411e67199e tempest-MultipleCreateTestJSON-1589015462 tempest-MultipleCreateTestJSON-1589015462-project-member] Acquiring lock "refresh_cache-7211d568-cc68-4e50-80ad-b78878d1deab" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 743.434140] env[61806]: DEBUG oslo_concurrency.lockutils [None req-3c3981d2-fa25-4f4b-87d4-a5411e67199e tempest-MultipleCreateTestJSON-1589015462 tempest-MultipleCreateTestJSON-1589015462-project-member] Acquired lock "refresh_cache-7211d568-cc68-4e50-80ad-b78878d1deab" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 743.434297] env[61806]: DEBUG nova.network.neutron [None req-3c3981d2-fa25-4f4b-87d4-a5411e67199e tempest-MultipleCreateTestJSON-1589015462 tempest-MultipleCreateTestJSON-1589015462-project-member] [instance: 7211d568-cc68-4e50-80ad-b78878d1deab] Building network info cache for instance {{(pid=61806) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 743.516223] env[61806]: DEBUG nova.compute.manager [None req-3bf756e0-c56e-4902-9c2b-b224a1d09a7c tempest-ListServersNegativeTestJSON-272187588 tempest-ListServersNegativeTestJSON-272187588-project-member] [instance: f2446d04-15b8-4b3d-8932-c37df67b0f8f] Start spawning the instance on the hypervisor. {{(pid=61806) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 743.540245] env[61806]: DEBUG nova.virt.hardware [None req-3bf756e0-c56e-4902-9c2b-b224a1d09a7c tempest-ListServersNegativeTestJSON-272187588 tempest-ListServersNegativeTestJSON-272187588-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-15T18:05:10Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-15T18:04:52Z,direct_url=,disk_format='vmdk',id=ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='84ba83fce2c349a6988173c5d6fc3b07',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-15T18:04:53Z,virtual_size=,visibility=), allow threads: False {{(pid=61806) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 743.540511] env[61806]: DEBUG nova.virt.hardware [None req-3bf756e0-c56e-4902-9c2b-b224a1d09a7c tempest-ListServersNegativeTestJSON-272187588 tempest-ListServersNegativeTestJSON-272187588-project-member] Flavor limits 0:0:0 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 743.540672] env[61806]: DEBUG nova.virt.hardware [None req-3bf756e0-c56e-4902-9c2b-b224a1d09a7c tempest-ListServersNegativeTestJSON-272187588 tempest-ListServersNegativeTestJSON-272187588-project-member] Image limits 0:0:0 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 743.540851] env[61806]: DEBUG nova.virt.hardware [None req-3bf756e0-c56e-4902-9c2b-b224a1d09a7c tempest-ListServersNegativeTestJSON-272187588 tempest-ListServersNegativeTestJSON-272187588-project-member] Flavor pref 0:0:0 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 743.540998] env[61806]: DEBUG nova.virt.hardware [None req-3bf756e0-c56e-4902-9c2b-b224a1d09a7c tempest-ListServersNegativeTestJSON-272187588 tempest-ListServersNegativeTestJSON-272187588-project-member] Image pref 0:0:0 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 743.541161] env[61806]: DEBUG nova.virt.hardware [None req-3bf756e0-c56e-4902-9c2b-b224a1d09a7c tempest-ListServersNegativeTestJSON-272187588 tempest-ListServersNegativeTestJSON-272187588-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 743.541363] env[61806]: DEBUG nova.virt.hardware [None req-3bf756e0-c56e-4902-9c2b-b224a1d09a7c tempest-ListServersNegativeTestJSON-272187588 tempest-ListServersNegativeTestJSON-272187588-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61806) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 743.541565] env[61806]: DEBUG nova.virt.hardware [None req-3bf756e0-c56e-4902-9c2b-b224a1d09a7c tempest-ListServersNegativeTestJSON-272187588 tempest-ListServersNegativeTestJSON-272187588-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61806) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 743.541748] env[61806]: DEBUG nova.virt.hardware [None req-3bf756e0-c56e-4902-9c2b-b224a1d09a7c tempest-ListServersNegativeTestJSON-272187588 tempest-ListServersNegativeTestJSON-272187588-project-member] Got 1 possible topologies {{(pid=61806) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 743.541911] env[61806]: DEBUG nova.virt.hardware [None req-3bf756e0-c56e-4902-9c2b-b224a1d09a7c tempest-ListServersNegativeTestJSON-272187588 tempest-ListServersNegativeTestJSON-272187588-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61806) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 743.542104] env[61806]: DEBUG nova.virt.hardware [None req-3bf756e0-c56e-4902-9c2b-b224a1d09a7c tempest-ListServersNegativeTestJSON-272187588 tempest-ListServersNegativeTestJSON-272187588-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61806) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 743.543197] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-91af1229-292a-4281-9bbf-02122fe654b6 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 743.551095] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9fe9206b-ad50-45fd-9adb-8c067d810c91 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 743.564489] env[61806]: ERROR nova.compute.manager [None req-3bf756e0-c56e-4902-9c2b-b224a1d09a7c tempest-ListServersNegativeTestJSON-272187588 tempest-ListServersNegativeTestJSON-272187588-project-member] [instance: f2446d04-15b8-4b3d-8932-c37df67b0f8f] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port ad624cd6-204b-4189-a855-10dca931bd79, please check neutron logs for more information. [ 743.564489] env[61806]: ERROR nova.compute.manager [instance: f2446d04-15b8-4b3d-8932-c37df67b0f8f] Traceback (most recent call last): [ 743.564489] env[61806]: ERROR nova.compute.manager [instance: f2446d04-15b8-4b3d-8932-c37df67b0f8f] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 743.564489] env[61806]: ERROR nova.compute.manager [instance: f2446d04-15b8-4b3d-8932-c37df67b0f8f] yield resources [ 743.564489] env[61806]: ERROR nova.compute.manager [instance: f2446d04-15b8-4b3d-8932-c37df67b0f8f] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 743.564489] env[61806]: ERROR nova.compute.manager [instance: f2446d04-15b8-4b3d-8932-c37df67b0f8f] self.driver.spawn(context, instance, image_meta, [ 743.564489] env[61806]: ERROR nova.compute.manager [instance: f2446d04-15b8-4b3d-8932-c37df67b0f8f] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 743.564489] env[61806]: ERROR nova.compute.manager [instance: f2446d04-15b8-4b3d-8932-c37df67b0f8f] self._vmops.spawn(context, instance, image_meta, injected_files, [ 743.564489] env[61806]: ERROR nova.compute.manager [instance: f2446d04-15b8-4b3d-8932-c37df67b0f8f] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 743.564489] env[61806]: ERROR nova.compute.manager [instance: f2446d04-15b8-4b3d-8932-c37df67b0f8f] vm_ref = self.build_virtual_machine(instance, [ 743.564489] env[61806]: ERROR nova.compute.manager [instance: f2446d04-15b8-4b3d-8932-c37df67b0f8f] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 743.564924] env[61806]: ERROR nova.compute.manager [instance: f2446d04-15b8-4b3d-8932-c37df67b0f8f] vif_infos = vmwarevif.get_vif_info(self._session, [ 743.564924] env[61806]: ERROR nova.compute.manager [instance: f2446d04-15b8-4b3d-8932-c37df67b0f8f] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 743.564924] env[61806]: ERROR nova.compute.manager [instance: f2446d04-15b8-4b3d-8932-c37df67b0f8f] for vif in network_info: [ 743.564924] env[61806]: ERROR nova.compute.manager [instance: f2446d04-15b8-4b3d-8932-c37df67b0f8f] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 743.564924] env[61806]: ERROR nova.compute.manager [instance: f2446d04-15b8-4b3d-8932-c37df67b0f8f] return self._sync_wrapper(fn, *args, **kwargs) [ 743.564924] env[61806]: ERROR nova.compute.manager [instance: f2446d04-15b8-4b3d-8932-c37df67b0f8f] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 743.564924] env[61806]: ERROR nova.compute.manager [instance: f2446d04-15b8-4b3d-8932-c37df67b0f8f] self.wait() [ 743.564924] env[61806]: ERROR nova.compute.manager [instance: f2446d04-15b8-4b3d-8932-c37df67b0f8f] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 743.564924] env[61806]: ERROR nova.compute.manager [instance: f2446d04-15b8-4b3d-8932-c37df67b0f8f] self[:] = self._gt.wait() [ 743.564924] env[61806]: ERROR nova.compute.manager [instance: f2446d04-15b8-4b3d-8932-c37df67b0f8f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 743.564924] env[61806]: ERROR nova.compute.manager [instance: f2446d04-15b8-4b3d-8932-c37df67b0f8f] return self._exit_event.wait() [ 743.564924] env[61806]: ERROR nova.compute.manager [instance: f2446d04-15b8-4b3d-8932-c37df67b0f8f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 743.564924] env[61806]: ERROR nova.compute.manager [instance: f2446d04-15b8-4b3d-8932-c37df67b0f8f] current.throw(*self._exc) [ 743.565366] env[61806]: ERROR nova.compute.manager [instance: f2446d04-15b8-4b3d-8932-c37df67b0f8f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 743.565366] env[61806]: ERROR nova.compute.manager [instance: f2446d04-15b8-4b3d-8932-c37df67b0f8f] result = function(*args, **kwargs) [ 743.565366] env[61806]: ERROR nova.compute.manager [instance: f2446d04-15b8-4b3d-8932-c37df67b0f8f] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 743.565366] env[61806]: ERROR nova.compute.manager [instance: f2446d04-15b8-4b3d-8932-c37df67b0f8f] return func(*args, **kwargs) [ 743.565366] env[61806]: ERROR nova.compute.manager [instance: f2446d04-15b8-4b3d-8932-c37df67b0f8f] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 743.565366] env[61806]: ERROR nova.compute.manager [instance: f2446d04-15b8-4b3d-8932-c37df67b0f8f] raise e [ 743.565366] env[61806]: ERROR nova.compute.manager [instance: f2446d04-15b8-4b3d-8932-c37df67b0f8f] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 743.565366] env[61806]: ERROR nova.compute.manager [instance: f2446d04-15b8-4b3d-8932-c37df67b0f8f] nwinfo = self.network_api.allocate_for_instance( [ 743.565366] env[61806]: ERROR nova.compute.manager [instance: f2446d04-15b8-4b3d-8932-c37df67b0f8f] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 743.565366] env[61806]: ERROR nova.compute.manager [instance: f2446d04-15b8-4b3d-8932-c37df67b0f8f] created_port_ids = self._update_ports_for_instance( [ 743.565366] env[61806]: ERROR nova.compute.manager [instance: f2446d04-15b8-4b3d-8932-c37df67b0f8f] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 743.565366] env[61806]: ERROR nova.compute.manager [instance: f2446d04-15b8-4b3d-8932-c37df67b0f8f] with excutils.save_and_reraise_exception(): [ 743.565366] env[61806]: ERROR nova.compute.manager [instance: f2446d04-15b8-4b3d-8932-c37df67b0f8f] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 743.566084] env[61806]: ERROR nova.compute.manager [instance: f2446d04-15b8-4b3d-8932-c37df67b0f8f] self.force_reraise() [ 743.566084] env[61806]: ERROR nova.compute.manager [instance: f2446d04-15b8-4b3d-8932-c37df67b0f8f] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 743.566084] env[61806]: ERROR nova.compute.manager [instance: f2446d04-15b8-4b3d-8932-c37df67b0f8f] raise self.value [ 743.566084] env[61806]: ERROR nova.compute.manager [instance: f2446d04-15b8-4b3d-8932-c37df67b0f8f] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 743.566084] env[61806]: ERROR nova.compute.manager [instance: f2446d04-15b8-4b3d-8932-c37df67b0f8f] updated_port = self._update_port( [ 743.566084] env[61806]: ERROR nova.compute.manager [instance: f2446d04-15b8-4b3d-8932-c37df67b0f8f] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 743.566084] env[61806]: ERROR nova.compute.manager [instance: f2446d04-15b8-4b3d-8932-c37df67b0f8f] _ensure_no_port_binding_failure(port) [ 743.566084] env[61806]: ERROR nova.compute.manager [instance: f2446d04-15b8-4b3d-8932-c37df67b0f8f] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 743.566084] env[61806]: ERROR nova.compute.manager [instance: f2446d04-15b8-4b3d-8932-c37df67b0f8f] raise exception.PortBindingFailed(port_id=port['id']) [ 743.566084] env[61806]: ERROR nova.compute.manager [instance: f2446d04-15b8-4b3d-8932-c37df67b0f8f] nova.exception.PortBindingFailed: Binding failed for port ad624cd6-204b-4189-a855-10dca931bd79, please check neutron logs for more information. [ 743.566084] env[61806]: ERROR nova.compute.manager [instance: f2446d04-15b8-4b3d-8932-c37df67b0f8f] [ 743.566084] env[61806]: INFO nova.compute.manager [None req-3bf756e0-c56e-4902-9c2b-b224a1d09a7c tempest-ListServersNegativeTestJSON-272187588 tempest-ListServersNegativeTestJSON-272187588-project-member] [instance: f2446d04-15b8-4b3d-8932-c37df67b0f8f] Terminating instance [ 743.566739] env[61806]: DEBUG oslo_concurrency.lockutils [None req-3bf756e0-c56e-4902-9c2b-b224a1d09a7c tempest-ListServersNegativeTestJSON-272187588 tempest-ListServersNegativeTestJSON-272187588-project-member] Acquiring lock "refresh_cache-f2446d04-15b8-4b3d-8932-c37df67b0f8f" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 743.794041] env[61806]: DEBUG nova.network.neutron [req-471db471-fc87-4586-83cc-b32bd7c6934a req-5d30ec46-1a46-4bc6-97ca-7483698ad66d service nova] [instance: f2446d04-15b8-4b3d-8932-c37df67b0f8f] Instance cache missing network info. {{(pid=61806) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 743.880272] env[61806]: DEBUG nova.network.neutron [req-471db471-fc87-4586-83cc-b32bd7c6934a req-5d30ec46-1a46-4bc6-97ca-7483698ad66d service nova] [instance: f2446d04-15b8-4b3d-8932-c37df67b0f8f] Updating instance_info_cache with network_info: [] {{(pid=61806) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 743.959412] env[61806]: DEBUG nova.network.neutron [None req-3c3981d2-fa25-4f4b-87d4-a5411e67199e tempest-MultipleCreateTestJSON-1589015462 tempest-MultipleCreateTestJSON-1589015462-project-member] [instance: 7211d568-cc68-4e50-80ad-b78878d1deab] Instance cache missing network info. {{(pid=61806) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 744.030330] env[61806]: DEBUG nova.network.neutron [None req-3c3981d2-fa25-4f4b-87d4-a5411e67199e tempest-MultipleCreateTestJSON-1589015462 tempest-MultipleCreateTestJSON-1589015462-project-member] [instance: 7211d568-cc68-4e50-80ad-b78878d1deab] Updating instance_info_cache with network_info: [] {{(pid=61806) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 744.383361] env[61806]: DEBUG oslo_concurrency.lockutils [req-471db471-fc87-4586-83cc-b32bd7c6934a req-5d30ec46-1a46-4bc6-97ca-7483698ad66d service nova] Releasing lock "refresh_cache-f2446d04-15b8-4b3d-8932-c37df67b0f8f" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 744.383946] env[61806]: DEBUG oslo_concurrency.lockutils [None req-3bf756e0-c56e-4902-9c2b-b224a1d09a7c tempest-ListServersNegativeTestJSON-272187588 tempest-ListServersNegativeTestJSON-272187588-project-member] Acquired lock "refresh_cache-f2446d04-15b8-4b3d-8932-c37df67b0f8f" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 744.385028] env[61806]: DEBUG nova.network.neutron [None req-3bf756e0-c56e-4902-9c2b-b224a1d09a7c tempest-ListServersNegativeTestJSON-272187588 tempest-ListServersNegativeTestJSON-272187588-project-member] [instance: f2446d04-15b8-4b3d-8932-c37df67b0f8f] Building network info cache for instance {{(pid=61806) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 744.533369] env[61806]: DEBUG oslo_concurrency.lockutils [None req-3c3981d2-fa25-4f4b-87d4-a5411e67199e tempest-MultipleCreateTestJSON-1589015462 tempest-MultipleCreateTestJSON-1589015462-project-member] Releasing lock "refresh_cache-7211d568-cc68-4e50-80ad-b78878d1deab" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 744.533602] env[61806]: DEBUG nova.compute.manager [None req-3c3981d2-fa25-4f4b-87d4-a5411e67199e tempest-MultipleCreateTestJSON-1589015462 tempest-MultipleCreateTestJSON-1589015462-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61806) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 744.533813] env[61806]: DEBUG nova.compute.manager [None req-3c3981d2-fa25-4f4b-87d4-a5411e67199e tempest-MultipleCreateTestJSON-1589015462 tempest-MultipleCreateTestJSON-1589015462-project-member] [instance: 7211d568-cc68-4e50-80ad-b78878d1deab] Deallocating network for instance {{(pid=61806) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 744.533956] env[61806]: DEBUG nova.network.neutron [None req-3c3981d2-fa25-4f4b-87d4-a5411e67199e tempest-MultipleCreateTestJSON-1589015462 tempest-MultipleCreateTestJSON-1589015462-project-member] [instance: 7211d568-cc68-4e50-80ad-b78878d1deab] deallocate_for_instance() {{(pid=61806) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 744.548300] env[61806]: DEBUG nova.network.neutron [None req-3c3981d2-fa25-4f4b-87d4-a5411e67199e tempest-MultipleCreateTestJSON-1589015462 tempest-MultipleCreateTestJSON-1589015462-project-member] [instance: 7211d568-cc68-4e50-80ad-b78878d1deab] Instance cache missing network info. {{(pid=61806) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 744.697791] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3b529ac2-8f5c-4d78-8bc8-3ad2124c75b4 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 744.705321] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-25beb589-96ab-4d06-9cd7-b8d74f13c4e3 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 744.736704] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4e3318a5-a7f9-4089-b73d-e88fc858cae1 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 744.744394] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-999fe3a0-a8f3-40d5-a3d2-77108f3a5962 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 744.757190] env[61806]: DEBUG nova.compute.provider_tree [None req-3bf756e0-c56e-4902-9c2b-b224a1d09a7c tempest-ListServersNegativeTestJSON-272187588 tempest-ListServersNegativeTestJSON-272187588-project-member] Inventory has not changed in ProviderTree for provider: a2858be1-fd22-4e08-979e-87ad25293407 {{(pid=61806) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 744.905239] env[61806]: DEBUG nova.network.neutron [None req-3bf756e0-c56e-4902-9c2b-b224a1d09a7c tempest-ListServersNegativeTestJSON-272187588 tempest-ListServersNegativeTestJSON-272187588-project-member] [instance: f2446d04-15b8-4b3d-8932-c37df67b0f8f] Instance cache missing network info. {{(pid=61806) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 744.983157] env[61806]: DEBUG nova.network.neutron [None req-3bf756e0-c56e-4902-9c2b-b224a1d09a7c tempest-ListServersNegativeTestJSON-272187588 tempest-ListServersNegativeTestJSON-272187588-project-member] [instance: f2446d04-15b8-4b3d-8932-c37df67b0f8f] Updating instance_info_cache with network_info: [] {{(pid=61806) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 745.053354] env[61806]: DEBUG nova.network.neutron [None req-3c3981d2-fa25-4f4b-87d4-a5411e67199e tempest-MultipleCreateTestJSON-1589015462 tempest-MultipleCreateTestJSON-1589015462-project-member] [instance: 7211d568-cc68-4e50-80ad-b78878d1deab] Updating instance_info_cache with network_info: [] {{(pid=61806) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 745.260173] env[61806]: DEBUG nova.scheduler.client.report [None req-3bf756e0-c56e-4902-9c2b-b224a1d09a7c tempest-ListServersNegativeTestJSON-272187588 tempest-ListServersNegativeTestJSON-272187588-project-member] Inventory has not changed for provider a2858be1-fd22-4e08-979e-87ad25293407 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 139, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61806) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 745.300962] env[61806]: DEBUG nova.compute.manager [req-cf1cfd35-99b9-4ab0-b2eb-cf2f72bc8ad3 req-45f8965b-3a5b-4d6c-a2ed-01f71da2482d service nova] [instance: f2446d04-15b8-4b3d-8932-c37df67b0f8f] Received event network-vif-deleted-ad624cd6-204b-4189-a855-10dca931bd79 {{(pid=61806) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 745.486343] env[61806]: DEBUG oslo_concurrency.lockutils [None req-3bf756e0-c56e-4902-9c2b-b224a1d09a7c tempest-ListServersNegativeTestJSON-272187588 tempest-ListServersNegativeTestJSON-272187588-project-member] Releasing lock "refresh_cache-f2446d04-15b8-4b3d-8932-c37df67b0f8f" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 745.486696] env[61806]: DEBUG nova.compute.manager [None req-3bf756e0-c56e-4902-9c2b-b224a1d09a7c tempest-ListServersNegativeTestJSON-272187588 tempest-ListServersNegativeTestJSON-272187588-project-member] [instance: f2446d04-15b8-4b3d-8932-c37df67b0f8f] Start destroying the instance on the hypervisor. {{(pid=61806) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 745.486899] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-3bf756e0-c56e-4902-9c2b-b224a1d09a7c tempest-ListServersNegativeTestJSON-272187588 tempest-ListServersNegativeTestJSON-272187588-project-member] [instance: f2446d04-15b8-4b3d-8932-c37df67b0f8f] Destroying instance {{(pid=61806) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 745.487225] env[61806]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-e2e63d90-46fd-4b7f-abd8-4df061f5ca5a {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 745.496600] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9bea031a-1518-476e-8960-9ecff25bd63a {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 745.517750] env[61806]: WARNING nova.virt.vmwareapi.vmops [None req-3bf756e0-c56e-4902-9c2b-b224a1d09a7c tempest-ListServersNegativeTestJSON-272187588 tempest-ListServersNegativeTestJSON-272187588-project-member] [instance: f2446d04-15b8-4b3d-8932-c37df67b0f8f] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance f2446d04-15b8-4b3d-8932-c37df67b0f8f could not be found. [ 745.517984] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-3bf756e0-c56e-4902-9c2b-b224a1d09a7c tempest-ListServersNegativeTestJSON-272187588 tempest-ListServersNegativeTestJSON-272187588-project-member] [instance: f2446d04-15b8-4b3d-8932-c37df67b0f8f] Instance destroyed {{(pid=61806) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 745.518185] env[61806]: INFO nova.compute.manager [None req-3bf756e0-c56e-4902-9c2b-b224a1d09a7c tempest-ListServersNegativeTestJSON-272187588 tempest-ListServersNegativeTestJSON-272187588-project-member] [instance: f2446d04-15b8-4b3d-8932-c37df67b0f8f] Took 0.03 seconds to destroy the instance on the hypervisor. [ 745.518431] env[61806]: DEBUG oslo.service.loopingcall [None req-3bf756e0-c56e-4902-9c2b-b224a1d09a7c tempest-ListServersNegativeTestJSON-272187588 tempest-ListServersNegativeTestJSON-272187588-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61806) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 745.518649] env[61806]: DEBUG nova.compute.manager [-] [instance: f2446d04-15b8-4b3d-8932-c37df67b0f8f] Deallocating network for instance {{(pid=61806) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 745.518742] env[61806]: DEBUG nova.network.neutron [-] [instance: f2446d04-15b8-4b3d-8932-c37df67b0f8f] deallocate_for_instance() {{(pid=61806) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 745.532979] env[61806]: DEBUG nova.network.neutron [-] [instance: f2446d04-15b8-4b3d-8932-c37df67b0f8f] Instance cache missing network info. {{(pid=61806) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 745.555928] env[61806]: INFO nova.compute.manager [None req-3c3981d2-fa25-4f4b-87d4-a5411e67199e tempest-MultipleCreateTestJSON-1589015462 tempest-MultipleCreateTestJSON-1589015462-project-member] [instance: 7211d568-cc68-4e50-80ad-b78878d1deab] Took 1.02 seconds to deallocate network for instance. [ 745.765312] env[61806]: DEBUG oslo_concurrency.lockutils [None req-3bf756e0-c56e-4902-9c2b-b224a1d09a7c tempest-ListServersNegativeTestJSON-272187588 tempest-ListServersNegativeTestJSON-272187588-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.337s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 745.765844] env[61806]: DEBUG nova.compute.manager [None req-3bf756e0-c56e-4902-9c2b-b224a1d09a7c tempest-ListServersNegativeTestJSON-272187588 tempest-ListServersNegativeTestJSON-272187588-project-member] [instance: ea8d39a1-066d-4eb4-888d-776b0e45f684] Start building networks asynchronously for instance. {{(pid=61806) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 745.768373] env[61806]: DEBUG oslo_concurrency.lockutils [None req-3bf756e0-c56e-4902-9c2b-b224a1d09a7c tempest-ListServersNegativeTestJSON-272187588 tempest-ListServersNegativeTestJSON-272187588-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 15.206s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 745.770265] env[61806]: INFO nova.compute.claims [None req-3bf756e0-c56e-4902-9c2b-b224a1d09a7c tempest-ListServersNegativeTestJSON-272187588 tempest-ListServersNegativeTestJSON-272187588-project-member] [instance: 198dc1a4-050d-47b2-8bc4-30af5534a789] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 746.035900] env[61806]: DEBUG nova.network.neutron [-] [instance: f2446d04-15b8-4b3d-8932-c37df67b0f8f] Updating instance_info_cache with network_info: [] {{(pid=61806) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 746.275048] env[61806]: DEBUG nova.compute.utils [None req-3bf756e0-c56e-4902-9c2b-b224a1d09a7c tempest-ListServersNegativeTestJSON-272187588 tempest-ListServersNegativeTestJSON-272187588-project-member] Using /dev/sd instead of None {{(pid=61806) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 746.278532] env[61806]: DEBUG nova.compute.manager [None req-3bf756e0-c56e-4902-9c2b-b224a1d09a7c tempest-ListServersNegativeTestJSON-272187588 tempest-ListServersNegativeTestJSON-272187588-project-member] [instance: ea8d39a1-066d-4eb4-888d-776b0e45f684] Allocating IP information in the background. {{(pid=61806) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 746.278722] env[61806]: DEBUG nova.network.neutron [None req-3bf756e0-c56e-4902-9c2b-b224a1d09a7c tempest-ListServersNegativeTestJSON-272187588 tempest-ListServersNegativeTestJSON-272187588-project-member] [instance: ea8d39a1-066d-4eb4-888d-776b0e45f684] allocate_for_instance() {{(pid=61806) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 746.315836] env[61806]: DEBUG nova.policy [None req-3bf756e0-c56e-4902-9c2b-b224a1d09a7c tempest-ListServersNegativeTestJSON-272187588 tempest-ListServersNegativeTestJSON-272187588-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '6228fec9ded14ef9974634c9a872374e', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'fb3d07cc8f764d15bc6fab0f299da5e8', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61806) authorize /opt/stack/nova/nova/policy.py:201}} [ 746.538836] env[61806]: INFO nova.compute.manager [-] [instance: f2446d04-15b8-4b3d-8932-c37df67b0f8f] Took 1.02 seconds to deallocate network for instance. [ 746.543023] env[61806]: DEBUG nova.compute.claims [None req-3bf756e0-c56e-4902-9c2b-b224a1d09a7c tempest-ListServersNegativeTestJSON-272187588 tempest-ListServersNegativeTestJSON-272187588-project-member] [instance: f2446d04-15b8-4b3d-8932-c37df67b0f8f] Aborting claim: {{(pid=61806) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 746.543023] env[61806]: DEBUG oslo_concurrency.lockutils [None req-3bf756e0-c56e-4902-9c2b-b224a1d09a7c tempest-ListServersNegativeTestJSON-272187588 tempest-ListServersNegativeTestJSON-272187588-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 746.592012] env[61806]: INFO nova.scheduler.client.report [None req-3c3981d2-fa25-4f4b-87d4-a5411e67199e tempest-MultipleCreateTestJSON-1589015462 tempest-MultipleCreateTestJSON-1589015462-project-member] Deleted allocations for instance 7211d568-cc68-4e50-80ad-b78878d1deab [ 746.599014] env[61806]: DEBUG nova.network.neutron [None req-3bf756e0-c56e-4902-9c2b-b224a1d09a7c tempest-ListServersNegativeTestJSON-272187588 tempest-ListServersNegativeTestJSON-272187588-project-member] [instance: ea8d39a1-066d-4eb4-888d-776b0e45f684] Successfully created port: d1f266fb-3da7-4f92-b3cb-81aa9f3b924c {{(pid=61806) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 746.779998] env[61806]: DEBUG nova.compute.manager [None req-3bf756e0-c56e-4902-9c2b-b224a1d09a7c tempest-ListServersNegativeTestJSON-272187588 tempest-ListServersNegativeTestJSON-272187588-project-member] [instance: ea8d39a1-066d-4eb4-888d-776b0e45f684] Start building block device mappings for instance. {{(pid=61806) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 747.068116] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-685f5bf7-fb02-4624-9285-f7e426a9c791 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 747.076980] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b4d3ad3e-12a9-4640-8bd0-3106db8092a8 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 747.107438] env[61806]: DEBUG oslo_concurrency.lockutils [None req-3c3981d2-fa25-4f4b-87d4-a5411e67199e tempest-MultipleCreateTestJSON-1589015462 tempest-MultipleCreateTestJSON-1589015462-project-member] Lock "7211d568-cc68-4e50-80ad-b78878d1deab" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 157.436s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 747.109332] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f525e498-c7b5-4077-8653-a68d3dc27916 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 747.117141] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f4948015-52c1-4e82-ade4-40d728e07a64 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 747.132374] env[61806]: DEBUG nova.compute.provider_tree [None req-3bf756e0-c56e-4902-9c2b-b224a1d09a7c tempest-ListServersNegativeTestJSON-272187588 tempest-ListServersNegativeTestJSON-272187588-project-member] Inventory has not changed in ProviderTree for provider: a2858be1-fd22-4e08-979e-87ad25293407 {{(pid=61806) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 747.440677] env[61806]: DEBUG nova.compute.manager [req-8fb626b3-e86a-4aed-bb44-d3816464e40d req-135b21d2-37be-48a2-9675-c271dc5135e7 service nova] [instance: ea8d39a1-066d-4eb4-888d-776b0e45f684] Received event network-changed-d1f266fb-3da7-4f92-b3cb-81aa9f3b924c {{(pid=61806) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 747.440875] env[61806]: DEBUG nova.compute.manager [req-8fb626b3-e86a-4aed-bb44-d3816464e40d req-135b21d2-37be-48a2-9675-c271dc5135e7 service nova] [instance: ea8d39a1-066d-4eb4-888d-776b0e45f684] Refreshing instance network info cache due to event network-changed-d1f266fb-3da7-4f92-b3cb-81aa9f3b924c. {{(pid=61806) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 747.441075] env[61806]: DEBUG oslo_concurrency.lockutils [req-8fb626b3-e86a-4aed-bb44-d3816464e40d req-135b21d2-37be-48a2-9675-c271dc5135e7 service nova] Acquiring lock "refresh_cache-ea8d39a1-066d-4eb4-888d-776b0e45f684" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 747.441222] env[61806]: DEBUG oslo_concurrency.lockutils [req-8fb626b3-e86a-4aed-bb44-d3816464e40d req-135b21d2-37be-48a2-9675-c271dc5135e7 service nova] Acquired lock "refresh_cache-ea8d39a1-066d-4eb4-888d-776b0e45f684" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 747.441379] env[61806]: DEBUG nova.network.neutron [req-8fb626b3-e86a-4aed-bb44-d3816464e40d req-135b21d2-37be-48a2-9675-c271dc5135e7 service nova] [instance: ea8d39a1-066d-4eb4-888d-776b0e45f684] Refreshing network info cache for port d1f266fb-3da7-4f92-b3cb-81aa9f3b924c {{(pid=61806) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 747.613406] env[61806]: ERROR nova.compute.manager [None req-3bf756e0-c56e-4902-9c2b-b224a1d09a7c tempest-ListServersNegativeTestJSON-272187588 tempest-ListServersNegativeTestJSON-272187588-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port d1f266fb-3da7-4f92-b3cb-81aa9f3b924c, please check neutron logs for more information. [ 747.613406] env[61806]: ERROR nova.compute.manager Traceback (most recent call last): [ 747.613406] env[61806]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 747.613406] env[61806]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 747.613406] env[61806]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 747.613406] env[61806]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 747.613406] env[61806]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 747.613406] env[61806]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 747.613406] env[61806]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 747.613406] env[61806]: ERROR nova.compute.manager self.force_reraise() [ 747.613406] env[61806]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 747.613406] env[61806]: ERROR nova.compute.manager raise self.value [ 747.613406] env[61806]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 747.613406] env[61806]: ERROR nova.compute.manager updated_port = self._update_port( [ 747.613406] env[61806]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 747.613406] env[61806]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 747.614211] env[61806]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 747.614211] env[61806]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 747.614211] env[61806]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port d1f266fb-3da7-4f92-b3cb-81aa9f3b924c, please check neutron logs for more information. [ 747.614211] env[61806]: ERROR nova.compute.manager [ 747.614211] env[61806]: Traceback (most recent call last): [ 747.614211] env[61806]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 747.614211] env[61806]: listener.cb(fileno) [ 747.614211] env[61806]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 747.614211] env[61806]: result = function(*args, **kwargs) [ 747.614211] env[61806]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 747.614211] env[61806]: return func(*args, **kwargs) [ 747.614211] env[61806]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 747.614211] env[61806]: raise e [ 747.614211] env[61806]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 747.614211] env[61806]: nwinfo = self.network_api.allocate_for_instance( [ 747.614211] env[61806]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 747.614211] env[61806]: created_port_ids = self._update_ports_for_instance( [ 747.614211] env[61806]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 747.614211] env[61806]: with excutils.save_and_reraise_exception(): [ 747.614211] env[61806]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 747.614211] env[61806]: self.force_reraise() [ 747.614211] env[61806]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 747.614211] env[61806]: raise self.value [ 747.614211] env[61806]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 747.614211] env[61806]: updated_port = self._update_port( [ 747.614211] env[61806]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 747.614211] env[61806]: _ensure_no_port_binding_failure(port) [ 747.614211] env[61806]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 747.614211] env[61806]: raise exception.PortBindingFailed(port_id=port['id']) [ 747.615206] env[61806]: nova.exception.PortBindingFailed: Binding failed for port d1f266fb-3da7-4f92-b3cb-81aa9f3b924c, please check neutron logs for more information. [ 747.615206] env[61806]: Removing descriptor: 14 [ 747.615206] env[61806]: DEBUG nova.compute.manager [None req-35a8f95d-1ddd-492f-a93c-adb382770732 tempest-ServersV294TestFqdnHostnames-1709116086 tempest-ServersV294TestFqdnHostnames-1709116086-project-member] [instance: c25ddf7c-d0cc-4b73-96da-1dcd6012f072] Starting instance... {{(pid=61806) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 747.635086] env[61806]: DEBUG nova.scheduler.client.report [None req-3bf756e0-c56e-4902-9c2b-b224a1d09a7c tempest-ListServersNegativeTestJSON-272187588 tempest-ListServersNegativeTestJSON-272187588-project-member] Inventory has not changed for provider a2858be1-fd22-4e08-979e-87ad25293407 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 139, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61806) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 747.793191] env[61806]: DEBUG nova.compute.manager [None req-3bf756e0-c56e-4902-9c2b-b224a1d09a7c tempest-ListServersNegativeTestJSON-272187588 tempest-ListServersNegativeTestJSON-272187588-project-member] [instance: ea8d39a1-066d-4eb4-888d-776b0e45f684] Start spawning the instance on the hypervisor. {{(pid=61806) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 747.821043] env[61806]: DEBUG nova.virt.hardware [None req-3bf756e0-c56e-4902-9c2b-b224a1d09a7c tempest-ListServersNegativeTestJSON-272187588 tempest-ListServersNegativeTestJSON-272187588-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-15T18:05:10Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-15T18:04:52Z,direct_url=,disk_format='vmdk',id=ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='84ba83fce2c349a6988173c5d6fc3b07',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-15T18:04:53Z,virtual_size=,visibility=), allow threads: False {{(pid=61806) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 747.821043] env[61806]: DEBUG nova.virt.hardware [None req-3bf756e0-c56e-4902-9c2b-b224a1d09a7c tempest-ListServersNegativeTestJSON-272187588 tempest-ListServersNegativeTestJSON-272187588-project-member] Flavor limits 0:0:0 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 747.821043] env[61806]: DEBUG nova.virt.hardware [None req-3bf756e0-c56e-4902-9c2b-b224a1d09a7c tempest-ListServersNegativeTestJSON-272187588 tempest-ListServersNegativeTestJSON-272187588-project-member] Image limits 0:0:0 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 747.821265] env[61806]: DEBUG nova.virt.hardware [None req-3bf756e0-c56e-4902-9c2b-b224a1d09a7c tempest-ListServersNegativeTestJSON-272187588 tempest-ListServersNegativeTestJSON-272187588-project-member] Flavor pref 0:0:0 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 747.821265] env[61806]: DEBUG nova.virt.hardware [None req-3bf756e0-c56e-4902-9c2b-b224a1d09a7c tempest-ListServersNegativeTestJSON-272187588 tempest-ListServersNegativeTestJSON-272187588-project-member] Image pref 0:0:0 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 747.821265] env[61806]: DEBUG nova.virt.hardware [None req-3bf756e0-c56e-4902-9c2b-b224a1d09a7c tempest-ListServersNegativeTestJSON-272187588 tempest-ListServersNegativeTestJSON-272187588-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 747.821265] env[61806]: DEBUG nova.virt.hardware [None req-3bf756e0-c56e-4902-9c2b-b224a1d09a7c tempest-ListServersNegativeTestJSON-272187588 tempest-ListServersNegativeTestJSON-272187588-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61806) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 747.821265] env[61806]: DEBUG nova.virt.hardware [None req-3bf756e0-c56e-4902-9c2b-b224a1d09a7c tempest-ListServersNegativeTestJSON-272187588 tempest-ListServersNegativeTestJSON-272187588-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61806) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 747.821469] env[61806]: DEBUG nova.virt.hardware [None req-3bf756e0-c56e-4902-9c2b-b224a1d09a7c tempest-ListServersNegativeTestJSON-272187588 tempest-ListServersNegativeTestJSON-272187588-project-member] Got 1 possible topologies {{(pid=61806) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 747.821722] env[61806]: DEBUG nova.virt.hardware [None req-3bf756e0-c56e-4902-9c2b-b224a1d09a7c tempest-ListServersNegativeTestJSON-272187588 tempest-ListServersNegativeTestJSON-272187588-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61806) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 747.822050] env[61806]: DEBUG nova.virt.hardware [None req-3bf756e0-c56e-4902-9c2b-b224a1d09a7c tempest-ListServersNegativeTestJSON-272187588 tempest-ListServersNegativeTestJSON-272187588-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61806) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 747.823072] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9b86ff6a-3503-4a00-ab86-407b95b4d796 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 747.832011] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f5dbcc13-b11c-4bef-a652-83e0d824ccf8 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 747.846011] env[61806]: ERROR nova.compute.manager [None req-3bf756e0-c56e-4902-9c2b-b224a1d09a7c tempest-ListServersNegativeTestJSON-272187588 tempest-ListServersNegativeTestJSON-272187588-project-member] [instance: ea8d39a1-066d-4eb4-888d-776b0e45f684] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port d1f266fb-3da7-4f92-b3cb-81aa9f3b924c, please check neutron logs for more information. [ 747.846011] env[61806]: ERROR nova.compute.manager [instance: ea8d39a1-066d-4eb4-888d-776b0e45f684] Traceback (most recent call last): [ 747.846011] env[61806]: ERROR nova.compute.manager [instance: ea8d39a1-066d-4eb4-888d-776b0e45f684] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 747.846011] env[61806]: ERROR nova.compute.manager [instance: ea8d39a1-066d-4eb4-888d-776b0e45f684] yield resources [ 747.846011] env[61806]: ERROR nova.compute.manager [instance: ea8d39a1-066d-4eb4-888d-776b0e45f684] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 747.846011] env[61806]: ERROR nova.compute.manager [instance: ea8d39a1-066d-4eb4-888d-776b0e45f684] self.driver.spawn(context, instance, image_meta, [ 747.846011] env[61806]: ERROR nova.compute.manager [instance: ea8d39a1-066d-4eb4-888d-776b0e45f684] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 747.846011] env[61806]: ERROR nova.compute.manager [instance: ea8d39a1-066d-4eb4-888d-776b0e45f684] self._vmops.spawn(context, instance, image_meta, injected_files, [ 747.846011] env[61806]: ERROR nova.compute.manager [instance: ea8d39a1-066d-4eb4-888d-776b0e45f684] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 747.846011] env[61806]: ERROR nova.compute.manager [instance: ea8d39a1-066d-4eb4-888d-776b0e45f684] vm_ref = self.build_virtual_machine(instance, [ 747.846011] env[61806]: ERROR nova.compute.manager [instance: ea8d39a1-066d-4eb4-888d-776b0e45f684] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 747.846465] env[61806]: ERROR nova.compute.manager [instance: ea8d39a1-066d-4eb4-888d-776b0e45f684] vif_infos = vmwarevif.get_vif_info(self._session, [ 747.846465] env[61806]: ERROR nova.compute.manager [instance: ea8d39a1-066d-4eb4-888d-776b0e45f684] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 747.846465] env[61806]: ERROR nova.compute.manager [instance: ea8d39a1-066d-4eb4-888d-776b0e45f684] for vif in network_info: [ 747.846465] env[61806]: ERROR nova.compute.manager [instance: ea8d39a1-066d-4eb4-888d-776b0e45f684] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 747.846465] env[61806]: ERROR nova.compute.manager [instance: ea8d39a1-066d-4eb4-888d-776b0e45f684] return self._sync_wrapper(fn, *args, **kwargs) [ 747.846465] env[61806]: ERROR nova.compute.manager [instance: ea8d39a1-066d-4eb4-888d-776b0e45f684] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 747.846465] env[61806]: ERROR nova.compute.manager [instance: ea8d39a1-066d-4eb4-888d-776b0e45f684] self.wait() [ 747.846465] env[61806]: ERROR nova.compute.manager [instance: ea8d39a1-066d-4eb4-888d-776b0e45f684] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 747.846465] env[61806]: ERROR nova.compute.manager [instance: ea8d39a1-066d-4eb4-888d-776b0e45f684] self[:] = self._gt.wait() [ 747.846465] env[61806]: ERROR nova.compute.manager [instance: ea8d39a1-066d-4eb4-888d-776b0e45f684] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 747.846465] env[61806]: ERROR nova.compute.manager [instance: ea8d39a1-066d-4eb4-888d-776b0e45f684] return self._exit_event.wait() [ 747.846465] env[61806]: ERROR nova.compute.manager [instance: ea8d39a1-066d-4eb4-888d-776b0e45f684] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 747.846465] env[61806]: ERROR nova.compute.manager [instance: ea8d39a1-066d-4eb4-888d-776b0e45f684] current.throw(*self._exc) [ 747.846928] env[61806]: ERROR nova.compute.manager [instance: ea8d39a1-066d-4eb4-888d-776b0e45f684] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 747.846928] env[61806]: ERROR nova.compute.manager [instance: ea8d39a1-066d-4eb4-888d-776b0e45f684] result = function(*args, **kwargs) [ 747.846928] env[61806]: ERROR nova.compute.manager [instance: ea8d39a1-066d-4eb4-888d-776b0e45f684] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 747.846928] env[61806]: ERROR nova.compute.manager [instance: ea8d39a1-066d-4eb4-888d-776b0e45f684] return func(*args, **kwargs) [ 747.846928] env[61806]: ERROR nova.compute.manager [instance: ea8d39a1-066d-4eb4-888d-776b0e45f684] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 747.846928] env[61806]: ERROR nova.compute.manager [instance: ea8d39a1-066d-4eb4-888d-776b0e45f684] raise e [ 747.846928] env[61806]: ERROR nova.compute.manager [instance: ea8d39a1-066d-4eb4-888d-776b0e45f684] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 747.846928] env[61806]: ERROR nova.compute.manager [instance: ea8d39a1-066d-4eb4-888d-776b0e45f684] nwinfo = self.network_api.allocate_for_instance( [ 747.846928] env[61806]: ERROR nova.compute.manager [instance: ea8d39a1-066d-4eb4-888d-776b0e45f684] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 747.846928] env[61806]: ERROR nova.compute.manager [instance: ea8d39a1-066d-4eb4-888d-776b0e45f684] created_port_ids = self._update_ports_for_instance( [ 747.846928] env[61806]: ERROR nova.compute.manager [instance: ea8d39a1-066d-4eb4-888d-776b0e45f684] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 747.846928] env[61806]: ERROR nova.compute.manager [instance: ea8d39a1-066d-4eb4-888d-776b0e45f684] with excutils.save_and_reraise_exception(): [ 747.846928] env[61806]: ERROR nova.compute.manager [instance: ea8d39a1-066d-4eb4-888d-776b0e45f684] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 747.847380] env[61806]: ERROR nova.compute.manager [instance: ea8d39a1-066d-4eb4-888d-776b0e45f684] self.force_reraise() [ 747.847380] env[61806]: ERROR nova.compute.manager [instance: ea8d39a1-066d-4eb4-888d-776b0e45f684] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 747.847380] env[61806]: ERROR nova.compute.manager [instance: ea8d39a1-066d-4eb4-888d-776b0e45f684] raise self.value [ 747.847380] env[61806]: ERROR nova.compute.manager [instance: ea8d39a1-066d-4eb4-888d-776b0e45f684] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 747.847380] env[61806]: ERROR nova.compute.manager [instance: ea8d39a1-066d-4eb4-888d-776b0e45f684] updated_port = self._update_port( [ 747.847380] env[61806]: ERROR nova.compute.manager [instance: ea8d39a1-066d-4eb4-888d-776b0e45f684] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 747.847380] env[61806]: ERROR nova.compute.manager [instance: ea8d39a1-066d-4eb4-888d-776b0e45f684] _ensure_no_port_binding_failure(port) [ 747.847380] env[61806]: ERROR nova.compute.manager [instance: ea8d39a1-066d-4eb4-888d-776b0e45f684] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 747.847380] env[61806]: ERROR nova.compute.manager [instance: ea8d39a1-066d-4eb4-888d-776b0e45f684] raise exception.PortBindingFailed(port_id=port['id']) [ 747.847380] env[61806]: ERROR nova.compute.manager [instance: ea8d39a1-066d-4eb4-888d-776b0e45f684] nova.exception.PortBindingFailed: Binding failed for port d1f266fb-3da7-4f92-b3cb-81aa9f3b924c, please check neutron logs for more information. [ 747.847380] env[61806]: ERROR nova.compute.manager [instance: ea8d39a1-066d-4eb4-888d-776b0e45f684] [ 747.847380] env[61806]: INFO nova.compute.manager [None req-3bf756e0-c56e-4902-9c2b-b224a1d09a7c tempest-ListServersNegativeTestJSON-272187588 tempest-ListServersNegativeTestJSON-272187588-project-member] [instance: ea8d39a1-066d-4eb4-888d-776b0e45f684] Terminating instance [ 747.848574] env[61806]: DEBUG oslo_concurrency.lockutils [None req-3bf756e0-c56e-4902-9c2b-b224a1d09a7c tempest-ListServersNegativeTestJSON-272187588 tempest-ListServersNegativeTestJSON-272187588-project-member] Acquiring lock "refresh_cache-ea8d39a1-066d-4eb4-888d-776b0e45f684" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 747.963194] env[61806]: DEBUG nova.network.neutron [req-8fb626b3-e86a-4aed-bb44-d3816464e40d req-135b21d2-37be-48a2-9675-c271dc5135e7 service nova] [instance: ea8d39a1-066d-4eb4-888d-776b0e45f684] Instance cache missing network info. {{(pid=61806) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 748.037373] env[61806]: DEBUG nova.network.neutron [req-8fb626b3-e86a-4aed-bb44-d3816464e40d req-135b21d2-37be-48a2-9675-c271dc5135e7 service nova] [instance: ea8d39a1-066d-4eb4-888d-776b0e45f684] Updating instance_info_cache with network_info: [] {{(pid=61806) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 748.138623] env[61806]: DEBUG oslo_concurrency.lockutils [None req-35a8f95d-1ddd-492f-a93c-adb382770732 tempest-ServersV294TestFqdnHostnames-1709116086 tempest-ServersV294TestFqdnHostnames-1709116086-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 748.140496] env[61806]: DEBUG oslo_concurrency.lockutils [None req-3bf756e0-c56e-4902-9c2b-b224a1d09a7c tempest-ListServersNegativeTestJSON-272187588 tempest-ListServersNegativeTestJSON-272187588-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.372s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 748.140980] env[61806]: DEBUG nova.compute.manager [None req-3bf756e0-c56e-4902-9c2b-b224a1d09a7c tempest-ListServersNegativeTestJSON-272187588 tempest-ListServersNegativeTestJSON-272187588-project-member] [instance: 198dc1a4-050d-47b2-8bc4-30af5534a789] Start building networks asynchronously for instance. {{(pid=61806) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 748.143393] env[61806]: DEBUG oslo_concurrency.lockutils [None req-f0377b51-dab9-4f68-8592-ca0f31b4ba51 tempest-FloatingIPsAssociationNegativeTestJSON-30494360 tempest-FloatingIPsAssociationNegativeTestJSON-30494360-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 15.649s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 748.144723] env[61806]: INFO nova.compute.claims [None req-f0377b51-dab9-4f68-8592-ca0f31b4ba51 tempest-FloatingIPsAssociationNegativeTestJSON-30494360 tempest-FloatingIPsAssociationNegativeTestJSON-30494360-project-member] [instance: 37c07b8d-e2a0-4aa6-b7ba-194feed837e5] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 748.539743] env[61806]: DEBUG oslo_concurrency.lockutils [req-8fb626b3-e86a-4aed-bb44-d3816464e40d req-135b21d2-37be-48a2-9675-c271dc5135e7 service nova] Releasing lock "refresh_cache-ea8d39a1-066d-4eb4-888d-776b0e45f684" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 748.540210] env[61806]: DEBUG oslo_concurrency.lockutils [None req-3bf756e0-c56e-4902-9c2b-b224a1d09a7c tempest-ListServersNegativeTestJSON-272187588 tempest-ListServersNegativeTestJSON-272187588-project-member] Acquired lock "refresh_cache-ea8d39a1-066d-4eb4-888d-776b0e45f684" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 748.540398] env[61806]: DEBUG nova.network.neutron [None req-3bf756e0-c56e-4902-9c2b-b224a1d09a7c tempest-ListServersNegativeTestJSON-272187588 tempest-ListServersNegativeTestJSON-272187588-project-member] [instance: ea8d39a1-066d-4eb4-888d-776b0e45f684] Building network info cache for instance {{(pid=61806) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 748.649898] env[61806]: DEBUG nova.compute.utils [None req-3bf756e0-c56e-4902-9c2b-b224a1d09a7c tempest-ListServersNegativeTestJSON-272187588 tempest-ListServersNegativeTestJSON-272187588-project-member] Using /dev/sd instead of None {{(pid=61806) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 748.653166] env[61806]: DEBUG nova.compute.manager [None req-3bf756e0-c56e-4902-9c2b-b224a1d09a7c tempest-ListServersNegativeTestJSON-272187588 tempest-ListServersNegativeTestJSON-272187588-project-member] [instance: 198dc1a4-050d-47b2-8bc4-30af5534a789] Allocating IP information in the background. {{(pid=61806) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 748.653345] env[61806]: DEBUG nova.network.neutron [None req-3bf756e0-c56e-4902-9c2b-b224a1d09a7c tempest-ListServersNegativeTestJSON-272187588 tempest-ListServersNegativeTestJSON-272187588-project-member] [instance: 198dc1a4-050d-47b2-8bc4-30af5534a789] allocate_for_instance() {{(pid=61806) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 748.708470] env[61806]: DEBUG nova.policy [None req-3bf756e0-c56e-4902-9c2b-b224a1d09a7c tempest-ListServersNegativeTestJSON-272187588 tempest-ListServersNegativeTestJSON-272187588-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '6228fec9ded14ef9974634c9a872374e', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'fb3d07cc8f764d15bc6fab0f299da5e8', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61806) authorize /opt/stack/nova/nova/policy.py:201}} [ 748.968271] env[61806]: DEBUG nova.network.neutron [None req-3bf756e0-c56e-4902-9c2b-b224a1d09a7c tempest-ListServersNegativeTestJSON-272187588 tempest-ListServersNegativeTestJSON-272187588-project-member] [instance: 198dc1a4-050d-47b2-8bc4-30af5534a789] Successfully created port: 1a279290-dcd8-45fb-a136-2fcbc132e8d7 {{(pid=61806) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 749.063392] env[61806]: DEBUG nova.network.neutron [None req-3bf756e0-c56e-4902-9c2b-b224a1d09a7c tempest-ListServersNegativeTestJSON-272187588 tempest-ListServersNegativeTestJSON-272187588-project-member] [instance: ea8d39a1-066d-4eb4-888d-776b0e45f684] Instance cache missing network info. {{(pid=61806) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 749.137112] env[61806]: DEBUG nova.network.neutron [None req-3bf756e0-c56e-4902-9c2b-b224a1d09a7c tempest-ListServersNegativeTestJSON-272187588 tempest-ListServersNegativeTestJSON-272187588-project-member] [instance: ea8d39a1-066d-4eb4-888d-776b0e45f684] Updating instance_info_cache with network_info: [] {{(pid=61806) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 749.153972] env[61806]: DEBUG nova.compute.manager [None req-3bf756e0-c56e-4902-9c2b-b224a1d09a7c tempest-ListServersNegativeTestJSON-272187588 tempest-ListServersNegativeTestJSON-272187588-project-member] [instance: 198dc1a4-050d-47b2-8bc4-30af5534a789] Start building block device mappings for instance. {{(pid=61806) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 749.467539] env[61806]: DEBUG nova.compute.manager [req-4bcedcea-47bd-4be9-ae79-cc18bf37a652 req-5afb1b26-af84-4521-bf98-0a46a927e282 service nova] [instance: ea8d39a1-066d-4eb4-888d-776b0e45f684] Received event network-vif-deleted-d1f266fb-3da7-4f92-b3cb-81aa9f3b924c {{(pid=61806) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 749.497032] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5f074a90-fb1f-4abc-a036-03ca77dd5774 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 749.504886] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6def25c7-0f01-415a-9786-64311e2a12b6 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 749.533788] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-94ffe2c3-0be0-46ec-a4de-c5218092a8fc {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 749.541038] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-82c54312-f2b0-41c5-a910-39b43ca3b917 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 749.554067] env[61806]: DEBUG nova.compute.provider_tree [None req-f0377b51-dab9-4f68-8592-ca0f31b4ba51 tempest-FloatingIPsAssociationNegativeTestJSON-30494360 tempest-FloatingIPsAssociationNegativeTestJSON-30494360-project-member] Inventory has not changed in ProviderTree for provider: a2858be1-fd22-4e08-979e-87ad25293407 {{(pid=61806) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 749.647158] env[61806]: DEBUG oslo_concurrency.lockutils [None req-3bf756e0-c56e-4902-9c2b-b224a1d09a7c tempest-ListServersNegativeTestJSON-272187588 tempest-ListServersNegativeTestJSON-272187588-project-member] Releasing lock "refresh_cache-ea8d39a1-066d-4eb4-888d-776b0e45f684" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 749.647583] env[61806]: DEBUG nova.compute.manager [None req-3bf756e0-c56e-4902-9c2b-b224a1d09a7c tempest-ListServersNegativeTestJSON-272187588 tempest-ListServersNegativeTestJSON-272187588-project-member] [instance: ea8d39a1-066d-4eb4-888d-776b0e45f684] Start destroying the instance on the hypervisor. {{(pid=61806) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 749.647780] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-3bf756e0-c56e-4902-9c2b-b224a1d09a7c tempest-ListServersNegativeTestJSON-272187588 tempest-ListServersNegativeTestJSON-272187588-project-member] [instance: ea8d39a1-066d-4eb4-888d-776b0e45f684] Destroying instance {{(pid=61806) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 749.648098] env[61806]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-810a6d48-ee21-459e-8296-feeb7ef210b1 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 749.656976] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8c71a076-35f7-4bee-b5fe-f55c7812da26 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 749.682617] env[61806]: WARNING nova.virt.vmwareapi.vmops [None req-3bf756e0-c56e-4902-9c2b-b224a1d09a7c tempest-ListServersNegativeTestJSON-272187588 tempest-ListServersNegativeTestJSON-272187588-project-member] [instance: ea8d39a1-066d-4eb4-888d-776b0e45f684] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance ea8d39a1-066d-4eb4-888d-776b0e45f684 could not be found. [ 749.682863] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-3bf756e0-c56e-4902-9c2b-b224a1d09a7c tempest-ListServersNegativeTestJSON-272187588 tempest-ListServersNegativeTestJSON-272187588-project-member] [instance: ea8d39a1-066d-4eb4-888d-776b0e45f684] Instance destroyed {{(pid=61806) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 749.683048] env[61806]: INFO nova.compute.manager [None req-3bf756e0-c56e-4902-9c2b-b224a1d09a7c tempest-ListServersNegativeTestJSON-272187588 tempest-ListServersNegativeTestJSON-272187588-project-member] [instance: ea8d39a1-066d-4eb4-888d-776b0e45f684] Took 0.04 seconds to destroy the instance on the hypervisor. [ 749.683285] env[61806]: DEBUG oslo.service.loopingcall [None req-3bf756e0-c56e-4902-9c2b-b224a1d09a7c tempest-ListServersNegativeTestJSON-272187588 tempest-ListServersNegativeTestJSON-272187588-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61806) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 749.683489] env[61806]: DEBUG nova.compute.manager [-] [instance: ea8d39a1-066d-4eb4-888d-776b0e45f684] Deallocating network for instance {{(pid=61806) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 749.683586] env[61806]: DEBUG nova.network.neutron [-] [instance: ea8d39a1-066d-4eb4-888d-776b0e45f684] deallocate_for_instance() {{(pid=61806) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 749.699835] env[61806]: DEBUG nova.network.neutron [-] [instance: ea8d39a1-066d-4eb4-888d-776b0e45f684] Instance cache missing network info. {{(pid=61806) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 749.919464] env[61806]: ERROR nova.compute.manager [None req-3bf756e0-c56e-4902-9c2b-b224a1d09a7c tempest-ListServersNegativeTestJSON-272187588 tempest-ListServersNegativeTestJSON-272187588-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 1a279290-dcd8-45fb-a136-2fcbc132e8d7, please check neutron logs for more information. [ 749.919464] env[61806]: ERROR nova.compute.manager Traceback (most recent call last): [ 749.919464] env[61806]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 749.919464] env[61806]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 749.919464] env[61806]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 749.919464] env[61806]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 749.919464] env[61806]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 749.919464] env[61806]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 749.919464] env[61806]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 749.919464] env[61806]: ERROR nova.compute.manager self.force_reraise() [ 749.919464] env[61806]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 749.919464] env[61806]: ERROR nova.compute.manager raise self.value [ 749.919464] env[61806]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 749.919464] env[61806]: ERROR nova.compute.manager updated_port = self._update_port( [ 749.919464] env[61806]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 749.919464] env[61806]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 749.920087] env[61806]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 749.920087] env[61806]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 749.920087] env[61806]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 1a279290-dcd8-45fb-a136-2fcbc132e8d7, please check neutron logs for more information. [ 749.920087] env[61806]: ERROR nova.compute.manager [ 749.920087] env[61806]: Traceback (most recent call last): [ 749.920087] env[61806]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 749.920087] env[61806]: listener.cb(fileno) [ 749.920087] env[61806]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 749.920087] env[61806]: result = function(*args, **kwargs) [ 749.920087] env[61806]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 749.920087] env[61806]: return func(*args, **kwargs) [ 749.920087] env[61806]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 749.920087] env[61806]: raise e [ 749.920087] env[61806]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 749.920087] env[61806]: nwinfo = self.network_api.allocate_for_instance( [ 749.920087] env[61806]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 749.920087] env[61806]: created_port_ids = self._update_ports_for_instance( [ 749.920087] env[61806]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 749.920087] env[61806]: with excutils.save_and_reraise_exception(): [ 749.920087] env[61806]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 749.920087] env[61806]: self.force_reraise() [ 749.920087] env[61806]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 749.920087] env[61806]: raise self.value [ 749.920087] env[61806]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 749.920087] env[61806]: updated_port = self._update_port( [ 749.920087] env[61806]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 749.920087] env[61806]: _ensure_no_port_binding_failure(port) [ 749.920087] env[61806]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 749.920087] env[61806]: raise exception.PortBindingFailed(port_id=port['id']) [ 749.921053] env[61806]: nova.exception.PortBindingFailed: Binding failed for port 1a279290-dcd8-45fb-a136-2fcbc132e8d7, please check neutron logs for more information. [ 749.921053] env[61806]: Removing descriptor: 14 [ 750.056828] env[61806]: DEBUG nova.scheduler.client.report [None req-f0377b51-dab9-4f68-8592-ca0f31b4ba51 tempest-FloatingIPsAssociationNegativeTestJSON-30494360 tempest-FloatingIPsAssociationNegativeTestJSON-30494360-project-member] Inventory has not changed for provider a2858be1-fd22-4e08-979e-87ad25293407 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 139, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61806) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 750.174366] env[61806]: DEBUG nova.compute.manager [None req-3bf756e0-c56e-4902-9c2b-b224a1d09a7c tempest-ListServersNegativeTestJSON-272187588 tempest-ListServersNegativeTestJSON-272187588-project-member] [instance: 198dc1a4-050d-47b2-8bc4-30af5534a789] Start spawning the instance on the hypervisor. {{(pid=61806) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 750.198625] env[61806]: DEBUG nova.virt.hardware [None req-3bf756e0-c56e-4902-9c2b-b224a1d09a7c tempest-ListServersNegativeTestJSON-272187588 tempest-ListServersNegativeTestJSON-272187588-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-15T18:05:10Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-15T18:04:52Z,direct_url=,disk_format='vmdk',id=ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='84ba83fce2c349a6988173c5d6fc3b07',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-15T18:04:53Z,virtual_size=,visibility=), allow threads: False {{(pid=61806) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 750.198886] env[61806]: DEBUG nova.virt.hardware [None req-3bf756e0-c56e-4902-9c2b-b224a1d09a7c tempest-ListServersNegativeTestJSON-272187588 tempest-ListServersNegativeTestJSON-272187588-project-member] Flavor limits 0:0:0 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 750.199063] env[61806]: DEBUG nova.virt.hardware [None req-3bf756e0-c56e-4902-9c2b-b224a1d09a7c tempest-ListServersNegativeTestJSON-272187588 tempest-ListServersNegativeTestJSON-272187588-project-member] Image limits 0:0:0 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 750.199306] env[61806]: DEBUG nova.virt.hardware [None req-3bf756e0-c56e-4902-9c2b-b224a1d09a7c tempest-ListServersNegativeTestJSON-272187588 tempest-ListServersNegativeTestJSON-272187588-project-member] Flavor pref 0:0:0 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 750.199462] env[61806]: DEBUG nova.virt.hardware [None req-3bf756e0-c56e-4902-9c2b-b224a1d09a7c tempest-ListServersNegativeTestJSON-272187588 tempest-ListServersNegativeTestJSON-272187588-project-member] Image pref 0:0:0 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 750.199635] env[61806]: DEBUG nova.virt.hardware [None req-3bf756e0-c56e-4902-9c2b-b224a1d09a7c tempest-ListServersNegativeTestJSON-272187588 tempest-ListServersNegativeTestJSON-272187588-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 750.199944] env[61806]: DEBUG nova.virt.hardware [None req-3bf756e0-c56e-4902-9c2b-b224a1d09a7c tempest-ListServersNegativeTestJSON-272187588 tempest-ListServersNegativeTestJSON-272187588-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61806) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 750.200142] env[61806]: DEBUG nova.virt.hardware [None req-3bf756e0-c56e-4902-9c2b-b224a1d09a7c tempest-ListServersNegativeTestJSON-272187588 tempest-ListServersNegativeTestJSON-272187588-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61806) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 750.200319] env[61806]: DEBUG nova.virt.hardware [None req-3bf756e0-c56e-4902-9c2b-b224a1d09a7c tempest-ListServersNegativeTestJSON-272187588 tempest-ListServersNegativeTestJSON-272187588-project-member] Got 1 possible topologies {{(pid=61806) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 750.200484] env[61806]: DEBUG nova.virt.hardware [None req-3bf756e0-c56e-4902-9c2b-b224a1d09a7c tempest-ListServersNegativeTestJSON-272187588 tempest-ListServersNegativeTestJSON-272187588-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61806) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 750.200657] env[61806]: DEBUG nova.virt.hardware [None req-3bf756e0-c56e-4902-9c2b-b224a1d09a7c tempest-ListServersNegativeTestJSON-272187588 tempest-ListServersNegativeTestJSON-272187588-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61806) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 750.201547] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0d228ad6-c231-4ea1-b481-6a38acfe870f {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 750.204222] env[61806]: DEBUG nova.network.neutron [-] [instance: ea8d39a1-066d-4eb4-888d-776b0e45f684] Updating instance_info_cache with network_info: [] {{(pid=61806) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 750.210909] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-19e216b6-03fa-4658-afcd-073ddd35689d {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 750.225281] env[61806]: ERROR nova.compute.manager [None req-3bf756e0-c56e-4902-9c2b-b224a1d09a7c tempest-ListServersNegativeTestJSON-272187588 tempest-ListServersNegativeTestJSON-272187588-project-member] [instance: 198dc1a4-050d-47b2-8bc4-30af5534a789] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 1a279290-dcd8-45fb-a136-2fcbc132e8d7, please check neutron logs for more information. [ 750.225281] env[61806]: ERROR nova.compute.manager [instance: 198dc1a4-050d-47b2-8bc4-30af5534a789] Traceback (most recent call last): [ 750.225281] env[61806]: ERROR nova.compute.manager [instance: 198dc1a4-050d-47b2-8bc4-30af5534a789] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 750.225281] env[61806]: ERROR nova.compute.manager [instance: 198dc1a4-050d-47b2-8bc4-30af5534a789] yield resources [ 750.225281] env[61806]: ERROR nova.compute.manager [instance: 198dc1a4-050d-47b2-8bc4-30af5534a789] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 750.225281] env[61806]: ERROR nova.compute.manager [instance: 198dc1a4-050d-47b2-8bc4-30af5534a789] self.driver.spawn(context, instance, image_meta, [ 750.225281] env[61806]: ERROR nova.compute.manager [instance: 198dc1a4-050d-47b2-8bc4-30af5534a789] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 750.225281] env[61806]: ERROR nova.compute.manager [instance: 198dc1a4-050d-47b2-8bc4-30af5534a789] self._vmops.spawn(context, instance, image_meta, injected_files, [ 750.225281] env[61806]: ERROR nova.compute.manager [instance: 198dc1a4-050d-47b2-8bc4-30af5534a789] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 750.225281] env[61806]: ERROR nova.compute.manager [instance: 198dc1a4-050d-47b2-8bc4-30af5534a789] vm_ref = self.build_virtual_machine(instance, [ 750.225281] env[61806]: ERROR nova.compute.manager [instance: 198dc1a4-050d-47b2-8bc4-30af5534a789] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 750.225801] env[61806]: ERROR nova.compute.manager [instance: 198dc1a4-050d-47b2-8bc4-30af5534a789] vif_infos = vmwarevif.get_vif_info(self._session, [ 750.225801] env[61806]: ERROR nova.compute.manager [instance: 198dc1a4-050d-47b2-8bc4-30af5534a789] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 750.225801] env[61806]: ERROR nova.compute.manager [instance: 198dc1a4-050d-47b2-8bc4-30af5534a789] for vif in network_info: [ 750.225801] env[61806]: ERROR nova.compute.manager [instance: 198dc1a4-050d-47b2-8bc4-30af5534a789] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 750.225801] env[61806]: ERROR nova.compute.manager [instance: 198dc1a4-050d-47b2-8bc4-30af5534a789] return self._sync_wrapper(fn, *args, **kwargs) [ 750.225801] env[61806]: ERROR nova.compute.manager [instance: 198dc1a4-050d-47b2-8bc4-30af5534a789] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 750.225801] env[61806]: ERROR nova.compute.manager [instance: 198dc1a4-050d-47b2-8bc4-30af5534a789] self.wait() [ 750.225801] env[61806]: ERROR nova.compute.manager [instance: 198dc1a4-050d-47b2-8bc4-30af5534a789] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 750.225801] env[61806]: ERROR nova.compute.manager [instance: 198dc1a4-050d-47b2-8bc4-30af5534a789] self[:] = self._gt.wait() [ 750.225801] env[61806]: ERROR nova.compute.manager [instance: 198dc1a4-050d-47b2-8bc4-30af5534a789] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 750.225801] env[61806]: ERROR nova.compute.manager [instance: 198dc1a4-050d-47b2-8bc4-30af5534a789] return self._exit_event.wait() [ 750.225801] env[61806]: ERROR nova.compute.manager [instance: 198dc1a4-050d-47b2-8bc4-30af5534a789] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 750.225801] env[61806]: ERROR nova.compute.manager [instance: 198dc1a4-050d-47b2-8bc4-30af5534a789] current.throw(*self._exc) [ 750.226383] env[61806]: ERROR nova.compute.manager [instance: 198dc1a4-050d-47b2-8bc4-30af5534a789] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 750.226383] env[61806]: ERROR nova.compute.manager [instance: 198dc1a4-050d-47b2-8bc4-30af5534a789] result = function(*args, **kwargs) [ 750.226383] env[61806]: ERROR nova.compute.manager [instance: 198dc1a4-050d-47b2-8bc4-30af5534a789] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 750.226383] env[61806]: ERROR nova.compute.manager [instance: 198dc1a4-050d-47b2-8bc4-30af5534a789] return func(*args, **kwargs) [ 750.226383] env[61806]: ERROR nova.compute.manager [instance: 198dc1a4-050d-47b2-8bc4-30af5534a789] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 750.226383] env[61806]: ERROR nova.compute.manager [instance: 198dc1a4-050d-47b2-8bc4-30af5534a789] raise e [ 750.226383] env[61806]: ERROR nova.compute.manager [instance: 198dc1a4-050d-47b2-8bc4-30af5534a789] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 750.226383] env[61806]: ERROR nova.compute.manager [instance: 198dc1a4-050d-47b2-8bc4-30af5534a789] nwinfo = self.network_api.allocate_for_instance( [ 750.226383] env[61806]: ERROR nova.compute.manager [instance: 198dc1a4-050d-47b2-8bc4-30af5534a789] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 750.226383] env[61806]: ERROR nova.compute.manager [instance: 198dc1a4-050d-47b2-8bc4-30af5534a789] created_port_ids = self._update_ports_for_instance( [ 750.226383] env[61806]: ERROR nova.compute.manager [instance: 198dc1a4-050d-47b2-8bc4-30af5534a789] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 750.226383] env[61806]: ERROR nova.compute.manager [instance: 198dc1a4-050d-47b2-8bc4-30af5534a789] with excutils.save_and_reraise_exception(): [ 750.226383] env[61806]: ERROR nova.compute.manager [instance: 198dc1a4-050d-47b2-8bc4-30af5534a789] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 750.227237] env[61806]: ERROR nova.compute.manager [instance: 198dc1a4-050d-47b2-8bc4-30af5534a789] self.force_reraise() [ 750.227237] env[61806]: ERROR nova.compute.manager [instance: 198dc1a4-050d-47b2-8bc4-30af5534a789] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 750.227237] env[61806]: ERROR nova.compute.manager [instance: 198dc1a4-050d-47b2-8bc4-30af5534a789] raise self.value [ 750.227237] env[61806]: ERROR nova.compute.manager [instance: 198dc1a4-050d-47b2-8bc4-30af5534a789] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 750.227237] env[61806]: ERROR nova.compute.manager [instance: 198dc1a4-050d-47b2-8bc4-30af5534a789] updated_port = self._update_port( [ 750.227237] env[61806]: ERROR nova.compute.manager [instance: 198dc1a4-050d-47b2-8bc4-30af5534a789] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 750.227237] env[61806]: ERROR nova.compute.manager [instance: 198dc1a4-050d-47b2-8bc4-30af5534a789] _ensure_no_port_binding_failure(port) [ 750.227237] env[61806]: ERROR nova.compute.manager [instance: 198dc1a4-050d-47b2-8bc4-30af5534a789] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 750.227237] env[61806]: ERROR nova.compute.manager [instance: 198dc1a4-050d-47b2-8bc4-30af5534a789] raise exception.PortBindingFailed(port_id=port['id']) [ 750.227237] env[61806]: ERROR nova.compute.manager [instance: 198dc1a4-050d-47b2-8bc4-30af5534a789] nova.exception.PortBindingFailed: Binding failed for port 1a279290-dcd8-45fb-a136-2fcbc132e8d7, please check neutron logs for more information. [ 750.227237] env[61806]: ERROR nova.compute.manager [instance: 198dc1a4-050d-47b2-8bc4-30af5534a789] [ 750.227237] env[61806]: INFO nova.compute.manager [None req-3bf756e0-c56e-4902-9c2b-b224a1d09a7c tempest-ListServersNegativeTestJSON-272187588 tempest-ListServersNegativeTestJSON-272187588-project-member] [instance: 198dc1a4-050d-47b2-8bc4-30af5534a789] Terminating instance [ 750.227705] env[61806]: DEBUG oslo_concurrency.lockutils [None req-3bf756e0-c56e-4902-9c2b-b224a1d09a7c tempest-ListServersNegativeTestJSON-272187588 tempest-ListServersNegativeTestJSON-272187588-project-member] Acquiring lock "refresh_cache-198dc1a4-050d-47b2-8bc4-30af5534a789" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 750.227705] env[61806]: DEBUG oslo_concurrency.lockutils [None req-3bf756e0-c56e-4902-9c2b-b224a1d09a7c tempest-ListServersNegativeTestJSON-272187588 tempest-ListServersNegativeTestJSON-272187588-project-member] Acquired lock "refresh_cache-198dc1a4-050d-47b2-8bc4-30af5534a789" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 750.227705] env[61806]: DEBUG nova.network.neutron [None req-3bf756e0-c56e-4902-9c2b-b224a1d09a7c tempest-ListServersNegativeTestJSON-272187588 tempest-ListServersNegativeTestJSON-272187588-project-member] [instance: 198dc1a4-050d-47b2-8bc4-30af5534a789] Building network info cache for instance {{(pid=61806) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 750.562478] env[61806]: DEBUG oslo_concurrency.lockutils [None req-f0377b51-dab9-4f68-8592-ca0f31b4ba51 tempest-FloatingIPsAssociationNegativeTestJSON-30494360 tempest-FloatingIPsAssociationNegativeTestJSON-30494360-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.419s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 750.563207] env[61806]: DEBUG nova.compute.manager [None req-f0377b51-dab9-4f68-8592-ca0f31b4ba51 tempest-FloatingIPsAssociationNegativeTestJSON-30494360 tempest-FloatingIPsAssociationNegativeTestJSON-30494360-project-member] [instance: 37c07b8d-e2a0-4aa6-b7ba-194feed837e5] Start building networks asynchronously for instance. {{(pid=61806) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 750.565825] env[61806]: DEBUG oslo_concurrency.lockutils [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 14.917s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 750.565934] env[61806]: DEBUG oslo_concurrency.lockutils [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 750.566375] env[61806]: DEBUG nova.compute.resource_tracker [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=61806) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 750.566375] env[61806]: DEBUG oslo_concurrency.lockutils [None req-3c3981d2-fa25-4f4b-87d4-a5411e67199e tempest-MultipleCreateTestJSON-1589015462 tempest-MultipleCreateTestJSON-1589015462-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 14.800s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 750.569637] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d3aa6315-d669-4b62-9e8b-932e09d4369c {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 750.578136] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-52129098-fba1-4528-8fa7-a4be8b33af38 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 750.592427] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-76ab2d0d-2c24-449e-bffa-9dd4d79e4b96 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 750.599830] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ed015e0f-c91c-46b1-a68b-58ad5bf84756 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 750.631062] env[61806]: DEBUG nova.compute.resource_tracker [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=181514MB free_disk=139GB free_vcpus=48 pci_devices=None {{(pid=61806) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 750.631249] env[61806]: DEBUG oslo_concurrency.lockutils [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 750.707906] env[61806]: INFO nova.compute.manager [-] [instance: ea8d39a1-066d-4eb4-888d-776b0e45f684] Took 1.02 seconds to deallocate network for instance. [ 750.710330] env[61806]: DEBUG nova.compute.claims [None req-3bf756e0-c56e-4902-9c2b-b224a1d09a7c tempest-ListServersNegativeTestJSON-272187588 tempest-ListServersNegativeTestJSON-272187588-project-member] [instance: ea8d39a1-066d-4eb4-888d-776b0e45f684] Aborting claim: {{(pid=61806) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 750.710506] env[61806]: DEBUG oslo_concurrency.lockutils [None req-3bf756e0-c56e-4902-9c2b-b224a1d09a7c tempest-ListServersNegativeTestJSON-272187588 tempest-ListServersNegativeTestJSON-272187588-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 750.744013] env[61806]: DEBUG nova.network.neutron [None req-3bf756e0-c56e-4902-9c2b-b224a1d09a7c tempest-ListServersNegativeTestJSON-272187588 tempest-ListServersNegativeTestJSON-272187588-project-member] [instance: 198dc1a4-050d-47b2-8bc4-30af5534a789] Instance cache missing network info. {{(pid=61806) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 750.830401] env[61806]: DEBUG nova.network.neutron [None req-3bf756e0-c56e-4902-9c2b-b224a1d09a7c tempest-ListServersNegativeTestJSON-272187588 tempest-ListServersNegativeTestJSON-272187588-project-member] [instance: 198dc1a4-050d-47b2-8bc4-30af5534a789] Updating instance_info_cache with network_info: [] {{(pid=61806) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 751.072041] env[61806]: DEBUG nova.compute.utils [None req-f0377b51-dab9-4f68-8592-ca0f31b4ba51 tempest-FloatingIPsAssociationNegativeTestJSON-30494360 tempest-FloatingIPsAssociationNegativeTestJSON-30494360-project-member] Using /dev/sd instead of None {{(pid=61806) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 751.072987] env[61806]: DEBUG nova.compute.manager [None req-f0377b51-dab9-4f68-8592-ca0f31b4ba51 tempest-FloatingIPsAssociationNegativeTestJSON-30494360 tempest-FloatingIPsAssociationNegativeTestJSON-30494360-project-member] [instance: 37c07b8d-e2a0-4aa6-b7ba-194feed837e5] Allocating IP information in the background. {{(pid=61806) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 751.073094] env[61806]: DEBUG nova.network.neutron [None req-f0377b51-dab9-4f68-8592-ca0f31b4ba51 tempest-FloatingIPsAssociationNegativeTestJSON-30494360 tempest-FloatingIPsAssociationNegativeTestJSON-30494360-project-member] [instance: 37c07b8d-e2a0-4aa6-b7ba-194feed837e5] allocate_for_instance() {{(pid=61806) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 751.114277] env[61806]: DEBUG nova.policy [None req-f0377b51-dab9-4f68-8592-ca0f31b4ba51 tempest-FloatingIPsAssociationNegativeTestJSON-30494360 tempest-FloatingIPsAssociationNegativeTestJSON-30494360-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '48f2b933112f41d58a16ab0f1e2b853d', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '66e1c771d77b487ab5bd1cdb2fc75f10', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61806) authorize /opt/stack/nova/nova/policy.py:201}} [ 751.334967] env[61806]: DEBUG oslo_concurrency.lockutils [None req-3bf756e0-c56e-4902-9c2b-b224a1d09a7c tempest-ListServersNegativeTestJSON-272187588 tempest-ListServersNegativeTestJSON-272187588-project-member] Releasing lock "refresh_cache-198dc1a4-050d-47b2-8bc4-30af5534a789" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 751.335701] env[61806]: DEBUG nova.compute.manager [None req-3bf756e0-c56e-4902-9c2b-b224a1d09a7c tempest-ListServersNegativeTestJSON-272187588 tempest-ListServersNegativeTestJSON-272187588-project-member] [instance: 198dc1a4-050d-47b2-8bc4-30af5534a789] Start destroying the instance on the hypervisor. {{(pid=61806) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 751.336248] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-3bf756e0-c56e-4902-9c2b-b224a1d09a7c tempest-ListServersNegativeTestJSON-272187588 tempest-ListServersNegativeTestJSON-272187588-project-member] [instance: 198dc1a4-050d-47b2-8bc4-30af5534a789] Destroying instance {{(pid=61806) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 751.336793] env[61806]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-1a7ba56a-d6af-4db7-b61f-8456912688de {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 751.348899] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4177e30d-f8fb-4764-9c80-0bfe68b76086 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 751.360103] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-435195cf-b01d-4890-8cae-90ea9018b1c7 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 751.367263] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c0d890eb-ca98-4685-bdc9-0b98861ec2de {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 751.374820] env[61806]: WARNING nova.virt.vmwareapi.vmops [None req-3bf756e0-c56e-4902-9c2b-b224a1d09a7c tempest-ListServersNegativeTestJSON-272187588 tempest-ListServersNegativeTestJSON-272187588-project-member] [instance: 198dc1a4-050d-47b2-8bc4-30af5534a789] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 198dc1a4-050d-47b2-8bc4-30af5534a789 could not be found. [ 751.374954] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-3bf756e0-c56e-4902-9c2b-b224a1d09a7c tempest-ListServersNegativeTestJSON-272187588 tempest-ListServersNegativeTestJSON-272187588-project-member] [instance: 198dc1a4-050d-47b2-8bc4-30af5534a789] Instance destroyed {{(pid=61806) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 751.375194] env[61806]: INFO nova.compute.manager [None req-3bf756e0-c56e-4902-9c2b-b224a1d09a7c tempest-ListServersNegativeTestJSON-272187588 tempest-ListServersNegativeTestJSON-272187588-project-member] [instance: 198dc1a4-050d-47b2-8bc4-30af5534a789] Took 0.04 seconds to destroy the instance on the hypervisor. [ 751.375497] env[61806]: DEBUG oslo.service.loopingcall [None req-3bf756e0-c56e-4902-9c2b-b224a1d09a7c tempest-ListServersNegativeTestJSON-272187588 tempest-ListServersNegativeTestJSON-272187588-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61806) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 751.376275] env[61806]: DEBUG nova.compute.manager [-] [instance: 198dc1a4-050d-47b2-8bc4-30af5534a789] Deallocating network for instance {{(pid=61806) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 751.376369] env[61806]: DEBUG nova.network.neutron [-] [instance: 198dc1a4-050d-47b2-8bc4-30af5534a789] deallocate_for_instance() {{(pid=61806) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 751.405185] env[61806]: DEBUG nova.network.neutron [None req-f0377b51-dab9-4f68-8592-ca0f31b4ba51 tempest-FloatingIPsAssociationNegativeTestJSON-30494360 tempest-FloatingIPsAssociationNegativeTestJSON-30494360-project-member] [instance: 37c07b8d-e2a0-4aa6-b7ba-194feed837e5] Successfully created port: 1b064d62-24f5-4892-937d-ec37537f02ac {{(pid=61806) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 751.407304] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8729b7ed-976a-47bf-884e-46ceb13454d6 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 751.414751] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c1880247-cd54-4a00-8940-79ac5e22cd82 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 751.420070] env[61806]: DEBUG nova.network.neutron [-] [instance: 198dc1a4-050d-47b2-8bc4-30af5534a789] Instance cache missing network info. {{(pid=61806) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 751.431345] env[61806]: DEBUG nova.compute.provider_tree [None req-3c3981d2-fa25-4f4b-87d4-a5411e67199e tempest-MultipleCreateTestJSON-1589015462 tempest-MultipleCreateTestJSON-1589015462-project-member] Inventory has not changed in ProviderTree for provider: a2858be1-fd22-4e08-979e-87ad25293407 {{(pid=61806) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 751.497850] env[61806]: DEBUG nova.compute.manager [req-c951326f-e784-41c4-8566-93d0a0877265 req-afd1da5f-d8ba-442f-94a7-5f46ce883f37 service nova] [instance: 198dc1a4-050d-47b2-8bc4-30af5534a789] Received event network-changed-1a279290-dcd8-45fb-a136-2fcbc132e8d7 {{(pid=61806) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 751.498221] env[61806]: DEBUG nova.compute.manager [req-c951326f-e784-41c4-8566-93d0a0877265 req-afd1da5f-d8ba-442f-94a7-5f46ce883f37 service nova] [instance: 198dc1a4-050d-47b2-8bc4-30af5534a789] Refreshing instance network info cache due to event network-changed-1a279290-dcd8-45fb-a136-2fcbc132e8d7. {{(pid=61806) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 751.498443] env[61806]: DEBUG oslo_concurrency.lockutils [req-c951326f-e784-41c4-8566-93d0a0877265 req-afd1da5f-d8ba-442f-94a7-5f46ce883f37 service nova] Acquiring lock "refresh_cache-198dc1a4-050d-47b2-8bc4-30af5534a789" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 751.498584] env[61806]: DEBUG oslo_concurrency.lockutils [req-c951326f-e784-41c4-8566-93d0a0877265 req-afd1da5f-d8ba-442f-94a7-5f46ce883f37 service nova] Acquired lock "refresh_cache-198dc1a4-050d-47b2-8bc4-30af5534a789" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 751.498743] env[61806]: DEBUG nova.network.neutron [req-c951326f-e784-41c4-8566-93d0a0877265 req-afd1da5f-d8ba-442f-94a7-5f46ce883f37 service nova] [instance: 198dc1a4-050d-47b2-8bc4-30af5534a789] Refreshing network info cache for port 1a279290-dcd8-45fb-a136-2fcbc132e8d7 {{(pid=61806) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 751.578831] env[61806]: DEBUG nova.compute.manager [None req-f0377b51-dab9-4f68-8592-ca0f31b4ba51 tempest-FloatingIPsAssociationNegativeTestJSON-30494360 tempest-FloatingIPsAssociationNegativeTestJSON-30494360-project-member] [instance: 37c07b8d-e2a0-4aa6-b7ba-194feed837e5] Start building block device mappings for instance. {{(pid=61806) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 751.922991] env[61806]: DEBUG nova.network.neutron [-] [instance: 198dc1a4-050d-47b2-8bc4-30af5534a789] Updating instance_info_cache with network_info: [] {{(pid=61806) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 751.935123] env[61806]: DEBUG nova.scheduler.client.report [None req-3c3981d2-fa25-4f4b-87d4-a5411e67199e tempest-MultipleCreateTestJSON-1589015462 tempest-MultipleCreateTestJSON-1589015462-project-member] Inventory has not changed for provider a2858be1-fd22-4e08-979e-87ad25293407 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 139, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61806) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 752.018462] env[61806]: DEBUG nova.network.neutron [req-c951326f-e784-41c4-8566-93d0a0877265 req-afd1da5f-d8ba-442f-94a7-5f46ce883f37 service nova] [instance: 198dc1a4-050d-47b2-8bc4-30af5534a789] Instance cache missing network info. {{(pid=61806) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 752.107274] env[61806]: DEBUG nova.network.neutron [req-c951326f-e784-41c4-8566-93d0a0877265 req-afd1da5f-d8ba-442f-94a7-5f46ce883f37 service nova] [instance: 198dc1a4-050d-47b2-8bc4-30af5534a789] Updating instance_info_cache with network_info: [] {{(pid=61806) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 752.283495] env[61806]: ERROR nova.compute.manager [None req-f0377b51-dab9-4f68-8592-ca0f31b4ba51 tempest-FloatingIPsAssociationNegativeTestJSON-30494360 tempest-FloatingIPsAssociationNegativeTestJSON-30494360-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 1b064d62-24f5-4892-937d-ec37537f02ac, please check neutron logs for more information. [ 752.283495] env[61806]: ERROR nova.compute.manager Traceback (most recent call last): [ 752.283495] env[61806]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 752.283495] env[61806]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 752.283495] env[61806]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 752.283495] env[61806]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 752.283495] env[61806]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 752.283495] env[61806]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 752.283495] env[61806]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 752.283495] env[61806]: ERROR nova.compute.manager self.force_reraise() [ 752.283495] env[61806]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 752.283495] env[61806]: ERROR nova.compute.manager raise self.value [ 752.283495] env[61806]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 752.283495] env[61806]: ERROR nova.compute.manager updated_port = self._update_port( [ 752.283495] env[61806]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 752.283495] env[61806]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 752.284110] env[61806]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 752.284110] env[61806]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 752.284110] env[61806]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 1b064d62-24f5-4892-937d-ec37537f02ac, please check neutron logs for more information. [ 752.284110] env[61806]: ERROR nova.compute.manager [ 752.284110] env[61806]: Traceback (most recent call last): [ 752.284110] env[61806]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 752.284110] env[61806]: listener.cb(fileno) [ 752.284110] env[61806]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 752.284110] env[61806]: result = function(*args, **kwargs) [ 752.284110] env[61806]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 752.284110] env[61806]: return func(*args, **kwargs) [ 752.284110] env[61806]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 752.284110] env[61806]: raise e [ 752.284110] env[61806]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 752.284110] env[61806]: nwinfo = self.network_api.allocate_for_instance( [ 752.284110] env[61806]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 752.284110] env[61806]: created_port_ids = self._update_ports_for_instance( [ 752.284110] env[61806]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 752.284110] env[61806]: with excutils.save_and_reraise_exception(): [ 752.284110] env[61806]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 752.284110] env[61806]: self.force_reraise() [ 752.284110] env[61806]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 752.284110] env[61806]: raise self.value [ 752.284110] env[61806]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 752.284110] env[61806]: updated_port = self._update_port( [ 752.284110] env[61806]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 752.284110] env[61806]: _ensure_no_port_binding_failure(port) [ 752.284110] env[61806]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 752.284110] env[61806]: raise exception.PortBindingFailed(port_id=port['id']) [ 752.285173] env[61806]: nova.exception.PortBindingFailed: Binding failed for port 1b064d62-24f5-4892-937d-ec37537f02ac, please check neutron logs for more information. [ 752.285173] env[61806]: Removing descriptor: 14 [ 752.426431] env[61806]: INFO nova.compute.manager [-] [instance: 198dc1a4-050d-47b2-8bc4-30af5534a789] Took 1.05 seconds to deallocate network for instance. [ 752.429012] env[61806]: DEBUG nova.compute.claims [None req-3bf756e0-c56e-4902-9c2b-b224a1d09a7c tempest-ListServersNegativeTestJSON-272187588 tempest-ListServersNegativeTestJSON-272187588-project-member] [instance: 198dc1a4-050d-47b2-8bc4-30af5534a789] Aborting claim: {{(pid=61806) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 752.429205] env[61806]: DEBUG oslo_concurrency.lockutils [None req-3bf756e0-c56e-4902-9c2b-b224a1d09a7c tempest-ListServersNegativeTestJSON-272187588 tempest-ListServersNegativeTestJSON-272187588-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 752.440143] env[61806]: DEBUG oslo_concurrency.lockutils [None req-3c3981d2-fa25-4f4b-87d4-a5411e67199e tempest-MultipleCreateTestJSON-1589015462 tempest-MultipleCreateTestJSON-1589015462-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.874s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 752.441029] env[61806]: ERROR nova.compute.manager [None req-3c3981d2-fa25-4f4b-87d4-a5411e67199e tempest-MultipleCreateTestJSON-1589015462 tempest-MultipleCreateTestJSON-1589015462-project-member] [instance: 5b22c2fc-525c-481b-b84e-e0a7f68f633e] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 6cacc512-7619-41e7-8fcb-b88df2c7cf8b, please check neutron logs for more information. [ 752.441029] env[61806]: ERROR nova.compute.manager [instance: 5b22c2fc-525c-481b-b84e-e0a7f68f633e] Traceback (most recent call last): [ 752.441029] env[61806]: ERROR nova.compute.manager [instance: 5b22c2fc-525c-481b-b84e-e0a7f68f633e] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 752.441029] env[61806]: ERROR nova.compute.manager [instance: 5b22c2fc-525c-481b-b84e-e0a7f68f633e] self.driver.spawn(context, instance, image_meta, [ 752.441029] env[61806]: ERROR nova.compute.manager [instance: 5b22c2fc-525c-481b-b84e-e0a7f68f633e] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 752.441029] env[61806]: ERROR nova.compute.manager [instance: 5b22c2fc-525c-481b-b84e-e0a7f68f633e] self._vmops.spawn(context, instance, image_meta, injected_files, [ 752.441029] env[61806]: ERROR nova.compute.manager [instance: 5b22c2fc-525c-481b-b84e-e0a7f68f633e] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 752.441029] env[61806]: ERROR nova.compute.manager [instance: 5b22c2fc-525c-481b-b84e-e0a7f68f633e] vm_ref = self.build_virtual_machine(instance, [ 752.441029] env[61806]: ERROR nova.compute.manager [instance: 5b22c2fc-525c-481b-b84e-e0a7f68f633e] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 752.441029] env[61806]: ERROR nova.compute.manager [instance: 5b22c2fc-525c-481b-b84e-e0a7f68f633e] vif_infos = vmwarevif.get_vif_info(self._session, [ 752.441029] env[61806]: ERROR nova.compute.manager [instance: 5b22c2fc-525c-481b-b84e-e0a7f68f633e] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 752.441478] env[61806]: ERROR nova.compute.manager [instance: 5b22c2fc-525c-481b-b84e-e0a7f68f633e] for vif in network_info: [ 752.441478] env[61806]: ERROR nova.compute.manager [instance: 5b22c2fc-525c-481b-b84e-e0a7f68f633e] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 752.441478] env[61806]: ERROR nova.compute.manager [instance: 5b22c2fc-525c-481b-b84e-e0a7f68f633e] return self._sync_wrapper(fn, *args, **kwargs) [ 752.441478] env[61806]: ERROR nova.compute.manager [instance: 5b22c2fc-525c-481b-b84e-e0a7f68f633e] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 752.441478] env[61806]: ERROR nova.compute.manager [instance: 5b22c2fc-525c-481b-b84e-e0a7f68f633e] self.wait() [ 752.441478] env[61806]: ERROR nova.compute.manager [instance: 5b22c2fc-525c-481b-b84e-e0a7f68f633e] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 752.441478] env[61806]: ERROR nova.compute.manager [instance: 5b22c2fc-525c-481b-b84e-e0a7f68f633e] self[:] = self._gt.wait() [ 752.441478] env[61806]: ERROR nova.compute.manager [instance: 5b22c2fc-525c-481b-b84e-e0a7f68f633e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 752.441478] env[61806]: ERROR nova.compute.manager [instance: 5b22c2fc-525c-481b-b84e-e0a7f68f633e] return self._exit_event.wait() [ 752.441478] env[61806]: ERROR nova.compute.manager [instance: 5b22c2fc-525c-481b-b84e-e0a7f68f633e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 752.441478] env[61806]: ERROR nova.compute.manager [instance: 5b22c2fc-525c-481b-b84e-e0a7f68f633e] result = hub.switch() [ 752.441478] env[61806]: ERROR nova.compute.manager [instance: 5b22c2fc-525c-481b-b84e-e0a7f68f633e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 752.441478] env[61806]: ERROR nova.compute.manager [instance: 5b22c2fc-525c-481b-b84e-e0a7f68f633e] return self.greenlet.switch() [ 752.441979] env[61806]: ERROR nova.compute.manager [instance: 5b22c2fc-525c-481b-b84e-e0a7f68f633e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 752.441979] env[61806]: ERROR nova.compute.manager [instance: 5b22c2fc-525c-481b-b84e-e0a7f68f633e] result = function(*args, **kwargs) [ 752.441979] env[61806]: ERROR nova.compute.manager [instance: 5b22c2fc-525c-481b-b84e-e0a7f68f633e] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 752.441979] env[61806]: ERROR nova.compute.manager [instance: 5b22c2fc-525c-481b-b84e-e0a7f68f633e] return func(*args, **kwargs) [ 752.441979] env[61806]: ERROR nova.compute.manager [instance: 5b22c2fc-525c-481b-b84e-e0a7f68f633e] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 752.441979] env[61806]: ERROR nova.compute.manager [instance: 5b22c2fc-525c-481b-b84e-e0a7f68f633e] raise e [ 752.441979] env[61806]: ERROR nova.compute.manager [instance: 5b22c2fc-525c-481b-b84e-e0a7f68f633e] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 752.441979] env[61806]: ERROR nova.compute.manager [instance: 5b22c2fc-525c-481b-b84e-e0a7f68f633e] nwinfo = self.network_api.allocate_for_instance( [ 752.441979] env[61806]: ERROR nova.compute.manager [instance: 5b22c2fc-525c-481b-b84e-e0a7f68f633e] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 752.441979] env[61806]: ERROR nova.compute.manager [instance: 5b22c2fc-525c-481b-b84e-e0a7f68f633e] created_port_ids = self._update_ports_for_instance( [ 752.441979] env[61806]: ERROR nova.compute.manager [instance: 5b22c2fc-525c-481b-b84e-e0a7f68f633e] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 752.441979] env[61806]: ERROR nova.compute.manager [instance: 5b22c2fc-525c-481b-b84e-e0a7f68f633e] with excutils.save_and_reraise_exception(): [ 752.441979] env[61806]: ERROR nova.compute.manager [instance: 5b22c2fc-525c-481b-b84e-e0a7f68f633e] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 752.442435] env[61806]: ERROR nova.compute.manager [instance: 5b22c2fc-525c-481b-b84e-e0a7f68f633e] self.force_reraise() [ 752.442435] env[61806]: ERROR nova.compute.manager [instance: 5b22c2fc-525c-481b-b84e-e0a7f68f633e] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 752.442435] env[61806]: ERROR nova.compute.manager [instance: 5b22c2fc-525c-481b-b84e-e0a7f68f633e] raise self.value [ 752.442435] env[61806]: ERROR nova.compute.manager [instance: 5b22c2fc-525c-481b-b84e-e0a7f68f633e] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 752.442435] env[61806]: ERROR nova.compute.manager [instance: 5b22c2fc-525c-481b-b84e-e0a7f68f633e] updated_port = self._update_port( [ 752.442435] env[61806]: ERROR nova.compute.manager [instance: 5b22c2fc-525c-481b-b84e-e0a7f68f633e] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 752.442435] env[61806]: ERROR nova.compute.manager [instance: 5b22c2fc-525c-481b-b84e-e0a7f68f633e] _ensure_no_port_binding_failure(port) [ 752.442435] env[61806]: ERROR nova.compute.manager [instance: 5b22c2fc-525c-481b-b84e-e0a7f68f633e] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 752.442435] env[61806]: ERROR nova.compute.manager [instance: 5b22c2fc-525c-481b-b84e-e0a7f68f633e] raise exception.PortBindingFailed(port_id=port['id']) [ 752.442435] env[61806]: ERROR nova.compute.manager [instance: 5b22c2fc-525c-481b-b84e-e0a7f68f633e] nova.exception.PortBindingFailed: Binding failed for port 6cacc512-7619-41e7-8fcb-b88df2c7cf8b, please check neutron logs for more information. [ 752.442435] env[61806]: ERROR nova.compute.manager [instance: 5b22c2fc-525c-481b-b84e-e0a7f68f633e] [ 752.442854] env[61806]: DEBUG nova.compute.utils [None req-3c3981d2-fa25-4f4b-87d4-a5411e67199e tempest-MultipleCreateTestJSON-1589015462 tempest-MultipleCreateTestJSON-1589015462-project-member] [instance: 5b22c2fc-525c-481b-b84e-e0a7f68f633e] Binding failed for port 6cacc512-7619-41e7-8fcb-b88df2c7cf8b, please check neutron logs for more information. {{(pid=61806) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 752.442854] env[61806]: DEBUG oslo_concurrency.lockutils [None req-da1d2f3c-f6ea-47f5-8c64-65c86bf2d676 tempest-ServerRescueTestJSON-811948373 tempest-ServerRescueTestJSON-811948373-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 15.478s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 752.444142] env[61806]: INFO nova.compute.claims [None req-da1d2f3c-f6ea-47f5-8c64-65c86bf2d676 tempest-ServerRescueTestJSON-811948373 tempest-ServerRescueTestJSON-811948373-project-member] [instance: c61df5d3-7f43-48d4-a2c0-a8972372c0ba] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 752.446674] env[61806]: DEBUG nova.compute.manager [None req-3c3981d2-fa25-4f4b-87d4-a5411e67199e tempest-MultipleCreateTestJSON-1589015462 tempest-MultipleCreateTestJSON-1589015462-project-member] [instance: 5b22c2fc-525c-481b-b84e-e0a7f68f633e] Build of instance 5b22c2fc-525c-481b-b84e-e0a7f68f633e was re-scheduled: Binding failed for port 6cacc512-7619-41e7-8fcb-b88df2c7cf8b, please check neutron logs for more information. {{(pid=61806) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 752.447130] env[61806]: DEBUG nova.compute.manager [None req-3c3981d2-fa25-4f4b-87d4-a5411e67199e tempest-MultipleCreateTestJSON-1589015462 tempest-MultipleCreateTestJSON-1589015462-project-member] [instance: 5b22c2fc-525c-481b-b84e-e0a7f68f633e] Unplugging VIFs for instance {{(pid=61806) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 752.447485] env[61806]: DEBUG oslo_concurrency.lockutils [None req-3c3981d2-fa25-4f4b-87d4-a5411e67199e tempest-MultipleCreateTestJSON-1589015462 tempest-MultipleCreateTestJSON-1589015462-project-member] Acquiring lock "refresh_cache-5b22c2fc-525c-481b-b84e-e0a7f68f633e" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 752.447485] env[61806]: DEBUG oslo_concurrency.lockutils [None req-3c3981d2-fa25-4f4b-87d4-a5411e67199e tempest-MultipleCreateTestJSON-1589015462 tempest-MultipleCreateTestJSON-1589015462-project-member] Acquired lock "refresh_cache-5b22c2fc-525c-481b-b84e-e0a7f68f633e" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 752.447635] env[61806]: DEBUG nova.network.neutron [None req-3c3981d2-fa25-4f4b-87d4-a5411e67199e tempest-MultipleCreateTestJSON-1589015462 tempest-MultipleCreateTestJSON-1589015462-project-member] [instance: 5b22c2fc-525c-481b-b84e-e0a7f68f633e] Building network info cache for instance {{(pid=61806) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 752.587781] env[61806]: DEBUG nova.compute.manager [None req-f0377b51-dab9-4f68-8592-ca0f31b4ba51 tempest-FloatingIPsAssociationNegativeTestJSON-30494360 tempest-FloatingIPsAssociationNegativeTestJSON-30494360-project-member] [instance: 37c07b8d-e2a0-4aa6-b7ba-194feed837e5] Start spawning the instance on the hypervisor. {{(pid=61806) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 752.609546] env[61806]: DEBUG oslo_concurrency.lockutils [req-c951326f-e784-41c4-8566-93d0a0877265 req-afd1da5f-d8ba-442f-94a7-5f46ce883f37 service nova] Releasing lock "refresh_cache-198dc1a4-050d-47b2-8bc4-30af5534a789" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 752.609849] env[61806]: DEBUG nova.compute.manager [req-c951326f-e784-41c4-8566-93d0a0877265 req-afd1da5f-d8ba-442f-94a7-5f46ce883f37 service nova] [instance: 198dc1a4-050d-47b2-8bc4-30af5534a789] Received event network-vif-deleted-1a279290-dcd8-45fb-a136-2fcbc132e8d7 {{(pid=61806) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 752.612891] env[61806]: DEBUG nova.virt.hardware [None req-f0377b51-dab9-4f68-8592-ca0f31b4ba51 tempest-FloatingIPsAssociationNegativeTestJSON-30494360 tempest-FloatingIPsAssociationNegativeTestJSON-30494360-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-15T18:05:10Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-15T18:04:52Z,direct_url=,disk_format='vmdk',id=ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='84ba83fce2c349a6988173c5d6fc3b07',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-15T18:04:53Z,virtual_size=,visibility=), allow threads: False {{(pid=61806) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 752.613147] env[61806]: DEBUG nova.virt.hardware [None req-f0377b51-dab9-4f68-8592-ca0f31b4ba51 tempest-FloatingIPsAssociationNegativeTestJSON-30494360 tempest-FloatingIPsAssociationNegativeTestJSON-30494360-project-member] Flavor limits 0:0:0 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 752.613305] env[61806]: DEBUG nova.virt.hardware [None req-f0377b51-dab9-4f68-8592-ca0f31b4ba51 tempest-FloatingIPsAssociationNegativeTestJSON-30494360 tempest-FloatingIPsAssociationNegativeTestJSON-30494360-project-member] Image limits 0:0:0 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 752.613486] env[61806]: DEBUG nova.virt.hardware [None req-f0377b51-dab9-4f68-8592-ca0f31b4ba51 tempest-FloatingIPsAssociationNegativeTestJSON-30494360 tempest-FloatingIPsAssociationNegativeTestJSON-30494360-project-member] Flavor pref 0:0:0 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 752.613634] env[61806]: DEBUG nova.virt.hardware [None req-f0377b51-dab9-4f68-8592-ca0f31b4ba51 tempest-FloatingIPsAssociationNegativeTestJSON-30494360 tempest-FloatingIPsAssociationNegativeTestJSON-30494360-project-member] Image pref 0:0:0 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 752.613785] env[61806]: DEBUG nova.virt.hardware [None req-f0377b51-dab9-4f68-8592-ca0f31b4ba51 tempest-FloatingIPsAssociationNegativeTestJSON-30494360 tempest-FloatingIPsAssociationNegativeTestJSON-30494360-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 752.613989] env[61806]: DEBUG nova.virt.hardware [None req-f0377b51-dab9-4f68-8592-ca0f31b4ba51 tempest-FloatingIPsAssociationNegativeTestJSON-30494360 tempest-FloatingIPsAssociationNegativeTestJSON-30494360-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61806) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 752.614161] env[61806]: DEBUG nova.virt.hardware [None req-f0377b51-dab9-4f68-8592-ca0f31b4ba51 tempest-FloatingIPsAssociationNegativeTestJSON-30494360 tempest-FloatingIPsAssociationNegativeTestJSON-30494360-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61806) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 752.614330] env[61806]: DEBUG nova.virt.hardware [None req-f0377b51-dab9-4f68-8592-ca0f31b4ba51 tempest-FloatingIPsAssociationNegativeTestJSON-30494360 tempest-FloatingIPsAssociationNegativeTestJSON-30494360-project-member] Got 1 possible topologies {{(pid=61806) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 752.614503] env[61806]: DEBUG nova.virt.hardware [None req-f0377b51-dab9-4f68-8592-ca0f31b4ba51 tempest-FloatingIPsAssociationNegativeTestJSON-30494360 tempest-FloatingIPsAssociationNegativeTestJSON-30494360-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61806) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 752.614679] env[61806]: DEBUG nova.virt.hardware [None req-f0377b51-dab9-4f68-8592-ca0f31b4ba51 tempest-FloatingIPsAssociationNegativeTestJSON-30494360 tempest-FloatingIPsAssociationNegativeTestJSON-30494360-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61806) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 752.615941] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7030de8a-420c-465c-9b49-a00b78acf8bf {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 752.624297] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-848e4310-441a-443c-bd0b-0742bcffeb1d {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 752.637920] env[61806]: ERROR nova.compute.manager [None req-f0377b51-dab9-4f68-8592-ca0f31b4ba51 tempest-FloatingIPsAssociationNegativeTestJSON-30494360 tempest-FloatingIPsAssociationNegativeTestJSON-30494360-project-member] [instance: 37c07b8d-e2a0-4aa6-b7ba-194feed837e5] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 1b064d62-24f5-4892-937d-ec37537f02ac, please check neutron logs for more information. [ 752.637920] env[61806]: ERROR nova.compute.manager [instance: 37c07b8d-e2a0-4aa6-b7ba-194feed837e5] Traceback (most recent call last): [ 752.637920] env[61806]: ERROR nova.compute.manager [instance: 37c07b8d-e2a0-4aa6-b7ba-194feed837e5] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 752.637920] env[61806]: ERROR nova.compute.manager [instance: 37c07b8d-e2a0-4aa6-b7ba-194feed837e5] yield resources [ 752.637920] env[61806]: ERROR nova.compute.manager [instance: 37c07b8d-e2a0-4aa6-b7ba-194feed837e5] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 752.637920] env[61806]: ERROR nova.compute.manager [instance: 37c07b8d-e2a0-4aa6-b7ba-194feed837e5] self.driver.spawn(context, instance, image_meta, [ 752.637920] env[61806]: ERROR nova.compute.manager [instance: 37c07b8d-e2a0-4aa6-b7ba-194feed837e5] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 752.637920] env[61806]: ERROR nova.compute.manager [instance: 37c07b8d-e2a0-4aa6-b7ba-194feed837e5] self._vmops.spawn(context, instance, image_meta, injected_files, [ 752.637920] env[61806]: ERROR nova.compute.manager [instance: 37c07b8d-e2a0-4aa6-b7ba-194feed837e5] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 752.637920] env[61806]: ERROR nova.compute.manager [instance: 37c07b8d-e2a0-4aa6-b7ba-194feed837e5] vm_ref = self.build_virtual_machine(instance, [ 752.637920] env[61806]: ERROR nova.compute.manager [instance: 37c07b8d-e2a0-4aa6-b7ba-194feed837e5] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 752.638448] env[61806]: ERROR nova.compute.manager [instance: 37c07b8d-e2a0-4aa6-b7ba-194feed837e5] vif_infos = vmwarevif.get_vif_info(self._session, [ 752.638448] env[61806]: ERROR nova.compute.manager [instance: 37c07b8d-e2a0-4aa6-b7ba-194feed837e5] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 752.638448] env[61806]: ERROR nova.compute.manager [instance: 37c07b8d-e2a0-4aa6-b7ba-194feed837e5] for vif in network_info: [ 752.638448] env[61806]: ERROR nova.compute.manager [instance: 37c07b8d-e2a0-4aa6-b7ba-194feed837e5] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 752.638448] env[61806]: ERROR nova.compute.manager [instance: 37c07b8d-e2a0-4aa6-b7ba-194feed837e5] return self._sync_wrapper(fn, *args, **kwargs) [ 752.638448] env[61806]: ERROR nova.compute.manager [instance: 37c07b8d-e2a0-4aa6-b7ba-194feed837e5] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 752.638448] env[61806]: ERROR nova.compute.manager [instance: 37c07b8d-e2a0-4aa6-b7ba-194feed837e5] self.wait() [ 752.638448] env[61806]: ERROR nova.compute.manager [instance: 37c07b8d-e2a0-4aa6-b7ba-194feed837e5] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 752.638448] env[61806]: ERROR nova.compute.manager [instance: 37c07b8d-e2a0-4aa6-b7ba-194feed837e5] self[:] = self._gt.wait() [ 752.638448] env[61806]: ERROR nova.compute.manager [instance: 37c07b8d-e2a0-4aa6-b7ba-194feed837e5] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 752.638448] env[61806]: ERROR nova.compute.manager [instance: 37c07b8d-e2a0-4aa6-b7ba-194feed837e5] return self._exit_event.wait() [ 752.638448] env[61806]: ERROR nova.compute.manager [instance: 37c07b8d-e2a0-4aa6-b7ba-194feed837e5] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 752.638448] env[61806]: ERROR nova.compute.manager [instance: 37c07b8d-e2a0-4aa6-b7ba-194feed837e5] current.throw(*self._exc) [ 752.638971] env[61806]: ERROR nova.compute.manager [instance: 37c07b8d-e2a0-4aa6-b7ba-194feed837e5] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 752.638971] env[61806]: ERROR nova.compute.manager [instance: 37c07b8d-e2a0-4aa6-b7ba-194feed837e5] result = function(*args, **kwargs) [ 752.638971] env[61806]: ERROR nova.compute.manager [instance: 37c07b8d-e2a0-4aa6-b7ba-194feed837e5] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 752.638971] env[61806]: ERROR nova.compute.manager [instance: 37c07b8d-e2a0-4aa6-b7ba-194feed837e5] return func(*args, **kwargs) [ 752.638971] env[61806]: ERROR nova.compute.manager [instance: 37c07b8d-e2a0-4aa6-b7ba-194feed837e5] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 752.638971] env[61806]: ERROR nova.compute.manager [instance: 37c07b8d-e2a0-4aa6-b7ba-194feed837e5] raise e [ 752.638971] env[61806]: ERROR nova.compute.manager [instance: 37c07b8d-e2a0-4aa6-b7ba-194feed837e5] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 752.638971] env[61806]: ERROR nova.compute.manager [instance: 37c07b8d-e2a0-4aa6-b7ba-194feed837e5] nwinfo = self.network_api.allocate_for_instance( [ 752.638971] env[61806]: ERROR nova.compute.manager [instance: 37c07b8d-e2a0-4aa6-b7ba-194feed837e5] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 752.638971] env[61806]: ERROR nova.compute.manager [instance: 37c07b8d-e2a0-4aa6-b7ba-194feed837e5] created_port_ids = self._update_ports_for_instance( [ 752.638971] env[61806]: ERROR nova.compute.manager [instance: 37c07b8d-e2a0-4aa6-b7ba-194feed837e5] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 752.638971] env[61806]: ERROR nova.compute.manager [instance: 37c07b8d-e2a0-4aa6-b7ba-194feed837e5] with excutils.save_and_reraise_exception(): [ 752.638971] env[61806]: ERROR nova.compute.manager [instance: 37c07b8d-e2a0-4aa6-b7ba-194feed837e5] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 752.639488] env[61806]: ERROR nova.compute.manager [instance: 37c07b8d-e2a0-4aa6-b7ba-194feed837e5] self.force_reraise() [ 752.639488] env[61806]: ERROR nova.compute.manager [instance: 37c07b8d-e2a0-4aa6-b7ba-194feed837e5] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 752.639488] env[61806]: ERROR nova.compute.manager [instance: 37c07b8d-e2a0-4aa6-b7ba-194feed837e5] raise self.value [ 752.639488] env[61806]: ERROR nova.compute.manager [instance: 37c07b8d-e2a0-4aa6-b7ba-194feed837e5] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 752.639488] env[61806]: ERROR nova.compute.manager [instance: 37c07b8d-e2a0-4aa6-b7ba-194feed837e5] updated_port = self._update_port( [ 752.639488] env[61806]: ERROR nova.compute.manager [instance: 37c07b8d-e2a0-4aa6-b7ba-194feed837e5] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 752.639488] env[61806]: ERROR nova.compute.manager [instance: 37c07b8d-e2a0-4aa6-b7ba-194feed837e5] _ensure_no_port_binding_failure(port) [ 752.639488] env[61806]: ERROR nova.compute.manager [instance: 37c07b8d-e2a0-4aa6-b7ba-194feed837e5] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 752.639488] env[61806]: ERROR nova.compute.manager [instance: 37c07b8d-e2a0-4aa6-b7ba-194feed837e5] raise exception.PortBindingFailed(port_id=port['id']) [ 752.639488] env[61806]: ERROR nova.compute.manager [instance: 37c07b8d-e2a0-4aa6-b7ba-194feed837e5] nova.exception.PortBindingFailed: Binding failed for port 1b064d62-24f5-4892-937d-ec37537f02ac, please check neutron logs for more information. [ 752.639488] env[61806]: ERROR nova.compute.manager [instance: 37c07b8d-e2a0-4aa6-b7ba-194feed837e5] [ 752.639488] env[61806]: INFO nova.compute.manager [None req-f0377b51-dab9-4f68-8592-ca0f31b4ba51 tempest-FloatingIPsAssociationNegativeTestJSON-30494360 tempest-FloatingIPsAssociationNegativeTestJSON-30494360-project-member] [instance: 37c07b8d-e2a0-4aa6-b7ba-194feed837e5] Terminating instance [ 752.640303] env[61806]: DEBUG oslo_concurrency.lockutils [None req-f0377b51-dab9-4f68-8592-ca0f31b4ba51 tempest-FloatingIPsAssociationNegativeTestJSON-30494360 tempest-FloatingIPsAssociationNegativeTestJSON-30494360-project-member] Acquiring lock "refresh_cache-37c07b8d-e2a0-4aa6-b7ba-194feed837e5" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 752.640460] env[61806]: DEBUG oslo_concurrency.lockutils [None req-f0377b51-dab9-4f68-8592-ca0f31b4ba51 tempest-FloatingIPsAssociationNegativeTestJSON-30494360 tempest-FloatingIPsAssociationNegativeTestJSON-30494360-project-member] Acquired lock "refresh_cache-37c07b8d-e2a0-4aa6-b7ba-194feed837e5" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 752.640629] env[61806]: DEBUG nova.network.neutron [None req-f0377b51-dab9-4f68-8592-ca0f31b4ba51 tempest-FloatingIPsAssociationNegativeTestJSON-30494360 tempest-FloatingIPsAssociationNegativeTestJSON-30494360-project-member] [instance: 37c07b8d-e2a0-4aa6-b7ba-194feed837e5] Building network info cache for instance {{(pid=61806) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 752.969216] env[61806]: DEBUG nova.network.neutron [None req-3c3981d2-fa25-4f4b-87d4-a5411e67199e tempest-MultipleCreateTestJSON-1589015462 tempest-MultipleCreateTestJSON-1589015462-project-member] [instance: 5b22c2fc-525c-481b-b84e-e0a7f68f633e] Instance cache missing network info. {{(pid=61806) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 753.045896] env[61806]: DEBUG nova.network.neutron [None req-3c3981d2-fa25-4f4b-87d4-a5411e67199e tempest-MultipleCreateTestJSON-1589015462 tempest-MultipleCreateTestJSON-1589015462-project-member] [instance: 5b22c2fc-525c-481b-b84e-e0a7f68f633e] Updating instance_info_cache with network_info: [] {{(pid=61806) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 753.157507] env[61806]: DEBUG nova.network.neutron [None req-f0377b51-dab9-4f68-8592-ca0f31b4ba51 tempest-FloatingIPsAssociationNegativeTestJSON-30494360 tempest-FloatingIPsAssociationNegativeTestJSON-30494360-project-member] [instance: 37c07b8d-e2a0-4aa6-b7ba-194feed837e5] Instance cache missing network info. {{(pid=61806) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 753.246059] env[61806]: DEBUG nova.network.neutron [None req-f0377b51-dab9-4f68-8592-ca0f31b4ba51 tempest-FloatingIPsAssociationNegativeTestJSON-30494360 tempest-FloatingIPsAssociationNegativeTestJSON-30494360-project-member] [instance: 37c07b8d-e2a0-4aa6-b7ba-194feed837e5] Updating instance_info_cache with network_info: [] {{(pid=61806) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 753.532017] env[61806]: DEBUG nova.compute.manager [req-2d2d6d01-c734-4e03-b2f2-f52fde8dca14 req-14c4320c-f9b5-491e-9558-80b521cd3edf service nova] [instance: 37c07b8d-e2a0-4aa6-b7ba-194feed837e5] Received event network-changed-1b064d62-24f5-4892-937d-ec37537f02ac {{(pid=61806) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 753.532285] env[61806]: DEBUG nova.compute.manager [req-2d2d6d01-c734-4e03-b2f2-f52fde8dca14 req-14c4320c-f9b5-491e-9558-80b521cd3edf service nova] [instance: 37c07b8d-e2a0-4aa6-b7ba-194feed837e5] Refreshing instance network info cache due to event network-changed-1b064d62-24f5-4892-937d-ec37537f02ac. {{(pid=61806) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 753.532427] env[61806]: DEBUG oslo_concurrency.lockutils [req-2d2d6d01-c734-4e03-b2f2-f52fde8dca14 req-14c4320c-f9b5-491e-9558-80b521cd3edf service nova] Acquiring lock "refresh_cache-37c07b8d-e2a0-4aa6-b7ba-194feed837e5" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 753.548576] env[61806]: DEBUG oslo_concurrency.lockutils [None req-3c3981d2-fa25-4f4b-87d4-a5411e67199e tempest-MultipleCreateTestJSON-1589015462 tempest-MultipleCreateTestJSON-1589015462-project-member] Releasing lock "refresh_cache-5b22c2fc-525c-481b-b84e-e0a7f68f633e" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 753.548798] env[61806]: DEBUG nova.compute.manager [None req-3c3981d2-fa25-4f4b-87d4-a5411e67199e tempest-MultipleCreateTestJSON-1589015462 tempest-MultipleCreateTestJSON-1589015462-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61806) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 753.548999] env[61806]: DEBUG nova.compute.manager [None req-3c3981d2-fa25-4f4b-87d4-a5411e67199e tempest-MultipleCreateTestJSON-1589015462 tempest-MultipleCreateTestJSON-1589015462-project-member] [instance: 5b22c2fc-525c-481b-b84e-e0a7f68f633e] Deallocating network for instance {{(pid=61806) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 753.549141] env[61806]: DEBUG nova.network.neutron [None req-3c3981d2-fa25-4f4b-87d4-a5411e67199e tempest-MultipleCreateTestJSON-1589015462 tempest-MultipleCreateTestJSON-1589015462-project-member] [instance: 5b22c2fc-525c-481b-b84e-e0a7f68f633e] deallocate_for_instance() {{(pid=61806) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 753.564010] env[61806]: DEBUG nova.network.neutron [None req-3c3981d2-fa25-4f4b-87d4-a5411e67199e tempest-MultipleCreateTestJSON-1589015462 tempest-MultipleCreateTestJSON-1589015462-project-member] [instance: 5b22c2fc-525c-481b-b84e-e0a7f68f633e] Instance cache missing network info. {{(pid=61806) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 753.705671] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-727d0522-eec1-4dd9-ac5a-b6de6bd4c373 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 753.713183] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-82b0edab-3c7d-4a74-aba0-3f57ae5ef4b1 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 753.742220] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ef22fbad-858d-4ef0-8f7c-e122dcf50e76 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 753.748175] env[61806]: DEBUG oslo_concurrency.lockutils [None req-f0377b51-dab9-4f68-8592-ca0f31b4ba51 tempest-FloatingIPsAssociationNegativeTestJSON-30494360 tempest-FloatingIPsAssociationNegativeTestJSON-30494360-project-member] Releasing lock "refresh_cache-37c07b8d-e2a0-4aa6-b7ba-194feed837e5" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 753.748571] env[61806]: DEBUG nova.compute.manager [None req-f0377b51-dab9-4f68-8592-ca0f31b4ba51 tempest-FloatingIPsAssociationNegativeTestJSON-30494360 tempest-FloatingIPsAssociationNegativeTestJSON-30494360-project-member] [instance: 37c07b8d-e2a0-4aa6-b7ba-194feed837e5] Start destroying the instance on the hypervisor. {{(pid=61806) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 753.748797] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-f0377b51-dab9-4f68-8592-ca0f31b4ba51 tempest-FloatingIPsAssociationNegativeTestJSON-30494360 tempest-FloatingIPsAssociationNegativeTestJSON-30494360-project-member] [instance: 37c07b8d-e2a0-4aa6-b7ba-194feed837e5] Destroying instance {{(pid=61806) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 753.749121] env[61806]: DEBUG oslo_concurrency.lockutils [req-2d2d6d01-c734-4e03-b2f2-f52fde8dca14 req-14c4320c-f9b5-491e-9558-80b521cd3edf service nova] Acquired lock "refresh_cache-37c07b8d-e2a0-4aa6-b7ba-194feed837e5" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 753.749293] env[61806]: DEBUG nova.network.neutron [req-2d2d6d01-c734-4e03-b2f2-f52fde8dca14 req-14c4320c-f9b5-491e-9558-80b521cd3edf service nova] [instance: 37c07b8d-e2a0-4aa6-b7ba-194feed837e5] Refreshing network info cache for port 1b064d62-24f5-4892-937d-ec37537f02ac {{(pid=61806) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 753.750294] env[61806]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-86529fd3-fd7e-4839-b5b2-14ab01f7f564 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 753.752896] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3c31a323-7352-4077-9de0-5932f9477269 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 753.766794] env[61806]: DEBUG nova.compute.provider_tree [None req-da1d2f3c-f6ea-47f5-8c64-65c86bf2d676 tempest-ServerRescueTestJSON-811948373 tempest-ServerRescueTestJSON-811948373-project-member] Inventory has not changed in ProviderTree for provider: a2858be1-fd22-4e08-979e-87ad25293407 {{(pid=61806) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 753.771157] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-81d1b9db-203f-4c25-9504-f84e8cf797fa {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 753.792755] env[61806]: WARNING nova.virt.vmwareapi.vmops [None req-f0377b51-dab9-4f68-8592-ca0f31b4ba51 tempest-FloatingIPsAssociationNegativeTestJSON-30494360 tempest-FloatingIPsAssociationNegativeTestJSON-30494360-project-member] [instance: 37c07b8d-e2a0-4aa6-b7ba-194feed837e5] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 37c07b8d-e2a0-4aa6-b7ba-194feed837e5 could not be found. [ 753.792968] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-f0377b51-dab9-4f68-8592-ca0f31b4ba51 tempest-FloatingIPsAssociationNegativeTestJSON-30494360 tempest-FloatingIPsAssociationNegativeTestJSON-30494360-project-member] [instance: 37c07b8d-e2a0-4aa6-b7ba-194feed837e5] Instance destroyed {{(pid=61806) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 753.793202] env[61806]: INFO nova.compute.manager [None req-f0377b51-dab9-4f68-8592-ca0f31b4ba51 tempest-FloatingIPsAssociationNegativeTestJSON-30494360 tempest-FloatingIPsAssociationNegativeTestJSON-30494360-project-member] [instance: 37c07b8d-e2a0-4aa6-b7ba-194feed837e5] Took 0.04 seconds to destroy the instance on the hypervisor. [ 753.793410] env[61806]: DEBUG oslo.service.loopingcall [None req-f0377b51-dab9-4f68-8592-ca0f31b4ba51 tempest-FloatingIPsAssociationNegativeTestJSON-30494360 tempest-FloatingIPsAssociationNegativeTestJSON-30494360-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61806) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 753.793618] env[61806]: DEBUG nova.compute.manager [-] [instance: 37c07b8d-e2a0-4aa6-b7ba-194feed837e5] Deallocating network for instance {{(pid=61806) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 753.793725] env[61806]: DEBUG nova.network.neutron [-] [instance: 37c07b8d-e2a0-4aa6-b7ba-194feed837e5] deallocate_for_instance() {{(pid=61806) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 753.808760] env[61806]: DEBUG nova.network.neutron [-] [instance: 37c07b8d-e2a0-4aa6-b7ba-194feed837e5] Instance cache missing network info. {{(pid=61806) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 754.070010] env[61806]: DEBUG nova.network.neutron [None req-3c3981d2-fa25-4f4b-87d4-a5411e67199e tempest-MultipleCreateTestJSON-1589015462 tempest-MultipleCreateTestJSON-1589015462-project-member] [instance: 5b22c2fc-525c-481b-b84e-e0a7f68f633e] Updating instance_info_cache with network_info: [] {{(pid=61806) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 754.269805] env[61806]: DEBUG nova.scheduler.client.report [None req-da1d2f3c-f6ea-47f5-8c64-65c86bf2d676 tempest-ServerRescueTestJSON-811948373 tempest-ServerRescueTestJSON-811948373-project-member] Inventory has not changed for provider a2858be1-fd22-4e08-979e-87ad25293407 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 139, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61806) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 754.274039] env[61806]: DEBUG nova.network.neutron [req-2d2d6d01-c734-4e03-b2f2-f52fde8dca14 req-14c4320c-f9b5-491e-9558-80b521cd3edf service nova] [instance: 37c07b8d-e2a0-4aa6-b7ba-194feed837e5] Instance cache missing network info. {{(pid=61806) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 754.312051] env[61806]: DEBUG nova.network.neutron [-] [instance: 37c07b8d-e2a0-4aa6-b7ba-194feed837e5] Updating instance_info_cache with network_info: [] {{(pid=61806) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 754.348071] env[61806]: DEBUG nova.network.neutron [req-2d2d6d01-c734-4e03-b2f2-f52fde8dca14 req-14c4320c-f9b5-491e-9558-80b521cd3edf service nova] [instance: 37c07b8d-e2a0-4aa6-b7ba-194feed837e5] Updating instance_info_cache with network_info: [] {{(pid=61806) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 754.573036] env[61806]: INFO nova.compute.manager [None req-3c3981d2-fa25-4f4b-87d4-a5411e67199e tempest-MultipleCreateTestJSON-1589015462 tempest-MultipleCreateTestJSON-1589015462-project-member] [instance: 5b22c2fc-525c-481b-b84e-e0a7f68f633e] Took 1.02 seconds to deallocate network for instance. [ 754.774275] env[61806]: DEBUG oslo_concurrency.lockutils [None req-da1d2f3c-f6ea-47f5-8c64-65c86bf2d676 tempest-ServerRescueTestJSON-811948373 tempest-ServerRescueTestJSON-811948373-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.331s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 754.774817] env[61806]: DEBUG nova.compute.manager [None req-da1d2f3c-f6ea-47f5-8c64-65c86bf2d676 tempest-ServerRescueTestJSON-811948373 tempest-ServerRescueTestJSON-811948373-project-member] [instance: c61df5d3-7f43-48d4-a2c0-a8972372c0ba] Start building networks asynchronously for instance. {{(pid=61806) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 754.777412] env[61806]: DEBUG oslo_concurrency.lockutils [None req-d17e7876-7c69-44af-acb2-32ed225ce945 tempest-ServerActionsTestOtherA-1637843849 tempest-ServerActionsTestOtherA-1637843849-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 14.283s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 754.813489] env[61806]: INFO nova.compute.manager [-] [instance: 37c07b8d-e2a0-4aa6-b7ba-194feed837e5] Took 1.02 seconds to deallocate network for instance. [ 754.815481] env[61806]: DEBUG nova.compute.claims [None req-f0377b51-dab9-4f68-8592-ca0f31b4ba51 tempest-FloatingIPsAssociationNegativeTestJSON-30494360 tempest-FloatingIPsAssociationNegativeTestJSON-30494360-project-member] [instance: 37c07b8d-e2a0-4aa6-b7ba-194feed837e5] Aborting claim: {{(pid=61806) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 754.815687] env[61806]: DEBUG oslo_concurrency.lockutils [None req-f0377b51-dab9-4f68-8592-ca0f31b4ba51 tempest-FloatingIPsAssociationNegativeTestJSON-30494360 tempest-FloatingIPsAssociationNegativeTestJSON-30494360-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 754.850571] env[61806]: DEBUG oslo_concurrency.lockutils [req-2d2d6d01-c734-4e03-b2f2-f52fde8dca14 req-14c4320c-f9b5-491e-9558-80b521cd3edf service nova] Releasing lock "refresh_cache-37c07b8d-e2a0-4aa6-b7ba-194feed837e5" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 754.850827] env[61806]: DEBUG nova.compute.manager [req-2d2d6d01-c734-4e03-b2f2-f52fde8dca14 req-14c4320c-f9b5-491e-9558-80b521cd3edf service nova] [instance: 37c07b8d-e2a0-4aa6-b7ba-194feed837e5] Received event network-vif-deleted-1b064d62-24f5-4892-937d-ec37537f02ac {{(pid=61806) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 755.283063] env[61806]: DEBUG nova.compute.utils [None req-da1d2f3c-f6ea-47f5-8c64-65c86bf2d676 tempest-ServerRescueTestJSON-811948373 tempest-ServerRescueTestJSON-811948373-project-member] Using /dev/sd instead of None {{(pid=61806) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 755.288310] env[61806]: DEBUG nova.compute.manager [None req-da1d2f3c-f6ea-47f5-8c64-65c86bf2d676 tempest-ServerRescueTestJSON-811948373 tempest-ServerRescueTestJSON-811948373-project-member] [instance: c61df5d3-7f43-48d4-a2c0-a8972372c0ba] Allocating IP information in the background. {{(pid=61806) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 755.288310] env[61806]: DEBUG nova.network.neutron [None req-da1d2f3c-f6ea-47f5-8c64-65c86bf2d676 tempest-ServerRescueTestJSON-811948373 tempest-ServerRescueTestJSON-811948373-project-member] [instance: c61df5d3-7f43-48d4-a2c0-a8972372c0ba] allocate_for_instance() {{(pid=61806) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 755.327901] env[61806]: DEBUG nova.policy [None req-da1d2f3c-f6ea-47f5-8c64-65c86bf2d676 tempest-ServerRescueTestJSON-811948373 tempest-ServerRescueTestJSON-811948373-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'cfa2da68390840bea25bb8265771c1ff', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '87bf663590b74f908db2794041707733', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61806) authorize /opt/stack/nova/nova/policy.py:201}} [ 755.547264] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d215b08a-802d-47ba-82ac-74ed8a6087b8 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 755.553447] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-297d1584-0d5d-44d5-816b-4e3298d524b5 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 755.587108] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4c9eab4c-326b-457f-b5fb-da0afcef2e61 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 755.594749] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-57879698-8581-482a-aab4-f7f888fc418c {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 755.599353] env[61806]: DEBUG nova.network.neutron [None req-da1d2f3c-f6ea-47f5-8c64-65c86bf2d676 tempest-ServerRescueTestJSON-811948373 tempest-ServerRescueTestJSON-811948373-project-member] [instance: c61df5d3-7f43-48d4-a2c0-a8972372c0ba] Successfully created port: b12a7881-ba2b-4a75-8119-9f5cb5b2cdb7 {{(pid=61806) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 755.610530] env[61806]: DEBUG nova.compute.provider_tree [None req-d17e7876-7c69-44af-acb2-32ed225ce945 tempest-ServerActionsTestOtherA-1637843849 tempest-ServerActionsTestOtherA-1637843849-project-member] Inventory has not changed in ProviderTree for provider: a2858be1-fd22-4e08-979e-87ad25293407 {{(pid=61806) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 755.612317] env[61806]: INFO nova.scheduler.client.report [None req-3c3981d2-fa25-4f4b-87d4-a5411e67199e tempest-MultipleCreateTestJSON-1589015462 tempest-MultipleCreateTestJSON-1589015462-project-member] Deleted allocations for instance 5b22c2fc-525c-481b-b84e-e0a7f68f633e [ 755.622707] env[61806]: DEBUG nova.scheduler.client.report [None req-d17e7876-7c69-44af-acb2-32ed225ce945 tempest-ServerActionsTestOtherA-1637843849 tempest-ServerActionsTestOtherA-1637843849-project-member] Inventory has not changed for provider a2858be1-fd22-4e08-979e-87ad25293407 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 139, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61806) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 755.788763] env[61806]: DEBUG nova.compute.manager [None req-da1d2f3c-f6ea-47f5-8c64-65c86bf2d676 tempest-ServerRescueTestJSON-811948373 tempest-ServerRescueTestJSON-811948373-project-member] [instance: c61df5d3-7f43-48d4-a2c0-a8972372c0ba] Start building block device mappings for instance. {{(pid=61806) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 756.124791] env[61806]: DEBUG oslo_concurrency.lockutils [None req-d17e7876-7c69-44af-acb2-32ed225ce945 tempest-ServerActionsTestOtherA-1637843849 tempest-ServerActionsTestOtherA-1637843849-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.347s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 756.125446] env[61806]: ERROR nova.compute.manager [None req-d17e7876-7c69-44af-acb2-32ed225ce945 tempest-ServerActionsTestOtherA-1637843849 tempest-ServerActionsTestOtherA-1637843849-project-member] [instance: 15476ac0-289a-4e04-aa9d-4244c658e962] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 76b9d660-f9cc-4d3f-ae94-28e2b3d3b46d, please check neutron logs for more information. [ 756.125446] env[61806]: ERROR nova.compute.manager [instance: 15476ac0-289a-4e04-aa9d-4244c658e962] Traceback (most recent call last): [ 756.125446] env[61806]: ERROR nova.compute.manager [instance: 15476ac0-289a-4e04-aa9d-4244c658e962] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 756.125446] env[61806]: ERROR nova.compute.manager [instance: 15476ac0-289a-4e04-aa9d-4244c658e962] self.driver.spawn(context, instance, image_meta, [ 756.125446] env[61806]: ERROR nova.compute.manager [instance: 15476ac0-289a-4e04-aa9d-4244c658e962] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 756.125446] env[61806]: ERROR nova.compute.manager [instance: 15476ac0-289a-4e04-aa9d-4244c658e962] self._vmops.spawn(context, instance, image_meta, injected_files, [ 756.125446] env[61806]: ERROR nova.compute.manager [instance: 15476ac0-289a-4e04-aa9d-4244c658e962] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 756.125446] env[61806]: ERROR nova.compute.manager [instance: 15476ac0-289a-4e04-aa9d-4244c658e962] vm_ref = self.build_virtual_machine(instance, [ 756.125446] env[61806]: ERROR nova.compute.manager [instance: 15476ac0-289a-4e04-aa9d-4244c658e962] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 756.125446] env[61806]: ERROR nova.compute.manager [instance: 15476ac0-289a-4e04-aa9d-4244c658e962] vif_infos = vmwarevif.get_vif_info(self._session, [ 756.125446] env[61806]: ERROR nova.compute.manager [instance: 15476ac0-289a-4e04-aa9d-4244c658e962] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 756.125882] env[61806]: ERROR nova.compute.manager [instance: 15476ac0-289a-4e04-aa9d-4244c658e962] for vif in network_info: [ 756.125882] env[61806]: ERROR nova.compute.manager [instance: 15476ac0-289a-4e04-aa9d-4244c658e962] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 756.125882] env[61806]: ERROR nova.compute.manager [instance: 15476ac0-289a-4e04-aa9d-4244c658e962] return self._sync_wrapper(fn, *args, **kwargs) [ 756.125882] env[61806]: ERROR nova.compute.manager [instance: 15476ac0-289a-4e04-aa9d-4244c658e962] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 756.125882] env[61806]: ERROR nova.compute.manager [instance: 15476ac0-289a-4e04-aa9d-4244c658e962] self.wait() [ 756.125882] env[61806]: ERROR nova.compute.manager [instance: 15476ac0-289a-4e04-aa9d-4244c658e962] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 756.125882] env[61806]: ERROR nova.compute.manager [instance: 15476ac0-289a-4e04-aa9d-4244c658e962] self[:] = self._gt.wait() [ 756.125882] env[61806]: ERROR nova.compute.manager [instance: 15476ac0-289a-4e04-aa9d-4244c658e962] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 756.125882] env[61806]: ERROR nova.compute.manager [instance: 15476ac0-289a-4e04-aa9d-4244c658e962] return self._exit_event.wait() [ 756.125882] env[61806]: ERROR nova.compute.manager [instance: 15476ac0-289a-4e04-aa9d-4244c658e962] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 756.125882] env[61806]: ERROR nova.compute.manager [instance: 15476ac0-289a-4e04-aa9d-4244c658e962] result = hub.switch() [ 756.125882] env[61806]: ERROR nova.compute.manager [instance: 15476ac0-289a-4e04-aa9d-4244c658e962] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 756.125882] env[61806]: ERROR nova.compute.manager [instance: 15476ac0-289a-4e04-aa9d-4244c658e962] return self.greenlet.switch() [ 756.126323] env[61806]: ERROR nova.compute.manager [instance: 15476ac0-289a-4e04-aa9d-4244c658e962] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 756.126323] env[61806]: ERROR nova.compute.manager [instance: 15476ac0-289a-4e04-aa9d-4244c658e962] result = function(*args, **kwargs) [ 756.126323] env[61806]: ERROR nova.compute.manager [instance: 15476ac0-289a-4e04-aa9d-4244c658e962] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 756.126323] env[61806]: ERROR nova.compute.manager [instance: 15476ac0-289a-4e04-aa9d-4244c658e962] return func(*args, **kwargs) [ 756.126323] env[61806]: ERROR nova.compute.manager [instance: 15476ac0-289a-4e04-aa9d-4244c658e962] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 756.126323] env[61806]: ERROR nova.compute.manager [instance: 15476ac0-289a-4e04-aa9d-4244c658e962] raise e [ 756.126323] env[61806]: ERROR nova.compute.manager [instance: 15476ac0-289a-4e04-aa9d-4244c658e962] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 756.126323] env[61806]: ERROR nova.compute.manager [instance: 15476ac0-289a-4e04-aa9d-4244c658e962] nwinfo = self.network_api.allocate_for_instance( [ 756.126323] env[61806]: ERROR nova.compute.manager [instance: 15476ac0-289a-4e04-aa9d-4244c658e962] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 756.126323] env[61806]: ERROR nova.compute.manager [instance: 15476ac0-289a-4e04-aa9d-4244c658e962] created_port_ids = self._update_ports_for_instance( [ 756.126323] env[61806]: ERROR nova.compute.manager [instance: 15476ac0-289a-4e04-aa9d-4244c658e962] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 756.126323] env[61806]: ERROR nova.compute.manager [instance: 15476ac0-289a-4e04-aa9d-4244c658e962] with excutils.save_and_reraise_exception(): [ 756.126323] env[61806]: ERROR nova.compute.manager [instance: 15476ac0-289a-4e04-aa9d-4244c658e962] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 756.126775] env[61806]: ERROR nova.compute.manager [instance: 15476ac0-289a-4e04-aa9d-4244c658e962] self.force_reraise() [ 756.126775] env[61806]: ERROR nova.compute.manager [instance: 15476ac0-289a-4e04-aa9d-4244c658e962] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 756.126775] env[61806]: ERROR nova.compute.manager [instance: 15476ac0-289a-4e04-aa9d-4244c658e962] raise self.value [ 756.126775] env[61806]: ERROR nova.compute.manager [instance: 15476ac0-289a-4e04-aa9d-4244c658e962] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 756.126775] env[61806]: ERROR nova.compute.manager [instance: 15476ac0-289a-4e04-aa9d-4244c658e962] updated_port = self._update_port( [ 756.126775] env[61806]: ERROR nova.compute.manager [instance: 15476ac0-289a-4e04-aa9d-4244c658e962] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 756.126775] env[61806]: ERROR nova.compute.manager [instance: 15476ac0-289a-4e04-aa9d-4244c658e962] _ensure_no_port_binding_failure(port) [ 756.126775] env[61806]: ERROR nova.compute.manager [instance: 15476ac0-289a-4e04-aa9d-4244c658e962] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 756.126775] env[61806]: ERROR nova.compute.manager [instance: 15476ac0-289a-4e04-aa9d-4244c658e962] raise exception.PortBindingFailed(port_id=port['id']) [ 756.126775] env[61806]: ERROR nova.compute.manager [instance: 15476ac0-289a-4e04-aa9d-4244c658e962] nova.exception.PortBindingFailed: Binding failed for port 76b9d660-f9cc-4d3f-ae94-28e2b3d3b46d, please check neutron logs for more information. [ 756.126775] env[61806]: ERROR nova.compute.manager [instance: 15476ac0-289a-4e04-aa9d-4244c658e962] [ 756.127161] env[61806]: DEBUG nova.compute.utils [None req-d17e7876-7c69-44af-acb2-32ed225ce945 tempest-ServerActionsTestOtherA-1637843849 tempest-ServerActionsTestOtherA-1637843849-project-member] [instance: 15476ac0-289a-4e04-aa9d-4244c658e962] Binding failed for port 76b9d660-f9cc-4d3f-ae94-28e2b3d3b46d, please check neutron logs for more information. {{(pid=61806) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 756.127463] env[61806]: DEBUG oslo_concurrency.lockutils [None req-3c3981d2-fa25-4f4b-87d4-a5411e67199e tempest-MultipleCreateTestJSON-1589015462 tempest-MultipleCreateTestJSON-1589015462-project-member] Lock "5b22c2fc-525c-481b-b84e-e0a7f68f633e" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 166.423s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 756.127703] env[61806]: DEBUG oslo_concurrency.lockutils [None req-b5a6ece2-b279-4bbd-a3d1-16582c4d8894 tempest-ServersTestMultiNic-1480543082 tempest-ServersTestMultiNic-1480543082-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 14.762s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 756.129132] env[61806]: INFO nova.compute.claims [None req-b5a6ece2-b279-4bbd-a3d1-16582c4d8894 tempest-ServersTestMultiNic-1480543082 tempest-ServersTestMultiNic-1480543082-project-member] [instance: 24c83131-1451-429f-84fe-dcd070898d65] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 756.136995] env[61806]: DEBUG nova.compute.manager [None req-d17e7876-7c69-44af-acb2-32ed225ce945 tempest-ServerActionsTestOtherA-1637843849 tempest-ServerActionsTestOtherA-1637843849-project-member] [instance: 15476ac0-289a-4e04-aa9d-4244c658e962] Build of instance 15476ac0-289a-4e04-aa9d-4244c658e962 was re-scheduled: Binding failed for port 76b9d660-f9cc-4d3f-ae94-28e2b3d3b46d, please check neutron logs for more information. {{(pid=61806) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 756.137463] env[61806]: DEBUG nova.compute.manager [None req-d17e7876-7c69-44af-acb2-32ed225ce945 tempest-ServerActionsTestOtherA-1637843849 tempest-ServerActionsTestOtherA-1637843849-project-member] [instance: 15476ac0-289a-4e04-aa9d-4244c658e962] Unplugging VIFs for instance {{(pid=61806) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 756.137692] env[61806]: DEBUG oslo_concurrency.lockutils [None req-d17e7876-7c69-44af-acb2-32ed225ce945 tempest-ServerActionsTestOtherA-1637843849 tempest-ServerActionsTestOtherA-1637843849-project-member] Acquiring lock "refresh_cache-15476ac0-289a-4e04-aa9d-4244c658e962" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 756.137829] env[61806]: DEBUG oslo_concurrency.lockutils [None req-d17e7876-7c69-44af-acb2-32ed225ce945 tempest-ServerActionsTestOtherA-1637843849 tempest-ServerActionsTestOtherA-1637843849-project-member] Acquired lock "refresh_cache-15476ac0-289a-4e04-aa9d-4244c658e962" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 756.137985] env[61806]: DEBUG nova.network.neutron [None req-d17e7876-7c69-44af-acb2-32ed225ce945 tempest-ServerActionsTestOtherA-1637843849 tempest-ServerActionsTestOtherA-1637843849-project-member] [instance: 15476ac0-289a-4e04-aa9d-4244c658e962] Building network info cache for instance {{(pid=61806) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 756.372845] env[61806]: DEBUG nova.compute.manager [req-031252e0-5eed-4343-ab2a-f31da0bcbfef req-529ab3b0-d0d7-459f-99f1-26e6991c5e9b service nova] [instance: c61df5d3-7f43-48d4-a2c0-a8972372c0ba] Received event network-changed-b12a7881-ba2b-4a75-8119-9f5cb5b2cdb7 {{(pid=61806) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 756.373135] env[61806]: DEBUG nova.compute.manager [req-031252e0-5eed-4343-ab2a-f31da0bcbfef req-529ab3b0-d0d7-459f-99f1-26e6991c5e9b service nova] [instance: c61df5d3-7f43-48d4-a2c0-a8972372c0ba] Refreshing instance network info cache due to event network-changed-b12a7881-ba2b-4a75-8119-9f5cb5b2cdb7. {{(pid=61806) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 756.373262] env[61806]: DEBUG oslo_concurrency.lockutils [req-031252e0-5eed-4343-ab2a-f31da0bcbfef req-529ab3b0-d0d7-459f-99f1-26e6991c5e9b service nova] Acquiring lock "refresh_cache-c61df5d3-7f43-48d4-a2c0-a8972372c0ba" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 756.373400] env[61806]: DEBUG oslo_concurrency.lockutils [req-031252e0-5eed-4343-ab2a-f31da0bcbfef req-529ab3b0-d0d7-459f-99f1-26e6991c5e9b service nova] Acquired lock "refresh_cache-c61df5d3-7f43-48d4-a2c0-a8972372c0ba" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 756.373968] env[61806]: DEBUG nova.network.neutron [req-031252e0-5eed-4343-ab2a-f31da0bcbfef req-529ab3b0-d0d7-459f-99f1-26e6991c5e9b service nova] [instance: c61df5d3-7f43-48d4-a2c0-a8972372c0ba] Refreshing network info cache for port b12a7881-ba2b-4a75-8119-9f5cb5b2cdb7 {{(pid=61806) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 756.579477] env[61806]: ERROR nova.compute.manager [None req-da1d2f3c-f6ea-47f5-8c64-65c86bf2d676 tempest-ServerRescueTestJSON-811948373 tempest-ServerRescueTestJSON-811948373-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port b12a7881-ba2b-4a75-8119-9f5cb5b2cdb7, please check neutron logs for more information. [ 756.579477] env[61806]: ERROR nova.compute.manager Traceback (most recent call last): [ 756.579477] env[61806]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 756.579477] env[61806]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 756.579477] env[61806]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 756.579477] env[61806]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 756.579477] env[61806]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 756.579477] env[61806]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 756.579477] env[61806]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 756.579477] env[61806]: ERROR nova.compute.manager self.force_reraise() [ 756.579477] env[61806]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 756.579477] env[61806]: ERROR nova.compute.manager raise self.value [ 756.579477] env[61806]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 756.579477] env[61806]: ERROR nova.compute.manager updated_port = self._update_port( [ 756.579477] env[61806]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 756.579477] env[61806]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 756.580097] env[61806]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 756.580097] env[61806]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 756.580097] env[61806]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port b12a7881-ba2b-4a75-8119-9f5cb5b2cdb7, please check neutron logs for more information. [ 756.580097] env[61806]: ERROR nova.compute.manager [ 756.580097] env[61806]: Traceback (most recent call last): [ 756.580097] env[61806]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 756.580097] env[61806]: listener.cb(fileno) [ 756.580097] env[61806]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 756.580097] env[61806]: result = function(*args, **kwargs) [ 756.580097] env[61806]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 756.580097] env[61806]: return func(*args, **kwargs) [ 756.580097] env[61806]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 756.580097] env[61806]: raise e [ 756.580097] env[61806]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 756.580097] env[61806]: nwinfo = self.network_api.allocate_for_instance( [ 756.580097] env[61806]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 756.580097] env[61806]: created_port_ids = self._update_ports_for_instance( [ 756.580097] env[61806]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 756.580097] env[61806]: with excutils.save_and_reraise_exception(): [ 756.580097] env[61806]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 756.580097] env[61806]: self.force_reraise() [ 756.580097] env[61806]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 756.580097] env[61806]: raise self.value [ 756.580097] env[61806]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 756.580097] env[61806]: updated_port = self._update_port( [ 756.580097] env[61806]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 756.580097] env[61806]: _ensure_no_port_binding_failure(port) [ 756.580097] env[61806]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 756.580097] env[61806]: raise exception.PortBindingFailed(port_id=port['id']) [ 756.581196] env[61806]: nova.exception.PortBindingFailed: Binding failed for port b12a7881-ba2b-4a75-8119-9f5cb5b2cdb7, please check neutron logs for more information. [ 756.581196] env[61806]: Removing descriptor: 14 [ 756.639893] env[61806]: DEBUG nova.compute.manager [None req-f91d1db1-2f8a-42e8-9cfa-f4915f7b77dc tempest-ImagesTestJSON-618852556 tempest-ImagesTestJSON-618852556-project-member] [instance: e0ef0a35-82a5-495b-9d5c-5805e8306390] Starting instance... {{(pid=61806) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 756.660408] env[61806]: DEBUG nova.network.neutron [None req-d17e7876-7c69-44af-acb2-32ed225ce945 tempest-ServerActionsTestOtherA-1637843849 tempest-ServerActionsTestOtherA-1637843849-project-member] [instance: 15476ac0-289a-4e04-aa9d-4244c658e962] Instance cache missing network info. {{(pid=61806) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 756.755230] env[61806]: DEBUG nova.network.neutron [None req-d17e7876-7c69-44af-acb2-32ed225ce945 tempest-ServerActionsTestOtherA-1637843849 tempest-ServerActionsTestOtherA-1637843849-project-member] [instance: 15476ac0-289a-4e04-aa9d-4244c658e962] Updating instance_info_cache with network_info: [] {{(pid=61806) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 756.799485] env[61806]: DEBUG nova.compute.manager [None req-da1d2f3c-f6ea-47f5-8c64-65c86bf2d676 tempest-ServerRescueTestJSON-811948373 tempest-ServerRescueTestJSON-811948373-project-member] [instance: c61df5d3-7f43-48d4-a2c0-a8972372c0ba] Start spawning the instance on the hypervisor. {{(pid=61806) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 756.822851] env[61806]: DEBUG nova.virt.hardware [None req-da1d2f3c-f6ea-47f5-8c64-65c86bf2d676 tempest-ServerRescueTestJSON-811948373 tempest-ServerRescueTestJSON-811948373-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-15T18:05:10Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-15T18:04:52Z,direct_url=,disk_format='vmdk',id=ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='84ba83fce2c349a6988173c5d6fc3b07',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-15T18:04:53Z,virtual_size=,visibility=), allow threads: False {{(pid=61806) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 756.823108] env[61806]: DEBUG nova.virt.hardware [None req-da1d2f3c-f6ea-47f5-8c64-65c86bf2d676 tempest-ServerRescueTestJSON-811948373 tempest-ServerRescueTestJSON-811948373-project-member] Flavor limits 0:0:0 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 756.823269] env[61806]: DEBUG nova.virt.hardware [None req-da1d2f3c-f6ea-47f5-8c64-65c86bf2d676 tempest-ServerRescueTestJSON-811948373 tempest-ServerRescueTestJSON-811948373-project-member] Image limits 0:0:0 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 756.823452] env[61806]: DEBUG nova.virt.hardware [None req-da1d2f3c-f6ea-47f5-8c64-65c86bf2d676 tempest-ServerRescueTestJSON-811948373 tempest-ServerRescueTestJSON-811948373-project-member] Flavor pref 0:0:0 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 756.823598] env[61806]: DEBUG nova.virt.hardware [None req-da1d2f3c-f6ea-47f5-8c64-65c86bf2d676 tempest-ServerRescueTestJSON-811948373 tempest-ServerRescueTestJSON-811948373-project-member] Image pref 0:0:0 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 756.823745] env[61806]: DEBUG nova.virt.hardware [None req-da1d2f3c-f6ea-47f5-8c64-65c86bf2d676 tempest-ServerRescueTestJSON-811948373 tempest-ServerRescueTestJSON-811948373-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 756.823949] env[61806]: DEBUG nova.virt.hardware [None req-da1d2f3c-f6ea-47f5-8c64-65c86bf2d676 tempest-ServerRescueTestJSON-811948373 tempest-ServerRescueTestJSON-811948373-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61806) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 756.824181] env[61806]: DEBUG nova.virt.hardware [None req-da1d2f3c-f6ea-47f5-8c64-65c86bf2d676 tempest-ServerRescueTestJSON-811948373 tempest-ServerRescueTestJSON-811948373-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61806) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 756.824363] env[61806]: DEBUG nova.virt.hardware [None req-da1d2f3c-f6ea-47f5-8c64-65c86bf2d676 tempest-ServerRescueTestJSON-811948373 tempest-ServerRescueTestJSON-811948373-project-member] Got 1 possible topologies {{(pid=61806) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 756.824528] env[61806]: DEBUG nova.virt.hardware [None req-da1d2f3c-f6ea-47f5-8c64-65c86bf2d676 tempest-ServerRescueTestJSON-811948373 tempest-ServerRescueTestJSON-811948373-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61806) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 756.824702] env[61806]: DEBUG nova.virt.hardware [None req-da1d2f3c-f6ea-47f5-8c64-65c86bf2d676 tempest-ServerRescueTestJSON-811948373 tempest-ServerRescueTestJSON-811948373-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61806) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 756.825577] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d88dec4c-ee82-4d53-ba4d-b7e90d106a64 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 756.834999] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3b80c1d7-d2dc-4a16-9855-859459305a6e {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 756.849669] env[61806]: ERROR nova.compute.manager [None req-da1d2f3c-f6ea-47f5-8c64-65c86bf2d676 tempest-ServerRescueTestJSON-811948373 tempest-ServerRescueTestJSON-811948373-project-member] [instance: c61df5d3-7f43-48d4-a2c0-a8972372c0ba] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port b12a7881-ba2b-4a75-8119-9f5cb5b2cdb7, please check neutron logs for more information. [ 756.849669] env[61806]: ERROR nova.compute.manager [instance: c61df5d3-7f43-48d4-a2c0-a8972372c0ba] Traceback (most recent call last): [ 756.849669] env[61806]: ERROR nova.compute.manager [instance: c61df5d3-7f43-48d4-a2c0-a8972372c0ba] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 756.849669] env[61806]: ERROR nova.compute.manager [instance: c61df5d3-7f43-48d4-a2c0-a8972372c0ba] yield resources [ 756.849669] env[61806]: ERROR nova.compute.manager [instance: c61df5d3-7f43-48d4-a2c0-a8972372c0ba] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 756.849669] env[61806]: ERROR nova.compute.manager [instance: c61df5d3-7f43-48d4-a2c0-a8972372c0ba] self.driver.spawn(context, instance, image_meta, [ 756.849669] env[61806]: ERROR nova.compute.manager [instance: c61df5d3-7f43-48d4-a2c0-a8972372c0ba] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 756.849669] env[61806]: ERROR nova.compute.manager [instance: c61df5d3-7f43-48d4-a2c0-a8972372c0ba] self._vmops.spawn(context, instance, image_meta, injected_files, [ 756.849669] env[61806]: ERROR nova.compute.manager [instance: c61df5d3-7f43-48d4-a2c0-a8972372c0ba] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 756.849669] env[61806]: ERROR nova.compute.manager [instance: c61df5d3-7f43-48d4-a2c0-a8972372c0ba] vm_ref = self.build_virtual_machine(instance, [ 756.849669] env[61806]: ERROR nova.compute.manager [instance: c61df5d3-7f43-48d4-a2c0-a8972372c0ba] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 756.850126] env[61806]: ERROR nova.compute.manager [instance: c61df5d3-7f43-48d4-a2c0-a8972372c0ba] vif_infos = vmwarevif.get_vif_info(self._session, [ 756.850126] env[61806]: ERROR nova.compute.manager [instance: c61df5d3-7f43-48d4-a2c0-a8972372c0ba] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 756.850126] env[61806]: ERROR nova.compute.manager [instance: c61df5d3-7f43-48d4-a2c0-a8972372c0ba] for vif in network_info: [ 756.850126] env[61806]: ERROR nova.compute.manager [instance: c61df5d3-7f43-48d4-a2c0-a8972372c0ba] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 756.850126] env[61806]: ERROR nova.compute.manager [instance: c61df5d3-7f43-48d4-a2c0-a8972372c0ba] return self._sync_wrapper(fn, *args, **kwargs) [ 756.850126] env[61806]: ERROR nova.compute.manager [instance: c61df5d3-7f43-48d4-a2c0-a8972372c0ba] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 756.850126] env[61806]: ERROR nova.compute.manager [instance: c61df5d3-7f43-48d4-a2c0-a8972372c0ba] self.wait() [ 756.850126] env[61806]: ERROR nova.compute.manager [instance: c61df5d3-7f43-48d4-a2c0-a8972372c0ba] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 756.850126] env[61806]: ERROR nova.compute.manager [instance: c61df5d3-7f43-48d4-a2c0-a8972372c0ba] self[:] = self._gt.wait() [ 756.850126] env[61806]: ERROR nova.compute.manager [instance: c61df5d3-7f43-48d4-a2c0-a8972372c0ba] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 756.850126] env[61806]: ERROR nova.compute.manager [instance: c61df5d3-7f43-48d4-a2c0-a8972372c0ba] return self._exit_event.wait() [ 756.850126] env[61806]: ERROR nova.compute.manager [instance: c61df5d3-7f43-48d4-a2c0-a8972372c0ba] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 756.850126] env[61806]: ERROR nova.compute.manager [instance: c61df5d3-7f43-48d4-a2c0-a8972372c0ba] current.throw(*self._exc) [ 756.850662] env[61806]: ERROR nova.compute.manager [instance: c61df5d3-7f43-48d4-a2c0-a8972372c0ba] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 756.850662] env[61806]: ERROR nova.compute.manager [instance: c61df5d3-7f43-48d4-a2c0-a8972372c0ba] result = function(*args, **kwargs) [ 756.850662] env[61806]: ERROR nova.compute.manager [instance: c61df5d3-7f43-48d4-a2c0-a8972372c0ba] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 756.850662] env[61806]: ERROR nova.compute.manager [instance: c61df5d3-7f43-48d4-a2c0-a8972372c0ba] return func(*args, **kwargs) [ 756.850662] env[61806]: ERROR nova.compute.manager [instance: c61df5d3-7f43-48d4-a2c0-a8972372c0ba] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 756.850662] env[61806]: ERROR nova.compute.manager [instance: c61df5d3-7f43-48d4-a2c0-a8972372c0ba] raise e [ 756.850662] env[61806]: ERROR nova.compute.manager [instance: c61df5d3-7f43-48d4-a2c0-a8972372c0ba] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 756.850662] env[61806]: ERROR nova.compute.manager [instance: c61df5d3-7f43-48d4-a2c0-a8972372c0ba] nwinfo = self.network_api.allocate_for_instance( [ 756.850662] env[61806]: ERROR nova.compute.manager [instance: c61df5d3-7f43-48d4-a2c0-a8972372c0ba] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 756.850662] env[61806]: ERROR nova.compute.manager [instance: c61df5d3-7f43-48d4-a2c0-a8972372c0ba] created_port_ids = self._update_ports_for_instance( [ 756.850662] env[61806]: ERROR nova.compute.manager [instance: c61df5d3-7f43-48d4-a2c0-a8972372c0ba] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 756.850662] env[61806]: ERROR nova.compute.manager [instance: c61df5d3-7f43-48d4-a2c0-a8972372c0ba] with excutils.save_and_reraise_exception(): [ 756.850662] env[61806]: ERROR nova.compute.manager [instance: c61df5d3-7f43-48d4-a2c0-a8972372c0ba] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 756.851171] env[61806]: ERROR nova.compute.manager [instance: c61df5d3-7f43-48d4-a2c0-a8972372c0ba] self.force_reraise() [ 756.851171] env[61806]: ERROR nova.compute.manager [instance: c61df5d3-7f43-48d4-a2c0-a8972372c0ba] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 756.851171] env[61806]: ERROR nova.compute.manager [instance: c61df5d3-7f43-48d4-a2c0-a8972372c0ba] raise self.value [ 756.851171] env[61806]: ERROR nova.compute.manager [instance: c61df5d3-7f43-48d4-a2c0-a8972372c0ba] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 756.851171] env[61806]: ERROR nova.compute.manager [instance: c61df5d3-7f43-48d4-a2c0-a8972372c0ba] updated_port = self._update_port( [ 756.851171] env[61806]: ERROR nova.compute.manager [instance: c61df5d3-7f43-48d4-a2c0-a8972372c0ba] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 756.851171] env[61806]: ERROR nova.compute.manager [instance: c61df5d3-7f43-48d4-a2c0-a8972372c0ba] _ensure_no_port_binding_failure(port) [ 756.851171] env[61806]: ERROR nova.compute.manager [instance: c61df5d3-7f43-48d4-a2c0-a8972372c0ba] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 756.851171] env[61806]: ERROR nova.compute.manager [instance: c61df5d3-7f43-48d4-a2c0-a8972372c0ba] raise exception.PortBindingFailed(port_id=port['id']) [ 756.851171] env[61806]: ERROR nova.compute.manager [instance: c61df5d3-7f43-48d4-a2c0-a8972372c0ba] nova.exception.PortBindingFailed: Binding failed for port b12a7881-ba2b-4a75-8119-9f5cb5b2cdb7, please check neutron logs for more information. [ 756.851171] env[61806]: ERROR nova.compute.manager [instance: c61df5d3-7f43-48d4-a2c0-a8972372c0ba] [ 756.851171] env[61806]: INFO nova.compute.manager [None req-da1d2f3c-f6ea-47f5-8c64-65c86bf2d676 tempest-ServerRescueTestJSON-811948373 tempest-ServerRescueTestJSON-811948373-project-member] [instance: c61df5d3-7f43-48d4-a2c0-a8972372c0ba] Terminating instance [ 756.852094] env[61806]: DEBUG oslo_concurrency.lockutils [None req-da1d2f3c-f6ea-47f5-8c64-65c86bf2d676 tempest-ServerRescueTestJSON-811948373 tempest-ServerRescueTestJSON-811948373-project-member] Acquiring lock "refresh_cache-c61df5d3-7f43-48d4-a2c0-a8972372c0ba" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 756.890472] env[61806]: DEBUG nova.network.neutron [req-031252e0-5eed-4343-ab2a-f31da0bcbfef req-529ab3b0-d0d7-459f-99f1-26e6991c5e9b service nova] [instance: c61df5d3-7f43-48d4-a2c0-a8972372c0ba] Instance cache missing network info. {{(pid=61806) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 756.959384] env[61806]: DEBUG nova.network.neutron [req-031252e0-5eed-4343-ab2a-f31da0bcbfef req-529ab3b0-d0d7-459f-99f1-26e6991c5e9b service nova] [instance: c61df5d3-7f43-48d4-a2c0-a8972372c0ba] Updating instance_info_cache with network_info: [] {{(pid=61806) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 757.154875] env[61806]: DEBUG oslo_concurrency.lockutils [None req-f91d1db1-2f8a-42e8-9cfa-f4915f7b77dc tempest-ImagesTestJSON-618852556 tempest-ImagesTestJSON-618852556-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 757.260526] env[61806]: DEBUG oslo_concurrency.lockutils [None req-d17e7876-7c69-44af-acb2-32ed225ce945 tempest-ServerActionsTestOtherA-1637843849 tempest-ServerActionsTestOtherA-1637843849-project-member] Releasing lock "refresh_cache-15476ac0-289a-4e04-aa9d-4244c658e962" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 757.260923] env[61806]: DEBUG nova.compute.manager [None req-d17e7876-7c69-44af-acb2-32ed225ce945 tempest-ServerActionsTestOtherA-1637843849 tempest-ServerActionsTestOtherA-1637843849-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61806) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 757.261138] env[61806]: DEBUG nova.compute.manager [None req-d17e7876-7c69-44af-acb2-32ed225ce945 tempest-ServerActionsTestOtherA-1637843849 tempest-ServerActionsTestOtherA-1637843849-project-member] [instance: 15476ac0-289a-4e04-aa9d-4244c658e962] Deallocating network for instance {{(pid=61806) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 757.261347] env[61806]: DEBUG nova.network.neutron [None req-d17e7876-7c69-44af-acb2-32ed225ce945 tempest-ServerActionsTestOtherA-1637843849 tempest-ServerActionsTestOtherA-1637843849-project-member] [instance: 15476ac0-289a-4e04-aa9d-4244c658e962] deallocate_for_instance() {{(pid=61806) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 757.278773] env[61806]: DEBUG oslo_concurrency.lockutils [None req-cbb90e6d-7dac-4c30-b860-1f0db7dd091e tempest-MultipleCreateTestJSON-1589015462 tempest-MultipleCreateTestJSON-1589015462-project-member] Acquiring lock "25218e83-c1ab-4b97-ade1-7c32b6f99305" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 757.279009] env[61806]: DEBUG oslo_concurrency.lockutils [None req-cbb90e6d-7dac-4c30-b860-1f0db7dd091e tempest-MultipleCreateTestJSON-1589015462 tempest-MultipleCreateTestJSON-1589015462-project-member] Lock "25218e83-c1ab-4b97-ade1-7c32b6f99305" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 757.284346] env[61806]: DEBUG nova.network.neutron [None req-d17e7876-7c69-44af-acb2-32ed225ce945 tempest-ServerActionsTestOtherA-1637843849 tempest-ServerActionsTestOtherA-1637843849-project-member] [instance: 15476ac0-289a-4e04-aa9d-4244c658e962] Instance cache missing network info. {{(pid=61806) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 757.308471] env[61806]: DEBUG oslo_concurrency.lockutils [None req-cbb90e6d-7dac-4c30-b860-1f0db7dd091e tempest-MultipleCreateTestJSON-1589015462 tempest-MultipleCreateTestJSON-1589015462-project-member] Acquiring lock "ef502a7d-7dc0-4319-8ce3-6a8701288237" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 757.308471] env[61806]: DEBUG oslo_concurrency.lockutils [None req-cbb90e6d-7dac-4c30-b860-1f0db7dd091e tempest-MultipleCreateTestJSON-1589015462 tempest-MultipleCreateTestJSON-1589015462-project-member] Lock "ef502a7d-7dc0-4319-8ce3-6a8701288237" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 757.444495] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-af9aeb0a-fae0-4f29-976c-3c626552c93a {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 757.452189] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a4269425-ae99-4848-8e34-941c7c43a049 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 757.481752] env[61806]: DEBUG oslo_concurrency.lockutils [req-031252e0-5eed-4343-ab2a-f31da0bcbfef req-529ab3b0-d0d7-459f-99f1-26e6991c5e9b service nova] Releasing lock "refresh_cache-c61df5d3-7f43-48d4-a2c0-a8972372c0ba" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 757.482297] env[61806]: DEBUG oslo_concurrency.lockutils [None req-da1d2f3c-f6ea-47f5-8c64-65c86bf2d676 tempest-ServerRescueTestJSON-811948373 tempest-ServerRescueTestJSON-811948373-project-member] Acquired lock "refresh_cache-c61df5d3-7f43-48d4-a2c0-a8972372c0ba" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 757.482485] env[61806]: DEBUG nova.network.neutron [None req-da1d2f3c-f6ea-47f5-8c64-65c86bf2d676 tempest-ServerRescueTestJSON-811948373 tempest-ServerRescueTestJSON-811948373-project-member] [instance: c61df5d3-7f43-48d4-a2c0-a8972372c0ba] Building network info cache for instance {{(pid=61806) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 757.484088] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-94ecffbd-810d-45ad-a5d6-25af29410948 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 757.491701] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-893737f3-bdf8-4021-ac4a-25e6d110176b {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 757.506599] env[61806]: DEBUG nova.compute.provider_tree [None req-b5a6ece2-b279-4bbd-a3d1-16582c4d8894 tempest-ServersTestMultiNic-1480543082 tempest-ServersTestMultiNic-1480543082-project-member] Inventory has not changed in ProviderTree for provider: a2858be1-fd22-4e08-979e-87ad25293407 {{(pid=61806) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 757.787373] env[61806]: DEBUG nova.network.neutron [None req-d17e7876-7c69-44af-acb2-32ed225ce945 tempest-ServerActionsTestOtherA-1637843849 tempest-ServerActionsTestOtherA-1637843849-project-member] [instance: 15476ac0-289a-4e04-aa9d-4244c658e962] Updating instance_info_cache with network_info: [] {{(pid=61806) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 758.003863] env[61806]: DEBUG nova.network.neutron [None req-da1d2f3c-f6ea-47f5-8c64-65c86bf2d676 tempest-ServerRescueTestJSON-811948373 tempest-ServerRescueTestJSON-811948373-project-member] [instance: c61df5d3-7f43-48d4-a2c0-a8972372c0ba] Instance cache missing network info. {{(pid=61806) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 758.010203] env[61806]: DEBUG nova.scheduler.client.report [None req-b5a6ece2-b279-4bbd-a3d1-16582c4d8894 tempest-ServersTestMultiNic-1480543082 tempest-ServersTestMultiNic-1480543082-project-member] Inventory has not changed for provider a2858be1-fd22-4e08-979e-87ad25293407 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 139, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61806) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 758.099836] env[61806]: DEBUG nova.network.neutron [None req-da1d2f3c-f6ea-47f5-8c64-65c86bf2d676 tempest-ServerRescueTestJSON-811948373 tempest-ServerRescueTestJSON-811948373-project-member] [instance: c61df5d3-7f43-48d4-a2c0-a8972372c0ba] Updating instance_info_cache with network_info: [] {{(pid=61806) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 758.289953] env[61806]: INFO nova.compute.manager [None req-d17e7876-7c69-44af-acb2-32ed225ce945 tempest-ServerActionsTestOtherA-1637843849 tempest-ServerActionsTestOtherA-1637843849-project-member] [instance: 15476ac0-289a-4e04-aa9d-4244c658e962] Took 1.03 seconds to deallocate network for instance. [ 758.397582] env[61806]: DEBUG nova.compute.manager [req-b628a79c-210e-4e78-8b4d-68290099ba91 req-7c1b9f31-255b-4380-b7f7-d0438016d596 service nova] [instance: c61df5d3-7f43-48d4-a2c0-a8972372c0ba] Received event network-vif-deleted-b12a7881-ba2b-4a75-8119-9f5cb5b2cdb7 {{(pid=61806) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 758.514166] env[61806]: DEBUG oslo_concurrency.lockutils [None req-b5a6ece2-b279-4bbd-a3d1-16582c4d8894 tempest-ServersTestMultiNic-1480543082 tempest-ServersTestMultiNic-1480543082-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.386s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 758.514676] env[61806]: DEBUG nova.compute.manager [None req-b5a6ece2-b279-4bbd-a3d1-16582c4d8894 tempest-ServersTestMultiNic-1480543082 tempest-ServersTestMultiNic-1480543082-project-member] [instance: 24c83131-1451-429f-84fe-dcd070898d65] Start building networks asynchronously for instance. {{(pid=61806) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 758.517494] env[61806]: DEBUG oslo_concurrency.lockutils [None req-1382c347-5a79-4289-8b32-002f296b7ae3 tempest-TenantUsagesTestJSON-255073140 tempest-TenantUsagesTestJSON-255073140-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 15.215s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 758.602101] env[61806]: DEBUG oslo_concurrency.lockutils [None req-da1d2f3c-f6ea-47f5-8c64-65c86bf2d676 tempest-ServerRescueTestJSON-811948373 tempest-ServerRescueTestJSON-811948373-project-member] Releasing lock "refresh_cache-c61df5d3-7f43-48d4-a2c0-a8972372c0ba" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 758.602551] env[61806]: DEBUG nova.compute.manager [None req-da1d2f3c-f6ea-47f5-8c64-65c86bf2d676 tempest-ServerRescueTestJSON-811948373 tempest-ServerRescueTestJSON-811948373-project-member] [instance: c61df5d3-7f43-48d4-a2c0-a8972372c0ba] Start destroying the instance on the hypervisor. {{(pid=61806) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 758.602752] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-da1d2f3c-f6ea-47f5-8c64-65c86bf2d676 tempest-ServerRescueTestJSON-811948373 tempest-ServerRescueTestJSON-811948373-project-member] [instance: c61df5d3-7f43-48d4-a2c0-a8972372c0ba] Destroying instance {{(pid=61806) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 758.603065] env[61806]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-8f9ef89a-e676-4e1b-a496-ead6c43802cd {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 758.612835] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-57088a60-0e94-49f5-8857-220d5b10299d {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 758.634490] env[61806]: WARNING nova.virt.vmwareapi.vmops [None req-da1d2f3c-f6ea-47f5-8c64-65c86bf2d676 tempest-ServerRescueTestJSON-811948373 tempest-ServerRescueTestJSON-811948373-project-member] [instance: c61df5d3-7f43-48d4-a2c0-a8972372c0ba] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance c61df5d3-7f43-48d4-a2c0-a8972372c0ba could not be found. [ 758.634706] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-da1d2f3c-f6ea-47f5-8c64-65c86bf2d676 tempest-ServerRescueTestJSON-811948373 tempest-ServerRescueTestJSON-811948373-project-member] [instance: c61df5d3-7f43-48d4-a2c0-a8972372c0ba] Instance destroyed {{(pid=61806) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 758.634915] env[61806]: INFO nova.compute.manager [None req-da1d2f3c-f6ea-47f5-8c64-65c86bf2d676 tempest-ServerRescueTestJSON-811948373 tempest-ServerRescueTestJSON-811948373-project-member] [instance: c61df5d3-7f43-48d4-a2c0-a8972372c0ba] Took 0.03 seconds to destroy the instance on the hypervisor. [ 758.635239] env[61806]: DEBUG oslo.service.loopingcall [None req-da1d2f3c-f6ea-47f5-8c64-65c86bf2d676 tempest-ServerRescueTestJSON-811948373 tempest-ServerRescueTestJSON-811948373-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61806) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 758.635467] env[61806]: DEBUG nova.compute.manager [-] [instance: c61df5d3-7f43-48d4-a2c0-a8972372c0ba] Deallocating network for instance {{(pid=61806) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 758.635562] env[61806]: DEBUG nova.network.neutron [-] [instance: c61df5d3-7f43-48d4-a2c0-a8972372c0ba] deallocate_for_instance() {{(pid=61806) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 758.704241] env[61806]: DEBUG nova.network.neutron [-] [instance: c61df5d3-7f43-48d4-a2c0-a8972372c0ba] Instance cache missing network info. {{(pid=61806) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 759.024325] env[61806]: DEBUG nova.compute.utils [None req-b5a6ece2-b279-4bbd-a3d1-16582c4d8894 tempest-ServersTestMultiNic-1480543082 tempest-ServersTestMultiNic-1480543082-project-member] Using /dev/sd instead of None {{(pid=61806) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 759.029028] env[61806]: DEBUG nova.compute.manager [None req-b5a6ece2-b279-4bbd-a3d1-16582c4d8894 tempest-ServersTestMultiNic-1480543082 tempest-ServersTestMultiNic-1480543082-project-member] [instance: 24c83131-1451-429f-84fe-dcd070898d65] Allocating IP information in the background. {{(pid=61806) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 759.029028] env[61806]: DEBUG nova.network.neutron [None req-b5a6ece2-b279-4bbd-a3d1-16582c4d8894 tempest-ServersTestMultiNic-1480543082 tempest-ServersTestMultiNic-1480543082-project-member] [instance: 24c83131-1451-429f-84fe-dcd070898d65] allocate_for_instance() {{(pid=61806) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 759.071149] env[61806]: DEBUG nova.policy [None req-b5a6ece2-b279-4bbd-a3d1-16582c4d8894 tempest-ServersTestMultiNic-1480543082 tempest-ServersTestMultiNic-1480543082-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '2e72eab1cc5c4fbba3e2d7dbfa0035ec', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '4abf9a81bb9649db910a44e5562dd342', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61806) authorize /opt/stack/nova/nova/policy.py:201}} [ 759.205694] env[61806]: DEBUG nova.network.neutron [-] [instance: c61df5d3-7f43-48d4-a2c0-a8972372c0ba] Updating instance_info_cache with network_info: [] {{(pid=61806) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 759.319093] env[61806]: INFO nova.scheduler.client.report [None req-d17e7876-7c69-44af-acb2-32ed225ce945 tempest-ServerActionsTestOtherA-1637843849 tempest-ServerActionsTestOtherA-1637843849-project-member] Deleted allocations for instance 15476ac0-289a-4e04-aa9d-4244c658e962 [ 759.329248] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0a2d8ec2-2530-4d2e-9dbe-5218ca275751 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 759.337716] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-469f8d49-c764-4566-a0d5-f1a9cd1a63d2 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 759.373759] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-44038886-7532-4b25-b861-65d693a7420b {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 759.382076] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-82262816-4793-408e-a678-ff533ff49167 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 759.388024] env[61806]: DEBUG nova.network.neutron [None req-b5a6ece2-b279-4bbd-a3d1-16582c4d8894 tempest-ServersTestMultiNic-1480543082 tempest-ServersTestMultiNic-1480543082-project-member] [instance: 24c83131-1451-429f-84fe-dcd070898d65] Successfully created port: 30e4bd97-a4da-443f-a523-14f33ef9b039 {{(pid=61806) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 759.400557] env[61806]: DEBUG nova.compute.provider_tree [None req-1382c347-5a79-4289-8b32-002f296b7ae3 tempest-TenantUsagesTestJSON-255073140 tempest-TenantUsagesTestJSON-255073140-project-member] Inventory has not changed in ProviderTree for provider: a2858be1-fd22-4e08-979e-87ad25293407 {{(pid=61806) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 759.531862] env[61806]: DEBUG nova.compute.manager [None req-b5a6ece2-b279-4bbd-a3d1-16582c4d8894 tempest-ServersTestMultiNic-1480543082 tempest-ServersTestMultiNic-1480543082-project-member] [instance: 24c83131-1451-429f-84fe-dcd070898d65] Start building block device mappings for instance. {{(pid=61806) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 759.687839] env[61806]: DEBUG nova.network.neutron [None req-b5a6ece2-b279-4bbd-a3d1-16582c4d8894 tempest-ServersTestMultiNic-1480543082 tempest-ServersTestMultiNic-1480543082-project-member] [instance: 24c83131-1451-429f-84fe-dcd070898d65] Successfully created port: cbccb99e-b009-4ba4-8376-d82f644957bd {{(pid=61806) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 759.711069] env[61806]: INFO nova.compute.manager [-] [instance: c61df5d3-7f43-48d4-a2c0-a8972372c0ba] Took 1.08 seconds to deallocate network for instance. [ 759.713572] env[61806]: DEBUG nova.compute.claims [None req-da1d2f3c-f6ea-47f5-8c64-65c86bf2d676 tempest-ServerRescueTestJSON-811948373 tempest-ServerRescueTestJSON-811948373-project-member] [instance: c61df5d3-7f43-48d4-a2c0-a8972372c0ba] Aborting claim: {{(pid=61806) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 759.713782] env[61806]: DEBUG oslo_concurrency.lockutils [None req-da1d2f3c-f6ea-47f5-8c64-65c86bf2d676 tempest-ServerRescueTestJSON-811948373 tempest-ServerRescueTestJSON-811948373-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 759.833219] env[61806]: DEBUG oslo_concurrency.lockutils [None req-d17e7876-7c69-44af-acb2-32ed225ce945 tempest-ServerActionsTestOtherA-1637843849 tempest-ServerActionsTestOtherA-1637843849-project-member] Lock "15476ac0-289a-4e04-aa9d-4244c658e962" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 159.028s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 759.907155] env[61806]: DEBUG nova.scheduler.client.report [None req-1382c347-5a79-4289-8b32-002f296b7ae3 tempest-TenantUsagesTestJSON-255073140 tempest-TenantUsagesTestJSON-255073140-project-member] Inventory has not changed for provider a2858be1-fd22-4e08-979e-87ad25293407 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 139, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61806) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 760.337394] env[61806]: DEBUG nova.compute.manager [None req-0b87073d-3de3-486a-876d-8ca5794993c1 tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] [instance: 2a3f81e3-b84f-4370-a598-2f6a607c3d47] Starting instance... {{(pid=61806) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 760.409596] env[61806]: DEBUG oslo_concurrency.lockutils [None req-1382c347-5a79-4289-8b32-002f296b7ae3 tempest-TenantUsagesTestJSON-255073140 tempest-TenantUsagesTestJSON-255073140-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.892s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 760.410258] env[61806]: ERROR nova.compute.manager [None req-1382c347-5a79-4289-8b32-002f296b7ae3 tempest-TenantUsagesTestJSON-255073140 tempest-TenantUsagesTestJSON-255073140-project-member] [instance: 949c1050-5003-4519-b24c-8904de21a676] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 1add6960-077d-421a-b55c-d82b1d6d40eb, please check neutron logs for more information. [ 760.410258] env[61806]: ERROR nova.compute.manager [instance: 949c1050-5003-4519-b24c-8904de21a676] Traceback (most recent call last): [ 760.410258] env[61806]: ERROR nova.compute.manager [instance: 949c1050-5003-4519-b24c-8904de21a676] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 760.410258] env[61806]: ERROR nova.compute.manager [instance: 949c1050-5003-4519-b24c-8904de21a676] self.driver.spawn(context, instance, image_meta, [ 760.410258] env[61806]: ERROR nova.compute.manager [instance: 949c1050-5003-4519-b24c-8904de21a676] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 760.410258] env[61806]: ERROR nova.compute.manager [instance: 949c1050-5003-4519-b24c-8904de21a676] self._vmops.spawn(context, instance, image_meta, injected_files, [ 760.410258] env[61806]: ERROR nova.compute.manager [instance: 949c1050-5003-4519-b24c-8904de21a676] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 760.410258] env[61806]: ERROR nova.compute.manager [instance: 949c1050-5003-4519-b24c-8904de21a676] vm_ref = self.build_virtual_machine(instance, [ 760.410258] env[61806]: ERROR nova.compute.manager [instance: 949c1050-5003-4519-b24c-8904de21a676] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 760.410258] env[61806]: ERROR nova.compute.manager [instance: 949c1050-5003-4519-b24c-8904de21a676] vif_infos = vmwarevif.get_vif_info(self._session, [ 760.410258] env[61806]: ERROR nova.compute.manager [instance: 949c1050-5003-4519-b24c-8904de21a676] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 760.410722] env[61806]: ERROR nova.compute.manager [instance: 949c1050-5003-4519-b24c-8904de21a676] for vif in network_info: [ 760.410722] env[61806]: ERROR nova.compute.manager [instance: 949c1050-5003-4519-b24c-8904de21a676] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 760.410722] env[61806]: ERROR nova.compute.manager [instance: 949c1050-5003-4519-b24c-8904de21a676] return self._sync_wrapper(fn, *args, **kwargs) [ 760.410722] env[61806]: ERROR nova.compute.manager [instance: 949c1050-5003-4519-b24c-8904de21a676] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 760.410722] env[61806]: ERROR nova.compute.manager [instance: 949c1050-5003-4519-b24c-8904de21a676] self.wait() [ 760.410722] env[61806]: ERROR nova.compute.manager [instance: 949c1050-5003-4519-b24c-8904de21a676] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 760.410722] env[61806]: ERROR nova.compute.manager [instance: 949c1050-5003-4519-b24c-8904de21a676] self[:] = self._gt.wait() [ 760.410722] env[61806]: ERROR nova.compute.manager [instance: 949c1050-5003-4519-b24c-8904de21a676] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 760.410722] env[61806]: ERROR nova.compute.manager [instance: 949c1050-5003-4519-b24c-8904de21a676] return self._exit_event.wait() [ 760.410722] env[61806]: ERROR nova.compute.manager [instance: 949c1050-5003-4519-b24c-8904de21a676] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 760.410722] env[61806]: ERROR nova.compute.manager [instance: 949c1050-5003-4519-b24c-8904de21a676] current.throw(*self._exc) [ 760.410722] env[61806]: ERROR nova.compute.manager [instance: 949c1050-5003-4519-b24c-8904de21a676] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 760.410722] env[61806]: ERROR nova.compute.manager [instance: 949c1050-5003-4519-b24c-8904de21a676] result = function(*args, **kwargs) [ 760.411162] env[61806]: ERROR nova.compute.manager [instance: 949c1050-5003-4519-b24c-8904de21a676] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 760.411162] env[61806]: ERROR nova.compute.manager [instance: 949c1050-5003-4519-b24c-8904de21a676] return func(*args, **kwargs) [ 760.411162] env[61806]: ERROR nova.compute.manager [instance: 949c1050-5003-4519-b24c-8904de21a676] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 760.411162] env[61806]: ERROR nova.compute.manager [instance: 949c1050-5003-4519-b24c-8904de21a676] raise e [ 760.411162] env[61806]: ERROR nova.compute.manager [instance: 949c1050-5003-4519-b24c-8904de21a676] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 760.411162] env[61806]: ERROR nova.compute.manager [instance: 949c1050-5003-4519-b24c-8904de21a676] nwinfo = self.network_api.allocate_for_instance( [ 760.411162] env[61806]: ERROR nova.compute.manager [instance: 949c1050-5003-4519-b24c-8904de21a676] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 760.411162] env[61806]: ERROR nova.compute.manager [instance: 949c1050-5003-4519-b24c-8904de21a676] created_port_ids = self._update_ports_for_instance( [ 760.411162] env[61806]: ERROR nova.compute.manager [instance: 949c1050-5003-4519-b24c-8904de21a676] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 760.411162] env[61806]: ERROR nova.compute.manager [instance: 949c1050-5003-4519-b24c-8904de21a676] with excutils.save_and_reraise_exception(): [ 760.411162] env[61806]: ERROR nova.compute.manager [instance: 949c1050-5003-4519-b24c-8904de21a676] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 760.411162] env[61806]: ERROR nova.compute.manager [instance: 949c1050-5003-4519-b24c-8904de21a676] self.force_reraise() [ 760.411162] env[61806]: ERROR nova.compute.manager [instance: 949c1050-5003-4519-b24c-8904de21a676] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 760.411687] env[61806]: ERROR nova.compute.manager [instance: 949c1050-5003-4519-b24c-8904de21a676] raise self.value [ 760.411687] env[61806]: ERROR nova.compute.manager [instance: 949c1050-5003-4519-b24c-8904de21a676] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 760.411687] env[61806]: ERROR nova.compute.manager [instance: 949c1050-5003-4519-b24c-8904de21a676] updated_port = self._update_port( [ 760.411687] env[61806]: ERROR nova.compute.manager [instance: 949c1050-5003-4519-b24c-8904de21a676] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 760.411687] env[61806]: ERROR nova.compute.manager [instance: 949c1050-5003-4519-b24c-8904de21a676] _ensure_no_port_binding_failure(port) [ 760.411687] env[61806]: ERROR nova.compute.manager [instance: 949c1050-5003-4519-b24c-8904de21a676] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 760.411687] env[61806]: ERROR nova.compute.manager [instance: 949c1050-5003-4519-b24c-8904de21a676] raise exception.PortBindingFailed(port_id=port['id']) [ 760.411687] env[61806]: ERROR nova.compute.manager [instance: 949c1050-5003-4519-b24c-8904de21a676] nova.exception.PortBindingFailed: Binding failed for port 1add6960-077d-421a-b55c-d82b1d6d40eb, please check neutron logs for more information. [ 760.411687] env[61806]: ERROR nova.compute.manager [instance: 949c1050-5003-4519-b24c-8904de21a676] [ 760.411687] env[61806]: DEBUG nova.compute.utils [None req-1382c347-5a79-4289-8b32-002f296b7ae3 tempest-TenantUsagesTestJSON-255073140 tempest-TenantUsagesTestJSON-255073140-project-member] [instance: 949c1050-5003-4519-b24c-8904de21a676] Binding failed for port 1add6960-077d-421a-b55c-d82b1d6d40eb, please check neutron logs for more information. {{(pid=61806) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 760.412416] env[61806]: DEBUG oslo_concurrency.lockutils [None req-3bf756e0-c56e-4902-9c2b-b224a1d09a7c tempest-ListServersNegativeTestJSON-272187588 tempest-ListServersNegativeTestJSON-272187588-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 13.871s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 760.415856] env[61806]: DEBUG nova.compute.manager [None req-1382c347-5a79-4289-8b32-002f296b7ae3 tempest-TenantUsagesTestJSON-255073140 tempest-TenantUsagesTestJSON-255073140-project-member] [instance: 949c1050-5003-4519-b24c-8904de21a676] Build of instance 949c1050-5003-4519-b24c-8904de21a676 was re-scheduled: Binding failed for port 1add6960-077d-421a-b55c-d82b1d6d40eb, please check neutron logs for more information. {{(pid=61806) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 760.416949] env[61806]: DEBUG nova.compute.manager [None req-1382c347-5a79-4289-8b32-002f296b7ae3 tempest-TenantUsagesTestJSON-255073140 tempest-TenantUsagesTestJSON-255073140-project-member] [instance: 949c1050-5003-4519-b24c-8904de21a676] Unplugging VIFs for instance {{(pid=61806) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 760.416949] env[61806]: DEBUG oslo_concurrency.lockutils [None req-1382c347-5a79-4289-8b32-002f296b7ae3 tempest-TenantUsagesTestJSON-255073140 tempest-TenantUsagesTestJSON-255073140-project-member] Acquiring lock "refresh_cache-949c1050-5003-4519-b24c-8904de21a676" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 760.416949] env[61806]: DEBUG oslo_concurrency.lockutils [None req-1382c347-5a79-4289-8b32-002f296b7ae3 tempest-TenantUsagesTestJSON-255073140 tempest-TenantUsagesTestJSON-255073140-project-member] Acquired lock "refresh_cache-949c1050-5003-4519-b24c-8904de21a676" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 760.416949] env[61806]: DEBUG nova.network.neutron [None req-1382c347-5a79-4289-8b32-002f296b7ae3 tempest-TenantUsagesTestJSON-255073140 tempest-TenantUsagesTestJSON-255073140-project-member] [instance: 949c1050-5003-4519-b24c-8904de21a676] Building network info cache for instance {{(pid=61806) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 760.492733] env[61806]: DEBUG nova.compute.manager [req-f944cc19-9c05-40b5-86f7-fb2e763454c0 req-3d1ce8ec-b917-494b-9cff-e9a6a2ccbae6 service nova] [instance: 24c83131-1451-429f-84fe-dcd070898d65] Received event network-changed-30e4bd97-a4da-443f-a523-14f33ef9b039 {{(pid=61806) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 760.492945] env[61806]: DEBUG nova.compute.manager [req-f944cc19-9c05-40b5-86f7-fb2e763454c0 req-3d1ce8ec-b917-494b-9cff-e9a6a2ccbae6 service nova] [instance: 24c83131-1451-429f-84fe-dcd070898d65] Refreshing instance network info cache due to event network-changed-30e4bd97-a4da-443f-a523-14f33ef9b039. {{(pid=61806) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 760.493394] env[61806]: DEBUG oslo_concurrency.lockutils [req-f944cc19-9c05-40b5-86f7-fb2e763454c0 req-3d1ce8ec-b917-494b-9cff-e9a6a2ccbae6 service nova] Acquiring lock "refresh_cache-24c83131-1451-429f-84fe-dcd070898d65" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 760.493548] env[61806]: DEBUG oslo_concurrency.lockutils [req-f944cc19-9c05-40b5-86f7-fb2e763454c0 req-3d1ce8ec-b917-494b-9cff-e9a6a2ccbae6 service nova] Acquired lock "refresh_cache-24c83131-1451-429f-84fe-dcd070898d65" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 760.493715] env[61806]: DEBUG nova.network.neutron [req-f944cc19-9c05-40b5-86f7-fb2e763454c0 req-3d1ce8ec-b917-494b-9cff-e9a6a2ccbae6 service nova] [instance: 24c83131-1451-429f-84fe-dcd070898d65] Refreshing network info cache for port 30e4bd97-a4da-443f-a523-14f33ef9b039 {{(pid=61806) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 760.541199] env[61806]: DEBUG nova.compute.manager [None req-b5a6ece2-b279-4bbd-a3d1-16582c4d8894 tempest-ServersTestMultiNic-1480543082 tempest-ServersTestMultiNic-1480543082-project-member] [instance: 24c83131-1451-429f-84fe-dcd070898d65] Start spawning the instance on the hypervisor. {{(pid=61806) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 760.569265] env[61806]: DEBUG nova.virt.hardware [None req-b5a6ece2-b279-4bbd-a3d1-16582c4d8894 tempest-ServersTestMultiNic-1480543082 tempest-ServersTestMultiNic-1480543082-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-15T18:05:10Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-15T18:04:52Z,direct_url=,disk_format='vmdk',id=ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='84ba83fce2c349a6988173c5d6fc3b07',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-15T18:04:53Z,virtual_size=,visibility=), allow threads: False {{(pid=61806) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 760.569449] env[61806]: DEBUG nova.virt.hardware [None req-b5a6ece2-b279-4bbd-a3d1-16582c4d8894 tempest-ServersTestMultiNic-1480543082 tempest-ServersTestMultiNic-1480543082-project-member] Flavor limits 0:0:0 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 760.569626] env[61806]: DEBUG nova.virt.hardware [None req-b5a6ece2-b279-4bbd-a3d1-16582c4d8894 tempest-ServersTestMultiNic-1480543082 tempest-ServersTestMultiNic-1480543082-project-member] Image limits 0:0:0 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 760.569836] env[61806]: DEBUG nova.virt.hardware [None req-b5a6ece2-b279-4bbd-a3d1-16582c4d8894 tempest-ServersTestMultiNic-1480543082 tempest-ServersTestMultiNic-1480543082-project-member] Flavor pref 0:0:0 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 760.569983] env[61806]: DEBUG nova.virt.hardware [None req-b5a6ece2-b279-4bbd-a3d1-16582c4d8894 tempest-ServersTestMultiNic-1480543082 tempest-ServersTestMultiNic-1480543082-project-member] Image pref 0:0:0 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 760.570246] env[61806]: DEBUG nova.virt.hardware [None req-b5a6ece2-b279-4bbd-a3d1-16582c4d8894 tempest-ServersTestMultiNic-1480543082 tempest-ServersTestMultiNic-1480543082-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 760.570458] env[61806]: DEBUG nova.virt.hardware [None req-b5a6ece2-b279-4bbd-a3d1-16582c4d8894 tempest-ServersTestMultiNic-1480543082 tempest-ServersTestMultiNic-1480543082-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61806) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 760.570611] env[61806]: DEBUG nova.virt.hardware [None req-b5a6ece2-b279-4bbd-a3d1-16582c4d8894 tempest-ServersTestMultiNic-1480543082 tempest-ServersTestMultiNic-1480543082-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61806) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 760.570769] env[61806]: DEBUG nova.virt.hardware [None req-b5a6ece2-b279-4bbd-a3d1-16582c4d8894 tempest-ServersTestMultiNic-1480543082 tempest-ServersTestMultiNic-1480543082-project-member] Got 1 possible topologies {{(pid=61806) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 760.570922] env[61806]: DEBUG nova.virt.hardware [None req-b5a6ece2-b279-4bbd-a3d1-16582c4d8894 tempest-ServersTestMultiNic-1480543082 tempest-ServersTestMultiNic-1480543082-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61806) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 760.571096] env[61806]: DEBUG nova.virt.hardware [None req-b5a6ece2-b279-4bbd-a3d1-16582c4d8894 tempest-ServersTestMultiNic-1480543082 tempest-ServersTestMultiNic-1480543082-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61806) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 760.571968] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c2cdeb85-9838-4a16-845d-fb54e4e736a7 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 760.580405] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ea10d077-5ba0-4691-a024-2ad8c6e08088 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 760.643190] env[61806]: ERROR nova.compute.manager [None req-b5a6ece2-b279-4bbd-a3d1-16582c4d8894 tempest-ServersTestMultiNic-1480543082 tempest-ServersTestMultiNic-1480543082-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 30e4bd97-a4da-443f-a523-14f33ef9b039, please check neutron logs for more information. [ 760.643190] env[61806]: ERROR nova.compute.manager Traceback (most recent call last): [ 760.643190] env[61806]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 760.643190] env[61806]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 760.643190] env[61806]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 760.643190] env[61806]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 760.643190] env[61806]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 760.643190] env[61806]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 760.643190] env[61806]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 760.643190] env[61806]: ERROR nova.compute.manager self.force_reraise() [ 760.643190] env[61806]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 760.643190] env[61806]: ERROR nova.compute.manager raise self.value [ 760.643190] env[61806]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 760.643190] env[61806]: ERROR nova.compute.manager updated_port = self._update_port( [ 760.643190] env[61806]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 760.643190] env[61806]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 760.643723] env[61806]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 760.643723] env[61806]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 760.643723] env[61806]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 30e4bd97-a4da-443f-a523-14f33ef9b039, please check neutron logs for more information. [ 760.643723] env[61806]: ERROR nova.compute.manager [ 760.643723] env[61806]: Traceback (most recent call last): [ 760.643723] env[61806]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 760.643723] env[61806]: listener.cb(fileno) [ 760.643723] env[61806]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 760.643723] env[61806]: result = function(*args, **kwargs) [ 760.643723] env[61806]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 760.643723] env[61806]: return func(*args, **kwargs) [ 760.643723] env[61806]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 760.643723] env[61806]: raise e [ 760.643723] env[61806]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 760.643723] env[61806]: nwinfo = self.network_api.allocate_for_instance( [ 760.643723] env[61806]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 760.643723] env[61806]: created_port_ids = self._update_ports_for_instance( [ 760.643723] env[61806]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 760.643723] env[61806]: with excutils.save_and_reraise_exception(): [ 760.643723] env[61806]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 760.643723] env[61806]: self.force_reraise() [ 760.643723] env[61806]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 760.643723] env[61806]: raise self.value [ 760.643723] env[61806]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 760.643723] env[61806]: updated_port = self._update_port( [ 760.643723] env[61806]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 760.643723] env[61806]: _ensure_no_port_binding_failure(port) [ 760.643723] env[61806]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 760.643723] env[61806]: raise exception.PortBindingFailed(port_id=port['id']) [ 760.644772] env[61806]: nova.exception.PortBindingFailed: Binding failed for port 30e4bd97-a4da-443f-a523-14f33ef9b039, please check neutron logs for more information. [ 760.644772] env[61806]: Removing descriptor: 14 [ 760.644772] env[61806]: ERROR nova.compute.manager [None req-b5a6ece2-b279-4bbd-a3d1-16582c4d8894 tempest-ServersTestMultiNic-1480543082 tempest-ServersTestMultiNic-1480543082-project-member] [instance: 24c83131-1451-429f-84fe-dcd070898d65] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 30e4bd97-a4da-443f-a523-14f33ef9b039, please check neutron logs for more information. [ 760.644772] env[61806]: ERROR nova.compute.manager [instance: 24c83131-1451-429f-84fe-dcd070898d65] Traceback (most recent call last): [ 760.644772] env[61806]: ERROR nova.compute.manager [instance: 24c83131-1451-429f-84fe-dcd070898d65] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 760.644772] env[61806]: ERROR nova.compute.manager [instance: 24c83131-1451-429f-84fe-dcd070898d65] yield resources [ 760.644772] env[61806]: ERROR nova.compute.manager [instance: 24c83131-1451-429f-84fe-dcd070898d65] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 760.644772] env[61806]: ERROR nova.compute.manager [instance: 24c83131-1451-429f-84fe-dcd070898d65] self.driver.spawn(context, instance, image_meta, [ 760.644772] env[61806]: ERROR nova.compute.manager [instance: 24c83131-1451-429f-84fe-dcd070898d65] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 760.644772] env[61806]: ERROR nova.compute.manager [instance: 24c83131-1451-429f-84fe-dcd070898d65] self._vmops.spawn(context, instance, image_meta, injected_files, [ 760.644772] env[61806]: ERROR nova.compute.manager [instance: 24c83131-1451-429f-84fe-dcd070898d65] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 760.644772] env[61806]: ERROR nova.compute.manager [instance: 24c83131-1451-429f-84fe-dcd070898d65] vm_ref = self.build_virtual_machine(instance, [ 760.645213] env[61806]: ERROR nova.compute.manager [instance: 24c83131-1451-429f-84fe-dcd070898d65] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 760.645213] env[61806]: ERROR nova.compute.manager [instance: 24c83131-1451-429f-84fe-dcd070898d65] vif_infos = vmwarevif.get_vif_info(self._session, [ 760.645213] env[61806]: ERROR nova.compute.manager [instance: 24c83131-1451-429f-84fe-dcd070898d65] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 760.645213] env[61806]: ERROR nova.compute.manager [instance: 24c83131-1451-429f-84fe-dcd070898d65] for vif in network_info: [ 760.645213] env[61806]: ERROR nova.compute.manager [instance: 24c83131-1451-429f-84fe-dcd070898d65] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 760.645213] env[61806]: ERROR nova.compute.manager [instance: 24c83131-1451-429f-84fe-dcd070898d65] return self._sync_wrapper(fn, *args, **kwargs) [ 760.645213] env[61806]: ERROR nova.compute.manager [instance: 24c83131-1451-429f-84fe-dcd070898d65] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 760.645213] env[61806]: ERROR nova.compute.manager [instance: 24c83131-1451-429f-84fe-dcd070898d65] self.wait() [ 760.645213] env[61806]: ERROR nova.compute.manager [instance: 24c83131-1451-429f-84fe-dcd070898d65] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 760.645213] env[61806]: ERROR nova.compute.manager [instance: 24c83131-1451-429f-84fe-dcd070898d65] self[:] = self._gt.wait() [ 760.645213] env[61806]: ERROR nova.compute.manager [instance: 24c83131-1451-429f-84fe-dcd070898d65] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 760.645213] env[61806]: ERROR nova.compute.manager [instance: 24c83131-1451-429f-84fe-dcd070898d65] return self._exit_event.wait() [ 760.645213] env[61806]: ERROR nova.compute.manager [instance: 24c83131-1451-429f-84fe-dcd070898d65] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 760.645665] env[61806]: ERROR nova.compute.manager [instance: 24c83131-1451-429f-84fe-dcd070898d65] result = hub.switch() [ 760.645665] env[61806]: ERROR nova.compute.manager [instance: 24c83131-1451-429f-84fe-dcd070898d65] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 760.645665] env[61806]: ERROR nova.compute.manager [instance: 24c83131-1451-429f-84fe-dcd070898d65] return self.greenlet.switch() [ 760.645665] env[61806]: ERROR nova.compute.manager [instance: 24c83131-1451-429f-84fe-dcd070898d65] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 760.645665] env[61806]: ERROR nova.compute.manager [instance: 24c83131-1451-429f-84fe-dcd070898d65] result = function(*args, **kwargs) [ 760.645665] env[61806]: ERROR nova.compute.manager [instance: 24c83131-1451-429f-84fe-dcd070898d65] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 760.645665] env[61806]: ERROR nova.compute.manager [instance: 24c83131-1451-429f-84fe-dcd070898d65] return func(*args, **kwargs) [ 760.645665] env[61806]: ERROR nova.compute.manager [instance: 24c83131-1451-429f-84fe-dcd070898d65] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 760.645665] env[61806]: ERROR nova.compute.manager [instance: 24c83131-1451-429f-84fe-dcd070898d65] raise e [ 760.645665] env[61806]: ERROR nova.compute.manager [instance: 24c83131-1451-429f-84fe-dcd070898d65] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 760.645665] env[61806]: ERROR nova.compute.manager [instance: 24c83131-1451-429f-84fe-dcd070898d65] nwinfo = self.network_api.allocate_for_instance( [ 760.645665] env[61806]: ERROR nova.compute.manager [instance: 24c83131-1451-429f-84fe-dcd070898d65] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 760.645665] env[61806]: ERROR nova.compute.manager [instance: 24c83131-1451-429f-84fe-dcd070898d65] created_port_ids = self._update_ports_for_instance( [ 760.646423] env[61806]: ERROR nova.compute.manager [instance: 24c83131-1451-429f-84fe-dcd070898d65] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 760.646423] env[61806]: ERROR nova.compute.manager [instance: 24c83131-1451-429f-84fe-dcd070898d65] with excutils.save_and_reraise_exception(): [ 760.646423] env[61806]: ERROR nova.compute.manager [instance: 24c83131-1451-429f-84fe-dcd070898d65] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 760.646423] env[61806]: ERROR nova.compute.manager [instance: 24c83131-1451-429f-84fe-dcd070898d65] self.force_reraise() [ 760.646423] env[61806]: ERROR nova.compute.manager [instance: 24c83131-1451-429f-84fe-dcd070898d65] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 760.646423] env[61806]: ERROR nova.compute.manager [instance: 24c83131-1451-429f-84fe-dcd070898d65] raise self.value [ 760.646423] env[61806]: ERROR nova.compute.manager [instance: 24c83131-1451-429f-84fe-dcd070898d65] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 760.646423] env[61806]: ERROR nova.compute.manager [instance: 24c83131-1451-429f-84fe-dcd070898d65] updated_port = self._update_port( [ 760.646423] env[61806]: ERROR nova.compute.manager [instance: 24c83131-1451-429f-84fe-dcd070898d65] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 760.646423] env[61806]: ERROR nova.compute.manager [instance: 24c83131-1451-429f-84fe-dcd070898d65] _ensure_no_port_binding_failure(port) [ 760.646423] env[61806]: ERROR nova.compute.manager [instance: 24c83131-1451-429f-84fe-dcd070898d65] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 760.646423] env[61806]: ERROR nova.compute.manager [instance: 24c83131-1451-429f-84fe-dcd070898d65] raise exception.PortBindingFailed(port_id=port['id']) [ 760.646864] env[61806]: ERROR nova.compute.manager [instance: 24c83131-1451-429f-84fe-dcd070898d65] nova.exception.PortBindingFailed: Binding failed for port 30e4bd97-a4da-443f-a523-14f33ef9b039, please check neutron logs for more information. [ 760.646864] env[61806]: ERROR nova.compute.manager [instance: 24c83131-1451-429f-84fe-dcd070898d65] [ 760.646864] env[61806]: INFO nova.compute.manager [None req-b5a6ece2-b279-4bbd-a3d1-16582c4d8894 tempest-ServersTestMultiNic-1480543082 tempest-ServersTestMultiNic-1480543082-project-member] [instance: 24c83131-1451-429f-84fe-dcd070898d65] Terminating instance [ 760.646864] env[61806]: DEBUG oslo_concurrency.lockutils [None req-b5a6ece2-b279-4bbd-a3d1-16582c4d8894 tempest-ServersTestMultiNic-1480543082 tempest-ServersTestMultiNic-1480543082-project-member] Acquiring lock "refresh_cache-24c83131-1451-429f-84fe-dcd070898d65" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 760.865046] env[61806]: DEBUG oslo_concurrency.lockutils [None req-0b87073d-3de3-486a-876d-8ca5794993c1 tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 760.952102] env[61806]: DEBUG nova.network.neutron [None req-1382c347-5a79-4289-8b32-002f296b7ae3 tempest-TenantUsagesTestJSON-255073140 tempest-TenantUsagesTestJSON-255073140-project-member] [instance: 949c1050-5003-4519-b24c-8904de21a676] Instance cache missing network info. {{(pid=61806) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 761.023578] env[61806]: DEBUG nova.network.neutron [None req-1382c347-5a79-4289-8b32-002f296b7ae3 tempest-TenantUsagesTestJSON-255073140 tempest-TenantUsagesTestJSON-255073140-project-member] [instance: 949c1050-5003-4519-b24c-8904de21a676] Updating instance_info_cache with network_info: [] {{(pid=61806) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 761.066757] env[61806]: DEBUG nova.network.neutron [req-f944cc19-9c05-40b5-86f7-fb2e763454c0 req-3d1ce8ec-b917-494b-9cff-e9a6a2ccbae6 service nova] [instance: 24c83131-1451-429f-84fe-dcd070898d65] Instance cache missing network info. {{(pid=61806) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 761.147032] env[61806]: DEBUG nova.network.neutron [req-f944cc19-9c05-40b5-86f7-fb2e763454c0 req-3d1ce8ec-b917-494b-9cff-e9a6a2ccbae6 service nova] [instance: 24c83131-1451-429f-84fe-dcd070898d65] Updating instance_info_cache with network_info: [] {{(pid=61806) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 761.205075] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c8bba314-b783-407f-b607-3043f8f0957b {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 761.212856] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e45a3ceb-fd92-4e4e-b437-df8e14c3910c {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 761.244722] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f225940b-08f4-467c-b8bf-84f08c2862a0 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 761.253207] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-88860b6d-b1d7-4e78-983d-24397b08fa93 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 761.266041] env[61806]: DEBUG nova.compute.provider_tree [None req-3bf756e0-c56e-4902-9c2b-b224a1d09a7c tempest-ListServersNegativeTestJSON-272187588 tempest-ListServersNegativeTestJSON-272187588-project-member] Inventory has not changed in ProviderTree for provider: a2858be1-fd22-4e08-979e-87ad25293407 {{(pid=61806) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 761.528059] env[61806]: DEBUG oslo_concurrency.lockutils [None req-1382c347-5a79-4289-8b32-002f296b7ae3 tempest-TenantUsagesTestJSON-255073140 tempest-TenantUsagesTestJSON-255073140-project-member] Releasing lock "refresh_cache-949c1050-5003-4519-b24c-8904de21a676" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 761.528323] env[61806]: DEBUG nova.compute.manager [None req-1382c347-5a79-4289-8b32-002f296b7ae3 tempest-TenantUsagesTestJSON-255073140 tempest-TenantUsagesTestJSON-255073140-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61806) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 761.528487] env[61806]: DEBUG nova.compute.manager [None req-1382c347-5a79-4289-8b32-002f296b7ae3 tempest-TenantUsagesTestJSON-255073140 tempest-TenantUsagesTestJSON-255073140-project-member] [instance: 949c1050-5003-4519-b24c-8904de21a676] Deallocating network for instance {{(pid=61806) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 761.528673] env[61806]: DEBUG nova.network.neutron [None req-1382c347-5a79-4289-8b32-002f296b7ae3 tempest-TenantUsagesTestJSON-255073140 tempest-TenantUsagesTestJSON-255073140-project-member] [instance: 949c1050-5003-4519-b24c-8904de21a676] deallocate_for_instance() {{(pid=61806) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 761.555723] env[61806]: DEBUG nova.network.neutron [None req-1382c347-5a79-4289-8b32-002f296b7ae3 tempest-TenantUsagesTestJSON-255073140 tempest-TenantUsagesTestJSON-255073140-project-member] [instance: 949c1050-5003-4519-b24c-8904de21a676] Instance cache missing network info. {{(pid=61806) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 761.651158] env[61806]: DEBUG oslo_concurrency.lockutils [req-f944cc19-9c05-40b5-86f7-fb2e763454c0 req-3d1ce8ec-b917-494b-9cff-e9a6a2ccbae6 service nova] Releasing lock "refresh_cache-24c83131-1451-429f-84fe-dcd070898d65" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 761.651158] env[61806]: DEBUG oslo_concurrency.lockutils [None req-b5a6ece2-b279-4bbd-a3d1-16582c4d8894 tempest-ServersTestMultiNic-1480543082 tempest-ServersTestMultiNic-1480543082-project-member] Acquired lock "refresh_cache-24c83131-1451-429f-84fe-dcd070898d65" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 761.651158] env[61806]: DEBUG nova.network.neutron [None req-b5a6ece2-b279-4bbd-a3d1-16582c4d8894 tempest-ServersTestMultiNic-1480543082 tempest-ServersTestMultiNic-1480543082-project-member] [instance: 24c83131-1451-429f-84fe-dcd070898d65] Building network info cache for instance {{(pid=61806) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 761.772181] env[61806]: DEBUG nova.scheduler.client.report [None req-3bf756e0-c56e-4902-9c2b-b224a1d09a7c tempest-ListServersNegativeTestJSON-272187588 tempest-ListServersNegativeTestJSON-272187588-project-member] Inventory has not changed for provider a2858be1-fd22-4e08-979e-87ad25293407 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 139, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61806) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 762.059121] env[61806]: DEBUG nova.network.neutron [None req-1382c347-5a79-4289-8b32-002f296b7ae3 tempest-TenantUsagesTestJSON-255073140 tempest-TenantUsagesTestJSON-255073140-project-member] [instance: 949c1050-5003-4519-b24c-8904de21a676] Updating instance_info_cache with network_info: [] {{(pid=61806) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 762.180648] env[61806]: DEBUG nova.network.neutron [None req-b5a6ece2-b279-4bbd-a3d1-16582c4d8894 tempest-ServersTestMultiNic-1480543082 tempest-ServersTestMultiNic-1480543082-project-member] [instance: 24c83131-1451-429f-84fe-dcd070898d65] Instance cache missing network info. {{(pid=61806) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 762.264599] env[61806]: DEBUG nova.network.neutron [None req-b5a6ece2-b279-4bbd-a3d1-16582c4d8894 tempest-ServersTestMultiNic-1480543082 tempest-ServersTestMultiNic-1480543082-project-member] [instance: 24c83131-1451-429f-84fe-dcd070898d65] Updating instance_info_cache with network_info: [] {{(pid=61806) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 762.276984] env[61806]: DEBUG oslo_concurrency.lockutils [None req-3bf756e0-c56e-4902-9c2b-b224a1d09a7c tempest-ListServersNegativeTestJSON-272187588 tempest-ListServersNegativeTestJSON-272187588-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.864s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 762.277384] env[61806]: ERROR nova.compute.manager [None req-3bf756e0-c56e-4902-9c2b-b224a1d09a7c tempest-ListServersNegativeTestJSON-272187588 tempest-ListServersNegativeTestJSON-272187588-project-member] [instance: f2446d04-15b8-4b3d-8932-c37df67b0f8f] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port ad624cd6-204b-4189-a855-10dca931bd79, please check neutron logs for more information. [ 762.277384] env[61806]: ERROR nova.compute.manager [instance: f2446d04-15b8-4b3d-8932-c37df67b0f8f] Traceback (most recent call last): [ 762.277384] env[61806]: ERROR nova.compute.manager [instance: f2446d04-15b8-4b3d-8932-c37df67b0f8f] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 762.277384] env[61806]: ERROR nova.compute.manager [instance: f2446d04-15b8-4b3d-8932-c37df67b0f8f] self.driver.spawn(context, instance, image_meta, [ 762.277384] env[61806]: ERROR nova.compute.manager [instance: f2446d04-15b8-4b3d-8932-c37df67b0f8f] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 762.277384] env[61806]: ERROR nova.compute.manager [instance: f2446d04-15b8-4b3d-8932-c37df67b0f8f] self._vmops.spawn(context, instance, image_meta, injected_files, [ 762.277384] env[61806]: ERROR nova.compute.manager [instance: f2446d04-15b8-4b3d-8932-c37df67b0f8f] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 762.277384] env[61806]: ERROR nova.compute.manager [instance: f2446d04-15b8-4b3d-8932-c37df67b0f8f] vm_ref = self.build_virtual_machine(instance, [ 762.277384] env[61806]: ERROR nova.compute.manager [instance: f2446d04-15b8-4b3d-8932-c37df67b0f8f] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 762.277384] env[61806]: ERROR nova.compute.manager [instance: f2446d04-15b8-4b3d-8932-c37df67b0f8f] vif_infos = vmwarevif.get_vif_info(self._session, [ 762.277384] env[61806]: ERROR nova.compute.manager [instance: f2446d04-15b8-4b3d-8932-c37df67b0f8f] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 762.277800] env[61806]: ERROR nova.compute.manager [instance: f2446d04-15b8-4b3d-8932-c37df67b0f8f] for vif in network_info: [ 762.277800] env[61806]: ERROR nova.compute.manager [instance: f2446d04-15b8-4b3d-8932-c37df67b0f8f] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 762.277800] env[61806]: ERROR nova.compute.manager [instance: f2446d04-15b8-4b3d-8932-c37df67b0f8f] return self._sync_wrapper(fn, *args, **kwargs) [ 762.277800] env[61806]: ERROR nova.compute.manager [instance: f2446d04-15b8-4b3d-8932-c37df67b0f8f] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 762.277800] env[61806]: ERROR nova.compute.manager [instance: f2446d04-15b8-4b3d-8932-c37df67b0f8f] self.wait() [ 762.277800] env[61806]: ERROR nova.compute.manager [instance: f2446d04-15b8-4b3d-8932-c37df67b0f8f] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 762.277800] env[61806]: ERROR nova.compute.manager [instance: f2446d04-15b8-4b3d-8932-c37df67b0f8f] self[:] = self._gt.wait() [ 762.277800] env[61806]: ERROR nova.compute.manager [instance: f2446d04-15b8-4b3d-8932-c37df67b0f8f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 762.277800] env[61806]: ERROR nova.compute.manager [instance: f2446d04-15b8-4b3d-8932-c37df67b0f8f] return self._exit_event.wait() [ 762.277800] env[61806]: ERROR nova.compute.manager [instance: f2446d04-15b8-4b3d-8932-c37df67b0f8f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 762.277800] env[61806]: ERROR nova.compute.manager [instance: f2446d04-15b8-4b3d-8932-c37df67b0f8f] current.throw(*self._exc) [ 762.277800] env[61806]: ERROR nova.compute.manager [instance: f2446d04-15b8-4b3d-8932-c37df67b0f8f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 762.277800] env[61806]: ERROR nova.compute.manager [instance: f2446d04-15b8-4b3d-8932-c37df67b0f8f] result = function(*args, **kwargs) [ 762.278404] env[61806]: ERROR nova.compute.manager [instance: f2446d04-15b8-4b3d-8932-c37df67b0f8f] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 762.278404] env[61806]: ERROR nova.compute.manager [instance: f2446d04-15b8-4b3d-8932-c37df67b0f8f] return func(*args, **kwargs) [ 762.278404] env[61806]: ERROR nova.compute.manager [instance: f2446d04-15b8-4b3d-8932-c37df67b0f8f] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 762.278404] env[61806]: ERROR nova.compute.manager [instance: f2446d04-15b8-4b3d-8932-c37df67b0f8f] raise e [ 762.278404] env[61806]: ERROR nova.compute.manager [instance: f2446d04-15b8-4b3d-8932-c37df67b0f8f] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 762.278404] env[61806]: ERROR nova.compute.manager [instance: f2446d04-15b8-4b3d-8932-c37df67b0f8f] nwinfo = self.network_api.allocate_for_instance( [ 762.278404] env[61806]: ERROR nova.compute.manager [instance: f2446d04-15b8-4b3d-8932-c37df67b0f8f] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 762.278404] env[61806]: ERROR nova.compute.manager [instance: f2446d04-15b8-4b3d-8932-c37df67b0f8f] created_port_ids = self._update_ports_for_instance( [ 762.278404] env[61806]: ERROR nova.compute.manager [instance: f2446d04-15b8-4b3d-8932-c37df67b0f8f] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 762.278404] env[61806]: ERROR nova.compute.manager [instance: f2446d04-15b8-4b3d-8932-c37df67b0f8f] with excutils.save_and_reraise_exception(): [ 762.278404] env[61806]: ERROR nova.compute.manager [instance: f2446d04-15b8-4b3d-8932-c37df67b0f8f] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 762.278404] env[61806]: ERROR nova.compute.manager [instance: f2446d04-15b8-4b3d-8932-c37df67b0f8f] self.force_reraise() [ 762.278404] env[61806]: ERROR nova.compute.manager [instance: f2446d04-15b8-4b3d-8932-c37df67b0f8f] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 762.278822] env[61806]: ERROR nova.compute.manager [instance: f2446d04-15b8-4b3d-8932-c37df67b0f8f] raise self.value [ 762.278822] env[61806]: ERROR nova.compute.manager [instance: f2446d04-15b8-4b3d-8932-c37df67b0f8f] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 762.278822] env[61806]: ERROR nova.compute.manager [instance: f2446d04-15b8-4b3d-8932-c37df67b0f8f] updated_port = self._update_port( [ 762.278822] env[61806]: ERROR nova.compute.manager [instance: f2446d04-15b8-4b3d-8932-c37df67b0f8f] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 762.278822] env[61806]: ERROR nova.compute.manager [instance: f2446d04-15b8-4b3d-8932-c37df67b0f8f] _ensure_no_port_binding_failure(port) [ 762.278822] env[61806]: ERROR nova.compute.manager [instance: f2446d04-15b8-4b3d-8932-c37df67b0f8f] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 762.278822] env[61806]: ERROR nova.compute.manager [instance: f2446d04-15b8-4b3d-8932-c37df67b0f8f] raise exception.PortBindingFailed(port_id=port['id']) [ 762.278822] env[61806]: ERROR nova.compute.manager [instance: f2446d04-15b8-4b3d-8932-c37df67b0f8f] nova.exception.PortBindingFailed: Binding failed for port ad624cd6-204b-4189-a855-10dca931bd79, please check neutron logs for more information. [ 762.278822] env[61806]: ERROR nova.compute.manager [instance: f2446d04-15b8-4b3d-8932-c37df67b0f8f] [ 762.278822] env[61806]: DEBUG nova.compute.utils [None req-3bf756e0-c56e-4902-9c2b-b224a1d09a7c tempest-ListServersNegativeTestJSON-272187588 tempest-ListServersNegativeTestJSON-272187588-project-member] [instance: f2446d04-15b8-4b3d-8932-c37df67b0f8f] Binding failed for port ad624cd6-204b-4189-a855-10dca931bd79, please check neutron logs for more information. {{(pid=61806) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 762.279398] env[61806]: DEBUG oslo_concurrency.lockutils [None req-35a8f95d-1ddd-492f-a93c-adb382770732 tempest-ServersV294TestFqdnHostnames-1709116086 tempest-ServersV294TestFqdnHostnames-1709116086-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 14.141s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 762.281063] env[61806]: INFO nova.compute.claims [None req-35a8f95d-1ddd-492f-a93c-adb382770732 tempest-ServersV294TestFqdnHostnames-1709116086 tempest-ServersV294TestFqdnHostnames-1709116086-project-member] [instance: c25ddf7c-d0cc-4b73-96da-1dcd6012f072] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 762.284281] env[61806]: DEBUG nova.compute.manager [None req-3bf756e0-c56e-4902-9c2b-b224a1d09a7c tempest-ListServersNegativeTestJSON-272187588 tempest-ListServersNegativeTestJSON-272187588-project-member] [instance: f2446d04-15b8-4b3d-8932-c37df67b0f8f] Build of instance f2446d04-15b8-4b3d-8932-c37df67b0f8f was re-scheduled: Binding failed for port ad624cd6-204b-4189-a855-10dca931bd79, please check neutron logs for more information. {{(pid=61806) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 762.284694] env[61806]: DEBUG nova.compute.manager [None req-3bf756e0-c56e-4902-9c2b-b224a1d09a7c tempest-ListServersNegativeTestJSON-272187588 tempest-ListServersNegativeTestJSON-272187588-project-member] [instance: f2446d04-15b8-4b3d-8932-c37df67b0f8f] Unplugging VIFs for instance {{(pid=61806) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 762.284926] env[61806]: DEBUG oslo_concurrency.lockutils [None req-3bf756e0-c56e-4902-9c2b-b224a1d09a7c tempest-ListServersNegativeTestJSON-272187588 tempest-ListServersNegativeTestJSON-272187588-project-member] Acquiring lock "refresh_cache-f2446d04-15b8-4b3d-8932-c37df67b0f8f" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 762.285091] env[61806]: DEBUG oslo_concurrency.lockutils [None req-3bf756e0-c56e-4902-9c2b-b224a1d09a7c tempest-ListServersNegativeTestJSON-272187588 tempest-ListServersNegativeTestJSON-272187588-project-member] Acquired lock "refresh_cache-f2446d04-15b8-4b3d-8932-c37df67b0f8f" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 762.285253] env[61806]: DEBUG nova.network.neutron [None req-3bf756e0-c56e-4902-9c2b-b224a1d09a7c tempest-ListServersNegativeTestJSON-272187588 tempest-ListServersNegativeTestJSON-272187588-project-member] [instance: f2446d04-15b8-4b3d-8932-c37df67b0f8f] Building network info cache for instance {{(pid=61806) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 762.517832] env[61806]: DEBUG nova.compute.manager [req-e0f8e0a9-f83c-4354-9c97-e1b3adc670e2 req-c29c087f-4c9d-4494-9386-53621a3629d7 service nova] [instance: 24c83131-1451-429f-84fe-dcd070898d65] Received event network-vif-deleted-30e4bd97-a4da-443f-a523-14f33ef9b039 {{(pid=61806) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 762.564796] env[61806]: INFO nova.compute.manager [None req-1382c347-5a79-4289-8b32-002f296b7ae3 tempest-TenantUsagesTestJSON-255073140 tempest-TenantUsagesTestJSON-255073140-project-member] [instance: 949c1050-5003-4519-b24c-8904de21a676] Took 1.04 seconds to deallocate network for instance. [ 762.766338] env[61806]: DEBUG oslo_concurrency.lockutils [None req-b5a6ece2-b279-4bbd-a3d1-16582c4d8894 tempest-ServersTestMultiNic-1480543082 tempest-ServersTestMultiNic-1480543082-project-member] Releasing lock "refresh_cache-24c83131-1451-429f-84fe-dcd070898d65" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 762.766808] env[61806]: DEBUG nova.compute.manager [None req-b5a6ece2-b279-4bbd-a3d1-16582c4d8894 tempest-ServersTestMultiNic-1480543082 tempest-ServersTestMultiNic-1480543082-project-member] [instance: 24c83131-1451-429f-84fe-dcd070898d65] Start destroying the instance on the hypervisor. {{(pid=61806) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 762.767046] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-b5a6ece2-b279-4bbd-a3d1-16582c4d8894 tempest-ServersTestMultiNic-1480543082 tempest-ServersTestMultiNic-1480543082-project-member] [instance: 24c83131-1451-429f-84fe-dcd070898d65] Destroying instance {{(pid=61806) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 762.767360] env[61806]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-7b5c52ea-3e2e-4997-89e3-9158b6affedf {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 762.776321] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9342ed04-5775-4225-b049-bd30115b1b52 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 762.801702] env[61806]: WARNING nova.virt.vmwareapi.vmops [None req-b5a6ece2-b279-4bbd-a3d1-16582c4d8894 tempest-ServersTestMultiNic-1480543082 tempest-ServersTestMultiNic-1480543082-project-member] [instance: 24c83131-1451-429f-84fe-dcd070898d65] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 24c83131-1451-429f-84fe-dcd070898d65 could not be found. [ 762.801911] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-b5a6ece2-b279-4bbd-a3d1-16582c4d8894 tempest-ServersTestMultiNic-1480543082 tempest-ServersTestMultiNic-1480543082-project-member] [instance: 24c83131-1451-429f-84fe-dcd070898d65] Instance destroyed {{(pid=61806) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 762.802085] env[61806]: INFO nova.compute.manager [None req-b5a6ece2-b279-4bbd-a3d1-16582c4d8894 tempest-ServersTestMultiNic-1480543082 tempest-ServersTestMultiNic-1480543082-project-member] [instance: 24c83131-1451-429f-84fe-dcd070898d65] Took 0.04 seconds to destroy the instance on the hypervisor. [ 762.802323] env[61806]: DEBUG oslo.service.loopingcall [None req-b5a6ece2-b279-4bbd-a3d1-16582c4d8894 tempest-ServersTestMultiNic-1480543082 tempest-ServersTestMultiNic-1480543082-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61806) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 762.802552] env[61806]: DEBUG nova.compute.manager [-] [instance: 24c83131-1451-429f-84fe-dcd070898d65] Deallocating network for instance {{(pid=61806) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 762.802688] env[61806]: DEBUG nova.network.neutron [-] [instance: 24c83131-1451-429f-84fe-dcd070898d65] deallocate_for_instance() {{(pid=61806) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 762.812643] env[61806]: DEBUG nova.network.neutron [None req-3bf756e0-c56e-4902-9c2b-b224a1d09a7c tempest-ListServersNegativeTestJSON-272187588 tempest-ListServersNegativeTestJSON-272187588-project-member] [instance: f2446d04-15b8-4b3d-8932-c37df67b0f8f] Instance cache missing network info. {{(pid=61806) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 762.830803] env[61806]: DEBUG nova.network.neutron [-] [instance: 24c83131-1451-429f-84fe-dcd070898d65] Instance cache missing network info. {{(pid=61806) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 762.894221] env[61806]: DEBUG nova.network.neutron [None req-3bf756e0-c56e-4902-9c2b-b224a1d09a7c tempest-ListServersNegativeTestJSON-272187588 tempest-ListServersNegativeTestJSON-272187588-project-member] [instance: f2446d04-15b8-4b3d-8932-c37df67b0f8f] Updating instance_info_cache with network_info: [] {{(pid=61806) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 763.397461] env[61806]: DEBUG oslo_concurrency.lockutils [None req-3bf756e0-c56e-4902-9c2b-b224a1d09a7c tempest-ListServersNegativeTestJSON-272187588 tempest-ListServersNegativeTestJSON-272187588-project-member] Releasing lock "refresh_cache-f2446d04-15b8-4b3d-8932-c37df67b0f8f" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 763.397879] env[61806]: DEBUG nova.compute.manager [None req-3bf756e0-c56e-4902-9c2b-b224a1d09a7c tempest-ListServersNegativeTestJSON-272187588 tempest-ListServersNegativeTestJSON-272187588-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61806) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 763.398108] env[61806]: DEBUG nova.compute.manager [None req-3bf756e0-c56e-4902-9c2b-b224a1d09a7c tempest-ListServersNegativeTestJSON-272187588 tempest-ListServersNegativeTestJSON-272187588-project-member] [instance: f2446d04-15b8-4b3d-8932-c37df67b0f8f] Deallocating network for instance {{(pid=61806) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 763.398318] env[61806]: DEBUG nova.network.neutron [None req-3bf756e0-c56e-4902-9c2b-b224a1d09a7c tempest-ListServersNegativeTestJSON-272187588 tempest-ListServersNegativeTestJSON-272187588-project-member] [instance: f2446d04-15b8-4b3d-8932-c37df67b0f8f] deallocate_for_instance() {{(pid=61806) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 763.416522] env[61806]: DEBUG nova.network.neutron [None req-3bf756e0-c56e-4902-9c2b-b224a1d09a7c tempest-ListServersNegativeTestJSON-272187588 tempest-ListServersNegativeTestJSON-272187588-project-member] [instance: f2446d04-15b8-4b3d-8932-c37df67b0f8f] Instance cache missing network info. {{(pid=61806) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 763.515512] env[61806]: DEBUG nova.network.neutron [-] [instance: 24c83131-1451-429f-84fe-dcd070898d65] Updating instance_info_cache with network_info: [] {{(pid=61806) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 763.566296] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cdb41d02-177f-46a6-b60d-ded39f903673 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 763.576285] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c65f3d79-ad70-4870-b5fa-672ca09e79de {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 763.627390] env[61806]: INFO nova.scheduler.client.report [None req-1382c347-5a79-4289-8b32-002f296b7ae3 tempest-TenantUsagesTestJSON-255073140 tempest-TenantUsagesTestJSON-255073140-project-member] Deleted allocations for instance 949c1050-5003-4519-b24c-8904de21a676 [ 763.637859] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b2affe23-eb53-4ed9-bc76-200f7407f0ba {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 763.646174] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dbbcf6fa-ba8f-486a-be5a-d54725276c73 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 763.659989] env[61806]: DEBUG nova.compute.provider_tree [None req-35a8f95d-1ddd-492f-a93c-adb382770732 tempest-ServersV294TestFqdnHostnames-1709116086 tempest-ServersV294TestFqdnHostnames-1709116086-project-member] Inventory has not changed in ProviderTree for provider: a2858be1-fd22-4e08-979e-87ad25293407 {{(pid=61806) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 763.918918] env[61806]: DEBUG nova.network.neutron [None req-3bf756e0-c56e-4902-9c2b-b224a1d09a7c tempest-ListServersNegativeTestJSON-272187588 tempest-ListServersNegativeTestJSON-272187588-project-member] [instance: f2446d04-15b8-4b3d-8932-c37df67b0f8f] Updating instance_info_cache with network_info: [] {{(pid=61806) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 764.018072] env[61806]: INFO nova.compute.manager [-] [instance: 24c83131-1451-429f-84fe-dcd070898d65] Took 1.22 seconds to deallocate network for instance. [ 764.020430] env[61806]: DEBUG nova.compute.claims [None req-b5a6ece2-b279-4bbd-a3d1-16582c4d8894 tempest-ServersTestMultiNic-1480543082 tempest-ServersTestMultiNic-1480543082-project-member] [instance: 24c83131-1451-429f-84fe-dcd070898d65] Aborting claim: {{(pid=61806) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 764.020595] env[61806]: DEBUG oslo_concurrency.lockutils [None req-b5a6ece2-b279-4bbd-a3d1-16582c4d8894 tempest-ServersTestMultiNic-1480543082 tempest-ServersTestMultiNic-1480543082-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 764.142950] env[61806]: DEBUG oslo_concurrency.lockutils [None req-1382c347-5a79-4289-8b32-002f296b7ae3 tempest-TenantUsagesTestJSON-255073140 tempest-TenantUsagesTestJSON-255073140-project-member] Lock "949c1050-5003-4519-b24c-8904de21a676" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 141.273s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 764.163015] env[61806]: DEBUG nova.scheduler.client.report [None req-35a8f95d-1ddd-492f-a93c-adb382770732 tempest-ServersV294TestFqdnHostnames-1709116086 tempest-ServersV294TestFqdnHostnames-1709116086-project-member] Inventory has not changed for provider a2858be1-fd22-4e08-979e-87ad25293407 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 139, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61806) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 764.423415] env[61806]: INFO nova.compute.manager [None req-3bf756e0-c56e-4902-9c2b-b224a1d09a7c tempest-ListServersNegativeTestJSON-272187588 tempest-ListServersNegativeTestJSON-272187588-project-member] [instance: f2446d04-15b8-4b3d-8932-c37df67b0f8f] Took 1.02 seconds to deallocate network for instance. [ 764.646266] env[61806]: DEBUG nova.compute.manager [None req-b3f33a6f-532a-458d-aec6-6ec54b0cadd5 tempest-ServerMetadataTestJSON-316746592 tempest-ServerMetadataTestJSON-316746592-project-member] [instance: 4184e263-6f56-4bc1-99b2-a2c460531516] Starting instance... {{(pid=61806) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 764.668063] env[61806]: DEBUG oslo_concurrency.lockutils [None req-35a8f95d-1ddd-492f-a93c-adb382770732 tempest-ServersV294TestFqdnHostnames-1709116086 tempest-ServersV294TestFqdnHostnames-1709116086-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.389s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 764.668736] env[61806]: DEBUG nova.compute.manager [None req-35a8f95d-1ddd-492f-a93c-adb382770732 tempest-ServersV294TestFqdnHostnames-1709116086 tempest-ServersV294TestFqdnHostnames-1709116086-project-member] [instance: c25ddf7c-d0cc-4b73-96da-1dcd6012f072] Start building networks asynchronously for instance. {{(pid=61806) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 764.671595] env[61806]: DEBUG oslo_concurrency.lockutils [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 14.040s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 765.176018] env[61806]: DEBUG oslo_concurrency.lockutils [None req-b3f33a6f-532a-458d-aec6-6ec54b0cadd5 tempest-ServerMetadataTestJSON-316746592 tempest-ServerMetadataTestJSON-316746592-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 765.176018] env[61806]: DEBUG nova.compute.utils [None req-35a8f95d-1ddd-492f-a93c-adb382770732 tempest-ServersV294TestFqdnHostnames-1709116086 tempest-ServersV294TestFqdnHostnames-1709116086-project-member] Using /dev/sd instead of None {{(pid=61806) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 765.179502] env[61806]: DEBUG nova.compute.manager [None req-35a8f95d-1ddd-492f-a93c-adb382770732 tempest-ServersV294TestFqdnHostnames-1709116086 tempest-ServersV294TestFqdnHostnames-1709116086-project-member] [instance: c25ddf7c-d0cc-4b73-96da-1dcd6012f072] Allocating IP information in the background. {{(pid=61806) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 765.179702] env[61806]: DEBUG nova.network.neutron [None req-35a8f95d-1ddd-492f-a93c-adb382770732 tempest-ServersV294TestFqdnHostnames-1709116086 tempest-ServersV294TestFqdnHostnames-1709116086-project-member] [instance: c25ddf7c-d0cc-4b73-96da-1dcd6012f072] allocate_for_instance() {{(pid=61806) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 765.240594] env[61806]: DEBUG nova.policy [None req-35a8f95d-1ddd-492f-a93c-adb382770732 tempest-ServersV294TestFqdnHostnames-1709116086 tempest-ServersV294TestFqdnHostnames-1709116086-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '5f3b242b2a8246a99568a2fee71b71cd', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '39440126d6374adda6ae8d5650f04177', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61806) authorize /opt/stack/nova/nova/policy.py:201}} [ 765.467339] env[61806]: INFO nova.scheduler.client.report [None req-3bf756e0-c56e-4902-9c2b-b224a1d09a7c tempest-ListServersNegativeTestJSON-272187588 tempest-ListServersNegativeTestJSON-272187588-project-member] Deleted allocations for instance f2446d04-15b8-4b3d-8932-c37df67b0f8f [ 765.544671] env[61806]: DEBUG nova.network.neutron [None req-35a8f95d-1ddd-492f-a93c-adb382770732 tempest-ServersV294TestFqdnHostnames-1709116086 tempest-ServersV294TestFqdnHostnames-1709116086-project-member] [instance: c25ddf7c-d0cc-4b73-96da-1dcd6012f072] Successfully created port: 59bb93fe-b4ce-4348-bd86-42344407e007 {{(pid=61806) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 765.681350] env[61806]: DEBUG nova.compute.manager [None req-35a8f95d-1ddd-492f-a93c-adb382770732 tempest-ServersV294TestFqdnHostnames-1709116086 tempest-ServersV294TestFqdnHostnames-1709116086-project-member] [instance: c25ddf7c-d0cc-4b73-96da-1dcd6012f072] Start building block device mappings for instance. {{(pid=61806) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 765.712683] env[61806]: DEBUG nova.compute.resource_tracker [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Instance ea8d39a1-066d-4eb4-888d-776b0e45f684 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61806) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 765.712683] env[61806]: DEBUG nova.compute.resource_tracker [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Instance 198dc1a4-050d-47b2-8bc4-30af5534a789 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61806) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 765.712683] env[61806]: DEBUG nova.compute.resource_tracker [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Instance 37c07b8d-e2a0-4aa6-b7ba-194feed837e5 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61806) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 765.712683] env[61806]: DEBUG nova.compute.resource_tracker [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Instance c61df5d3-7f43-48d4-a2c0-a8972372c0ba actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61806) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 765.712865] env[61806]: DEBUG nova.compute.resource_tracker [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Instance 24c83131-1451-429f-84fe-dcd070898d65 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61806) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 765.712865] env[61806]: DEBUG nova.compute.resource_tracker [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Instance c25ddf7c-d0cc-4b73-96da-1dcd6012f072 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61806) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 765.979625] env[61806]: DEBUG oslo_concurrency.lockutils [None req-3bf756e0-c56e-4902-9c2b-b224a1d09a7c tempest-ListServersNegativeTestJSON-272187588 tempest-ListServersNegativeTestJSON-272187588-project-member] Lock "f2446d04-15b8-4b3d-8932-c37df67b0f8f" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 137.949s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 766.220286] env[61806]: DEBUG nova.compute.resource_tracker [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Instance e0ef0a35-82a5-495b-9d5c-5805e8306390 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61806) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 766.335458] env[61806]: DEBUG nova.compute.manager [req-b29f7869-9fa5-4a8c-8fcf-36d831494d2d req-d884e4eb-f11b-4834-aea3-93681f4b19a8 service nova] [instance: c25ddf7c-d0cc-4b73-96da-1dcd6012f072] Received event network-changed-59bb93fe-b4ce-4348-bd86-42344407e007 {{(pid=61806) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 766.335659] env[61806]: DEBUG nova.compute.manager [req-b29f7869-9fa5-4a8c-8fcf-36d831494d2d req-d884e4eb-f11b-4834-aea3-93681f4b19a8 service nova] [instance: c25ddf7c-d0cc-4b73-96da-1dcd6012f072] Refreshing instance network info cache due to event network-changed-59bb93fe-b4ce-4348-bd86-42344407e007. {{(pid=61806) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 766.335865] env[61806]: DEBUG oslo_concurrency.lockutils [req-b29f7869-9fa5-4a8c-8fcf-36d831494d2d req-d884e4eb-f11b-4834-aea3-93681f4b19a8 service nova] Acquiring lock "refresh_cache-c25ddf7c-d0cc-4b73-96da-1dcd6012f072" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 766.336017] env[61806]: DEBUG oslo_concurrency.lockutils [req-b29f7869-9fa5-4a8c-8fcf-36d831494d2d req-d884e4eb-f11b-4834-aea3-93681f4b19a8 service nova] Acquired lock "refresh_cache-c25ddf7c-d0cc-4b73-96da-1dcd6012f072" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 766.336448] env[61806]: DEBUG nova.network.neutron [req-b29f7869-9fa5-4a8c-8fcf-36d831494d2d req-d884e4eb-f11b-4834-aea3-93681f4b19a8 service nova] [instance: c25ddf7c-d0cc-4b73-96da-1dcd6012f072] Refreshing network info cache for port 59bb93fe-b4ce-4348-bd86-42344407e007 {{(pid=61806) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 766.487031] env[61806]: DEBUG nova.compute.manager [None req-3d5b4b69-10ec-4d1a-aa3d-4dddf4155c99 tempest-ServerShowV247Test-1218443059 tempest-ServerShowV247Test-1218443059-project-member] [instance: 7c3fd2ff-ebd5-454f-a743-7fbae1088941] Starting instance... {{(pid=61806) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 766.575708] env[61806]: ERROR nova.compute.manager [None req-35a8f95d-1ddd-492f-a93c-adb382770732 tempest-ServersV294TestFqdnHostnames-1709116086 tempest-ServersV294TestFqdnHostnames-1709116086-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 59bb93fe-b4ce-4348-bd86-42344407e007, please check neutron logs for more information. [ 766.575708] env[61806]: ERROR nova.compute.manager Traceback (most recent call last): [ 766.575708] env[61806]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 766.575708] env[61806]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 766.575708] env[61806]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 766.575708] env[61806]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 766.575708] env[61806]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 766.575708] env[61806]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 766.575708] env[61806]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 766.575708] env[61806]: ERROR nova.compute.manager self.force_reraise() [ 766.575708] env[61806]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 766.575708] env[61806]: ERROR nova.compute.manager raise self.value [ 766.575708] env[61806]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 766.575708] env[61806]: ERROR nova.compute.manager updated_port = self._update_port( [ 766.575708] env[61806]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 766.575708] env[61806]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 766.576248] env[61806]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 766.576248] env[61806]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 766.576248] env[61806]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 59bb93fe-b4ce-4348-bd86-42344407e007, please check neutron logs for more information. [ 766.576248] env[61806]: ERROR nova.compute.manager [ 766.576248] env[61806]: Traceback (most recent call last): [ 766.576248] env[61806]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 766.576248] env[61806]: listener.cb(fileno) [ 766.576248] env[61806]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 766.576248] env[61806]: result = function(*args, **kwargs) [ 766.576248] env[61806]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 766.576248] env[61806]: return func(*args, **kwargs) [ 766.576248] env[61806]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 766.576248] env[61806]: raise e [ 766.576248] env[61806]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 766.576248] env[61806]: nwinfo = self.network_api.allocate_for_instance( [ 766.576248] env[61806]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 766.576248] env[61806]: created_port_ids = self._update_ports_for_instance( [ 766.576248] env[61806]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 766.576248] env[61806]: with excutils.save_and_reraise_exception(): [ 766.576248] env[61806]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 766.576248] env[61806]: self.force_reraise() [ 766.576248] env[61806]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 766.576248] env[61806]: raise self.value [ 766.576248] env[61806]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 766.576248] env[61806]: updated_port = self._update_port( [ 766.576248] env[61806]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 766.576248] env[61806]: _ensure_no_port_binding_failure(port) [ 766.576248] env[61806]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 766.576248] env[61806]: raise exception.PortBindingFailed(port_id=port['id']) [ 766.577188] env[61806]: nova.exception.PortBindingFailed: Binding failed for port 59bb93fe-b4ce-4348-bd86-42344407e007, please check neutron logs for more information. [ 766.577188] env[61806]: Removing descriptor: 14 [ 766.696947] env[61806]: DEBUG nova.compute.manager [None req-35a8f95d-1ddd-492f-a93c-adb382770732 tempest-ServersV294TestFqdnHostnames-1709116086 tempest-ServersV294TestFqdnHostnames-1709116086-project-member] [instance: c25ddf7c-d0cc-4b73-96da-1dcd6012f072] Start spawning the instance on the hypervisor. {{(pid=61806) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 766.722919] env[61806]: DEBUG nova.virt.hardware [None req-35a8f95d-1ddd-492f-a93c-adb382770732 tempest-ServersV294TestFqdnHostnames-1709116086 tempest-ServersV294TestFqdnHostnames-1709116086-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-15T18:05:10Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-15T18:04:52Z,direct_url=,disk_format='vmdk',id=ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='84ba83fce2c349a6988173c5d6fc3b07',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-15T18:04:53Z,virtual_size=,visibility=), allow threads: False {{(pid=61806) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 766.723180] env[61806]: DEBUG nova.virt.hardware [None req-35a8f95d-1ddd-492f-a93c-adb382770732 tempest-ServersV294TestFqdnHostnames-1709116086 tempest-ServersV294TestFqdnHostnames-1709116086-project-member] Flavor limits 0:0:0 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 766.723329] env[61806]: DEBUG nova.virt.hardware [None req-35a8f95d-1ddd-492f-a93c-adb382770732 tempest-ServersV294TestFqdnHostnames-1709116086 tempest-ServersV294TestFqdnHostnames-1709116086-project-member] Image limits 0:0:0 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 766.723515] env[61806]: DEBUG nova.virt.hardware [None req-35a8f95d-1ddd-492f-a93c-adb382770732 tempest-ServersV294TestFqdnHostnames-1709116086 tempest-ServersV294TestFqdnHostnames-1709116086-project-member] Flavor pref 0:0:0 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 766.723662] env[61806]: DEBUG nova.virt.hardware [None req-35a8f95d-1ddd-492f-a93c-adb382770732 tempest-ServersV294TestFqdnHostnames-1709116086 tempest-ServersV294TestFqdnHostnames-1709116086-project-member] Image pref 0:0:0 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 766.723829] env[61806]: DEBUG nova.virt.hardware [None req-35a8f95d-1ddd-492f-a93c-adb382770732 tempest-ServersV294TestFqdnHostnames-1709116086 tempest-ServersV294TestFqdnHostnames-1709116086-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 766.724159] env[61806]: DEBUG nova.virt.hardware [None req-35a8f95d-1ddd-492f-a93c-adb382770732 tempest-ServersV294TestFqdnHostnames-1709116086 tempest-ServersV294TestFqdnHostnames-1709116086-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61806) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 766.724340] env[61806]: DEBUG nova.virt.hardware [None req-35a8f95d-1ddd-492f-a93c-adb382770732 tempest-ServersV294TestFqdnHostnames-1709116086 tempest-ServersV294TestFqdnHostnames-1709116086-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61806) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 766.724512] env[61806]: DEBUG nova.virt.hardware [None req-35a8f95d-1ddd-492f-a93c-adb382770732 tempest-ServersV294TestFqdnHostnames-1709116086 tempest-ServersV294TestFqdnHostnames-1709116086-project-member] Got 1 possible topologies {{(pid=61806) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 766.724674] env[61806]: DEBUG nova.virt.hardware [None req-35a8f95d-1ddd-492f-a93c-adb382770732 tempest-ServersV294TestFqdnHostnames-1709116086 tempest-ServersV294TestFqdnHostnames-1709116086-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61806) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 766.724849] env[61806]: DEBUG nova.virt.hardware [None req-35a8f95d-1ddd-492f-a93c-adb382770732 tempest-ServersV294TestFqdnHostnames-1709116086 tempest-ServersV294TestFqdnHostnames-1709116086-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61806) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 766.725539] env[61806]: DEBUG nova.compute.resource_tracker [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Instance 2a3f81e3-b84f-4370-a598-2f6a607c3d47 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61806) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 766.730262] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3e8ee394-d9ea-4dd2-9def-16fa051f2647 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 766.735572] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fceb7848-f8c4-4a32-95f7-f3ea264f6a1f {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 766.749763] env[61806]: ERROR nova.compute.manager [None req-35a8f95d-1ddd-492f-a93c-adb382770732 tempest-ServersV294TestFqdnHostnames-1709116086 tempest-ServersV294TestFqdnHostnames-1709116086-project-member] [instance: c25ddf7c-d0cc-4b73-96da-1dcd6012f072] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 59bb93fe-b4ce-4348-bd86-42344407e007, please check neutron logs for more information. [ 766.749763] env[61806]: ERROR nova.compute.manager [instance: c25ddf7c-d0cc-4b73-96da-1dcd6012f072] Traceback (most recent call last): [ 766.749763] env[61806]: ERROR nova.compute.manager [instance: c25ddf7c-d0cc-4b73-96da-1dcd6012f072] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 766.749763] env[61806]: ERROR nova.compute.manager [instance: c25ddf7c-d0cc-4b73-96da-1dcd6012f072] yield resources [ 766.749763] env[61806]: ERROR nova.compute.manager [instance: c25ddf7c-d0cc-4b73-96da-1dcd6012f072] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 766.749763] env[61806]: ERROR nova.compute.manager [instance: c25ddf7c-d0cc-4b73-96da-1dcd6012f072] self.driver.spawn(context, instance, image_meta, [ 766.749763] env[61806]: ERROR nova.compute.manager [instance: c25ddf7c-d0cc-4b73-96da-1dcd6012f072] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 766.749763] env[61806]: ERROR nova.compute.manager [instance: c25ddf7c-d0cc-4b73-96da-1dcd6012f072] self._vmops.spawn(context, instance, image_meta, injected_files, [ 766.749763] env[61806]: ERROR nova.compute.manager [instance: c25ddf7c-d0cc-4b73-96da-1dcd6012f072] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 766.749763] env[61806]: ERROR nova.compute.manager [instance: c25ddf7c-d0cc-4b73-96da-1dcd6012f072] vm_ref = self.build_virtual_machine(instance, [ 766.749763] env[61806]: ERROR nova.compute.manager [instance: c25ddf7c-d0cc-4b73-96da-1dcd6012f072] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 766.750457] env[61806]: ERROR nova.compute.manager [instance: c25ddf7c-d0cc-4b73-96da-1dcd6012f072] vif_infos = vmwarevif.get_vif_info(self._session, [ 766.750457] env[61806]: ERROR nova.compute.manager [instance: c25ddf7c-d0cc-4b73-96da-1dcd6012f072] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 766.750457] env[61806]: ERROR nova.compute.manager [instance: c25ddf7c-d0cc-4b73-96da-1dcd6012f072] for vif in network_info: [ 766.750457] env[61806]: ERROR nova.compute.manager [instance: c25ddf7c-d0cc-4b73-96da-1dcd6012f072] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 766.750457] env[61806]: ERROR nova.compute.manager [instance: c25ddf7c-d0cc-4b73-96da-1dcd6012f072] return self._sync_wrapper(fn, *args, **kwargs) [ 766.750457] env[61806]: ERROR nova.compute.manager [instance: c25ddf7c-d0cc-4b73-96da-1dcd6012f072] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 766.750457] env[61806]: ERROR nova.compute.manager [instance: c25ddf7c-d0cc-4b73-96da-1dcd6012f072] self.wait() [ 766.750457] env[61806]: ERROR nova.compute.manager [instance: c25ddf7c-d0cc-4b73-96da-1dcd6012f072] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 766.750457] env[61806]: ERROR nova.compute.manager [instance: c25ddf7c-d0cc-4b73-96da-1dcd6012f072] self[:] = self._gt.wait() [ 766.750457] env[61806]: ERROR nova.compute.manager [instance: c25ddf7c-d0cc-4b73-96da-1dcd6012f072] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 766.750457] env[61806]: ERROR nova.compute.manager [instance: c25ddf7c-d0cc-4b73-96da-1dcd6012f072] return self._exit_event.wait() [ 766.750457] env[61806]: ERROR nova.compute.manager [instance: c25ddf7c-d0cc-4b73-96da-1dcd6012f072] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 766.750457] env[61806]: ERROR nova.compute.manager [instance: c25ddf7c-d0cc-4b73-96da-1dcd6012f072] current.throw(*self._exc) [ 766.751090] env[61806]: ERROR nova.compute.manager [instance: c25ddf7c-d0cc-4b73-96da-1dcd6012f072] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 766.751090] env[61806]: ERROR nova.compute.manager [instance: c25ddf7c-d0cc-4b73-96da-1dcd6012f072] result = function(*args, **kwargs) [ 766.751090] env[61806]: ERROR nova.compute.manager [instance: c25ddf7c-d0cc-4b73-96da-1dcd6012f072] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 766.751090] env[61806]: ERROR nova.compute.manager [instance: c25ddf7c-d0cc-4b73-96da-1dcd6012f072] return func(*args, **kwargs) [ 766.751090] env[61806]: ERROR nova.compute.manager [instance: c25ddf7c-d0cc-4b73-96da-1dcd6012f072] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 766.751090] env[61806]: ERROR nova.compute.manager [instance: c25ddf7c-d0cc-4b73-96da-1dcd6012f072] raise e [ 766.751090] env[61806]: ERROR nova.compute.manager [instance: c25ddf7c-d0cc-4b73-96da-1dcd6012f072] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 766.751090] env[61806]: ERROR nova.compute.manager [instance: c25ddf7c-d0cc-4b73-96da-1dcd6012f072] nwinfo = self.network_api.allocate_for_instance( [ 766.751090] env[61806]: ERROR nova.compute.manager [instance: c25ddf7c-d0cc-4b73-96da-1dcd6012f072] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 766.751090] env[61806]: ERROR nova.compute.manager [instance: c25ddf7c-d0cc-4b73-96da-1dcd6012f072] created_port_ids = self._update_ports_for_instance( [ 766.751090] env[61806]: ERROR nova.compute.manager [instance: c25ddf7c-d0cc-4b73-96da-1dcd6012f072] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 766.751090] env[61806]: ERROR nova.compute.manager [instance: c25ddf7c-d0cc-4b73-96da-1dcd6012f072] with excutils.save_and_reraise_exception(): [ 766.751090] env[61806]: ERROR nova.compute.manager [instance: c25ddf7c-d0cc-4b73-96da-1dcd6012f072] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 766.751899] env[61806]: ERROR nova.compute.manager [instance: c25ddf7c-d0cc-4b73-96da-1dcd6012f072] self.force_reraise() [ 766.751899] env[61806]: ERROR nova.compute.manager [instance: c25ddf7c-d0cc-4b73-96da-1dcd6012f072] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 766.751899] env[61806]: ERROR nova.compute.manager [instance: c25ddf7c-d0cc-4b73-96da-1dcd6012f072] raise self.value [ 766.751899] env[61806]: ERROR nova.compute.manager [instance: c25ddf7c-d0cc-4b73-96da-1dcd6012f072] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 766.751899] env[61806]: ERROR nova.compute.manager [instance: c25ddf7c-d0cc-4b73-96da-1dcd6012f072] updated_port = self._update_port( [ 766.751899] env[61806]: ERROR nova.compute.manager [instance: c25ddf7c-d0cc-4b73-96da-1dcd6012f072] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 766.751899] env[61806]: ERROR nova.compute.manager [instance: c25ddf7c-d0cc-4b73-96da-1dcd6012f072] _ensure_no_port_binding_failure(port) [ 766.751899] env[61806]: ERROR nova.compute.manager [instance: c25ddf7c-d0cc-4b73-96da-1dcd6012f072] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 766.751899] env[61806]: ERROR nova.compute.manager [instance: c25ddf7c-d0cc-4b73-96da-1dcd6012f072] raise exception.PortBindingFailed(port_id=port['id']) [ 766.751899] env[61806]: ERROR nova.compute.manager [instance: c25ddf7c-d0cc-4b73-96da-1dcd6012f072] nova.exception.PortBindingFailed: Binding failed for port 59bb93fe-b4ce-4348-bd86-42344407e007, please check neutron logs for more information. [ 766.751899] env[61806]: ERROR nova.compute.manager [instance: c25ddf7c-d0cc-4b73-96da-1dcd6012f072] [ 766.751899] env[61806]: INFO nova.compute.manager [None req-35a8f95d-1ddd-492f-a93c-adb382770732 tempest-ServersV294TestFqdnHostnames-1709116086 tempest-ServersV294TestFqdnHostnames-1709116086-project-member] [instance: c25ddf7c-d0cc-4b73-96da-1dcd6012f072] Terminating instance [ 766.752493] env[61806]: DEBUG oslo_concurrency.lockutils [None req-35a8f95d-1ddd-492f-a93c-adb382770732 tempest-ServersV294TestFqdnHostnames-1709116086 tempest-ServersV294TestFqdnHostnames-1709116086-project-member] Acquiring lock "refresh_cache-c25ddf7c-d0cc-4b73-96da-1dcd6012f072" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 766.853854] env[61806]: DEBUG nova.network.neutron [req-b29f7869-9fa5-4a8c-8fcf-36d831494d2d req-d884e4eb-f11b-4834-aea3-93681f4b19a8 service nova] [instance: c25ddf7c-d0cc-4b73-96da-1dcd6012f072] Instance cache missing network info. {{(pid=61806) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 766.933455] env[61806]: DEBUG nova.network.neutron [req-b29f7869-9fa5-4a8c-8fcf-36d831494d2d req-d884e4eb-f11b-4834-aea3-93681f4b19a8 service nova] [instance: c25ddf7c-d0cc-4b73-96da-1dcd6012f072] Updating instance_info_cache with network_info: [] {{(pid=61806) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 767.010165] env[61806]: DEBUG oslo_concurrency.lockutils [None req-3d5b4b69-10ec-4d1a-aa3d-4dddf4155c99 tempest-ServerShowV247Test-1218443059 tempest-ServerShowV247Test-1218443059-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 767.231028] env[61806]: DEBUG nova.compute.resource_tracker [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Instance 4184e263-6f56-4bc1-99b2-a2c460531516 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61806) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 767.436240] env[61806]: DEBUG oslo_concurrency.lockutils [req-b29f7869-9fa5-4a8c-8fcf-36d831494d2d req-d884e4eb-f11b-4834-aea3-93681f4b19a8 service nova] Releasing lock "refresh_cache-c25ddf7c-d0cc-4b73-96da-1dcd6012f072" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 767.436702] env[61806]: DEBUG oslo_concurrency.lockutils [None req-35a8f95d-1ddd-492f-a93c-adb382770732 tempest-ServersV294TestFqdnHostnames-1709116086 tempest-ServersV294TestFqdnHostnames-1709116086-project-member] Acquired lock "refresh_cache-c25ddf7c-d0cc-4b73-96da-1dcd6012f072" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 767.436888] env[61806]: DEBUG nova.network.neutron [None req-35a8f95d-1ddd-492f-a93c-adb382770732 tempest-ServersV294TestFqdnHostnames-1709116086 tempest-ServersV294TestFqdnHostnames-1709116086-project-member] [instance: c25ddf7c-d0cc-4b73-96da-1dcd6012f072] Building network info cache for instance {{(pid=61806) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 767.733872] env[61806]: DEBUG nova.compute.resource_tracker [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Instance 7c3fd2ff-ebd5-454f-a743-7fbae1088941 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61806) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 767.969212] env[61806]: DEBUG nova.network.neutron [None req-35a8f95d-1ddd-492f-a93c-adb382770732 tempest-ServersV294TestFqdnHostnames-1709116086 tempest-ServersV294TestFqdnHostnames-1709116086-project-member] [instance: c25ddf7c-d0cc-4b73-96da-1dcd6012f072] Instance cache missing network info. {{(pid=61806) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 768.091174] env[61806]: DEBUG nova.network.neutron [None req-35a8f95d-1ddd-492f-a93c-adb382770732 tempest-ServersV294TestFqdnHostnames-1709116086 tempest-ServersV294TestFqdnHostnames-1709116086-project-member] [instance: c25ddf7c-d0cc-4b73-96da-1dcd6012f072] Updating instance_info_cache with network_info: [] {{(pid=61806) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 768.238905] env[61806]: DEBUG nova.compute.resource_tracker [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Instance 4febd093-9f2b-494e-b175-e4693b0e3e0d has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61806) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 768.431804] env[61806]: DEBUG nova.compute.manager [req-415d0b10-3a70-43ef-9fcd-a4c54a2cfeaa req-862761b0-6d7a-4690-8aae-1fb29013a094 service nova] [instance: c25ddf7c-d0cc-4b73-96da-1dcd6012f072] Received event network-vif-deleted-59bb93fe-b4ce-4348-bd86-42344407e007 {{(pid=61806) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 768.594478] env[61806]: DEBUG oslo_concurrency.lockutils [None req-35a8f95d-1ddd-492f-a93c-adb382770732 tempest-ServersV294TestFqdnHostnames-1709116086 tempest-ServersV294TestFqdnHostnames-1709116086-project-member] Releasing lock "refresh_cache-c25ddf7c-d0cc-4b73-96da-1dcd6012f072" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 768.594922] env[61806]: DEBUG nova.compute.manager [None req-35a8f95d-1ddd-492f-a93c-adb382770732 tempest-ServersV294TestFqdnHostnames-1709116086 tempest-ServersV294TestFqdnHostnames-1709116086-project-member] [instance: c25ddf7c-d0cc-4b73-96da-1dcd6012f072] Start destroying the instance on the hypervisor. {{(pid=61806) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 768.595140] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-35a8f95d-1ddd-492f-a93c-adb382770732 tempest-ServersV294TestFqdnHostnames-1709116086 tempest-ServersV294TestFqdnHostnames-1709116086-project-member] [instance: c25ddf7c-d0cc-4b73-96da-1dcd6012f072] Destroying instance {{(pid=61806) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 768.595447] env[61806]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-91e233dc-71f3-4ea0-baa3-fc27b1ada9e9 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 768.604723] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-33e4af8f-cc54-4bc0-9aa1-aef943c1d766 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 768.626508] env[61806]: WARNING nova.virt.vmwareapi.vmops [None req-35a8f95d-1ddd-492f-a93c-adb382770732 tempest-ServersV294TestFqdnHostnames-1709116086 tempest-ServersV294TestFqdnHostnames-1709116086-project-member] [instance: c25ddf7c-d0cc-4b73-96da-1dcd6012f072] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance c25ddf7c-d0cc-4b73-96da-1dcd6012f072 could not be found. [ 768.626508] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-35a8f95d-1ddd-492f-a93c-adb382770732 tempest-ServersV294TestFqdnHostnames-1709116086 tempest-ServersV294TestFqdnHostnames-1709116086-project-member] [instance: c25ddf7c-d0cc-4b73-96da-1dcd6012f072] Instance destroyed {{(pid=61806) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 768.626665] env[61806]: INFO nova.compute.manager [None req-35a8f95d-1ddd-492f-a93c-adb382770732 tempest-ServersV294TestFqdnHostnames-1709116086 tempest-ServersV294TestFqdnHostnames-1709116086-project-member] [instance: c25ddf7c-d0cc-4b73-96da-1dcd6012f072] Took 0.03 seconds to destroy the instance on the hypervisor. [ 768.626828] env[61806]: DEBUG oslo.service.loopingcall [None req-35a8f95d-1ddd-492f-a93c-adb382770732 tempest-ServersV294TestFqdnHostnames-1709116086 tempest-ServersV294TestFqdnHostnames-1709116086-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61806) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 768.627062] env[61806]: DEBUG nova.compute.manager [-] [instance: c25ddf7c-d0cc-4b73-96da-1dcd6012f072] Deallocating network for instance {{(pid=61806) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 768.627161] env[61806]: DEBUG nova.network.neutron [-] [instance: c25ddf7c-d0cc-4b73-96da-1dcd6012f072] deallocate_for_instance() {{(pid=61806) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 768.642155] env[61806]: DEBUG nova.network.neutron [-] [instance: c25ddf7c-d0cc-4b73-96da-1dcd6012f072] Instance cache missing network info. {{(pid=61806) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 768.744981] env[61806]: DEBUG nova.compute.resource_tracker [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Instance a934d02d-26aa-4900-b473-a58489e5629e has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61806) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 768.808492] env[61806]: DEBUG oslo_concurrency.lockutils [None req-0ea3a38f-3871-4412-a6bc-d21c2086d262 tempest-AttachInterfacesV270Test-1974652806 tempest-AttachInterfacesV270Test-1974652806-project-member] Acquiring lock "117139cd-4ed4-4bea-8ad3-f4ac32c692e4" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 768.808690] env[61806]: DEBUG oslo_concurrency.lockutils [None req-0ea3a38f-3871-4412-a6bc-d21c2086d262 tempest-AttachInterfacesV270Test-1974652806 tempest-AttachInterfacesV270Test-1974652806-project-member] Lock "117139cd-4ed4-4bea-8ad3-f4ac32c692e4" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 769.144616] env[61806]: DEBUG nova.network.neutron [-] [instance: c25ddf7c-d0cc-4b73-96da-1dcd6012f072] Updating instance_info_cache with network_info: [] {{(pid=61806) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 769.248351] env[61806]: DEBUG nova.compute.resource_tracker [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Instance 616ec206-9804-469e-ab5c-41aea7f048aa has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61806) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 769.647745] env[61806]: INFO nova.compute.manager [-] [instance: c25ddf7c-d0cc-4b73-96da-1dcd6012f072] Took 1.02 seconds to deallocate network for instance. [ 769.650826] env[61806]: DEBUG nova.compute.claims [None req-35a8f95d-1ddd-492f-a93c-adb382770732 tempest-ServersV294TestFqdnHostnames-1709116086 tempest-ServersV294TestFqdnHostnames-1709116086-project-member] [instance: c25ddf7c-d0cc-4b73-96da-1dcd6012f072] Aborting claim: {{(pid=61806) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 769.651088] env[61806]: DEBUG oslo_concurrency.lockutils [None req-35a8f95d-1ddd-492f-a93c-adb382770732 tempest-ServersV294TestFqdnHostnames-1709116086 tempest-ServersV294TestFqdnHostnames-1709116086-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 769.751477] env[61806]: DEBUG nova.compute.resource_tracker [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Instance 0758901a-7093-41d3-b0e2-5c519333abdd has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61806) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 770.254548] env[61806]: DEBUG nova.compute.resource_tracker [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Instance 4373b735-31cf-4b53-b655-38555cf212a5 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61806) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 770.757231] env[61806]: DEBUG nova.compute.resource_tracker [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Instance a842df44-d8a9-4376-b9fc-5ca19a68a4b7 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61806) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 771.259817] env[61806]: DEBUG nova.compute.resource_tracker [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Instance 43dcce0d-4fdb-4ee1-8309-92e92d2331a9 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61806) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 771.763017] env[61806]: DEBUG nova.compute.resource_tracker [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Instance b41531f2-f28c-4d82-9682-0b557bbaa491 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61806) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 772.265730] env[61806]: DEBUG nova.compute.resource_tracker [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Instance 67d1bdd8-cf8f-4fb5-bd3c-42008e4d60e3 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61806) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 772.769736] env[61806]: DEBUG nova.compute.resource_tracker [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Instance 25218e83-c1ab-4b97-ade1-7c32b6f99305 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61806) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 773.272708] env[61806]: DEBUG nova.compute.resource_tracker [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Instance ef502a7d-7dc0-4319-8ce3-6a8701288237 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61806) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 773.272983] env[61806]: DEBUG nova.compute.resource_tracker [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Total usable vcpus: 48, total allocated vcpus: 6 {{(pid=61806) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 773.273152] env[61806]: DEBUG nova.compute.resource_tracker [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=1664MB phys_disk=200GB used_disk=6GB total_vcpus=48 used_vcpus=6 pci_stats=[] {{(pid=61806) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 773.518568] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c33bf4a8-597a-42ce-bcd8-186416b5d27d {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 773.526024] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bc6a819b-e335-446d-b786-968a523ed03c {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 773.554882] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-58fce5b6-a3d9-41e0-8189-f6fa424d0af1 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 773.562153] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c71a7658-b825-4499-9427-2bea9c297348 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 773.575031] env[61806]: DEBUG nova.compute.provider_tree [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Inventory has not changed in ProviderTree for provider: a2858be1-fd22-4e08-979e-87ad25293407 {{(pid=61806) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 774.078818] env[61806]: DEBUG nova.scheduler.client.report [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Inventory has not changed for provider a2858be1-fd22-4e08-979e-87ad25293407 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 139, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61806) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 774.583406] env[61806]: DEBUG nova.compute.resource_tracker [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=61806) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 774.583659] env[61806]: DEBUG oslo_concurrency.lockutils [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 9.912s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 774.583949] env[61806]: DEBUG oslo_concurrency.lockutils [None req-3bf756e0-c56e-4902-9c2b-b224a1d09a7c tempest-ListServersNegativeTestJSON-272187588 tempest-ListServersNegativeTestJSON-272187588-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 23.873s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 774.587325] env[61806]: DEBUG oslo_service.periodic_task [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Running periodic task ComputeManager._run_pending_deletes {{(pid=61806) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 774.587540] env[61806]: DEBUG nova.compute.manager [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Cleaning up deleted instances {{(pid=61806) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11226}} [ 775.094050] env[61806]: DEBUG nova.compute.manager [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] There are 2 instances to clean {{(pid=61806) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11235}} [ 775.094050] env[61806]: DEBUG nova.compute.manager [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] [instance: 9db80c3c-e2c3-4034-a79a-d7be924c4ff6] Instance has had 0 of 5 cleanup attempts {{(pid=61806) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 775.325915] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3f7c6af1-ba98-49a1-821c-fc670bc35561 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 775.334015] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a963fb8f-1f35-4d76-ac71-233a2a261139 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 775.363946] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6ce28a4d-3556-4590-b552-c83a8fa3a46e {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 775.371632] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f865d790-9298-4e5e-89e4-4d2e56063b4f {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 775.385460] env[61806]: DEBUG nova.compute.provider_tree [None req-3bf756e0-c56e-4902-9c2b-b224a1d09a7c tempest-ListServersNegativeTestJSON-272187588 tempest-ListServersNegativeTestJSON-272187588-project-member] Inventory has not changed in ProviderTree for provider: a2858be1-fd22-4e08-979e-87ad25293407 {{(pid=61806) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 775.596884] env[61806]: DEBUG nova.compute.manager [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] [instance: 8016bafc-b287-4df1-bd4b-a00210177a86] Instance has had 0 of 5 cleanup attempts {{(pid=61806) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 775.891550] env[61806]: DEBUG nova.scheduler.client.report [None req-3bf756e0-c56e-4902-9c2b-b224a1d09a7c tempest-ListServersNegativeTestJSON-272187588 tempest-ListServersNegativeTestJSON-272187588-project-member] Inventory has not changed for provider a2858be1-fd22-4e08-979e-87ad25293407 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 139, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61806) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 776.100401] env[61806]: DEBUG oslo_service.periodic_task [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Running periodic task ComputeManager._cleanup_incomplete_migrations {{(pid=61806) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 776.100691] env[61806]: DEBUG nova.compute.manager [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Cleaning up deleted instances with incomplete migration {{(pid=61806) _cleanup_incomplete_migrations /opt/stack/nova/nova/compute/manager.py:11264}} [ 776.396188] env[61806]: DEBUG oslo_concurrency.lockutils [None req-3bf756e0-c56e-4902-9c2b-b224a1d09a7c tempest-ListServersNegativeTestJSON-272187588 tempest-ListServersNegativeTestJSON-272187588-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.812s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 776.396882] env[61806]: ERROR nova.compute.manager [None req-3bf756e0-c56e-4902-9c2b-b224a1d09a7c tempest-ListServersNegativeTestJSON-272187588 tempest-ListServersNegativeTestJSON-272187588-project-member] [instance: ea8d39a1-066d-4eb4-888d-776b0e45f684] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port d1f266fb-3da7-4f92-b3cb-81aa9f3b924c, please check neutron logs for more information. [ 776.396882] env[61806]: ERROR nova.compute.manager [instance: ea8d39a1-066d-4eb4-888d-776b0e45f684] Traceback (most recent call last): [ 776.396882] env[61806]: ERROR nova.compute.manager [instance: ea8d39a1-066d-4eb4-888d-776b0e45f684] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 776.396882] env[61806]: ERROR nova.compute.manager [instance: ea8d39a1-066d-4eb4-888d-776b0e45f684] self.driver.spawn(context, instance, image_meta, [ 776.396882] env[61806]: ERROR nova.compute.manager [instance: ea8d39a1-066d-4eb4-888d-776b0e45f684] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 776.396882] env[61806]: ERROR nova.compute.manager [instance: ea8d39a1-066d-4eb4-888d-776b0e45f684] self._vmops.spawn(context, instance, image_meta, injected_files, [ 776.396882] env[61806]: ERROR nova.compute.manager [instance: ea8d39a1-066d-4eb4-888d-776b0e45f684] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 776.396882] env[61806]: ERROR nova.compute.manager [instance: ea8d39a1-066d-4eb4-888d-776b0e45f684] vm_ref = self.build_virtual_machine(instance, [ 776.396882] env[61806]: ERROR nova.compute.manager [instance: ea8d39a1-066d-4eb4-888d-776b0e45f684] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 776.396882] env[61806]: ERROR nova.compute.manager [instance: ea8d39a1-066d-4eb4-888d-776b0e45f684] vif_infos = vmwarevif.get_vif_info(self._session, [ 776.396882] env[61806]: ERROR nova.compute.manager [instance: ea8d39a1-066d-4eb4-888d-776b0e45f684] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 776.397425] env[61806]: ERROR nova.compute.manager [instance: ea8d39a1-066d-4eb4-888d-776b0e45f684] for vif in network_info: [ 776.397425] env[61806]: ERROR nova.compute.manager [instance: ea8d39a1-066d-4eb4-888d-776b0e45f684] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 776.397425] env[61806]: ERROR nova.compute.manager [instance: ea8d39a1-066d-4eb4-888d-776b0e45f684] return self._sync_wrapper(fn, *args, **kwargs) [ 776.397425] env[61806]: ERROR nova.compute.manager [instance: ea8d39a1-066d-4eb4-888d-776b0e45f684] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 776.397425] env[61806]: ERROR nova.compute.manager [instance: ea8d39a1-066d-4eb4-888d-776b0e45f684] self.wait() [ 776.397425] env[61806]: ERROR nova.compute.manager [instance: ea8d39a1-066d-4eb4-888d-776b0e45f684] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 776.397425] env[61806]: ERROR nova.compute.manager [instance: ea8d39a1-066d-4eb4-888d-776b0e45f684] self[:] = self._gt.wait() [ 776.397425] env[61806]: ERROR nova.compute.manager [instance: ea8d39a1-066d-4eb4-888d-776b0e45f684] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 776.397425] env[61806]: ERROR nova.compute.manager [instance: ea8d39a1-066d-4eb4-888d-776b0e45f684] return self._exit_event.wait() [ 776.397425] env[61806]: ERROR nova.compute.manager [instance: ea8d39a1-066d-4eb4-888d-776b0e45f684] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 776.397425] env[61806]: ERROR nova.compute.manager [instance: ea8d39a1-066d-4eb4-888d-776b0e45f684] current.throw(*self._exc) [ 776.397425] env[61806]: ERROR nova.compute.manager [instance: ea8d39a1-066d-4eb4-888d-776b0e45f684] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 776.397425] env[61806]: ERROR nova.compute.manager [instance: ea8d39a1-066d-4eb4-888d-776b0e45f684] result = function(*args, **kwargs) [ 776.397885] env[61806]: ERROR nova.compute.manager [instance: ea8d39a1-066d-4eb4-888d-776b0e45f684] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 776.397885] env[61806]: ERROR nova.compute.manager [instance: ea8d39a1-066d-4eb4-888d-776b0e45f684] return func(*args, **kwargs) [ 776.397885] env[61806]: ERROR nova.compute.manager [instance: ea8d39a1-066d-4eb4-888d-776b0e45f684] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 776.397885] env[61806]: ERROR nova.compute.manager [instance: ea8d39a1-066d-4eb4-888d-776b0e45f684] raise e [ 776.397885] env[61806]: ERROR nova.compute.manager [instance: ea8d39a1-066d-4eb4-888d-776b0e45f684] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 776.397885] env[61806]: ERROR nova.compute.manager [instance: ea8d39a1-066d-4eb4-888d-776b0e45f684] nwinfo = self.network_api.allocate_for_instance( [ 776.397885] env[61806]: ERROR nova.compute.manager [instance: ea8d39a1-066d-4eb4-888d-776b0e45f684] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 776.397885] env[61806]: ERROR nova.compute.manager [instance: ea8d39a1-066d-4eb4-888d-776b0e45f684] created_port_ids = self._update_ports_for_instance( [ 776.397885] env[61806]: ERROR nova.compute.manager [instance: ea8d39a1-066d-4eb4-888d-776b0e45f684] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 776.397885] env[61806]: ERROR nova.compute.manager [instance: ea8d39a1-066d-4eb4-888d-776b0e45f684] with excutils.save_and_reraise_exception(): [ 776.397885] env[61806]: ERROR nova.compute.manager [instance: ea8d39a1-066d-4eb4-888d-776b0e45f684] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 776.397885] env[61806]: ERROR nova.compute.manager [instance: ea8d39a1-066d-4eb4-888d-776b0e45f684] self.force_reraise() [ 776.397885] env[61806]: ERROR nova.compute.manager [instance: ea8d39a1-066d-4eb4-888d-776b0e45f684] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 776.398322] env[61806]: ERROR nova.compute.manager [instance: ea8d39a1-066d-4eb4-888d-776b0e45f684] raise self.value [ 776.398322] env[61806]: ERROR nova.compute.manager [instance: ea8d39a1-066d-4eb4-888d-776b0e45f684] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 776.398322] env[61806]: ERROR nova.compute.manager [instance: ea8d39a1-066d-4eb4-888d-776b0e45f684] updated_port = self._update_port( [ 776.398322] env[61806]: ERROR nova.compute.manager [instance: ea8d39a1-066d-4eb4-888d-776b0e45f684] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 776.398322] env[61806]: ERROR nova.compute.manager [instance: ea8d39a1-066d-4eb4-888d-776b0e45f684] _ensure_no_port_binding_failure(port) [ 776.398322] env[61806]: ERROR nova.compute.manager [instance: ea8d39a1-066d-4eb4-888d-776b0e45f684] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 776.398322] env[61806]: ERROR nova.compute.manager [instance: ea8d39a1-066d-4eb4-888d-776b0e45f684] raise exception.PortBindingFailed(port_id=port['id']) [ 776.398322] env[61806]: ERROR nova.compute.manager [instance: ea8d39a1-066d-4eb4-888d-776b0e45f684] nova.exception.PortBindingFailed: Binding failed for port d1f266fb-3da7-4f92-b3cb-81aa9f3b924c, please check neutron logs for more information. [ 776.398322] env[61806]: ERROR nova.compute.manager [instance: ea8d39a1-066d-4eb4-888d-776b0e45f684] [ 776.398322] env[61806]: DEBUG nova.compute.utils [None req-3bf756e0-c56e-4902-9c2b-b224a1d09a7c tempest-ListServersNegativeTestJSON-272187588 tempest-ListServersNegativeTestJSON-272187588-project-member] [instance: ea8d39a1-066d-4eb4-888d-776b0e45f684] Binding failed for port d1f266fb-3da7-4f92-b3cb-81aa9f3b924c, please check neutron logs for more information. {{(pid=61806) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 776.398862] env[61806]: DEBUG oslo_concurrency.lockutils [None req-3bf756e0-c56e-4902-9c2b-b224a1d09a7c tempest-ListServersNegativeTestJSON-272187588 tempest-ListServersNegativeTestJSON-272187588-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 23.970s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 776.401772] env[61806]: DEBUG nova.compute.manager [None req-3bf756e0-c56e-4902-9c2b-b224a1d09a7c tempest-ListServersNegativeTestJSON-272187588 tempest-ListServersNegativeTestJSON-272187588-project-member] [instance: ea8d39a1-066d-4eb4-888d-776b0e45f684] Build of instance ea8d39a1-066d-4eb4-888d-776b0e45f684 was re-scheduled: Binding failed for port d1f266fb-3da7-4f92-b3cb-81aa9f3b924c, please check neutron logs for more information. {{(pid=61806) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 776.402260] env[61806]: DEBUG nova.compute.manager [None req-3bf756e0-c56e-4902-9c2b-b224a1d09a7c tempest-ListServersNegativeTestJSON-272187588 tempest-ListServersNegativeTestJSON-272187588-project-member] [instance: ea8d39a1-066d-4eb4-888d-776b0e45f684] Unplugging VIFs for instance {{(pid=61806) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 776.402489] env[61806]: DEBUG oslo_concurrency.lockutils [None req-3bf756e0-c56e-4902-9c2b-b224a1d09a7c tempest-ListServersNegativeTestJSON-272187588 tempest-ListServersNegativeTestJSON-272187588-project-member] Acquiring lock "refresh_cache-ea8d39a1-066d-4eb4-888d-776b0e45f684" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 776.402636] env[61806]: DEBUG oslo_concurrency.lockutils [None req-3bf756e0-c56e-4902-9c2b-b224a1d09a7c tempest-ListServersNegativeTestJSON-272187588 tempest-ListServersNegativeTestJSON-272187588-project-member] Acquired lock "refresh_cache-ea8d39a1-066d-4eb4-888d-776b0e45f684" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 776.402797] env[61806]: DEBUG nova.network.neutron [None req-3bf756e0-c56e-4902-9c2b-b224a1d09a7c tempest-ListServersNegativeTestJSON-272187588 tempest-ListServersNegativeTestJSON-272187588-project-member] [instance: ea8d39a1-066d-4eb4-888d-776b0e45f684] Building network info cache for instance {{(pid=61806) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 776.603971] env[61806]: DEBUG oslo_service.periodic_task [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Running periodic task ComputeManager._cleanup_expired_console_auth_tokens {{(pid=61806) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 776.926793] env[61806]: DEBUG nova.network.neutron [None req-3bf756e0-c56e-4902-9c2b-b224a1d09a7c tempest-ListServersNegativeTestJSON-272187588 tempest-ListServersNegativeTestJSON-272187588-project-member] [instance: ea8d39a1-066d-4eb4-888d-776b0e45f684] Instance cache missing network info. {{(pid=61806) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 777.020905] env[61806]: DEBUG nova.network.neutron [None req-3bf756e0-c56e-4902-9c2b-b224a1d09a7c tempest-ListServersNegativeTestJSON-272187588 tempest-ListServersNegativeTestJSON-272187588-project-member] [instance: ea8d39a1-066d-4eb4-888d-776b0e45f684] Updating instance_info_cache with network_info: [] {{(pid=61806) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 777.146869] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c7d015b6-4df2-4426-a853-f7f006506645 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 777.154711] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a9ffc1d4-7eea-46bd-b149-c4b6a75b8100 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 777.184092] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-684ea430-5c67-4f92-a49d-89eb80ab2ae1 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 777.191681] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-412e969e-0e42-42a2-b887-c7b6e1823d51 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 777.204541] env[61806]: DEBUG nova.compute.provider_tree [None req-3bf756e0-c56e-4902-9c2b-b224a1d09a7c tempest-ListServersNegativeTestJSON-272187588 tempest-ListServersNegativeTestJSON-272187588-project-member] Inventory has not changed in ProviderTree for provider: a2858be1-fd22-4e08-979e-87ad25293407 {{(pid=61806) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 777.523447] env[61806]: DEBUG oslo_concurrency.lockutils [None req-3bf756e0-c56e-4902-9c2b-b224a1d09a7c tempest-ListServersNegativeTestJSON-272187588 tempest-ListServersNegativeTestJSON-272187588-project-member] Releasing lock "refresh_cache-ea8d39a1-066d-4eb4-888d-776b0e45f684" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 777.523694] env[61806]: DEBUG nova.compute.manager [None req-3bf756e0-c56e-4902-9c2b-b224a1d09a7c tempest-ListServersNegativeTestJSON-272187588 tempest-ListServersNegativeTestJSON-272187588-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61806) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 777.523879] env[61806]: DEBUG nova.compute.manager [None req-3bf756e0-c56e-4902-9c2b-b224a1d09a7c tempest-ListServersNegativeTestJSON-272187588 tempest-ListServersNegativeTestJSON-272187588-project-member] [instance: ea8d39a1-066d-4eb4-888d-776b0e45f684] Deallocating network for instance {{(pid=61806) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 777.524063] env[61806]: DEBUG nova.network.neutron [None req-3bf756e0-c56e-4902-9c2b-b224a1d09a7c tempest-ListServersNegativeTestJSON-272187588 tempest-ListServersNegativeTestJSON-272187588-project-member] [instance: ea8d39a1-066d-4eb4-888d-776b0e45f684] deallocate_for_instance() {{(pid=61806) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 777.538752] env[61806]: DEBUG nova.network.neutron [None req-3bf756e0-c56e-4902-9c2b-b224a1d09a7c tempest-ListServersNegativeTestJSON-272187588 tempest-ListServersNegativeTestJSON-272187588-project-member] [instance: ea8d39a1-066d-4eb4-888d-776b0e45f684] Instance cache missing network info. {{(pid=61806) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 777.707446] env[61806]: DEBUG nova.scheduler.client.report [None req-3bf756e0-c56e-4902-9c2b-b224a1d09a7c tempest-ListServersNegativeTestJSON-272187588 tempest-ListServersNegativeTestJSON-272187588-project-member] Inventory has not changed for provider a2858be1-fd22-4e08-979e-87ad25293407 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 139, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61806) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 778.041491] env[61806]: DEBUG nova.network.neutron [None req-3bf756e0-c56e-4902-9c2b-b224a1d09a7c tempest-ListServersNegativeTestJSON-272187588 tempest-ListServersNegativeTestJSON-272187588-project-member] [instance: ea8d39a1-066d-4eb4-888d-776b0e45f684] Updating instance_info_cache with network_info: [] {{(pid=61806) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 778.213033] env[61806]: DEBUG oslo_concurrency.lockutils [None req-3bf756e0-c56e-4902-9c2b-b224a1d09a7c tempest-ListServersNegativeTestJSON-272187588 tempest-ListServersNegativeTestJSON-272187588-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.814s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 778.213611] env[61806]: ERROR nova.compute.manager [None req-3bf756e0-c56e-4902-9c2b-b224a1d09a7c tempest-ListServersNegativeTestJSON-272187588 tempest-ListServersNegativeTestJSON-272187588-project-member] [instance: 198dc1a4-050d-47b2-8bc4-30af5534a789] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 1a279290-dcd8-45fb-a136-2fcbc132e8d7, please check neutron logs for more information. [ 778.213611] env[61806]: ERROR nova.compute.manager [instance: 198dc1a4-050d-47b2-8bc4-30af5534a789] Traceback (most recent call last): [ 778.213611] env[61806]: ERROR nova.compute.manager [instance: 198dc1a4-050d-47b2-8bc4-30af5534a789] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 778.213611] env[61806]: ERROR nova.compute.manager [instance: 198dc1a4-050d-47b2-8bc4-30af5534a789] self.driver.spawn(context, instance, image_meta, [ 778.213611] env[61806]: ERROR nova.compute.manager [instance: 198dc1a4-050d-47b2-8bc4-30af5534a789] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 778.213611] env[61806]: ERROR nova.compute.manager [instance: 198dc1a4-050d-47b2-8bc4-30af5534a789] self._vmops.spawn(context, instance, image_meta, injected_files, [ 778.213611] env[61806]: ERROR nova.compute.manager [instance: 198dc1a4-050d-47b2-8bc4-30af5534a789] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 778.213611] env[61806]: ERROR nova.compute.manager [instance: 198dc1a4-050d-47b2-8bc4-30af5534a789] vm_ref = self.build_virtual_machine(instance, [ 778.213611] env[61806]: ERROR nova.compute.manager [instance: 198dc1a4-050d-47b2-8bc4-30af5534a789] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 778.213611] env[61806]: ERROR nova.compute.manager [instance: 198dc1a4-050d-47b2-8bc4-30af5534a789] vif_infos = vmwarevif.get_vif_info(self._session, [ 778.213611] env[61806]: ERROR nova.compute.manager [instance: 198dc1a4-050d-47b2-8bc4-30af5534a789] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 778.213989] env[61806]: ERROR nova.compute.manager [instance: 198dc1a4-050d-47b2-8bc4-30af5534a789] for vif in network_info: [ 778.213989] env[61806]: ERROR nova.compute.manager [instance: 198dc1a4-050d-47b2-8bc4-30af5534a789] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 778.213989] env[61806]: ERROR nova.compute.manager [instance: 198dc1a4-050d-47b2-8bc4-30af5534a789] return self._sync_wrapper(fn, *args, **kwargs) [ 778.213989] env[61806]: ERROR nova.compute.manager [instance: 198dc1a4-050d-47b2-8bc4-30af5534a789] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 778.213989] env[61806]: ERROR nova.compute.manager [instance: 198dc1a4-050d-47b2-8bc4-30af5534a789] self.wait() [ 778.213989] env[61806]: ERROR nova.compute.manager [instance: 198dc1a4-050d-47b2-8bc4-30af5534a789] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 778.213989] env[61806]: ERROR nova.compute.manager [instance: 198dc1a4-050d-47b2-8bc4-30af5534a789] self[:] = self._gt.wait() [ 778.213989] env[61806]: ERROR nova.compute.manager [instance: 198dc1a4-050d-47b2-8bc4-30af5534a789] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 778.213989] env[61806]: ERROR nova.compute.manager [instance: 198dc1a4-050d-47b2-8bc4-30af5534a789] return self._exit_event.wait() [ 778.213989] env[61806]: ERROR nova.compute.manager [instance: 198dc1a4-050d-47b2-8bc4-30af5534a789] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 778.213989] env[61806]: ERROR nova.compute.manager [instance: 198dc1a4-050d-47b2-8bc4-30af5534a789] current.throw(*self._exc) [ 778.213989] env[61806]: ERROR nova.compute.manager [instance: 198dc1a4-050d-47b2-8bc4-30af5534a789] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 778.213989] env[61806]: ERROR nova.compute.manager [instance: 198dc1a4-050d-47b2-8bc4-30af5534a789] result = function(*args, **kwargs) [ 778.214383] env[61806]: ERROR nova.compute.manager [instance: 198dc1a4-050d-47b2-8bc4-30af5534a789] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 778.214383] env[61806]: ERROR nova.compute.manager [instance: 198dc1a4-050d-47b2-8bc4-30af5534a789] return func(*args, **kwargs) [ 778.214383] env[61806]: ERROR nova.compute.manager [instance: 198dc1a4-050d-47b2-8bc4-30af5534a789] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 778.214383] env[61806]: ERROR nova.compute.manager [instance: 198dc1a4-050d-47b2-8bc4-30af5534a789] raise e [ 778.214383] env[61806]: ERROR nova.compute.manager [instance: 198dc1a4-050d-47b2-8bc4-30af5534a789] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 778.214383] env[61806]: ERROR nova.compute.manager [instance: 198dc1a4-050d-47b2-8bc4-30af5534a789] nwinfo = self.network_api.allocate_for_instance( [ 778.214383] env[61806]: ERROR nova.compute.manager [instance: 198dc1a4-050d-47b2-8bc4-30af5534a789] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 778.214383] env[61806]: ERROR nova.compute.manager [instance: 198dc1a4-050d-47b2-8bc4-30af5534a789] created_port_ids = self._update_ports_for_instance( [ 778.214383] env[61806]: ERROR nova.compute.manager [instance: 198dc1a4-050d-47b2-8bc4-30af5534a789] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 778.214383] env[61806]: ERROR nova.compute.manager [instance: 198dc1a4-050d-47b2-8bc4-30af5534a789] with excutils.save_and_reraise_exception(): [ 778.214383] env[61806]: ERROR nova.compute.manager [instance: 198dc1a4-050d-47b2-8bc4-30af5534a789] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 778.214383] env[61806]: ERROR nova.compute.manager [instance: 198dc1a4-050d-47b2-8bc4-30af5534a789] self.force_reraise() [ 778.214383] env[61806]: ERROR nova.compute.manager [instance: 198dc1a4-050d-47b2-8bc4-30af5534a789] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 778.214765] env[61806]: ERROR nova.compute.manager [instance: 198dc1a4-050d-47b2-8bc4-30af5534a789] raise self.value [ 778.214765] env[61806]: ERROR nova.compute.manager [instance: 198dc1a4-050d-47b2-8bc4-30af5534a789] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 778.214765] env[61806]: ERROR nova.compute.manager [instance: 198dc1a4-050d-47b2-8bc4-30af5534a789] updated_port = self._update_port( [ 778.214765] env[61806]: ERROR nova.compute.manager [instance: 198dc1a4-050d-47b2-8bc4-30af5534a789] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 778.214765] env[61806]: ERROR nova.compute.manager [instance: 198dc1a4-050d-47b2-8bc4-30af5534a789] _ensure_no_port_binding_failure(port) [ 778.214765] env[61806]: ERROR nova.compute.manager [instance: 198dc1a4-050d-47b2-8bc4-30af5534a789] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 778.214765] env[61806]: ERROR nova.compute.manager [instance: 198dc1a4-050d-47b2-8bc4-30af5534a789] raise exception.PortBindingFailed(port_id=port['id']) [ 778.214765] env[61806]: ERROR nova.compute.manager [instance: 198dc1a4-050d-47b2-8bc4-30af5534a789] nova.exception.PortBindingFailed: Binding failed for port 1a279290-dcd8-45fb-a136-2fcbc132e8d7, please check neutron logs for more information. [ 778.214765] env[61806]: ERROR nova.compute.manager [instance: 198dc1a4-050d-47b2-8bc4-30af5534a789] [ 778.214765] env[61806]: DEBUG nova.compute.utils [None req-3bf756e0-c56e-4902-9c2b-b224a1d09a7c tempest-ListServersNegativeTestJSON-272187588 tempest-ListServersNegativeTestJSON-272187588-project-member] [instance: 198dc1a4-050d-47b2-8bc4-30af5534a789] Binding failed for port 1a279290-dcd8-45fb-a136-2fcbc132e8d7, please check neutron logs for more information. {{(pid=61806) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 778.215596] env[61806]: DEBUG oslo_concurrency.lockutils [None req-f0377b51-dab9-4f68-8592-ca0f31b4ba51 tempest-FloatingIPsAssociationNegativeTestJSON-30494360 tempest-FloatingIPsAssociationNegativeTestJSON-30494360-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 23.400s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 778.218377] env[61806]: DEBUG nova.compute.manager [None req-3bf756e0-c56e-4902-9c2b-b224a1d09a7c tempest-ListServersNegativeTestJSON-272187588 tempest-ListServersNegativeTestJSON-272187588-project-member] [instance: 198dc1a4-050d-47b2-8bc4-30af5534a789] Build of instance 198dc1a4-050d-47b2-8bc4-30af5534a789 was re-scheduled: Binding failed for port 1a279290-dcd8-45fb-a136-2fcbc132e8d7, please check neutron logs for more information. {{(pid=61806) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 778.218803] env[61806]: DEBUG nova.compute.manager [None req-3bf756e0-c56e-4902-9c2b-b224a1d09a7c tempest-ListServersNegativeTestJSON-272187588 tempest-ListServersNegativeTestJSON-272187588-project-member] [instance: 198dc1a4-050d-47b2-8bc4-30af5534a789] Unplugging VIFs for instance {{(pid=61806) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 778.219037] env[61806]: DEBUG oslo_concurrency.lockutils [None req-3bf756e0-c56e-4902-9c2b-b224a1d09a7c tempest-ListServersNegativeTestJSON-272187588 tempest-ListServersNegativeTestJSON-272187588-project-member] Acquiring lock "refresh_cache-198dc1a4-050d-47b2-8bc4-30af5534a789" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 778.219189] env[61806]: DEBUG oslo_concurrency.lockutils [None req-3bf756e0-c56e-4902-9c2b-b224a1d09a7c tempest-ListServersNegativeTestJSON-272187588 tempest-ListServersNegativeTestJSON-272187588-project-member] Acquired lock "refresh_cache-198dc1a4-050d-47b2-8bc4-30af5534a789" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 778.219350] env[61806]: DEBUG nova.network.neutron [None req-3bf756e0-c56e-4902-9c2b-b224a1d09a7c tempest-ListServersNegativeTestJSON-272187588 tempest-ListServersNegativeTestJSON-272187588-project-member] [instance: 198dc1a4-050d-47b2-8bc4-30af5534a789] Building network info cache for instance {{(pid=61806) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 778.544599] env[61806]: INFO nova.compute.manager [None req-3bf756e0-c56e-4902-9c2b-b224a1d09a7c tempest-ListServersNegativeTestJSON-272187588 tempest-ListServersNegativeTestJSON-272187588-project-member] [instance: ea8d39a1-066d-4eb4-888d-776b0e45f684] Took 1.02 seconds to deallocate network for instance. [ 778.739549] env[61806]: DEBUG nova.network.neutron [None req-3bf756e0-c56e-4902-9c2b-b224a1d09a7c tempest-ListServersNegativeTestJSON-272187588 tempest-ListServersNegativeTestJSON-272187588-project-member] [instance: 198dc1a4-050d-47b2-8bc4-30af5534a789] Instance cache missing network info. {{(pid=61806) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 778.828215] env[61806]: DEBUG nova.network.neutron [None req-3bf756e0-c56e-4902-9c2b-b224a1d09a7c tempest-ListServersNegativeTestJSON-272187588 tempest-ListServersNegativeTestJSON-272187588-project-member] [instance: 198dc1a4-050d-47b2-8bc4-30af5534a789] Updating instance_info_cache with network_info: [] {{(pid=61806) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 779.127949] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a58c0c8d-0562-4484-8826-077efc31b3d8 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 779.135707] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c68c4353-9d04-4ea9-9d8b-9ba2538d8d22 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 779.167035] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7f103d05-4c94-4f67-a8a8-00dbb05fb79e {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 779.174704] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bdeab1ee-ed9d-42a1-88e1-b4adaf76030c {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 779.191044] env[61806]: DEBUG nova.compute.provider_tree [None req-f0377b51-dab9-4f68-8592-ca0f31b4ba51 tempest-FloatingIPsAssociationNegativeTestJSON-30494360 tempest-FloatingIPsAssociationNegativeTestJSON-30494360-project-member] Inventory has not changed in ProviderTree for provider: a2858be1-fd22-4e08-979e-87ad25293407 {{(pid=61806) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 779.331263] env[61806]: DEBUG oslo_concurrency.lockutils [None req-3bf756e0-c56e-4902-9c2b-b224a1d09a7c tempest-ListServersNegativeTestJSON-272187588 tempest-ListServersNegativeTestJSON-272187588-project-member] Releasing lock "refresh_cache-198dc1a4-050d-47b2-8bc4-30af5534a789" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 779.331567] env[61806]: DEBUG nova.compute.manager [None req-3bf756e0-c56e-4902-9c2b-b224a1d09a7c tempest-ListServersNegativeTestJSON-272187588 tempest-ListServersNegativeTestJSON-272187588-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61806) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 779.331677] env[61806]: DEBUG nova.compute.manager [None req-3bf756e0-c56e-4902-9c2b-b224a1d09a7c tempest-ListServersNegativeTestJSON-272187588 tempest-ListServersNegativeTestJSON-272187588-project-member] [instance: 198dc1a4-050d-47b2-8bc4-30af5534a789] Deallocating network for instance {{(pid=61806) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 779.331842] env[61806]: DEBUG nova.network.neutron [None req-3bf756e0-c56e-4902-9c2b-b224a1d09a7c tempest-ListServersNegativeTestJSON-272187588 tempest-ListServersNegativeTestJSON-272187588-project-member] [instance: 198dc1a4-050d-47b2-8bc4-30af5534a789] deallocate_for_instance() {{(pid=61806) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 779.345705] env[61806]: DEBUG nova.network.neutron [None req-3bf756e0-c56e-4902-9c2b-b224a1d09a7c tempest-ListServersNegativeTestJSON-272187588 tempest-ListServersNegativeTestJSON-272187588-project-member] [instance: 198dc1a4-050d-47b2-8bc4-30af5534a789] Instance cache missing network info. {{(pid=61806) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 779.577905] env[61806]: INFO nova.scheduler.client.report [None req-3bf756e0-c56e-4902-9c2b-b224a1d09a7c tempest-ListServersNegativeTestJSON-272187588 tempest-ListServersNegativeTestJSON-272187588-project-member] Deleted allocations for instance ea8d39a1-066d-4eb4-888d-776b0e45f684 [ 779.691991] env[61806]: DEBUG nova.scheduler.client.report [None req-f0377b51-dab9-4f68-8592-ca0f31b4ba51 tempest-FloatingIPsAssociationNegativeTestJSON-30494360 tempest-FloatingIPsAssociationNegativeTestJSON-30494360-project-member] Inventory has not changed for provider a2858be1-fd22-4e08-979e-87ad25293407 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 139, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61806) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 779.849057] env[61806]: DEBUG nova.network.neutron [None req-3bf756e0-c56e-4902-9c2b-b224a1d09a7c tempest-ListServersNegativeTestJSON-272187588 tempest-ListServersNegativeTestJSON-272187588-project-member] [instance: 198dc1a4-050d-47b2-8bc4-30af5534a789] Updating instance_info_cache with network_info: [] {{(pid=61806) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 780.086663] env[61806]: DEBUG oslo_concurrency.lockutils [None req-3bf756e0-c56e-4902-9c2b-b224a1d09a7c tempest-ListServersNegativeTestJSON-272187588 tempest-ListServersNegativeTestJSON-272187588-project-member] Lock "ea8d39a1-066d-4eb4-888d-776b0e45f684" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 152.017s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 780.196821] env[61806]: DEBUG oslo_concurrency.lockutils [None req-f0377b51-dab9-4f68-8592-ca0f31b4ba51 tempest-FloatingIPsAssociationNegativeTestJSON-30494360 tempest-FloatingIPsAssociationNegativeTestJSON-30494360-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.981s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 780.197460] env[61806]: ERROR nova.compute.manager [None req-f0377b51-dab9-4f68-8592-ca0f31b4ba51 tempest-FloatingIPsAssociationNegativeTestJSON-30494360 tempest-FloatingIPsAssociationNegativeTestJSON-30494360-project-member] [instance: 37c07b8d-e2a0-4aa6-b7ba-194feed837e5] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 1b064d62-24f5-4892-937d-ec37537f02ac, please check neutron logs for more information. [ 780.197460] env[61806]: ERROR nova.compute.manager [instance: 37c07b8d-e2a0-4aa6-b7ba-194feed837e5] Traceback (most recent call last): [ 780.197460] env[61806]: ERROR nova.compute.manager [instance: 37c07b8d-e2a0-4aa6-b7ba-194feed837e5] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 780.197460] env[61806]: ERROR nova.compute.manager [instance: 37c07b8d-e2a0-4aa6-b7ba-194feed837e5] self.driver.spawn(context, instance, image_meta, [ 780.197460] env[61806]: ERROR nova.compute.manager [instance: 37c07b8d-e2a0-4aa6-b7ba-194feed837e5] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 780.197460] env[61806]: ERROR nova.compute.manager [instance: 37c07b8d-e2a0-4aa6-b7ba-194feed837e5] self._vmops.spawn(context, instance, image_meta, injected_files, [ 780.197460] env[61806]: ERROR nova.compute.manager [instance: 37c07b8d-e2a0-4aa6-b7ba-194feed837e5] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 780.197460] env[61806]: ERROR nova.compute.manager [instance: 37c07b8d-e2a0-4aa6-b7ba-194feed837e5] vm_ref = self.build_virtual_machine(instance, [ 780.197460] env[61806]: ERROR nova.compute.manager [instance: 37c07b8d-e2a0-4aa6-b7ba-194feed837e5] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 780.197460] env[61806]: ERROR nova.compute.manager [instance: 37c07b8d-e2a0-4aa6-b7ba-194feed837e5] vif_infos = vmwarevif.get_vif_info(self._session, [ 780.197460] env[61806]: ERROR nova.compute.manager [instance: 37c07b8d-e2a0-4aa6-b7ba-194feed837e5] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 780.197835] env[61806]: ERROR nova.compute.manager [instance: 37c07b8d-e2a0-4aa6-b7ba-194feed837e5] for vif in network_info: [ 780.197835] env[61806]: ERROR nova.compute.manager [instance: 37c07b8d-e2a0-4aa6-b7ba-194feed837e5] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 780.197835] env[61806]: ERROR nova.compute.manager [instance: 37c07b8d-e2a0-4aa6-b7ba-194feed837e5] return self._sync_wrapper(fn, *args, **kwargs) [ 780.197835] env[61806]: ERROR nova.compute.manager [instance: 37c07b8d-e2a0-4aa6-b7ba-194feed837e5] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 780.197835] env[61806]: ERROR nova.compute.manager [instance: 37c07b8d-e2a0-4aa6-b7ba-194feed837e5] self.wait() [ 780.197835] env[61806]: ERROR nova.compute.manager [instance: 37c07b8d-e2a0-4aa6-b7ba-194feed837e5] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 780.197835] env[61806]: ERROR nova.compute.manager [instance: 37c07b8d-e2a0-4aa6-b7ba-194feed837e5] self[:] = self._gt.wait() [ 780.197835] env[61806]: ERROR nova.compute.manager [instance: 37c07b8d-e2a0-4aa6-b7ba-194feed837e5] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 780.197835] env[61806]: ERROR nova.compute.manager [instance: 37c07b8d-e2a0-4aa6-b7ba-194feed837e5] return self._exit_event.wait() [ 780.197835] env[61806]: ERROR nova.compute.manager [instance: 37c07b8d-e2a0-4aa6-b7ba-194feed837e5] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 780.197835] env[61806]: ERROR nova.compute.manager [instance: 37c07b8d-e2a0-4aa6-b7ba-194feed837e5] current.throw(*self._exc) [ 780.197835] env[61806]: ERROR nova.compute.manager [instance: 37c07b8d-e2a0-4aa6-b7ba-194feed837e5] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 780.197835] env[61806]: ERROR nova.compute.manager [instance: 37c07b8d-e2a0-4aa6-b7ba-194feed837e5] result = function(*args, **kwargs) [ 780.198249] env[61806]: ERROR nova.compute.manager [instance: 37c07b8d-e2a0-4aa6-b7ba-194feed837e5] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 780.198249] env[61806]: ERROR nova.compute.manager [instance: 37c07b8d-e2a0-4aa6-b7ba-194feed837e5] return func(*args, **kwargs) [ 780.198249] env[61806]: ERROR nova.compute.manager [instance: 37c07b8d-e2a0-4aa6-b7ba-194feed837e5] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 780.198249] env[61806]: ERROR nova.compute.manager [instance: 37c07b8d-e2a0-4aa6-b7ba-194feed837e5] raise e [ 780.198249] env[61806]: ERROR nova.compute.manager [instance: 37c07b8d-e2a0-4aa6-b7ba-194feed837e5] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 780.198249] env[61806]: ERROR nova.compute.manager [instance: 37c07b8d-e2a0-4aa6-b7ba-194feed837e5] nwinfo = self.network_api.allocate_for_instance( [ 780.198249] env[61806]: ERROR nova.compute.manager [instance: 37c07b8d-e2a0-4aa6-b7ba-194feed837e5] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 780.198249] env[61806]: ERROR nova.compute.manager [instance: 37c07b8d-e2a0-4aa6-b7ba-194feed837e5] created_port_ids = self._update_ports_for_instance( [ 780.198249] env[61806]: ERROR nova.compute.manager [instance: 37c07b8d-e2a0-4aa6-b7ba-194feed837e5] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 780.198249] env[61806]: ERROR nova.compute.manager [instance: 37c07b8d-e2a0-4aa6-b7ba-194feed837e5] with excutils.save_and_reraise_exception(): [ 780.198249] env[61806]: ERROR nova.compute.manager [instance: 37c07b8d-e2a0-4aa6-b7ba-194feed837e5] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 780.198249] env[61806]: ERROR nova.compute.manager [instance: 37c07b8d-e2a0-4aa6-b7ba-194feed837e5] self.force_reraise() [ 780.198249] env[61806]: ERROR nova.compute.manager [instance: 37c07b8d-e2a0-4aa6-b7ba-194feed837e5] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 780.198691] env[61806]: ERROR nova.compute.manager [instance: 37c07b8d-e2a0-4aa6-b7ba-194feed837e5] raise self.value [ 780.198691] env[61806]: ERROR nova.compute.manager [instance: 37c07b8d-e2a0-4aa6-b7ba-194feed837e5] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 780.198691] env[61806]: ERROR nova.compute.manager [instance: 37c07b8d-e2a0-4aa6-b7ba-194feed837e5] updated_port = self._update_port( [ 780.198691] env[61806]: ERROR nova.compute.manager [instance: 37c07b8d-e2a0-4aa6-b7ba-194feed837e5] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 780.198691] env[61806]: ERROR nova.compute.manager [instance: 37c07b8d-e2a0-4aa6-b7ba-194feed837e5] _ensure_no_port_binding_failure(port) [ 780.198691] env[61806]: ERROR nova.compute.manager [instance: 37c07b8d-e2a0-4aa6-b7ba-194feed837e5] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 780.198691] env[61806]: ERROR nova.compute.manager [instance: 37c07b8d-e2a0-4aa6-b7ba-194feed837e5] raise exception.PortBindingFailed(port_id=port['id']) [ 780.198691] env[61806]: ERROR nova.compute.manager [instance: 37c07b8d-e2a0-4aa6-b7ba-194feed837e5] nova.exception.PortBindingFailed: Binding failed for port 1b064d62-24f5-4892-937d-ec37537f02ac, please check neutron logs for more information. [ 780.198691] env[61806]: ERROR nova.compute.manager [instance: 37c07b8d-e2a0-4aa6-b7ba-194feed837e5] [ 780.198691] env[61806]: DEBUG nova.compute.utils [None req-f0377b51-dab9-4f68-8592-ca0f31b4ba51 tempest-FloatingIPsAssociationNegativeTestJSON-30494360 tempest-FloatingIPsAssociationNegativeTestJSON-30494360-project-member] [instance: 37c07b8d-e2a0-4aa6-b7ba-194feed837e5] Binding failed for port 1b064d62-24f5-4892-937d-ec37537f02ac, please check neutron logs for more information. {{(pid=61806) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 780.199391] env[61806]: DEBUG oslo_concurrency.lockutils [None req-f91d1db1-2f8a-42e8-9cfa-f4915f7b77dc tempest-ImagesTestJSON-618852556 tempest-ImagesTestJSON-618852556-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 23.045s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 780.200867] env[61806]: INFO nova.compute.claims [None req-f91d1db1-2f8a-42e8-9cfa-f4915f7b77dc tempest-ImagesTestJSON-618852556 tempest-ImagesTestJSON-618852556-project-member] [instance: e0ef0a35-82a5-495b-9d5c-5805e8306390] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 780.203745] env[61806]: DEBUG nova.compute.manager [None req-f0377b51-dab9-4f68-8592-ca0f31b4ba51 tempest-FloatingIPsAssociationNegativeTestJSON-30494360 tempest-FloatingIPsAssociationNegativeTestJSON-30494360-project-member] [instance: 37c07b8d-e2a0-4aa6-b7ba-194feed837e5] Build of instance 37c07b8d-e2a0-4aa6-b7ba-194feed837e5 was re-scheduled: Binding failed for port 1b064d62-24f5-4892-937d-ec37537f02ac, please check neutron logs for more information. {{(pid=61806) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 780.203885] env[61806]: DEBUG nova.compute.manager [None req-f0377b51-dab9-4f68-8592-ca0f31b4ba51 tempest-FloatingIPsAssociationNegativeTestJSON-30494360 tempest-FloatingIPsAssociationNegativeTestJSON-30494360-project-member] [instance: 37c07b8d-e2a0-4aa6-b7ba-194feed837e5] Unplugging VIFs for instance {{(pid=61806) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 780.203988] env[61806]: DEBUG oslo_concurrency.lockutils [None req-f0377b51-dab9-4f68-8592-ca0f31b4ba51 tempest-FloatingIPsAssociationNegativeTestJSON-30494360 tempest-FloatingIPsAssociationNegativeTestJSON-30494360-project-member] Acquiring lock "refresh_cache-37c07b8d-e2a0-4aa6-b7ba-194feed837e5" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 780.204157] env[61806]: DEBUG oslo_concurrency.lockutils [None req-f0377b51-dab9-4f68-8592-ca0f31b4ba51 tempest-FloatingIPsAssociationNegativeTestJSON-30494360 tempest-FloatingIPsAssociationNegativeTestJSON-30494360-project-member] Acquired lock "refresh_cache-37c07b8d-e2a0-4aa6-b7ba-194feed837e5" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 780.204316] env[61806]: DEBUG nova.network.neutron [None req-f0377b51-dab9-4f68-8592-ca0f31b4ba51 tempest-FloatingIPsAssociationNegativeTestJSON-30494360 tempest-FloatingIPsAssociationNegativeTestJSON-30494360-project-member] [instance: 37c07b8d-e2a0-4aa6-b7ba-194feed837e5] Building network info cache for instance {{(pid=61806) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 780.352054] env[61806]: INFO nova.compute.manager [None req-3bf756e0-c56e-4902-9c2b-b224a1d09a7c tempest-ListServersNegativeTestJSON-272187588 tempest-ListServersNegativeTestJSON-272187588-project-member] [instance: 198dc1a4-050d-47b2-8bc4-30af5534a789] Took 1.02 seconds to deallocate network for instance. [ 780.589327] env[61806]: DEBUG nova.compute.manager [None req-6a30dd3c-3e08-4933-b4cc-6fae094056e7 tempest-ServerShowV247Test-1218443059 tempest-ServerShowV247Test-1218443059-project-member] [instance: 4febd093-9f2b-494e-b175-e4693b0e3e0d] Starting instance... {{(pid=61806) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 780.723402] env[61806]: DEBUG nova.network.neutron [None req-f0377b51-dab9-4f68-8592-ca0f31b4ba51 tempest-FloatingIPsAssociationNegativeTestJSON-30494360 tempest-FloatingIPsAssociationNegativeTestJSON-30494360-project-member] [instance: 37c07b8d-e2a0-4aa6-b7ba-194feed837e5] Instance cache missing network info. {{(pid=61806) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 780.808106] env[61806]: DEBUG nova.network.neutron [None req-f0377b51-dab9-4f68-8592-ca0f31b4ba51 tempest-FloatingIPsAssociationNegativeTestJSON-30494360 tempest-FloatingIPsAssociationNegativeTestJSON-30494360-project-member] [instance: 37c07b8d-e2a0-4aa6-b7ba-194feed837e5] Updating instance_info_cache with network_info: [] {{(pid=61806) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 781.111142] env[61806]: DEBUG oslo_concurrency.lockutils [None req-6a30dd3c-3e08-4933-b4cc-6fae094056e7 tempest-ServerShowV247Test-1218443059 tempest-ServerShowV247Test-1218443059-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 781.311060] env[61806]: DEBUG oslo_concurrency.lockutils [None req-f0377b51-dab9-4f68-8592-ca0f31b4ba51 tempest-FloatingIPsAssociationNegativeTestJSON-30494360 tempest-FloatingIPsAssociationNegativeTestJSON-30494360-project-member] Releasing lock "refresh_cache-37c07b8d-e2a0-4aa6-b7ba-194feed837e5" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 781.311060] env[61806]: DEBUG nova.compute.manager [None req-f0377b51-dab9-4f68-8592-ca0f31b4ba51 tempest-FloatingIPsAssociationNegativeTestJSON-30494360 tempest-FloatingIPsAssociationNegativeTestJSON-30494360-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61806) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 781.311171] env[61806]: DEBUG nova.compute.manager [None req-f0377b51-dab9-4f68-8592-ca0f31b4ba51 tempest-FloatingIPsAssociationNegativeTestJSON-30494360 tempest-FloatingIPsAssociationNegativeTestJSON-30494360-project-member] [instance: 37c07b8d-e2a0-4aa6-b7ba-194feed837e5] Deallocating network for instance {{(pid=61806) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 781.311299] env[61806]: DEBUG nova.network.neutron [None req-f0377b51-dab9-4f68-8592-ca0f31b4ba51 tempest-FloatingIPsAssociationNegativeTestJSON-30494360 tempest-FloatingIPsAssociationNegativeTestJSON-30494360-project-member] [instance: 37c07b8d-e2a0-4aa6-b7ba-194feed837e5] deallocate_for_instance() {{(pid=61806) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 781.328506] env[61806]: DEBUG nova.network.neutron [None req-f0377b51-dab9-4f68-8592-ca0f31b4ba51 tempest-FloatingIPsAssociationNegativeTestJSON-30494360 tempest-FloatingIPsAssociationNegativeTestJSON-30494360-project-member] [instance: 37c07b8d-e2a0-4aa6-b7ba-194feed837e5] Instance cache missing network info. {{(pid=61806) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 781.380464] env[61806]: INFO nova.scheduler.client.report [None req-3bf756e0-c56e-4902-9c2b-b224a1d09a7c tempest-ListServersNegativeTestJSON-272187588 tempest-ListServersNegativeTestJSON-272187588-project-member] Deleted allocations for instance 198dc1a4-050d-47b2-8bc4-30af5534a789 [ 781.439665] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-721a52d2-c271-4776-bbd8-b68270ed8fcf {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 781.447523] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f19f808b-ecc0-4a91-aba6-e9419d619551 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 781.476963] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7436ea8b-d7a1-4d24-998c-11a6d5543401 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 781.483629] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-86530669-06e6-4102-9668-d5edda2ffae9 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 781.496063] env[61806]: DEBUG nova.compute.provider_tree [None req-f91d1db1-2f8a-42e8-9cfa-f4915f7b77dc tempest-ImagesTestJSON-618852556 tempest-ImagesTestJSON-618852556-project-member] Inventory has not changed in ProviderTree for provider: a2858be1-fd22-4e08-979e-87ad25293407 {{(pid=61806) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 781.832026] env[61806]: DEBUG nova.network.neutron [None req-f0377b51-dab9-4f68-8592-ca0f31b4ba51 tempest-FloatingIPsAssociationNegativeTestJSON-30494360 tempest-FloatingIPsAssociationNegativeTestJSON-30494360-project-member] [instance: 37c07b8d-e2a0-4aa6-b7ba-194feed837e5] Updating instance_info_cache with network_info: [] {{(pid=61806) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 781.890570] env[61806]: DEBUG oslo_concurrency.lockutils [None req-3bf756e0-c56e-4902-9c2b-b224a1d09a7c tempest-ListServersNegativeTestJSON-272187588 tempest-ListServersNegativeTestJSON-272187588-project-member] Lock "198dc1a4-050d-47b2-8bc4-30af5534a789" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 153.760s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 781.999117] env[61806]: DEBUG nova.scheduler.client.report [None req-f91d1db1-2f8a-42e8-9cfa-f4915f7b77dc tempest-ImagesTestJSON-618852556 tempest-ImagesTestJSON-618852556-project-member] Inventory has not changed for provider a2858be1-fd22-4e08-979e-87ad25293407 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 139, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61806) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 782.334980] env[61806]: INFO nova.compute.manager [None req-f0377b51-dab9-4f68-8592-ca0f31b4ba51 tempest-FloatingIPsAssociationNegativeTestJSON-30494360 tempest-FloatingIPsAssociationNegativeTestJSON-30494360-project-member] [instance: 37c07b8d-e2a0-4aa6-b7ba-194feed837e5] Took 1.02 seconds to deallocate network for instance. [ 782.393189] env[61806]: DEBUG nova.compute.manager [None req-f7f6aee3-3d33-4e4b-b451-fa3af2e320d4 tempest-ServerDiagnosticsV248Test-463927155 tempest-ServerDiagnosticsV248Test-463927155-project-member] [instance: a934d02d-26aa-4900-b473-a58489e5629e] Starting instance... {{(pid=61806) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 782.507020] env[61806]: DEBUG oslo_concurrency.lockutils [None req-f91d1db1-2f8a-42e8-9cfa-f4915f7b77dc tempest-ImagesTestJSON-618852556 tempest-ImagesTestJSON-618852556-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.304s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 782.507020] env[61806]: DEBUG nova.compute.manager [None req-f91d1db1-2f8a-42e8-9cfa-f4915f7b77dc tempest-ImagesTestJSON-618852556 tempest-ImagesTestJSON-618852556-project-member] [instance: e0ef0a35-82a5-495b-9d5c-5805e8306390] Start building networks asynchronously for instance. {{(pid=61806) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 782.507020] env[61806]: DEBUG oslo_concurrency.lockutils [None req-da1d2f3c-f6ea-47f5-8c64-65c86bf2d676 tempest-ServerRescueTestJSON-811948373 tempest-ServerRescueTestJSON-811948373-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 22.793s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 782.913803] env[61806]: DEBUG oslo_concurrency.lockutils [None req-f7f6aee3-3d33-4e4b-b451-fa3af2e320d4 tempest-ServerDiagnosticsV248Test-463927155 tempest-ServerDiagnosticsV248Test-463927155-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 783.011996] env[61806]: DEBUG nova.compute.utils [None req-f91d1db1-2f8a-42e8-9cfa-f4915f7b77dc tempest-ImagesTestJSON-618852556 tempest-ImagesTestJSON-618852556-project-member] Using /dev/sd instead of None {{(pid=61806) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 783.017926] env[61806]: DEBUG nova.compute.manager [None req-f91d1db1-2f8a-42e8-9cfa-f4915f7b77dc tempest-ImagesTestJSON-618852556 tempest-ImagesTestJSON-618852556-project-member] [instance: e0ef0a35-82a5-495b-9d5c-5805e8306390] Allocating IP information in the background. {{(pid=61806) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 783.017926] env[61806]: DEBUG nova.network.neutron [None req-f91d1db1-2f8a-42e8-9cfa-f4915f7b77dc tempest-ImagesTestJSON-618852556 tempest-ImagesTestJSON-618852556-project-member] [instance: e0ef0a35-82a5-495b-9d5c-5805e8306390] allocate_for_instance() {{(pid=61806) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 783.060893] env[61806]: DEBUG nova.policy [None req-f91d1db1-2f8a-42e8-9cfa-f4915f7b77dc tempest-ImagesTestJSON-618852556 tempest-ImagesTestJSON-618852556-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'e6fa11a9b63f4cd6b04baf3115431167', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '4b9aa91aff1d4008ac5096902b77f852', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61806) authorize /opt/stack/nova/nova/policy.py:201}} [ 783.306157] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eebedb0b-6c4b-4a05-8d0e-5eb18d2fe7b9 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 783.315114] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d0fd3eb8-d644-493c-8bfd-98cebb8d3a67 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 783.354778] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3bb7a0da-bef2-4d72-b474-444a35cf3f1d {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 783.362657] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b803c0b1-1501-498e-a509-2980c7612d3e {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 783.383023] env[61806]: DEBUG nova.compute.provider_tree [None req-da1d2f3c-f6ea-47f5-8c64-65c86bf2d676 tempest-ServerRescueTestJSON-811948373 tempest-ServerRescueTestJSON-811948373-project-member] Inventory has not changed in ProviderTree for provider: a2858be1-fd22-4e08-979e-87ad25293407 {{(pid=61806) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 783.385872] env[61806]: INFO nova.scheduler.client.report [None req-f0377b51-dab9-4f68-8592-ca0f31b4ba51 tempest-FloatingIPsAssociationNegativeTestJSON-30494360 tempest-FloatingIPsAssociationNegativeTestJSON-30494360-project-member] Deleted allocations for instance 37c07b8d-e2a0-4aa6-b7ba-194feed837e5 [ 783.407197] env[61806]: DEBUG nova.network.neutron [None req-f91d1db1-2f8a-42e8-9cfa-f4915f7b77dc tempest-ImagesTestJSON-618852556 tempest-ImagesTestJSON-618852556-project-member] [instance: e0ef0a35-82a5-495b-9d5c-5805e8306390] Successfully created port: c42059da-e422-46df-be37-f9a0816b97c8 {{(pid=61806) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 783.517792] env[61806]: DEBUG nova.compute.manager [None req-f91d1db1-2f8a-42e8-9cfa-f4915f7b77dc tempest-ImagesTestJSON-618852556 tempest-ImagesTestJSON-618852556-project-member] [instance: e0ef0a35-82a5-495b-9d5c-5805e8306390] Start building block device mappings for instance. {{(pid=61806) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 783.893950] env[61806]: DEBUG nova.scheduler.client.report [None req-da1d2f3c-f6ea-47f5-8c64-65c86bf2d676 tempest-ServerRescueTestJSON-811948373 tempest-ServerRescueTestJSON-811948373-project-member] Inventory has not changed for provider a2858be1-fd22-4e08-979e-87ad25293407 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 139, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61806) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 783.898979] env[61806]: DEBUG oslo_concurrency.lockutils [None req-f0377b51-dab9-4f68-8592-ca0f31b4ba51 tempest-FloatingIPsAssociationNegativeTestJSON-30494360 tempest-FloatingIPsAssociationNegativeTestJSON-30494360-project-member] Lock "37c07b8d-e2a0-4aa6-b7ba-194feed837e5" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 154.740s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 784.400820] env[61806]: DEBUG oslo_concurrency.lockutils [None req-da1d2f3c-f6ea-47f5-8c64-65c86bf2d676 tempest-ServerRescueTestJSON-811948373 tempest-ServerRescueTestJSON-811948373-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.894s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 784.401734] env[61806]: ERROR nova.compute.manager [None req-da1d2f3c-f6ea-47f5-8c64-65c86bf2d676 tempest-ServerRescueTestJSON-811948373 tempest-ServerRescueTestJSON-811948373-project-member] [instance: c61df5d3-7f43-48d4-a2c0-a8972372c0ba] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port b12a7881-ba2b-4a75-8119-9f5cb5b2cdb7, please check neutron logs for more information. [ 784.401734] env[61806]: ERROR nova.compute.manager [instance: c61df5d3-7f43-48d4-a2c0-a8972372c0ba] Traceback (most recent call last): [ 784.401734] env[61806]: ERROR nova.compute.manager [instance: c61df5d3-7f43-48d4-a2c0-a8972372c0ba] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 784.401734] env[61806]: ERROR nova.compute.manager [instance: c61df5d3-7f43-48d4-a2c0-a8972372c0ba] self.driver.spawn(context, instance, image_meta, [ 784.401734] env[61806]: ERROR nova.compute.manager [instance: c61df5d3-7f43-48d4-a2c0-a8972372c0ba] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 784.401734] env[61806]: ERROR nova.compute.manager [instance: c61df5d3-7f43-48d4-a2c0-a8972372c0ba] self._vmops.spawn(context, instance, image_meta, injected_files, [ 784.401734] env[61806]: ERROR nova.compute.manager [instance: c61df5d3-7f43-48d4-a2c0-a8972372c0ba] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 784.401734] env[61806]: ERROR nova.compute.manager [instance: c61df5d3-7f43-48d4-a2c0-a8972372c0ba] vm_ref = self.build_virtual_machine(instance, [ 784.401734] env[61806]: ERROR nova.compute.manager [instance: c61df5d3-7f43-48d4-a2c0-a8972372c0ba] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 784.401734] env[61806]: ERROR nova.compute.manager [instance: c61df5d3-7f43-48d4-a2c0-a8972372c0ba] vif_infos = vmwarevif.get_vif_info(self._session, [ 784.401734] env[61806]: ERROR nova.compute.manager [instance: c61df5d3-7f43-48d4-a2c0-a8972372c0ba] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 784.402114] env[61806]: ERROR nova.compute.manager [instance: c61df5d3-7f43-48d4-a2c0-a8972372c0ba] for vif in network_info: [ 784.402114] env[61806]: ERROR nova.compute.manager [instance: c61df5d3-7f43-48d4-a2c0-a8972372c0ba] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 784.402114] env[61806]: ERROR nova.compute.manager [instance: c61df5d3-7f43-48d4-a2c0-a8972372c0ba] return self._sync_wrapper(fn, *args, **kwargs) [ 784.402114] env[61806]: ERROR nova.compute.manager [instance: c61df5d3-7f43-48d4-a2c0-a8972372c0ba] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 784.402114] env[61806]: ERROR nova.compute.manager [instance: c61df5d3-7f43-48d4-a2c0-a8972372c0ba] self.wait() [ 784.402114] env[61806]: ERROR nova.compute.manager [instance: c61df5d3-7f43-48d4-a2c0-a8972372c0ba] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 784.402114] env[61806]: ERROR nova.compute.manager [instance: c61df5d3-7f43-48d4-a2c0-a8972372c0ba] self[:] = self._gt.wait() [ 784.402114] env[61806]: ERROR nova.compute.manager [instance: c61df5d3-7f43-48d4-a2c0-a8972372c0ba] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 784.402114] env[61806]: ERROR nova.compute.manager [instance: c61df5d3-7f43-48d4-a2c0-a8972372c0ba] return self._exit_event.wait() [ 784.402114] env[61806]: ERROR nova.compute.manager [instance: c61df5d3-7f43-48d4-a2c0-a8972372c0ba] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 784.402114] env[61806]: ERROR nova.compute.manager [instance: c61df5d3-7f43-48d4-a2c0-a8972372c0ba] current.throw(*self._exc) [ 784.402114] env[61806]: ERROR nova.compute.manager [instance: c61df5d3-7f43-48d4-a2c0-a8972372c0ba] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 784.402114] env[61806]: ERROR nova.compute.manager [instance: c61df5d3-7f43-48d4-a2c0-a8972372c0ba] result = function(*args, **kwargs) [ 784.402464] env[61806]: ERROR nova.compute.manager [instance: c61df5d3-7f43-48d4-a2c0-a8972372c0ba] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 784.402464] env[61806]: ERROR nova.compute.manager [instance: c61df5d3-7f43-48d4-a2c0-a8972372c0ba] return func(*args, **kwargs) [ 784.402464] env[61806]: ERROR nova.compute.manager [instance: c61df5d3-7f43-48d4-a2c0-a8972372c0ba] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 784.402464] env[61806]: ERROR nova.compute.manager [instance: c61df5d3-7f43-48d4-a2c0-a8972372c0ba] raise e [ 784.402464] env[61806]: ERROR nova.compute.manager [instance: c61df5d3-7f43-48d4-a2c0-a8972372c0ba] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 784.402464] env[61806]: ERROR nova.compute.manager [instance: c61df5d3-7f43-48d4-a2c0-a8972372c0ba] nwinfo = self.network_api.allocate_for_instance( [ 784.402464] env[61806]: ERROR nova.compute.manager [instance: c61df5d3-7f43-48d4-a2c0-a8972372c0ba] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 784.402464] env[61806]: ERROR nova.compute.manager [instance: c61df5d3-7f43-48d4-a2c0-a8972372c0ba] created_port_ids = self._update_ports_for_instance( [ 784.402464] env[61806]: ERROR nova.compute.manager [instance: c61df5d3-7f43-48d4-a2c0-a8972372c0ba] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 784.402464] env[61806]: ERROR nova.compute.manager [instance: c61df5d3-7f43-48d4-a2c0-a8972372c0ba] with excutils.save_and_reraise_exception(): [ 784.402464] env[61806]: ERROR nova.compute.manager [instance: c61df5d3-7f43-48d4-a2c0-a8972372c0ba] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 784.402464] env[61806]: ERROR nova.compute.manager [instance: c61df5d3-7f43-48d4-a2c0-a8972372c0ba] self.force_reraise() [ 784.402464] env[61806]: ERROR nova.compute.manager [instance: c61df5d3-7f43-48d4-a2c0-a8972372c0ba] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 784.402718] env[61806]: ERROR nova.compute.manager [instance: c61df5d3-7f43-48d4-a2c0-a8972372c0ba] raise self.value [ 784.402718] env[61806]: ERROR nova.compute.manager [instance: c61df5d3-7f43-48d4-a2c0-a8972372c0ba] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 784.402718] env[61806]: ERROR nova.compute.manager [instance: c61df5d3-7f43-48d4-a2c0-a8972372c0ba] updated_port = self._update_port( [ 784.402718] env[61806]: ERROR nova.compute.manager [instance: c61df5d3-7f43-48d4-a2c0-a8972372c0ba] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 784.402718] env[61806]: ERROR nova.compute.manager [instance: c61df5d3-7f43-48d4-a2c0-a8972372c0ba] _ensure_no_port_binding_failure(port) [ 784.402718] env[61806]: ERROR nova.compute.manager [instance: c61df5d3-7f43-48d4-a2c0-a8972372c0ba] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 784.402718] env[61806]: ERROR nova.compute.manager [instance: c61df5d3-7f43-48d4-a2c0-a8972372c0ba] raise exception.PortBindingFailed(port_id=port['id']) [ 784.402718] env[61806]: ERROR nova.compute.manager [instance: c61df5d3-7f43-48d4-a2c0-a8972372c0ba] nova.exception.PortBindingFailed: Binding failed for port b12a7881-ba2b-4a75-8119-9f5cb5b2cdb7, please check neutron logs for more information. [ 784.402718] env[61806]: ERROR nova.compute.manager [instance: c61df5d3-7f43-48d4-a2c0-a8972372c0ba] [ 784.402718] env[61806]: DEBUG nova.compute.utils [None req-da1d2f3c-f6ea-47f5-8c64-65c86bf2d676 tempest-ServerRescueTestJSON-811948373 tempest-ServerRescueTestJSON-811948373-project-member] [instance: c61df5d3-7f43-48d4-a2c0-a8972372c0ba] Binding failed for port b12a7881-ba2b-4a75-8119-9f5cb5b2cdb7, please check neutron logs for more information. {{(pid=61806) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 784.404805] env[61806]: DEBUG oslo_concurrency.lockutils [None req-0b87073d-3de3-486a-876d-8ca5794993c1 tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 23.540s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 784.406532] env[61806]: INFO nova.compute.claims [None req-0b87073d-3de3-486a-876d-8ca5794993c1 tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] [instance: 2a3f81e3-b84f-4370-a598-2f6a607c3d47] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 784.409460] env[61806]: DEBUG nova.compute.manager [None req-38045444-a393-40c6-8f1e-16b0707aea48 tempest-ServerShowV257Test-1159790492 tempest-ServerShowV257Test-1159790492-project-member] [instance: 616ec206-9804-469e-ab5c-41aea7f048aa] Starting instance... {{(pid=61806) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 784.412269] env[61806]: DEBUG nova.compute.manager [None req-da1d2f3c-f6ea-47f5-8c64-65c86bf2d676 tempest-ServerRescueTestJSON-811948373 tempest-ServerRescueTestJSON-811948373-project-member] [instance: c61df5d3-7f43-48d4-a2c0-a8972372c0ba] Build of instance c61df5d3-7f43-48d4-a2c0-a8972372c0ba was re-scheduled: Binding failed for port b12a7881-ba2b-4a75-8119-9f5cb5b2cdb7, please check neutron logs for more information. {{(pid=61806) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 784.412779] env[61806]: DEBUG nova.compute.manager [None req-da1d2f3c-f6ea-47f5-8c64-65c86bf2d676 tempest-ServerRescueTestJSON-811948373 tempest-ServerRescueTestJSON-811948373-project-member] [instance: c61df5d3-7f43-48d4-a2c0-a8972372c0ba] Unplugging VIFs for instance {{(pid=61806) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 784.413105] env[61806]: DEBUG oslo_concurrency.lockutils [None req-da1d2f3c-f6ea-47f5-8c64-65c86bf2d676 tempest-ServerRescueTestJSON-811948373 tempest-ServerRescueTestJSON-811948373-project-member] Acquiring lock "refresh_cache-c61df5d3-7f43-48d4-a2c0-a8972372c0ba" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 784.413358] env[61806]: DEBUG oslo_concurrency.lockutils [None req-da1d2f3c-f6ea-47f5-8c64-65c86bf2d676 tempest-ServerRescueTestJSON-811948373 tempest-ServerRescueTestJSON-811948373-project-member] Acquired lock "refresh_cache-c61df5d3-7f43-48d4-a2c0-a8972372c0ba" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 784.413595] env[61806]: DEBUG nova.network.neutron [None req-da1d2f3c-f6ea-47f5-8c64-65c86bf2d676 tempest-ServerRescueTestJSON-811948373 tempest-ServerRescueTestJSON-811948373-project-member] [instance: c61df5d3-7f43-48d4-a2c0-a8972372c0ba] Building network info cache for instance {{(pid=61806) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 784.532977] env[61806]: DEBUG nova.compute.manager [None req-f91d1db1-2f8a-42e8-9cfa-f4915f7b77dc tempest-ImagesTestJSON-618852556 tempest-ImagesTestJSON-618852556-project-member] [instance: e0ef0a35-82a5-495b-9d5c-5805e8306390] Start spawning the instance on the hypervisor. {{(pid=61806) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 784.559144] env[61806]: DEBUG nova.virt.hardware [None req-f91d1db1-2f8a-42e8-9cfa-f4915f7b77dc tempest-ImagesTestJSON-618852556 tempest-ImagesTestJSON-618852556-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-15T18:05:10Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-15T18:04:52Z,direct_url=,disk_format='vmdk',id=ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='84ba83fce2c349a6988173c5d6fc3b07',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-15T18:04:53Z,virtual_size=,visibility=), allow threads: False {{(pid=61806) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 784.559144] env[61806]: DEBUG nova.virt.hardware [None req-f91d1db1-2f8a-42e8-9cfa-f4915f7b77dc tempest-ImagesTestJSON-618852556 tempest-ImagesTestJSON-618852556-project-member] Flavor limits 0:0:0 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 784.559144] env[61806]: DEBUG nova.virt.hardware [None req-f91d1db1-2f8a-42e8-9cfa-f4915f7b77dc tempest-ImagesTestJSON-618852556 tempest-ImagesTestJSON-618852556-project-member] Image limits 0:0:0 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 784.559144] env[61806]: DEBUG nova.virt.hardware [None req-f91d1db1-2f8a-42e8-9cfa-f4915f7b77dc tempest-ImagesTestJSON-618852556 tempest-ImagesTestJSON-618852556-project-member] Flavor pref 0:0:0 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 784.559308] env[61806]: DEBUG nova.virt.hardware [None req-f91d1db1-2f8a-42e8-9cfa-f4915f7b77dc tempest-ImagesTestJSON-618852556 tempest-ImagesTestJSON-618852556-project-member] Image pref 0:0:0 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 784.559308] env[61806]: DEBUG nova.virt.hardware [None req-f91d1db1-2f8a-42e8-9cfa-f4915f7b77dc tempest-ImagesTestJSON-618852556 tempest-ImagesTestJSON-618852556-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 784.559361] env[61806]: DEBUG nova.virt.hardware [None req-f91d1db1-2f8a-42e8-9cfa-f4915f7b77dc tempest-ImagesTestJSON-618852556 tempest-ImagesTestJSON-618852556-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61806) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 784.559508] env[61806]: DEBUG nova.virt.hardware [None req-f91d1db1-2f8a-42e8-9cfa-f4915f7b77dc tempest-ImagesTestJSON-618852556 tempest-ImagesTestJSON-618852556-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61806) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 784.559670] env[61806]: DEBUG nova.virt.hardware [None req-f91d1db1-2f8a-42e8-9cfa-f4915f7b77dc tempest-ImagesTestJSON-618852556 tempest-ImagesTestJSON-618852556-project-member] Got 1 possible topologies {{(pid=61806) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 784.560119] env[61806]: DEBUG nova.virt.hardware [None req-f91d1db1-2f8a-42e8-9cfa-f4915f7b77dc tempest-ImagesTestJSON-618852556 tempest-ImagesTestJSON-618852556-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61806) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 784.560119] env[61806]: DEBUG nova.virt.hardware [None req-f91d1db1-2f8a-42e8-9cfa-f4915f7b77dc tempest-ImagesTestJSON-618852556 tempest-ImagesTestJSON-618852556-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61806) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 784.560962] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4e2b3cbc-eb40-4adb-a088-12adb2d9395b {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 784.569197] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0e01e93f-4261-4666-a379-1c70c89c36c5 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 784.935413] env[61806]: DEBUG oslo_concurrency.lockutils [None req-38045444-a393-40c6-8f1e-16b0707aea48 tempest-ServerShowV257Test-1159790492 tempest-ServerShowV257Test-1159790492-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 784.946672] env[61806]: DEBUG nova.network.neutron [None req-da1d2f3c-f6ea-47f5-8c64-65c86bf2d676 tempest-ServerRescueTestJSON-811948373 tempest-ServerRescueTestJSON-811948373-project-member] [instance: c61df5d3-7f43-48d4-a2c0-a8972372c0ba] Instance cache missing network info. {{(pid=61806) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 785.118859] env[61806]: DEBUG nova.network.neutron [None req-da1d2f3c-f6ea-47f5-8c64-65c86bf2d676 tempest-ServerRescueTestJSON-811948373 tempest-ServerRescueTestJSON-811948373-project-member] [instance: c61df5d3-7f43-48d4-a2c0-a8972372c0ba] Updating instance_info_cache with network_info: [] {{(pid=61806) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 785.217423] env[61806]: DEBUG nova.compute.manager [req-182dc07d-a476-4996-a405-c9deda3b3913 req-2731fbc6-f8ec-4659-82da-8342e4718ec8 service nova] [instance: e0ef0a35-82a5-495b-9d5c-5805e8306390] Received event network-vif-plugged-c42059da-e422-46df-be37-f9a0816b97c8 {{(pid=61806) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 785.217585] env[61806]: DEBUG oslo_concurrency.lockutils [req-182dc07d-a476-4996-a405-c9deda3b3913 req-2731fbc6-f8ec-4659-82da-8342e4718ec8 service nova] Acquiring lock "e0ef0a35-82a5-495b-9d5c-5805e8306390-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 785.217723] env[61806]: DEBUG oslo_concurrency.lockutils [req-182dc07d-a476-4996-a405-c9deda3b3913 req-2731fbc6-f8ec-4659-82da-8342e4718ec8 service nova] Lock "e0ef0a35-82a5-495b-9d5c-5805e8306390-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 785.217887] env[61806]: DEBUG oslo_concurrency.lockutils [req-182dc07d-a476-4996-a405-c9deda3b3913 req-2731fbc6-f8ec-4659-82da-8342e4718ec8 service nova] Lock "e0ef0a35-82a5-495b-9d5c-5805e8306390-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 785.219328] env[61806]: DEBUG nova.compute.manager [req-182dc07d-a476-4996-a405-c9deda3b3913 req-2731fbc6-f8ec-4659-82da-8342e4718ec8 service nova] [instance: e0ef0a35-82a5-495b-9d5c-5805e8306390] No waiting events found dispatching network-vif-plugged-c42059da-e422-46df-be37-f9a0816b97c8 {{(pid=61806) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 785.219548] env[61806]: WARNING nova.compute.manager [req-182dc07d-a476-4996-a405-c9deda3b3913 req-2731fbc6-f8ec-4659-82da-8342e4718ec8 service nova] [instance: e0ef0a35-82a5-495b-9d5c-5805e8306390] Received unexpected event network-vif-plugged-c42059da-e422-46df-be37-f9a0816b97c8 for instance with vm_state building and task_state spawning. [ 785.331204] env[61806]: DEBUG nova.network.neutron [None req-f91d1db1-2f8a-42e8-9cfa-f4915f7b77dc tempest-ImagesTestJSON-618852556 tempest-ImagesTestJSON-618852556-project-member] [instance: e0ef0a35-82a5-495b-9d5c-5805e8306390] Successfully updated port: c42059da-e422-46df-be37-f9a0816b97c8 {{(pid=61806) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 785.621059] env[61806]: DEBUG oslo_concurrency.lockutils [None req-da1d2f3c-f6ea-47f5-8c64-65c86bf2d676 tempest-ServerRescueTestJSON-811948373 tempest-ServerRescueTestJSON-811948373-project-member] Releasing lock "refresh_cache-c61df5d3-7f43-48d4-a2c0-a8972372c0ba" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 785.621773] env[61806]: DEBUG nova.compute.manager [None req-da1d2f3c-f6ea-47f5-8c64-65c86bf2d676 tempest-ServerRescueTestJSON-811948373 tempest-ServerRescueTestJSON-811948373-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61806) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 785.621773] env[61806]: DEBUG nova.compute.manager [None req-da1d2f3c-f6ea-47f5-8c64-65c86bf2d676 tempest-ServerRescueTestJSON-811948373 tempest-ServerRescueTestJSON-811948373-project-member] [instance: c61df5d3-7f43-48d4-a2c0-a8972372c0ba] Deallocating network for instance {{(pid=61806) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 785.621773] env[61806]: DEBUG nova.network.neutron [None req-da1d2f3c-f6ea-47f5-8c64-65c86bf2d676 tempest-ServerRescueTestJSON-811948373 tempest-ServerRescueTestJSON-811948373-project-member] [instance: c61df5d3-7f43-48d4-a2c0-a8972372c0ba] deallocate_for_instance() {{(pid=61806) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 785.637763] env[61806]: DEBUG nova.network.neutron [None req-da1d2f3c-f6ea-47f5-8c64-65c86bf2d676 tempest-ServerRescueTestJSON-811948373 tempest-ServerRescueTestJSON-811948373-project-member] [instance: c61df5d3-7f43-48d4-a2c0-a8972372c0ba] Instance cache missing network info. {{(pid=61806) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 785.793471] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4e2663a6-71d3-49ed-bf3f-bc51975d67ee {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 785.801026] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-62919261-e9ee-4efa-b38f-87b9249b1c1e {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 785.838555] env[61806]: DEBUG oslo_concurrency.lockutils [None req-f91d1db1-2f8a-42e8-9cfa-f4915f7b77dc tempest-ImagesTestJSON-618852556 tempest-ImagesTestJSON-618852556-project-member] Acquiring lock "refresh_cache-e0ef0a35-82a5-495b-9d5c-5805e8306390" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 785.838816] env[61806]: DEBUG oslo_concurrency.lockutils [None req-f91d1db1-2f8a-42e8-9cfa-f4915f7b77dc tempest-ImagesTestJSON-618852556 tempest-ImagesTestJSON-618852556-project-member] Acquired lock "refresh_cache-e0ef0a35-82a5-495b-9d5c-5805e8306390" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 785.838869] env[61806]: DEBUG nova.network.neutron [None req-f91d1db1-2f8a-42e8-9cfa-f4915f7b77dc tempest-ImagesTestJSON-618852556 tempest-ImagesTestJSON-618852556-project-member] [instance: e0ef0a35-82a5-495b-9d5c-5805e8306390] Building network info cache for instance {{(pid=61806) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 785.840724] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c7a16549-d270-4859-b98e-a7b076e70738 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 785.850058] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d1bca5af-7745-4f9d-9330-cce84b601a4c {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 785.865955] env[61806]: DEBUG nova.compute.provider_tree [None req-0b87073d-3de3-486a-876d-8ca5794993c1 tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] Inventory has not changed in ProviderTree for provider: a2858be1-fd22-4e08-979e-87ad25293407 {{(pid=61806) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 786.140769] env[61806]: DEBUG nova.network.neutron [None req-da1d2f3c-f6ea-47f5-8c64-65c86bf2d676 tempest-ServerRescueTestJSON-811948373 tempest-ServerRescueTestJSON-811948373-project-member] [instance: c61df5d3-7f43-48d4-a2c0-a8972372c0ba] Updating instance_info_cache with network_info: [] {{(pid=61806) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 786.372329] env[61806]: DEBUG nova.scheduler.client.report [None req-0b87073d-3de3-486a-876d-8ca5794993c1 tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] Inventory has not changed for provider a2858be1-fd22-4e08-979e-87ad25293407 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 139, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61806) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 786.396802] env[61806]: DEBUG nova.network.neutron [None req-f91d1db1-2f8a-42e8-9cfa-f4915f7b77dc tempest-ImagesTestJSON-618852556 tempest-ImagesTestJSON-618852556-project-member] [instance: e0ef0a35-82a5-495b-9d5c-5805e8306390] Instance cache missing network info. {{(pid=61806) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 786.644899] env[61806]: INFO nova.compute.manager [None req-da1d2f3c-f6ea-47f5-8c64-65c86bf2d676 tempest-ServerRescueTestJSON-811948373 tempest-ServerRescueTestJSON-811948373-project-member] [instance: c61df5d3-7f43-48d4-a2c0-a8972372c0ba] Took 1.02 seconds to deallocate network for instance. [ 786.729599] env[61806]: DEBUG nova.network.neutron [None req-f91d1db1-2f8a-42e8-9cfa-f4915f7b77dc tempest-ImagesTestJSON-618852556 tempest-ImagesTestJSON-618852556-project-member] [instance: e0ef0a35-82a5-495b-9d5c-5805e8306390] Updating instance_info_cache with network_info: [{"id": "c42059da-e422-46df-be37-f9a0816b97c8", "address": "fa:16:3e:60:a9:54", "network": {"id": "288ebc87-a30f-439b-9885-9cdda68a9591", "bridge": "br-int", "label": "tempest-ImagesTestJSON-914411285-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4b9aa91aff1d4008ac5096902b77f852", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8cbc9b8f-ce19-4262-bf4d-88cd4f259a1c", "external-id": "nsx-vlan-transportzone-630", "segmentation_id": 630, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc42059da-e4", "ovs_interfaceid": "c42059da-e422-46df-be37-f9a0816b97c8", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61806) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 786.886047] env[61806]: DEBUG oslo_concurrency.lockutils [None req-0b87073d-3de3-486a-876d-8ca5794993c1 tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.479s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 786.886047] env[61806]: DEBUG nova.compute.manager [None req-0b87073d-3de3-486a-876d-8ca5794993c1 tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] [instance: 2a3f81e3-b84f-4370-a598-2f6a607c3d47] Start building networks asynchronously for instance. {{(pid=61806) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 786.889254] env[61806]: DEBUG oslo_concurrency.lockutils [None req-b5a6ece2-b279-4bbd-a3d1-16582c4d8894 tempest-ServersTestMultiNic-1480543082 tempest-ServersTestMultiNic-1480543082-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 22.869s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 787.234308] env[61806]: DEBUG oslo_concurrency.lockutils [None req-f91d1db1-2f8a-42e8-9cfa-f4915f7b77dc tempest-ImagesTestJSON-618852556 tempest-ImagesTestJSON-618852556-project-member] Releasing lock "refresh_cache-e0ef0a35-82a5-495b-9d5c-5805e8306390" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 787.234655] env[61806]: DEBUG nova.compute.manager [None req-f91d1db1-2f8a-42e8-9cfa-f4915f7b77dc tempest-ImagesTestJSON-618852556 tempest-ImagesTestJSON-618852556-project-member] [instance: e0ef0a35-82a5-495b-9d5c-5805e8306390] Instance network_info: |[{"id": "c42059da-e422-46df-be37-f9a0816b97c8", "address": "fa:16:3e:60:a9:54", "network": {"id": "288ebc87-a30f-439b-9885-9cdda68a9591", "bridge": "br-int", "label": "tempest-ImagesTestJSON-914411285-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4b9aa91aff1d4008ac5096902b77f852", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8cbc9b8f-ce19-4262-bf4d-88cd4f259a1c", "external-id": "nsx-vlan-transportzone-630", "segmentation_id": 630, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc42059da-e4", "ovs_interfaceid": "c42059da-e422-46df-be37-f9a0816b97c8", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61806) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 787.235259] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-f91d1db1-2f8a-42e8-9cfa-f4915f7b77dc tempest-ImagesTestJSON-618852556 tempest-ImagesTestJSON-618852556-project-member] [instance: e0ef0a35-82a5-495b-9d5c-5805e8306390] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:60:a9:54', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '8cbc9b8f-ce19-4262-bf4d-88cd4f259a1c', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'c42059da-e422-46df-be37-f9a0816b97c8', 'vif_model': 'vmxnet3'}] {{(pid=61806) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 787.245462] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [None req-f91d1db1-2f8a-42e8-9cfa-f4915f7b77dc tempest-ImagesTestJSON-618852556 tempest-ImagesTestJSON-618852556-project-member] Creating folder: Project (4b9aa91aff1d4008ac5096902b77f852). Parent ref: group-v277609. {{(pid=61806) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 787.245762] env[61806]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-bc5a90f3-9fa3-45fc-b7cc-0d89248b00c7 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 787.250674] env[61806]: DEBUG nova.compute.manager [req-e49b5cf4-8b82-4aa7-bde0-e4546ccf209e req-0d786fad-d52d-4df0-b521-32e171ae6bab service nova] [instance: e0ef0a35-82a5-495b-9d5c-5805e8306390] Received event network-changed-c42059da-e422-46df-be37-f9a0816b97c8 {{(pid=61806) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 787.250674] env[61806]: DEBUG nova.compute.manager [req-e49b5cf4-8b82-4aa7-bde0-e4546ccf209e req-0d786fad-d52d-4df0-b521-32e171ae6bab service nova] [instance: e0ef0a35-82a5-495b-9d5c-5805e8306390] Refreshing instance network info cache due to event network-changed-c42059da-e422-46df-be37-f9a0816b97c8. {{(pid=61806) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 787.250674] env[61806]: DEBUG oslo_concurrency.lockutils [req-e49b5cf4-8b82-4aa7-bde0-e4546ccf209e req-0d786fad-d52d-4df0-b521-32e171ae6bab service nova] Acquiring lock "refresh_cache-e0ef0a35-82a5-495b-9d5c-5805e8306390" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 787.250799] env[61806]: DEBUG oslo_concurrency.lockutils [req-e49b5cf4-8b82-4aa7-bde0-e4546ccf209e req-0d786fad-d52d-4df0-b521-32e171ae6bab service nova] Acquired lock "refresh_cache-e0ef0a35-82a5-495b-9d5c-5805e8306390" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 787.250926] env[61806]: DEBUG nova.network.neutron [req-e49b5cf4-8b82-4aa7-bde0-e4546ccf209e req-0d786fad-d52d-4df0-b521-32e171ae6bab service nova] [instance: e0ef0a35-82a5-495b-9d5c-5805e8306390] Refreshing network info cache for port c42059da-e422-46df-be37-f9a0816b97c8 {{(pid=61806) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 787.257846] env[61806]: INFO nova.virt.vmwareapi.vm_util [None req-f91d1db1-2f8a-42e8-9cfa-f4915f7b77dc tempest-ImagesTestJSON-618852556 tempest-ImagesTestJSON-618852556-project-member] Created folder: Project (4b9aa91aff1d4008ac5096902b77f852) in parent group-v277609. [ 787.258162] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [None req-f91d1db1-2f8a-42e8-9cfa-f4915f7b77dc tempest-ImagesTestJSON-618852556 tempest-ImagesTestJSON-618852556-project-member] Creating folder: Instances. Parent ref: group-v277627. {{(pid=61806) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 787.258279] env[61806]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-332ee7aa-b729-4717-9faa-3293f85be504 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 787.270182] env[61806]: INFO nova.virt.vmwareapi.vm_util [None req-f91d1db1-2f8a-42e8-9cfa-f4915f7b77dc tempest-ImagesTestJSON-618852556 tempest-ImagesTestJSON-618852556-project-member] Created folder: Instances in parent group-v277627. [ 787.270430] env[61806]: DEBUG oslo.service.loopingcall [None req-f91d1db1-2f8a-42e8-9cfa-f4915f7b77dc tempest-ImagesTestJSON-618852556 tempest-ImagesTestJSON-618852556-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61806) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 787.270614] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: e0ef0a35-82a5-495b-9d5c-5805e8306390] Creating VM on the ESX host {{(pid=61806) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 787.271289] env[61806]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-176501a6-abe2-4485-bf96-625ba9cd28dd {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 787.293414] env[61806]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 787.293414] env[61806]: value = "task-1294520" [ 787.293414] env[61806]: _type = "Task" [ 787.293414] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 787.301320] env[61806]: DEBUG oslo_vmware.api [-] Task: {'id': task-1294520, 'name': CreateVM_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 787.397458] env[61806]: DEBUG nova.compute.utils [None req-0b87073d-3de3-486a-876d-8ca5794993c1 tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] Using /dev/sd instead of None {{(pid=61806) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 787.402103] env[61806]: DEBUG nova.compute.manager [None req-0b87073d-3de3-486a-876d-8ca5794993c1 tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] [instance: 2a3f81e3-b84f-4370-a598-2f6a607c3d47] Allocating IP information in the background. {{(pid=61806) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 787.402307] env[61806]: DEBUG nova.network.neutron [None req-0b87073d-3de3-486a-876d-8ca5794993c1 tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] [instance: 2a3f81e3-b84f-4370-a598-2f6a607c3d47] allocate_for_instance() {{(pid=61806) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 787.509080] env[61806]: DEBUG nova.policy [None req-0b87073d-3de3-486a-876d-8ca5794993c1 tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'b1eac9072ccb4b3bbd23f240e941a76e', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '6e391e1267014fafbb8f5a3211299819', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61806) authorize /opt/stack/nova/nova/policy.py:201}} [ 787.679458] env[61806]: INFO nova.scheduler.client.report [None req-da1d2f3c-f6ea-47f5-8c64-65c86bf2d676 tempest-ServerRescueTestJSON-811948373 tempest-ServerRescueTestJSON-811948373-project-member] Deleted allocations for instance c61df5d3-7f43-48d4-a2c0-a8972372c0ba [ 787.765515] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-29eaebec-1b4a-4e05-b87d-e54a09ee2539 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 787.779252] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bc16e33e-59e3-462c-a945-1f7a256de41b {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 787.817499] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7c5ab4b4-f538-4c7c-b55d-0e5b4a3829cd {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 787.824908] env[61806]: DEBUG oslo_vmware.api [-] Task: {'id': task-1294520, 'name': CreateVM_Task, 'duration_secs': 0.35837} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 787.827000] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: e0ef0a35-82a5-495b-9d5c-5805e8306390] Created VM on the ESX host {{(pid=61806) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 787.830725] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3154d383-c9e7-4ab5-9b1a-32d592e01437 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 787.843793] env[61806]: DEBUG nova.compute.provider_tree [None req-b5a6ece2-b279-4bbd-a3d1-16582c4d8894 tempest-ServersTestMultiNic-1480543082 tempest-ServersTestMultiNic-1480543082-project-member] Inventory has not changed in ProviderTree for provider: a2858be1-fd22-4e08-979e-87ad25293407 {{(pid=61806) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 787.846641] env[61806]: DEBUG oslo_concurrency.lockutils [None req-f91d1db1-2f8a-42e8-9cfa-f4915f7b77dc tempest-ImagesTestJSON-618852556 tempest-ImagesTestJSON-618852556-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 787.846811] env[61806]: DEBUG oslo_concurrency.lockutils [None req-f91d1db1-2f8a-42e8-9cfa-f4915f7b77dc tempest-ImagesTestJSON-618852556 tempest-ImagesTestJSON-618852556-project-member] Acquired lock "[datastore1] devstack-image-cache_base/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 787.847147] env[61806]: DEBUG oslo_concurrency.lockutils [None req-f91d1db1-2f8a-42e8-9cfa-f4915f7b77dc tempest-ImagesTestJSON-618852556 tempest-ImagesTestJSON-618852556-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 787.847396] env[61806]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-cfffe199-0b19-4f79-9b93-84333c8b6df0 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 787.851867] env[61806]: DEBUG oslo_vmware.api [None req-f91d1db1-2f8a-42e8-9cfa-f4915f7b77dc tempest-ImagesTestJSON-618852556 tempest-ImagesTestJSON-618852556-project-member] Waiting for the task: (returnval){ [ 787.851867] env[61806]: value = "session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]52ffb999-ced0-feec-af30-fb8f89e51cd4" [ 787.851867] env[61806]: _type = "Task" [ 787.851867] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 787.859228] env[61806]: DEBUG oslo_vmware.api [None req-f91d1db1-2f8a-42e8-9cfa-f4915f7b77dc tempest-ImagesTestJSON-618852556 tempest-ImagesTestJSON-618852556-project-member] Task: {'id': session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]52ffb999-ced0-feec-af30-fb8f89e51cd4, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 787.906149] env[61806]: DEBUG nova.compute.manager [None req-0b87073d-3de3-486a-876d-8ca5794993c1 tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] [instance: 2a3f81e3-b84f-4370-a598-2f6a607c3d47] Start building block device mappings for instance. {{(pid=61806) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 788.191463] env[61806]: DEBUG oslo_concurrency.lockutils [None req-da1d2f3c-f6ea-47f5-8c64-65c86bf2d676 tempest-ServerRescueTestJSON-811948373 tempest-ServerRescueTestJSON-811948373-project-member] Lock "c61df5d3-7f43-48d4-a2c0-a8972372c0ba" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 158.988s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 788.208648] env[61806]: DEBUG nova.network.neutron [None req-0b87073d-3de3-486a-876d-8ca5794993c1 tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] [instance: 2a3f81e3-b84f-4370-a598-2f6a607c3d47] Successfully created port: a839bcaa-9990-4bad-be1f-1f06eb1c978e {{(pid=61806) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 788.268860] env[61806]: DEBUG nova.network.neutron [req-e49b5cf4-8b82-4aa7-bde0-e4546ccf209e req-0d786fad-d52d-4df0-b521-32e171ae6bab service nova] [instance: e0ef0a35-82a5-495b-9d5c-5805e8306390] Updated VIF entry in instance network info cache for port c42059da-e422-46df-be37-f9a0816b97c8. {{(pid=61806) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 788.269228] env[61806]: DEBUG nova.network.neutron [req-e49b5cf4-8b82-4aa7-bde0-e4546ccf209e req-0d786fad-d52d-4df0-b521-32e171ae6bab service nova] [instance: e0ef0a35-82a5-495b-9d5c-5805e8306390] Updating instance_info_cache with network_info: [{"id": "c42059da-e422-46df-be37-f9a0816b97c8", "address": "fa:16:3e:60:a9:54", "network": {"id": "288ebc87-a30f-439b-9885-9cdda68a9591", "bridge": "br-int", "label": "tempest-ImagesTestJSON-914411285-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4b9aa91aff1d4008ac5096902b77f852", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8cbc9b8f-ce19-4262-bf4d-88cd4f259a1c", "external-id": "nsx-vlan-transportzone-630", "segmentation_id": 630, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc42059da-e4", "ovs_interfaceid": "c42059da-e422-46df-be37-f9a0816b97c8", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61806) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 788.346302] env[61806]: DEBUG nova.scheduler.client.report [None req-b5a6ece2-b279-4bbd-a3d1-16582c4d8894 tempest-ServersTestMultiNic-1480543082 tempest-ServersTestMultiNic-1480543082-project-member] Inventory has not changed for provider a2858be1-fd22-4e08-979e-87ad25293407 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 139, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61806) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 788.361831] env[61806]: DEBUG oslo_vmware.api [None req-f91d1db1-2f8a-42e8-9cfa-f4915f7b77dc tempest-ImagesTestJSON-618852556 tempest-ImagesTestJSON-618852556-project-member] Task: {'id': session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]52ffb999-ced0-feec-af30-fb8f89e51cd4, 'name': SearchDatastore_Task, 'duration_secs': 0.367806} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 788.362170] env[61806]: DEBUG oslo_concurrency.lockutils [None req-f91d1db1-2f8a-42e8-9cfa-f4915f7b77dc tempest-ImagesTestJSON-618852556 tempest-ImagesTestJSON-618852556-project-member] Releasing lock "[datastore1] devstack-image-cache_base/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 788.362404] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-f91d1db1-2f8a-42e8-9cfa-f4915f7b77dc tempest-ImagesTestJSON-618852556 tempest-ImagesTestJSON-618852556-project-member] [instance: e0ef0a35-82a5-495b-9d5c-5805e8306390] Processing image ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa {{(pid=61806) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 788.362633] env[61806]: DEBUG oslo_concurrency.lockutils [None req-f91d1db1-2f8a-42e8-9cfa-f4915f7b77dc tempest-ImagesTestJSON-618852556 tempest-ImagesTestJSON-618852556-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa.vmdk" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 788.362778] env[61806]: DEBUG oslo_concurrency.lockutils [None req-f91d1db1-2f8a-42e8-9cfa-f4915f7b77dc tempest-ImagesTestJSON-618852556 tempest-ImagesTestJSON-618852556-project-member] Acquired lock "[datastore1] devstack-image-cache_base/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa.vmdk" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 788.362999] env[61806]: DEBUG nova.virt.vmwareapi.ds_util [None req-f91d1db1-2f8a-42e8-9cfa-f4915f7b77dc tempest-ImagesTestJSON-618852556 tempest-ImagesTestJSON-618852556-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61806) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 788.363829] env[61806]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-1eed0d9c-e7dd-4946-90a4-8d4279032396 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 788.372340] env[61806]: DEBUG nova.virt.vmwareapi.ds_util [None req-f91d1db1-2f8a-42e8-9cfa-f4915f7b77dc tempest-ImagesTestJSON-618852556 tempest-ImagesTestJSON-618852556-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61806) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 788.372600] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-f91d1db1-2f8a-42e8-9cfa-f4915f7b77dc tempest-ImagesTestJSON-618852556 tempest-ImagesTestJSON-618852556-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61806) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 788.373453] env[61806]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d2d896e3-8662-458d-bced-367f5c4375ef {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 788.378314] env[61806]: DEBUG oslo_vmware.api [None req-f91d1db1-2f8a-42e8-9cfa-f4915f7b77dc tempest-ImagesTestJSON-618852556 tempest-ImagesTestJSON-618852556-project-member] Waiting for the task: (returnval){ [ 788.378314] env[61806]: value = "session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]52e943f4-1fa5-c019-bb8a-231c204dc929" [ 788.378314] env[61806]: _type = "Task" [ 788.378314] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 788.386072] env[61806]: DEBUG oslo_vmware.api [None req-f91d1db1-2f8a-42e8-9cfa-f4915f7b77dc tempest-ImagesTestJSON-618852556 tempest-ImagesTestJSON-618852556-project-member] Task: {'id': session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]52e943f4-1fa5-c019-bb8a-231c204dc929, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 788.694811] env[61806]: DEBUG nova.compute.manager [None req-dbeadda9-7825-4678-af18-5f977e49124f tempest-ServerActionsV293TestJSON-906151111 tempest-ServerActionsV293TestJSON-906151111-project-member] [instance: 0758901a-7093-41d3-b0e2-5c519333abdd] Starting instance... {{(pid=61806) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 788.772957] env[61806]: DEBUG oslo_concurrency.lockutils [req-e49b5cf4-8b82-4aa7-bde0-e4546ccf209e req-0d786fad-d52d-4df0-b521-32e171ae6bab service nova] Releasing lock "refresh_cache-e0ef0a35-82a5-495b-9d5c-5805e8306390" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 788.851116] env[61806]: DEBUG oslo_concurrency.lockutils [None req-b5a6ece2-b279-4bbd-a3d1-16582c4d8894 tempest-ServersTestMultiNic-1480543082 tempest-ServersTestMultiNic-1480543082-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.962s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 788.851862] env[61806]: ERROR nova.compute.manager [None req-b5a6ece2-b279-4bbd-a3d1-16582c4d8894 tempest-ServersTestMultiNic-1480543082 tempest-ServersTestMultiNic-1480543082-project-member] [instance: 24c83131-1451-429f-84fe-dcd070898d65] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 30e4bd97-a4da-443f-a523-14f33ef9b039, please check neutron logs for more information. [ 788.851862] env[61806]: ERROR nova.compute.manager [instance: 24c83131-1451-429f-84fe-dcd070898d65] Traceback (most recent call last): [ 788.851862] env[61806]: ERROR nova.compute.manager [instance: 24c83131-1451-429f-84fe-dcd070898d65] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 788.851862] env[61806]: ERROR nova.compute.manager [instance: 24c83131-1451-429f-84fe-dcd070898d65] self.driver.spawn(context, instance, image_meta, [ 788.851862] env[61806]: ERROR nova.compute.manager [instance: 24c83131-1451-429f-84fe-dcd070898d65] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 788.851862] env[61806]: ERROR nova.compute.manager [instance: 24c83131-1451-429f-84fe-dcd070898d65] self._vmops.spawn(context, instance, image_meta, injected_files, [ 788.851862] env[61806]: ERROR nova.compute.manager [instance: 24c83131-1451-429f-84fe-dcd070898d65] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 788.851862] env[61806]: ERROR nova.compute.manager [instance: 24c83131-1451-429f-84fe-dcd070898d65] vm_ref = self.build_virtual_machine(instance, [ 788.851862] env[61806]: ERROR nova.compute.manager [instance: 24c83131-1451-429f-84fe-dcd070898d65] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 788.851862] env[61806]: ERROR nova.compute.manager [instance: 24c83131-1451-429f-84fe-dcd070898d65] vif_infos = vmwarevif.get_vif_info(self._session, [ 788.851862] env[61806]: ERROR nova.compute.manager [instance: 24c83131-1451-429f-84fe-dcd070898d65] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 788.852160] env[61806]: ERROR nova.compute.manager [instance: 24c83131-1451-429f-84fe-dcd070898d65] for vif in network_info: [ 788.852160] env[61806]: ERROR nova.compute.manager [instance: 24c83131-1451-429f-84fe-dcd070898d65] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 788.852160] env[61806]: ERROR nova.compute.manager [instance: 24c83131-1451-429f-84fe-dcd070898d65] return self._sync_wrapper(fn, *args, **kwargs) [ 788.852160] env[61806]: ERROR nova.compute.manager [instance: 24c83131-1451-429f-84fe-dcd070898d65] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 788.852160] env[61806]: ERROR nova.compute.manager [instance: 24c83131-1451-429f-84fe-dcd070898d65] self.wait() [ 788.852160] env[61806]: ERROR nova.compute.manager [instance: 24c83131-1451-429f-84fe-dcd070898d65] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 788.852160] env[61806]: ERROR nova.compute.manager [instance: 24c83131-1451-429f-84fe-dcd070898d65] self[:] = self._gt.wait() [ 788.852160] env[61806]: ERROR nova.compute.manager [instance: 24c83131-1451-429f-84fe-dcd070898d65] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 788.852160] env[61806]: ERROR nova.compute.manager [instance: 24c83131-1451-429f-84fe-dcd070898d65] return self._exit_event.wait() [ 788.852160] env[61806]: ERROR nova.compute.manager [instance: 24c83131-1451-429f-84fe-dcd070898d65] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 788.852160] env[61806]: ERROR nova.compute.manager [instance: 24c83131-1451-429f-84fe-dcd070898d65] result = hub.switch() [ 788.852160] env[61806]: ERROR nova.compute.manager [instance: 24c83131-1451-429f-84fe-dcd070898d65] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 788.852160] env[61806]: ERROR nova.compute.manager [instance: 24c83131-1451-429f-84fe-dcd070898d65] return self.greenlet.switch() [ 788.852428] env[61806]: ERROR nova.compute.manager [instance: 24c83131-1451-429f-84fe-dcd070898d65] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 788.852428] env[61806]: ERROR nova.compute.manager [instance: 24c83131-1451-429f-84fe-dcd070898d65] result = function(*args, **kwargs) [ 788.852428] env[61806]: ERROR nova.compute.manager [instance: 24c83131-1451-429f-84fe-dcd070898d65] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 788.852428] env[61806]: ERROR nova.compute.manager [instance: 24c83131-1451-429f-84fe-dcd070898d65] return func(*args, **kwargs) [ 788.852428] env[61806]: ERROR nova.compute.manager [instance: 24c83131-1451-429f-84fe-dcd070898d65] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 788.852428] env[61806]: ERROR nova.compute.manager [instance: 24c83131-1451-429f-84fe-dcd070898d65] raise e [ 788.852428] env[61806]: ERROR nova.compute.manager [instance: 24c83131-1451-429f-84fe-dcd070898d65] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 788.852428] env[61806]: ERROR nova.compute.manager [instance: 24c83131-1451-429f-84fe-dcd070898d65] nwinfo = self.network_api.allocate_for_instance( [ 788.852428] env[61806]: ERROR nova.compute.manager [instance: 24c83131-1451-429f-84fe-dcd070898d65] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 788.852428] env[61806]: ERROR nova.compute.manager [instance: 24c83131-1451-429f-84fe-dcd070898d65] created_port_ids = self._update_ports_for_instance( [ 788.852428] env[61806]: ERROR nova.compute.manager [instance: 24c83131-1451-429f-84fe-dcd070898d65] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 788.852428] env[61806]: ERROR nova.compute.manager [instance: 24c83131-1451-429f-84fe-dcd070898d65] with excutils.save_and_reraise_exception(): [ 788.852428] env[61806]: ERROR nova.compute.manager [instance: 24c83131-1451-429f-84fe-dcd070898d65] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 788.852779] env[61806]: ERROR nova.compute.manager [instance: 24c83131-1451-429f-84fe-dcd070898d65] self.force_reraise() [ 788.852779] env[61806]: ERROR nova.compute.manager [instance: 24c83131-1451-429f-84fe-dcd070898d65] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 788.852779] env[61806]: ERROR nova.compute.manager [instance: 24c83131-1451-429f-84fe-dcd070898d65] raise self.value [ 788.852779] env[61806]: ERROR nova.compute.manager [instance: 24c83131-1451-429f-84fe-dcd070898d65] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 788.852779] env[61806]: ERROR nova.compute.manager [instance: 24c83131-1451-429f-84fe-dcd070898d65] updated_port = self._update_port( [ 788.852779] env[61806]: ERROR nova.compute.manager [instance: 24c83131-1451-429f-84fe-dcd070898d65] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 788.852779] env[61806]: ERROR nova.compute.manager [instance: 24c83131-1451-429f-84fe-dcd070898d65] _ensure_no_port_binding_failure(port) [ 788.852779] env[61806]: ERROR nova.compute.manager [instance: 24c83131-1451-429f-84fe-dcd070898d65] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 788.852779] env[61806]: ERROR nova.compute.manager [instance: 24c83131-1451-429f-84fe-dcd070898d65] raise exception.PortBindingFailed(port_id=port['id']) [ 788.852779] env[61806]: ERROR nova.compute.manager [instance: 24c83131-1451-429f-84fe-dcd070898d65] nova.exception.PortBindingFailed: Binding failed for port 30e4bd97-a4da-443f-a523-14f33ef9b039, please check neutron logs for more information. [ 788.852779] env[61806]: ERROR nova.compute.manager [instance: 24c83131-1451-429f-84fe-dcd070898d65] [ 788.853039] env[61806]: DEBUG nova.compute.utils [None req-b5a6ece2-b279-4bbd-a3d1-16582c4d8894 tempest-ServersTestMultiNic-1480543082 tempest-ServersTestMultiNic-1480543082-project-member] [instance: 24c83131-1451-429f-84fe-dcd070898d65] Binding failed for port 30e4bd97-a4da-443f-a523-14f33ef9b039, please check neutron logs for more information. {{(pid=61806) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 788.853758] env[61806]: DEBUG oslo_concurrency.lockutils [None req-b3f33a6f-532a-458d-aec6-6ec54b0cadd5 tempest-ServerMetadataTestJSON-316746592 tempest-ServerMetadataTestJSON-316746592-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 23.680s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 788.857370] env[61806]: INFO nova.compute.claims [None req-b3f33a6f-532a-458d-aec6-6ec54b0cadd5 tempest-ServerMetadataTestJSON-316746592 tempest-ServerMetadataTestJSON-316746592-project-member] [instance: 4184e263-6f56-4bc1-99b2-a2c460531516] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 788.861363] env[61806]: DEBUG nova.compute.manager [None req-b5a6ece2-b279-4bbd-a3d1-16582c4d8894 tempest-ServersTestMultiNic-1480543082 tempest-ServersTestMultiNic-1480543082-project-member] [instance: 24c83131-1451-429f-84fe-dcd070898d65] Build of instance 24c83131-1451-429f-84fe-dcd070898d65 was re-scheduled: Binding failed for port 30e4bd97-a4da-443f-a523-14f33ef9b039, please check neutron logs for more information. {{(pid=61806) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 788.861813] env[61806]: DEBUG nova.compute.manager [None req-b5a6ece2-b279-4bbd-a3d1-16582c4d8894 tempest-ServersTestMultiNic-1480543082 tempest-ServersTestMultiNic-1480543082-project-member] [instance: 24c83131-1451-429f-84fe-dcd070898d65] Unplugging VIFs for instance {{(pid=61806) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 788.862071] env[61806]: DEBUG oslo_concurrency.lockutils [None req-b5a6ece2-b279-4bbd-a3d1-16582c4d8894 tempest-ServersTestMultiNic-1480543082 tempest-ServersTestMultiNic-1480543082-project-member] Acquiring lock "refresh_cache-24c83131-1451-429f-84fe-dcd070898d65" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 788.862236] env[61806]: DEBUG oslo_concurrency.lockutils [None req-b5a6ece2-b279-4bbd-a3d1-16582c4d8894 tempest-ServersTestMultiNic-1480543082 tempest-ServersTestMultiNic-1480543082-project-member] Acquired lock "refresh_cache-24c83131-1451-429f-84fe-dcd070898d65" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 788.863817] env[61806]: DEBUG nova.network.neutron [None req-b5a6ece2-b279-4bbd-a3d1-16582c4d8894 tempest-ServersTestMultiNic-1480543082 tempest-ServersTestMultiNic-1480543082-project-member] [instance: 24c83131-1451-429f-84fe-dcd070898d65] Building network info cache for instance {{(pid=61806) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 788.890287] env[61806]: DEBUG oslo_vmware.api [None req-f91d1db1-2f8a-42e8-9cfa-f4915f7b77dc tempest-ImagesTestJSON-618852556 tempest-ImagesTestJSON-618852556-project-member] Task: {'id': session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]52e943f4-1fa5-c019-bb8a-231c204dc929, 'name': SearchDatastore_Task, 'duration_secs': 0.009015} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 788.890287] env[61806]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-6d247283-f3d1-4b7c-b606-0d9810f59c89 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 788.899269] env[61806]: DEBUG oslo_vmware.api [None req-f91d1db1-2f8a-42e8-9cfa-f4915f7b77dc tempest-ImagesTestJSON-618852556 tempest-ImagesTestJSON-618852556-project-member] Waiting for the task: (returnval){ [ 788.899269] env[61806]: value = "session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]5242c13a-c992-9d07-d82b-3fdf1d6d1941" [ 788.899269] env[61806]: _type = "Task" [ 788.899269] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 788.914243] env[61806]: DEBUG oslo_vmware.api [None req-f91d1db1-2f8a-42e8-9cfa-f4915f7b77dc tempest-ImagesTestJSON-618852556 tempest-ImagesTestJSON-618852556-project-member] Task: {'id': session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]5242c13a-c992-9d07-d82b-3fdf1d6d1941, 'name': SearchDatastore_Task, 'duration_secs': 0.008907} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 788.917422] env[61806]: DEBUG nova.compute.manager [None req-0b87073d-3de3-486a-876d-8ca5794993c1 tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] [instance: 2a3f81e3-b84f-4370-a598-2f6a607c3d47] Start spawning the instance on the hypervisor. {{(pid=61806) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 788.917422] env[61806]: DEBUG oslo_concurrency.lockutils [None req-f91d1db1-2f8a-42e8-9cfa-f4915f7b77dc tempest-ImagesTestJSON-618852556 tempest-ImagesTestJSON-618852556-project-member] Releasing lock "[datastore1] devstack-image-cache_base/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa.vmdk" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 788.917551] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [None req-f91d1db1-2f8a-42e8-9cfa-f4915f7b77dc tempest-ImagesTestJSON-618852556 tempest-ImagesTestJSON-618852556-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa.vmdk to [datastore1] e0ef0a35-82a5-495b-9d5c-5805e8306390/e0ef0a35-82a5-495b-9d5c-5805e8306390.vmdk {{(pid=61806) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 788.918686] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-c4da824c-2088-4de4-b5e7-f459e7f38f92 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 788.925819] env[61806]: DEBUG oslo_vmware.api [None req-f91d1db1-2f8a-42e8-9cfa-f4915f7b77dc tempest-ImagesTestJSON-618852556 tempest-ImagesTestJSON-618852556-project-member] Waiting for the task: (returnval){ [ 788.925819] env[61806]: value = "task-1294521" [ 788.925819] env[61806]: _type = "Task" [ 788.925819] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 788.936692] env[61806]: DEBUG oslo_vmware.api [None req-f91d1db1-2f8a-42e8-9cfa-f4915f7b77dc tempest-ImagesTestJSON-618852556 tempest-ImagesTestJSON-618852556-project-member] Task: {'id': task-1294521, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 788.954532] env[61806]: DEBUG nova.virt.hardware [None req-0b87073d-3de3-486a-876d-8ca5794993c1 tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-15T18:05:10Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-15T18:04:52Z,direct_url=,disk_format='vmdk',id=ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='84ba83fce2c349a6988173c5d6fc3b07',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-15T18:04:53Z,virtual_size=,visibility=), allow threads: False {{(pid=61806) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 788.954942] env[61806]: DEBUG nova.virt.hardware [None req-0b87073d-3de3-486a-876d-8ca5794993c1 tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] Flavor limits 0:0:0 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 788.954942] env[61806]: DEBUG nova.virt.hardware [None req-0b87073d-3de3-486a-876d-8ca5794993c1 tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] Image limits 0:0:0 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 788.955135] env[61806]: DEBUG nova.virt.hardware [None req-0b87073d-3de3-486a-876d-8ca5794993c1 tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] Flavor pref 0:0:0 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 788.955286] env[61806]: DEBUG nova.virt.hardware [None req-0b87073d-3de3-486a-876d-8ca5794993c1 tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] Image pref 0:0:0 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 788.955435] env[61806]: DEBUG nova.virt.hardware [None req-0b87073d-3de3-486a-876d-8ca5794993c1 tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 788.955642] env[61806]: DEBUG nova.virt.hardware [None req-0b87073d-3de3-486a-876d-8ca5794993c1 tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61806) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 788.955803] env[61806]: DEBUG nova.virt.hardware [None req-0b87073d-3de3-486a-876d-8ca5794993c1 tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61806) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 788.955971] env[61806]: DEBUG nova.virt.hardware [None req-0b87073d-3de3-486a-876d-8ca5794993c1 tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] Got 1 possible topologies {{(pid=61806) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 788.956423] env[61806]: DEBUG nova.virt.hardware [None req-0b87073d-3de3-486a-876d-8ca5794993c1 tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61806) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 788.956631] env[61806]: DEBUG nova.virt.hardware [None req-0b87073d-3de3-486a-876d-8ca5794993c1 tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61806) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 788.958571] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f56573a2-c8d0-44aa-a6d4-a2674370ee21 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 788.970437] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bbb17408-0eda-400d-80f2-f78089a1a348 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 789.216983] env[61806]: DEBUG oslo_concurrency.lockutils [None req-dbeadda9-7825-4678-af18-5f977e49124f tempest-ServerActionsV293TestJSON-906151111 tempest-ServerActionsV293TestJSON-906151111-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 789.394801] env[61806]: DEBUG nova.network.neutron [None req-b5a6ece2-b279-4bbd-a3d1-16582c4d8894 tempest-ServersTestMultiNic-1480543082 tempest-ServersTestMultiNic-1480543082-project-member] [instance: 24c83131-1451-429f-84fe-dcd070898d65] Instance cache missing network info. {{(pid=61806) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 789.438099] env[61806]: DEBUG oslo_vmware.api [None req-f91d1db1-2f8a-42e8-9cfa-f4915f7b77dc tempest-ImagesTestJSON-618852556 tempest-ImagesTestJSON-618852556-project-member] Task: {'id': task-1294521, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.477411} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 789.438389] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [None req-f91d1db1-2f8a-42e8-9cfa-f4915f7b77dc tempest-ImagesTestJSON-618852556 tempest-ImagesTestJSON-618852556-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa.vmdk to [datastore1] e0ef0a35-82a5-495b-9d5c-5805e8306390/e0ef0a35-82a5-495b-9d5c-5805e8306390.vmdk {{(pid=61806) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 789.439341] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-f91d1db1-2f8a-42e8-9cfa-f4915f7b77dc tempest-ImagesTestJSON-618852556 tempest-ImagesTestJSON-618852556-project-member] [instance: e0ef0a35-82a5-495b-9d5c-5805e8306390] Extending root virtual disk to 1048576 {{(pid=61806) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 789.439341] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-c902cb82-784b-475d-a3bf-cf1ac8865189 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 789.446570] env[61806]: DEBUG oslo_vmware.api [None req-f91d1db1-2f8a-42e8-9cfa-f4915f7b77dc tempest-ImagesTestJSON-618852556 tempest-ImagesTestJSON-618852556-project-member] Waiting for the task: (returnval){ [ 789.446570] env[61806]: value = "task-1294522" [ 789.446570] env[61806]: _type = "Task" [ 789.446570] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 789.455356] env[61806]: DEBUG oslo_vmware.api [None req-f91d1db1-2f8a-42e8-9cfa-f4915f7b77dc tempest-ImagesTestJSON-618852556 tempest-ImagesTestJSON-618852556-project-member] Task: {'id': task-1294522, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 789.480711] env[61806]: DEBUG oslo_concurrency.lockutils [None req-4e0c8443-0fd4-459b-a4c5-7cf373aa1089 tempest-ServerAddressesNegativeTestJSON-1526246390 tempest-ServerAddressesNegativeTestJSON-1526246390-project-member] Acquiring lock "c8033551-6591-4e37-ae78-4efe7145b10b" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 789.480711] env[61806]: DEBUG oslo_concurrency.lockutils [None req-4e0c8443-0fd4-459b-a4c5-7cf373aa1089 tempest-ServerAddressesNegativeTestJSON-1526246390 tempest-ServerAddressesNegativeTestJSON-1526246390-project-member] Lock "c8033551-6591-4e37-ae78-4efe7145b10b" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 789.565076] env[61806]: DEBUG nova.network.neutron [None req-b5a6ece2-b279-4bbd-a3d1-16582c4d8894 tempest-ServersTestMultiNic-1480543082 tempest-ServersTestMultiNic-1480543082-project-member] [instance: 24c83131-1451-429f-84fe-dcd070898d65] Updating instance_info_cache with network_info: [] {{(pid=61806) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 789.964260] env[61806]: DEBUG oslo_vmware.api [None req-f91d1db1-2f8a-42e8-9cfa-f4915f7b77dc tempest-ImagesTestJSON-618852556 tempest-ImagesTestJSON-618852556-project-member] Task: {'id': task-1294522, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.060986} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 789.967583] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-f91d1db1-2f8a-42e8-9cfa-f4915f7b77dc tempest-ImagesTestJSON-618852556 tempest-ImagesTestJSON-618852556-project-member] [instance: e0ef0a35-82a5-495b-9d5c-5805e8306390] Extended root virtual disk {{(pid=61806) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 789.969157] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-278998cf-5e52-4a73-93ee-da03454ce16c {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 789.999194] env[61806]: DEBUG nova.virt.vmwareapi.volumeops [None req-f91d1db1-2f8a-42e8-9cfa-f4915f7b77dc tempest-ImagesTestJSON-618852556 tempest-ImagesTestJSON-618852556-project-member] [instance: e0ef0a35-82a5-495b-9d5c-5805e8306390] Reconfiguring VM instance instance-00000035 to attach disk [datastore1] e0ef0a35-82a5-495b-9d5c-5805e8306390/e0ef0a35-82a5-495b-9d5c-5805e8306390.vmdk or device None with type sparse {{(pid=61806) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 790.002369] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-c9deae70-9a5a-402d-b31c-59463007ac3a {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 790.028801] env[61806]: DEBUG oslo_vmware.api [None req-f91d1db1-2f8a-42e8-9cfa-f4915f7b77dc tempest-ImagesTestJSON-618852556 tempest-ImagesTestJSON-618852556-project-member] Waiting for the task: (returnval){ [ 790.028801] env[61806]: value = "task-1294523" [ 790.028801] env[61806]: _type = "Task" [ 790.028801] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 790.038665] env[61806]: DEBUG oslo_vmware.api [None req-f91d1db1-2f8a-42e8-9cfa-f4915f7b77dc tempest-ImagesTestJSON-618852556 tempest-ImagesTestJSON-618852556-project-member] Task: {'id': task-1294523, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 790.070993] env[61806]: DEBUG oslo_concurrency.lockutils [None req-b5a6ece2-b279-4bbd-a3d1-16582c4d8894 tempest-ServersTestMultiNic-1480543082 tempest-ServersTestMultiNic-1480543082-project-member] Releasing lock "refresh_cache-24c83131-1451-429f-84fe-dcd070898d65" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 790.071264] env[61806]: DEBUG nova.compute.manager [None req-b5a6ece2-b279-4bbd-a3d1-16582c4d8894 tempest-ServersTestMultiNic-1480543082 tempest-ServersTestMultiNic-1480543082-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61806) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 790.071436] env[61806]: DEBUG nova.compute.manager [None req-b5a6ece2-b279-4bbd-a3d1-16582c4d8894 tempest-ServersTestMultiNic-1480543082 tempest-ServersTestMultiNic-1480543082-project-member] [instance: 24c83131-1451-429f-84fe-dcd070898d65] Deallocating network for instance {{(pid=61806) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 790.071602] env[61806]: DEBUG nova.network.neutron [None req-b5a6ece2-b279-4bbd-a3d1-16582c4d8894 tempest-ServersTestMultiNic-1480543082 tempest-ServersTestMultiNic-1480543082-project-member] [instance: 24c83131-1451-429f-84fe-dcd070898d65] deallocate_for_instance() {{(pid=61806) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 790.129884] env[61806]: DEBUG nova.network.neutron [None req-b5a6ece2-b279-4bbd-a3d1-16582c4d8894 tempest-ServersTestMultiNic-1480543082 tempest-ServersTestMultiNic-1480543082-project-member] [instance: 24c83131-1451-429f-84fe-dcd070898d65] Instance cache missing network info. {{(pid=61806) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 790.307232] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ea981af9-731f-41db-826d-40b8e2e8e6a3 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 790.315684] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4b7d1ba7-1101-4038-a45d-b3261c6082c5 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 790.350885] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c263051b-5ee5-4159-95f8-76b0b047b55e {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 790.354463] env[61806]: DEBUG nova.network.neutron [None req-0b87073d-3de3-486a-876d-8ca5794993c1 tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] [instance: 2a3f81e3-b84f-4370-a598-2f6a607c3d47] Successfully updated port: a839bcaa-9990-4bad-be1f-1f06eb1c978e {{(pid=61806) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 790.364625] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-764f0f99-a662-4dc5-8220-f6e9950e46f4 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 790.384949] env[61806]: DEBUG nova.compute.provider_tree [None req-b3f33a6f-532a-458d-aec6-6ec54b0cadd5 tempest-ServerMetadataTestJSON-316746592 tempest-ServerMetadataTestJSON-316746592-project-member] Inventory has not changed in ProviderTree for provider: a2858be1-fd22-4e08-979e-87ad25293407 {{(pid=61806) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 790.418460] env[61806]: DEBUG nova.compute.manager [req-26874d32-8e78-42d4-91d8-f31993bbf8dc req-da3779e2-d948-4f66-ba6b-1a151e3f64f9 service nova] [instance: 2a3f81e3-b84f-4370-a598-2f6a607c3d47] Received event network-vif-plugged-a839bcaa-9990-4bad-be1f-1f06eb1c978e {{(pid=61806) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 790.418460] env[61806]: DEBUG oslo_concurrency.lockutils [req-26874d32-8e78-42d4-91d8-f31993bbf8dc req-da3779e2-d948-4f66-ba6b-1a151e3f64f9 service nova] Acquiring lock "2a3f81e3-b84f-4370-a598-2f6a607c3d47-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 790.419509] env[61806]: DEBUG oslo_concurrency.lockutils [req-26874d32-8e78-42d4-91d8-f31993bbf8dc req-da3779e2-d948-4f66-ba6b-1a151e3f64f9 service nova] Lock "2a3f81e3-b84f-4370-a598-2f6a607c3d47-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 790.419509] env[61806]: DEBUG oslo_concurrency.lockutils [req-26874d32-8e78-42d4-91d8-f31993bbf8dc req-da3779e2-d948-4f66-ba6b-1a151e3f64f9 service nova] Lock "2a3f81e3-b84f-4370-a598-2f6a607c3d47-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 790.419509] env[61806]: DEBUG nova.compute.manager [req-26874d32-8e78-42d4-91d8-f31993bbf8dc req-da3779e2-d948-4f66-ba6b-1a151e3f64f9 service nova] [instance: 2a3f81e3-b84f-4370-a598-2f6a607c3d47] No waiting events found dispatching network-vif-plugged-a839bcaa-9990-4bad-be1f-1f06eb1c978e {{(pid=61806) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 790.420961] env[61806]: WARNING nova.compute.manager [req-26874d32-8e78-42d4-91d8-f31993bbf8dc req-da3779e2-d948-4f66-ba6b-1a151e3f64f9 service nova] [instance: 2a3f81e3-b84f-4370-a598-2f6a607c3d47] Received unexpected event network-vif-plugged-a839bcaa-9990-4bad-be1f-1f06eb1c978e for instance with vm_state building and task_state spawning. [ 790.540556] env[61806]: DEBUG oslo_vmware.api [None req-f91d1db1-2f8a-42e8-9cfa-f4915f7b77dc tempest-ImagesTestJSON-618852556 tempest-ImagesTestJSON-618852556-project-member] Task: {'id': task-1294523, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 790.633601] env[61806]: DEBUG nova.network.neutron [None req-b5a6ece2-b279-4bbd-a3d1-16582c4d8894 tempest-ServersTestMultiNic-1480543082 tempest-ServersTestMultiNic-1480543082-project-member] [instance: 24c83131-1451-429f-84fe-dcd070898d65] Updating instance_info_cache with network_info: [] {{(pid=61806) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 790.859189] env[61806]: DEBUG oslo_concurrency.lockutils [None req-0b87073d-3de3-486a-876d-8ca5794993c1 tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] Acquiring lock "refresh_cache-2a3f81e3-b84f-4370-a598-2f6a607c3d47" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 790.860263] env[61806]: DEBUG oslo_concurrency.lockutils [None req-0b87073d-3de3-486a-876d-8ca5794993c1 tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] Acquired lock "refresh_cache-2a3f81e3-b84f-4370-a598-2f6a607c3d47" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 790.860663] env[61806]: DEBUG nova.network.neutron [None req-0b87073d-3de3-486a-876d-8ca5794993c1 tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] [instance: 2a3f81e3-b84f-4370-a598-2f6a607c3d47] Building network info cache for instance {{(pid=61806) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 790.889337] env[61806]: DEBUG nova.scheduler.client.report [None req-b3f33a6f-532a-458d-aec6-6ec54b0cadd5 tempest-ServerMetadataTestJSON-316746592 tempest-ServerMetadataTestJSON-316746592-project-member] Inventory has not changed for provider a2858be1-fd22-4e08-979e-87ad25293407 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 139, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61806) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 791.043736] env[61806]: DEBUG oslo_vmware.api [None req-f91d1db1-2f8a-42e8-9cfa-f4915f7b77dc tempest-ImagesTestJSON-618852556 tempest-ImagesTestJSON-618852556-project-member] Task: {'id': task-1294523, 'name': ReconfigVM_Task, 'duration_secs': 0.851973} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 791.045110] env[61806]: DEBUG nova.virt.vmwareapi.volumeops [None req-f91d1db1-2f8a-42e8-9cfa-f4915f7b77dc tempest-ImagesTestJSON-618852556 tempest-ImagesTestJSON-618852556-project-member] [instance: e0ef0a35-82a5-495b-9d5c-5805e8306390] Reconfigured VM instance instance-00000035 to attach disk [datastore1] e0ef0a35-82a5-495b-9d5c-5805e8306390/e0ef0a35-82a5-495b-9d5c-5805e8306390.vmdk or device None with type sparse {{(pid=61806) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 791.045436] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-17f991b6-6d85-4ae0-82f1-228adc3fa6cb {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 791.056538] env[61806]: DEBUG oslo_vmware.api [None req-f91d1db1-2f8a-42e8-9cfa-f4915f7b77dc tempest-ImagesTestJSON-618852556 tempest-ImagesTestJSON-618852556-project-member] Waiting for the task: (returnval){ [ 791.056538] env[61806]: value = "task-1294524" [ 791.056538] env[61806]: _type = "Task" [ 791.056538] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 791.064869] env[61806]: DEBUG oslo_vmware.api [None req-f91d1db1-2f8a-42e8-9cfa-f4915f7b77dc tempest-ImagesTestJSON-618852556 tempest-ImagesTestJSON-618852556-project-member] Task: {'id': task-1294524, 'name': Rename_Task} progress is 5%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 791.139717] env[61806]: INFO nova.compute.manager [None req-b5a6ece2-b279-4bbd-a3d1-16582c4d8894 tempest-ServersTestMultiNic-1480543082 tempest-ServersTestMultiNic-1480543082-project-member] [instance: 24c83131-1451-429f-84fe-dcd070898d65] Took 1.07 seconds to deallocate network for instance. [ 791.395437] env[61806]: DEBUG oslo_concurrency.lockutils [None req-b3f33a6f-532a-458d-aec6-6ec54b0cadd5 tempest-ServerMetadataTestJSON-316746592 tempest-ServerMetadataTestJSON-316746592-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.542s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 791.395975] env[61806]: DEBUG nova.compute.manager [None req-b3f33a6f-532a-458d-aec6-6ec54b0cadd5 tempest-ServerMetadataTestJSON-316746592 tempest-ServerMetadataTestJSON-316746592-project-member] [instance: 4184e263-6f56-4bc1-99b2-a2c460531516] Start building networks asynchronously for instance. {{(pid=61806) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 791.398563] env[61806]: DEBUG oslo_concurrency.lockutils [None req-3d5b4b69-10ec-4d1a-aa3d-4dddf4155c99 tempest-ServerShowV247Test-1218443059 tempest-ServerShowV247Test-1218443059-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 24.389s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 791.399947] env[61806]: INFO nova.compute.claims [None req-3d5b4b69-10ec-4d1a-aa3d-4dddf4155c99 tempest-ServerShowV247Test-1218443059 tempest-ServerShowV247Test-1218443059-project-member] [instance: 7c3fd2ff-ebd5-454f-a743-7fbae1088941] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 791.403162] env[61806]: DEBUG nova.network.neutron [None req-0b87073d-3de3-486a-876d-8ca5794993c1 tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] [instance: 2a3f81e3-b84f-4370-a598-2f6a607c3d47] Instance cache missing network info. {{(pid=61806) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 791.568821] env[61806]: DEBUG oslo_vmware.api [None req-f91d1db1-2f8a-42e8-9cfa-f4915f7b77dc tempest-ImagesTestJSON-618852556 tempest-ImagesTestJSON-618852556-project-member] Task: {'id': task-1294524, 'name': Rename_Task, 'duration_secs': 0.164042} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 791.569794] env[61806]: DEBUG nova.network.neutron [None req-0b87073d-3de3-486a-876d-8ca5794993c1 tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] [instance: 2a3f81e3-b84f-4370-a598-2f6a607c3d47] Updating instance_info_cache with network_info: [{"id": "a839bcaa-9990-4bad-be1f-1f06eb1c978e", "address": "fa:16:3e:0b:8a:a6", "network": {"id": "a7ffda72-e57c-41fa-a7ed-d7b12084bc9d", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-62794865-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "6e391e1267014fafbb8f5a3211299819", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8b3cfeb1-f262-4fd9-b506-8e9c0733e2d8", "external-id": "nsx-vlan-transportzone-119", "segmentation_id": 119, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa839bcaa-99", "ovs_interfaceid": "a839bcaa-9990-4bad-be1f-1f06eb1c978e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61806) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 791.572044] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [None req-f91d1db1-2f8a-42e8-9cfa-f4915f7b77dc tempest-ImagesTestJSON-618852556 tempest-ImagesTestJSON-618852556-project-member] [instance: e0ef0a35-82a5-495b-9d5c-5805e8306390] Powering on the VM {{(pid=61806) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 791.572161] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-d1a2509a-391a-435e-bc0b-ed957c5c6f9b {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 791.580774] env[61806]: DEBUG oslo_vmware.api [None req-f91d1db1-2f8a-42e8-9cfa-f4915f7b77dc tempest-ImagesTestJSON-618852556 tempest-ImagesTestJSON-618852556-project-member] Waiting for the task: (returnval){ [ 791.580774] env[61806]: value = "task-1294525" [ 791.580774] env[61806]: _type = "Task" [ 791.580774] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 791.591182] env[61806]: DEBUG oslo_vmware.api [None req-f91d1db1-2f8a-42e8-9cfa-f4915f7b77dc tempest-ImagesTestJSON-618852556 tempest-ImagesTestJSON-618852556-project-member] Task: {'id': task-1294525, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 791.905552] env[61806]: DEBUG nova.compute.utils [None req-b3f33a6f-532a-458d-aec6-6ec54b0cadd5 tempest-ServerMetadataTestJSON-316746592 tempest-ServerMetadataTestJSON-316746592-project-member] Using /dev/sd instead of None {{(pid=61806) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 791.907243] env[61806]: DEBUG nova.compute.manager [None req-b3f33a6f-532a-458d-aec6-6ec54b0cadd5 tempest-ServerMetadataTestJSON-316746592 tempest-ServerMetadataTestJSON-316746592-project-member] [instance: 4184e263-6f56-4bc1-99b2-a2c460531516] Allocating IP information in the background. {{(pid=61806) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 791.907510] env[61806]: DEBUG nova.network.neutron [None req-b3f33a6f-532a-458d-aec6-6ec54b0cadd5 tempest-ServerMetadataTestJSON-316746592 tempest-ServerMetadataTestJSON-316746592-project-member] [instance: 4184e263-6f56-4bc1-99b2-a2c460531516] allocate_for_instance() {{(pid=61806) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 791.949701] env[61806]: DEBUG nova.policy [None req-b3f33a6f-532a-458d-aec6-6ec54b0cadd5 tempest-ServerMetadataTestJSON-316746592 tempest-ServerMetadataTestJSON-316746592-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'de6e82a6e7964f50a82a6911361d3f50', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '0dd7568d384e47b791d785730e401649', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61806) authorize /opt/stack/nova/nova/policy.py:201}} [ 792.072942] env[61806]: DEBUG oslo_concurrency.lockutils [None req-0b87073d-3de3-486a-876d-8ca5794993c1 tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] Releasing lock "refresh_cache-2a3f81e3-b84f-4370-a598-2f6a607c3d47" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 792.073337] env[61806]: DEBUG nova.compute.manager [None req-0b87073d-3de3-486a-876d-8ca5794993c1 tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] [instance: 2a3f81e3-b84f-4370-a598-2f6a607c3d47] Instance network_info: |[{"id": "a839bcaa-9990-4bad-be1f-1f06eb1c978e", "address": "fa:16:3e:0b:8a:a6", "network": {"id": "a7ffda72-e57c-41fa-a7ed-d7b12084bc9d", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-62794865-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "6e391e1267014fafbb8f5a3211299819", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8b3cfeb1-f262-4fd9-b506-8e9c0733e2d8", "external-id": "nsx-vlan-transportzone-119", "segmentation_id": 119, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa839bcaa-99", "ovs_interfaceid": "a839bcaa-9990-4bad-be1f-1f06eb1c978e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61806) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 792.073799] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-0b87073d-3de3-486a-876d-8ca5794993c1 tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] [instance: 2a3f81e3-b84f-4370-a598-2f6a607c3d47] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:0b:8a:a6', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '8b3cfeb1-f262-4fd9-b506-8e9c0733e2d8', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'a839bcaa-9990-4bad-be1f-1f06eb1c978e', 'vif_model': 'vmxnet3'}] {{(pid=61806) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 792.084565] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [None req-0b87073d-3de3-486a-876d-8ca5794993c1 tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] Creating folder: Project (6e391e1267014fafbb8f5a3211299819). Parent ref: group-v277609. {{(pid=61806) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 792.084946] env[61806]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-8ca99ede-5295-4f03-a337-a3ab121f5238 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 792.097504] env[61806]: DEBUG oslo_vmware.api [None req-f91d1db1-2f8a-42e8-9cfa-f4915f7b77dc tempest-ImagesTestJSON-618852556 tempest-ImagesTestJSON-618852556-project-member] Task: {'id': task-1294525, 'name': PowerOnVM_Task} progress is 87%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 792.100218] env[61806]: INFO nova.virt.vmwareapi.vm_util [None req-0b87073d-3de3-486a-876d-8ca5794993c1 tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] Created folder: Project (6e391e1267014fafbb8f5a3211299819) in parent group-v277609. [ 792.100409] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [None req-0b87073d-3de3-486a-876d-8ca5794993c1 tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] Creating folder: Instances. Parent ref: group-v277630. {{(pid=61806) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 792.100838] env[61806]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-eecce961-efad-4421-a8f0-4191f8f04d5e {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 792.109101] env[61806]: INFO nova.virt.vmwareapi.vm_util [None req-0b87073d-3de3-486a-876d-8ca5794993c1 tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] Created folder: Instances in parent group-v277630. [ 792.109376] env[61806]: DEBUG oslo.service.loopingcall [None req-0b87073d-3de3-486a-876d-8ca5794993c1 tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61806) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 792.109613] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 2a3f81e3-b84f-4370-a598-2f6a607c3d47] Creating VM on the ESX host {{(pid=61806) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 792.109745] env[61806]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-85a7af52-6992-4bf2-85eb-93164afd1939 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 792.129473] env[61806]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 792.129473] env[61806]: value = "task-1294528" [ 792.129473] env[61806]: _type = "Task" [ 792.129473] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 792.137011] env[61806]: DEBUG oslo_vmware.api [-] Task: {'id': task-1294528, 'name': CreateVM_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 792.189317] env[61806]: INFO nova.scheduler.client.report [None req-b5a6ece2-b279-4bbd-a3d1-16582c4d8894 tempest-ServersTestMultiNic-1480543082 tempest-ServersTestMultiNic-1480543082-project-member] Deleted allocations for instance 24c83131-1451-429f-84fe-dcd070898d65 [ 792.297977] env[61806]: DEBUG nova.network.neutron [None req-b3f33a6f-532a-458d-aec6-6ec54b0cadd5 tempest-ServerMetadataTestJSON-316746592 tempest-ServerMetadataTestJSON-316746592-project-member] [instance: 4184e263-6f56-4bc1-99b2-a2c460531516] Successfully created port: 4427e5e3-c696-45c2-b1a8-c89329397198 {{(pid=61806) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 792.413201] env[61806]: DEBUG nova.compute.manager [None req-b3f33a6f-532a-458d-aec6-6ec54b0cadd5 tempest-ServerMetadataTestJSON-316746592 tempest-ServerMetadataTestJSON-316746592-project-member] [instance: 4184e263-6f56-4bc1-99b2-a2c460531516] Start building block device mappings for instance. {{(pid=61806) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 792.467792] env[61806]: DEBUG nova.compute.manager [req-ff40df20-ff93-4355-8c3e-5e00ad42eaf8 req-e8654a9e-35db-41fa-b45e-fa918fbd2c63 service nova] [instance: 2a3f81e3-b84f-4370-a598-2f6a607c3d47] Received event network-changed-a839bcaa-9990-4bad-be1f-1f06eb1c978e {{(pid=61806) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 792.467792] env[61806]: DEBUG nova.compute.manager [req-ff40df20-ff93-4355-8c3e-5e00ad42eaf8 req-e8654a9e-35db-41fa-b45e-fa918fbd2c63 service nova] [instance: 2a3f81e3-b84f-4370-a598-2f6a607c3d47] Refreshing instance network info cache due to event network-changed-a839bcaa-9990-4bad-be1f-1f06eb1c978e. {{(pid=61806) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 792.467792] env[61806]: DEBUG oslo_concurrency.lockutils [req-ff40df20-ff93-4355-8c3e-5e00ad42eaf8 req-e8654a9e-35db-41fa-b45e-fa918fbd2c63 service nova] Acquiring lock "refresh_cache-2a3f81e3-b84f-4370-a598-2f6a607c3d47" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 792.467792] env[61806]: DEBUG oslo_concurrency.lockutils [req-ff40df20-ff93-4355-8c3e-5e00ad42eaf8 req-e8654a9e-35db-41fa-b45e-fa918fbd2c63 service nova] Acquired lock "refresh_cache-2a3f81e3-b84f-4370-a598-2f6a607c3d47" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 792.467792] env[61806]: DEBUG nova.network.neutron [req-ff40df20-ff93-4355-8c3e-5e00ad42eaf8 req-e8654a9e-35db-41fa-b45e-fa918fbd2c63 service nova] [instance: 2a3f81e3-b84f-4370-a598-2f6a607c3d47] Refreshing network info cache for port a839bcaa-9990-4bad-be1f-1f06eb1c978e {{(pid=61806) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 792.595255] env[61806]: DEBUG oslo_vmware.api [None req-f91d1db1-2f8a-42e8-9cfa-f4915f7b77dc tempest-ImagesTestJSON-618852556 tempest-ImagesTestJSON-618852556-project-member] Task: {'id': task-1294525, 'name': PowerOnVM_Task, 'duration_secs': 0.623314} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 792.597903] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [None req-f91d1db1-2f8a-42e8-9cfa-f4915f7b77dc tempest-ImagesTestJSON-618852556 tempest-ImagesTestJSON-618852556-project-member] [instance: e0ef0a35-82a5-495b-9d5c-5805e8306390] Powered on the VM {{(pid=61806) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 792.597903] env[61806]: INFO nova.compute.manager [None req-f91d1db1-2f8a-42e8-9cfa-f4915f7b77dc tempest-ImagesTestJSON-618852556 tempest-ImagesTestJSON-618852556-project-member] [instance: e0ef0a35-82a5-495b-9d5c-5805e8306390] Took 8.06 seconds to spawn the instance on the hypervisor. [ 792.597903] env[61806]: DEBUG nova.compute.manager [None req-f91d1db1-2f8a-42e8-9cfa-f4915f7b77dc tempest-ImagesTestJSON-618852556 tempest-ImagesTestJSON-618852556-project-member] [instance: e0ef0a35-82a5-495b-9d5c-5805e8306390] Checking state {{(pid=61806) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 792.597903] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-83ae9453-fa09-4945-bbee-292d46b0ec8b {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 792.644702] env[61806]: DEBUG oslo_vmware.api [-] Task: {'id': task-1294528, 'name': CreateVM_Task, 'duration_secs': 0.360471} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 792.645573] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 2a3f81e3-b84f-4370-a598-2f6a607c3d47] Created VM on the ESX host {{(pid=61806) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 792.646303] env[61806]: DEBUG oslo_concurrency.lockutils [None req-0b87073d-3de3-486a-876d-8ca5794993c1 tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 792.646556] env[61806]: DEBUG oslo_concurrency.lockutils [None req-0b87073d-3de3-486a-876d-8ca5794993c1 tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] Acquired lock "[datastore1] devstack-image-cache_base/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 792.648450] env[61806]: DEBUG oslo_concurrency.lockutils [None req-0b87073d-3de3-486a-876d-8ca5794993c1 tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 792.648450] env[61806]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-fa2691b9-b4ee-47ed-83bb-2f1837cdb54c {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 792.654819] env[61806]: DEBUG oslo_vmware.api [None req-0b87073d-3de3-486a-876d-8ca5794993c1 tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] Waiting for the task: (returnval){ [ 792.654819] env[61806]: value = "session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]527ea654-7174-6801-5392-cf7fc77f4789" [ 792.654819] env[61806]: _type = "Task" [ 792.654819] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 792.664429] env[61806]: DEBUG oslo_vmware.api [None req-0b87073d-3de3-486a-876d-8ca5794993c1 tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] Task: {'id': session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]527ea654-7174-6801-5392-cf7fc77f4789, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 792.703969] env[61806]: DEBUG oslo_concurrency.lockutils [None req-b5a6ece2-b279-4bbd-a3d1-16582c4d8894 tempest-ServersTestMultiNic-1480543082 tempest-ServersTestMultiNic-1480543082-project-member] Lock "24c83131-1451-429f-84fe-dcd070898d65" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 157.366s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 792.738169] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b473253a-2b6c-4786-8be0-fbc7664be763 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 792.745512] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-11570407-95a8-4a1b-9210-9cfa08205e26 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 792.774951] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ceed4737-7f6e-45d7-84be-417476902acf {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 792.782178] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-85e92f13-699f-4ca0-9b2b-1eaecc2e7947 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 792.795552] env[61806]: DEBUG nova.compute.provider_tree [None req-3d5b4b69-10ec-4d1a-aa3d-4dddf4155c99 tempest-ServerShowV247Test-1218443059 tempest-ServerShowV247Test-1218443059-project-member] Inventory has not changed in ProviderTree for provider: a2858be1-fd22-4e08-979e-87ad25293407 {{(pid=61806) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 793.118086] env[61806]: INFO nova.compute.manager [None req-f91d1db1-2f8a-42e8-9cfa-f4915f7b77dc tempest-ImagesTestJSON-618852556 tempest-ImagesTestJSON-618852556-project-member] [instance: e0ef0a35-82a5-495b-9d5c-5805e8306390] Took 35.98 seconds to build instance. [ 793.165122] env[61806]: DEBUG oslo_vmware.api [None req-0b87073d-3de3-486a-876d-8ca5794993c1 tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] Task: {'id': session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]527ea654-7174-6801-5392-cf7fc77f4789, 'name': SearchDatastore_Task, 'duration_secs': 0.01059} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 793.165454] env[61806]: DEBUG oslo_concurrency.lockutils [None req-0b87073d-3de3-486a-876d-8ca5794993c1 tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] Releasing lock "[datastore1] devstack-image-cache_base/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 793.165682] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-0b87073d-3de3-486a-876d-8ca5794993c1 tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] [instance: 2a3f81e3-b84f-4370-a598-2f6a607c3d47] Processing image ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa {{(pid=61806) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 793.165975] env[61806]: DEBUG oslo_concurrency.lockutils [None req-0b87073d-3de3-486a-876d-8ca5794993c1 tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa.vmdk" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 793.166068] env[61806]: DEBUG oslo_concurrency.lockutils [None req-0b87073d-3de3-486a-876d-8ca5794993c1 tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] Acquired lock "[datastore1] devstack-image-cache_base/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa.vmdk" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 793.166255] env[61806]: DEBUG nova.virt.vmwareapi.ds_util [None req-0b87073d-3de3-486a-876d-8ca5794993c1 tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61806) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 793.166797] env[61806]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-8e483294-1dc0-4e01-8c90-e5be64143d50 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 793.171709] env[61806]: DEBUG nova.network.neutron [req-ff40df20-ff93-4355-8c3e-5e00ad42eaf8 req-e8654a9e-35db-41fa-b45e-fa918fbd2c63 service nova] [instance: 2a3f81e3-b84f-4370-a598-2f6a607c3d47] Updated VIF entry in instance network info cache for port a839bcaa-9990-4bad-be1f-1f06eb1c978e. {{(pid=61806) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 793.171963] env[61806]: DEBUG nova.network.neutron [req-ff40df20-ff93-4355-8c3e-5e00ad42eaf8 req-e8654a9e-35db-41fa-b45e-fa918fbd2c63 service nova] [instance: 2a3f81e3-b84f-4370-a598-2f6a607c3d47] Updating instance_info_cache with network_info: [{"id": "a839bcaa-9990-4bad-be1f-1f06eb1c978e", "address": "fa:16:3e:0b:8a:a6", "network": {"id": "a7ffda72-e57c-41fa-a7ed-d7b12084bc9d", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-62794865-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "6e391e1267014fafbb8f5a3211299819", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8b3cfeb1-f262-4fd9-b506-8e9c0733e2d8", "external-id": "nsx-vlan-transportzone-119", "segmentation_id": 119, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa839bcaa-99", "ovs_interfaceid": "a839bcaa-9990-4bad-be1f-1f06eb1c978e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61806) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 793.175287] env[61806]: DEBUG nova.virt.vmwareapi.ds_util [None req-0b87073d-3de3-486a-876d-8ca5794993c1 tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61806) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 793.175454] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-0b87073d-3de3-486a-876d-8ca5794993c1 tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61806) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 793.176230] env[61806]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-706b8097-5043-4615-82f9-8b403e6443cb {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 793.181261] env[61806]: DEBUG oslo_vmware.api [None req-0b87073d-3de3-486a-876d-8ca5794993c1 tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] Waiting for the task: (returnval){ [ 793.181261] env[61806]: value = "session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]5283c709-ab1a-c7ee-87c6-f9d1be56df59" [ 793.181261] env[61806]: _type = "Task" [ 793.181261] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 793.190649] env[61806]: DEBUG oslo_vmware.api [None req-0b87073d-3de3-486a-876d-8ca5794993c1 tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] Task: {'id': session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]5283c709-ab1a-c7ee-87c6-f9d1be56df59, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 793.206790] env[61806]: DEBUG nova.compute.manager [None req-5af76806-1cdf-4002-962d-34b8992ee9dc tempest-VolumesAdminNegativeTest-1387177911 tempest-VolumesAdminNegativeTest-1387177911-project-member] [instance: 4373b735-31cf-4b53-b655-38555cf212a5] Starting instance... {{(pid=61806) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 793.298547] env[61806]: DEBUG nova.scheduler.client.report [None req-3d5b4b69-10ec-4d1a-aa3d-4dddf4155c99 tempest-ServerShowV247Test-1218443059 tempest-ServerShowV247Test-1218443059-project-member] Inventory has not changed for provider a2858be1-fd22-4e08-979e-87ad25293407 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 139, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61806) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 793.431027] env[61806]: DEBUG nova.compute.manager [None req-b3f33a6f-532a-458d-aec6-6ec54b0cadd5 tempest-ServerMetadataTestJSON-316746592 tempest-ServerMetadataTestJSON-316746592-project-member] [instance: 4184e263-6f56-4bc1-99b2-a2c460531516] Start spawning the instance on the hypervisor. {{(pid=61806) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 793.452800] env[61806]: DEBUG nova.virt.hardware [None req-b3f33a6f-532a-458d-aec6-6ec54b0cadd5 tempest-ServerMetadataTestJSON-316746592 tempest-ServerMetadataTestJSON-316746592-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-15T18:05:10Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-15T18:04:52Z,direct_url=,disk_format='vmdk',id=ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='84ba83fce2c349a6988173c5d6fc3b07',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-15T18:04:53Z,virtual_size=,visibility=), allow threads: False {{(pid=61806) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 793.453075] env[61806]: DEBUG nova.virt.hardware [None req-b3f33a6f-532a-458d-aec6-6ec54b0cadd5 tempest-ServerMetadataTestJSON-316746592 tempest-ServerMetadataTestJSON-316746592-project-member] Flavor limits 0:0:0 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 793.453253] env[61806]: DEBUG nova.virt.hardware [None req-b3f33a6f-532a-458d-aec6-6ec54b0cadd5 tempest-ServerMetadataTestJSON-316746592 tempest-ServerMetadataTestJSON-316746592-project-member] Image limits 0:0:0 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 793.453439] env[61806]: DEBUG nova.virt.hardware [None req-b3f33a6f-532a-458d-aec6-6ec54b0cadd5 tempest-ServerMetadataTestJSON-316746592 tempest-ServerMetadataTestJSON-316746592-project-member] Flavor pref 0:0:0 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 793.453589] env[61806]: DEBUG nova.virt.hardware [None req-b3f33a6f-532a-458d-aec6-6ec54b0cadd5 tempest-ServerMetadataTestJSON-316746592 tempest-ServerMetadataTestJSON-316746592-project-member] Image pref 0:0:0 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 793.453738] env[61806]: DEBUG nova.virt.hardware [None req-b3f33a6f-532a-458d-aec6-6ec54b0cadd5 tempest-ServerMetadataTestJSON-316746592 tempest-ServerMetadataTestJSON-316746592-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 793.453945] env[61806]: DEBUG nova.virt.hardware [None req-b3f33a6f-532a-458d-aec6-6ec54b0cadd5 tempest-ServerMetadataTestJSON-316746592 tempest-ServerMetadataTestJSON-316746592-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61806) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 793.454122] env[61806]: DEBUG nova.virt.hardware [None req-b3f33a6f-532a-458d-aec6-6ec54b0cadd5 tempest-ServerMetadataTestJSON-316746592 tempest-ServerMetadataTestJSON-316746592-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61806) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 793.455143] env[61806]: DEBUG nova.virt.hardware [None req-b3f33a6f-532a-458d-aec6-6ec54b0cadd5 tempest-ServerMetadataTestJSON-316746592 tempest-ServerMetadataTestJSON-316746592-project-member] Got 1 possible topologies {{(pid=61806) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 793.455143] env[61806]: DEBUG nova.virt.hardware [None req-b3f33a6f-532a-458d-aec6-6ec54b0cadd5 tempest-ServerMetadataTestJSON-316746592 tempest-ServerMetadataTestJSON-316746592-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61806) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 793.455143] env[61806]: DEBUG nova.virt.hardware [None req-b3f33a6f-532a-458d-aec6-6ec54b0cadd5 tempest-ServerMetadataTestJSON-316746592 tempest-ServerMetadataTestJSON-316746592-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61806) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 793.456132] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3dfc8b5c-d834-415b-8784-176db4f96cd3 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 793.463434] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3dc5886b-1831-4e9d-a4b4-78015c7d7cce {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 793.620472] env[61806]: DEBUG oslo_concurrency.lockutils [None req-f91d1db1-2f8a-42e8-9cfa-f4915f7b77dc tempest-ImagesTestJSON-618852556 tempest-ImagesTestJSON-618852556-project-member] Lock "e0ef0a35-82a5-495b-9d5c-5805e8306390" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 153.200s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 793.674815] env[61806]: DEBUG oslo_concurrency.lockutils [req-ff40df20-ff93-4355-8c3e-5e00ad42eaf8 req-e8654a9e-35db-41fa-b45e-fa918fbd2c63 service nova] Releasing lock "refresh_cache-2a3f81e3-b84f-4370-a598-2f6a607c3d47" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 793.691813] env[61806]: DEBUG oslo_vmware.api [None req-0b87073d-3de3-486a-876d-8ca5794993c1 tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] Task: {'id': session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]5283c709-ab1a-c7ee-87c6-f9d1be56df59, 'name': SearchDatastore_Task, 'duration_secs': 0.011314} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 793.692737] env[61806]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-1ab67e6b-539f-4520-95c6-fcaa9563c8cc {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 793.698129] env[61806]: DEBUG oslo_vmware.api [None req-0b87073d-3de3-486a-876d-8ca5794993c1 tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] Waiting for the task: (returnval){ [ 793.698129] env[61806]: value = "session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]525fa399-13a7-5577-6a8d-31879f94d73b" [ 793.698129] env[61806]: _type = "Task" [ 793.698129] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 793.706557] env[61806]: DEBUG oslo_vmware.api [None req-0b87073d-3de3-486a-876d-8ca5794993c1 tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] Task: {'id': session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]525fa399-13a7-5577-6a8d-31879f94d73b, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 793.729995] env[61806]: DEBUG oslo_concurrency.lockutils [None req-5af76806-1cdf-4002-962d-34b8992ee9dc tempest-VolumesAdminNegativeTest-1387177911 tempest-VolumesAdminNegativeTest-1387177911-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 793.803274] env[61806]: DEBUG oslo_concurrency.lockutils [None req-3d5b4b69-10ec-4d1a-aa3d-4dddf4155c99 tempest-ServerShowV247Test-1218443059 tempest-ServerShowV247Test-1218443059-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.405s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 793.803778] env[61806]: DEBUG nova.compute.manager [None req-3d5b4b69-10ec-4d1a-aa3d-4dddf4155c99 tempest-ServerShowV247Test-1218443059 tempest-ServerShowV247Test-1218443059-project-member] [instance: 7c3fd2ff-ebd5-454f-a743-7fbae1088941] Start building networks asynchronously for instance. {{(pid=61806) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 793.806339] env[61806]: DEBUG oslo_concurrency.lockutils [None req-35a8f95d-1ddd-492f-a93c-adb382770732 tempest-ServersV294TestFqdnHostnames-1709116086 tempest-ServersV294TestFqdnHostnames-1709116086-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 24.155s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 794.002165] env[61806]: DEBUG nova.network.neutron [None req-b3f33a6f-532a-458d-aec6-6ec54b0cadd5 tempest-ServerMetadataTestJSON-316746592 tempest-ServerMetadataTestJSON-316746592-project-member] [instance: 4184e263-6f56-4bc1-99b2-a2c460531516] Successfully updated port: 4427e5e3-c696-45c2-b1a8-c89329397198 {{(pid=61806) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 794.122833] env[61806]: DEBUG nova.compute.manager [None req-61b80e06-de2f-412b-be27-b9318eeb5c25 tempest-ServerAddressesTestJSON-661730579 tempest-ServerAddressesTestJSON-661730579-project-member] [instance: a842df44-d8a9-4376-b9fc-5ca19a68a4b7] Starting instance... {{(pid=61806) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 794.209714] env[61806]: DEBUG oslo_vmware.api [None req-0b87073d-3de3-486a-876d-8ca5794993c1 tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] Task: {'id': session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]525fa399-13a7-5577-6a8d-31879f94d73b, 'name': SearchDatastore_Task, 'duration_secs': 0.009168} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 794.210008] env[61806]: DEBUG oslo_concurrency.lockutils [None req-0b87073d-3de3-486a-876d-8ca5794993c1 tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] Releasing lock "[datastore1] devstack-image-cache_base/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa.vmdk" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 794.210278] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [None req-0b87073d-3de3-486a-876d-8ca5794993c1 tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa.vmdk to [datastore1] 2a3f81e3-b84f-4370-a598-2f6a607c3d47/2a3f81e3-b84f-4370-a598-2f6a607c3d47.vmdk {{(pid=61806) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 794.210543] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-b375f097-a0a0-4f33-b552-503c56afe51c {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 794.217600] env[61806]: DEBUG oslo_vmware.api [None req-0b87073d-3de3-486a-876d-8ca5794993c1 tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] Waiting for the task: (returnval){ [ 794.217600] env[61806]: value = "task-1294529" [ 794.217600] env[61806]: _type = "Task" [ 794.217600] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 794.225401] env[61806]: DEBUG oslo_vmware.api [None req-0b87073d-3de3-486a-876d-8ca5794993c1 tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] Task: {'id': task-1294529, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 794.286950] env[61806]: DEBUG nova.compute.manager [None req-9a79bd3c-e872-4196-bd5a-cef8933311f9 tempest-ImagesTestJSON-618852556 tempest-ImagesTestJSON-618852556-project-member] [instance: e0ef0a35-82a5-495b-9d5c-5805e8306390] Checking state {{(pid=61806) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 794.287874] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-16823c0c-2861-497e-896d-084948d50f4b {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 794.310648] env[61806]: DEBUG nova.compute.utils [None req-3d5b4b69-10ec-4d1a-aa3d-4dddf4155c99 tempest-ServerShowV247Test-1218443059 tempest-ServerShowV247Test-1218443059-project-member] Using /dev/sd instead of None {{(pid=61806) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 794.316076] env[61806]: DEBUG nova.compute.manager [None req-3d5b4b69-10ec-4d1a-aa3d-4dddf4155c99 tempest-ServerShowV247Test-1218443059 tempest-ServerShowV247Test-1218443059-project-member] [instance: 7c3fd2ff-ebd5-454f-a743-7fbae1088941] Not allocating networking since 'none' was specified. {{(pid=61806) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1977}} [ 794.508780] env[61806]: DEBUG oslo_concurrency.lockutils [None req-b3f33a6f-532a-458d-aec6-6ec54b0cadd5 tempest-ServerMetadataTestJSON-316746592 tempest-ServerMetadataTestJSON-316746592-project-member] Acquiring lock "refresh_cache-4184e263-6f56-4bc1-99b2-a2c460531516" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 794.508907] env[61806]: DEBUG oslo_concurrency.lockutils [None req-b3f33a6f-532a-458d-aec6-6ec54b0cadd5 tempest-ServerMetadataTestJSON-316746592 tempest-ServerMetadataTestJSON-316746592-project-member] Acquired lock "refresh_cache-4184e263-6f56-4bc1-99b2-a2c460531516" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 794.509048] env[61806]: DEBUG nova.network.neutron [None req-b3f33a6f-532a-458d-aec6-6ec54b0cadd5 tempest-ServerMetadataTestJSON-316746592 tempest-ServerMetadataTestJSON-316746592-project-member] [instance: 4184e263-6f56-4bc1-99b2-a2c460531516] Building network info cache for instance {{(pid=61806) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 794.533068] env[61806]: DEBUG nova.compute.manager [req-84302afb-be07-48a8-8e9c-cf00600a752d req-cc329977-dc03-4c53-ae85-dc84bf7d15e9 service nova] [instance: 4184e263-6f56-4bc1-99b2-a2c460531516] Received event network-vif-plugged-4427e5e3-c696-45c2-b1a8-c89329397198 {{(pid=61806) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 794.533432] env[61806]: DEBUG oslo_concurrency.lockutils [req-84302afb-be07-48a8-8e9c-cf00600a752d req-cc329977-dc03-4c53-ae85-dc84bf7d15e9 service nova] Acquiring lock "4184e263-6f56-4bc1-99b2-a2c460531516-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 794.533745] env[61806]: DEBUG oslo_concurrency.lockutils [req-84302afb-be07-48a8-8e9c-cf00600a752d req-cc329977-dc03-4c53-ae85-dc84bf7d15e9 service nova] Lock "4184e263-6f56-4bc1-99b2-a2c460531516-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 794.534007] env[61806]: DEBUG oslo_concurrency.lockutils [req-84302afb-be07-48a8-8e9c-cf00600a752d req-cc329977-dc03-4c53-ae85-dc84bf7d15e9 service nova] Lock "4184e263-6f56-4bc1-99b2-a2c460531516-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 794.534297] env[61806]: DEBUG nova.compute.manager [req-84302afb-be07-48a8-8e9c-cf00600a752d req-cc329977-dc03-4c53-ae85-dc84bf7d15e9 service nova] [instance: 4184e263-6f56-4bc1-99b2-a2c460531516] No waiting events found dispatching network-vif-plugged-4427e5e3-c696-45c2-b1a8-c89329397198 {{(pid=61806) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 794.534534] env[61806]: WARNING nova.compute.manager [req-84302afb-be07-48a8-8e9c-cf00600a752d req-cc329977-dc03-4c53-ae85-dc84bf7d15e9 service nova] [instance: 4184e263-6f56-4bc1-99b2-a2c460531516] Received unexpected event network-vif-plugged-4427e5e3-c696-45c2-b1a8-c89329397198 for instance with vm_state building and task_state spawning. [ 794.534722] env[61806]: DEBUG nova.compute.manager [req-84302afb-be07-48a8-8e9c-cf00600a752d req-cc329977-dc03-4c53-ae85-dc84bf7d15e9 service nova] [instance: 4184e263-6f56-4bc1-99b2-a2c460531516] Received event network-changed-4427e5e3-c696-45c2-b1a8-c89329397198 {{(pid=61806) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 794.534887] env[61806]: DEBUG nova.compute.manager [req-84302afb-be07-48a8-8e9c-cf00600a752d req-cc329977-dc03-4c53-ae85-dc84bf7d15e9 service nova] [instance: 4184e263-6f56-4bc1-99b2-a2c460531516] Refreshing instance network info cache due to event network-changed-4427e5e3-c696-45c2-b1a8-c89329397198. {{(pid=61806) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 794.535123] env[61806]: DEBUG oslo_concurrency.lockutils [req-84302afb-be07-48a8-8e9c-cf00600a752d req-cc329977-dc03-4c53-ae85-dc84bf7d15e9 service nova] Acquiring lock "refresh_cache-4184e263-6f56-4bc1-99b2-a2c460531516" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 794.587197] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7fe093fd-05a8-455e-b4ff-397a09f8b795 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 794.596541] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2f4eb0c6-5ffe-4d0b-9e23-8defa88a6c7a {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 794.633823] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0167d662-ec18-42cf-842c-02b38f807122 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 794.645664] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f915d90d-7468-4205-aeaa-62c994fa6290 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 794.661271] env[61806]: DEBUG nova.compute.provider_tree [None req-35a8f95d-1ddd-492f-a93c-adb382770732 tempest-ServersV294TestFqdnHostnames-1709116086 tempest-ServersV294TestFqdnHostnames-1709116086-project-member] Inventory has not changed in ProviderTree for provider: a2858be1-fd22-4e08-979e-87ad25293407 {{(pid=61806) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 794.663529] env[61806]: DEBUG oslo_concurrency.lockutils [None req-61b80e06-de2f-412b-be27-b9318eeb5c25 tempest-ServerAddressesTestJSON-661730579 tempest-ServerAddressesTestJSON-661730579-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 794.728924] env[61806]: DEBUG oslo_vmware.api [None req-0b87073d-3de3-486a-876d-8ca5794993c1 tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] Task: {'id': task-1294529, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.501284} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 794.729239] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [None req-0b87073d-3de3-486a-876d-8ca5794993c1 tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa.vmdk to [datastore1] 2a3f81e3-b84f-4370-a598-2f6a607c3d47/2a3f81e3-b84f-4370-a598-2f6a607c3d47.vmdk {{(pid=61806) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 794.729499] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-0b87073d-3de3-486a-876d-8ca5794993c1 tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] [instance: 2a3f81e3-b84f-4370-a598-2f6a607c3d47] Extending root virtual disk to 1048576 {{(pid=61806) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 794.729788] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-2c9b068a-1f1d-4023-84bd-757bd36c6b1a {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 794.736405] env[61806]: DEBUG oslo_vmware.api [None req-0b87073d-3de3-486a-876d-8ca5794993c1 tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] Waiting for the task: (returnval){ [ 794.736405] env[61806]: value = "task-1294530" [ 794.736405] env[61806]: _type = "Task" [ 794.736405] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 794.745444] env[61806]: DEBUG oslo_vmware.api [None req-0b87073d-3de3-486a-876d-8ca5794993c1 tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] Task: {'id': task-1294530, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 794.807502] env[61806]: INFO nova.compute.manager [None req-9a79bd3c-e872-4196-bd5a-cef8933311f9 tempest-ImagesTestJSON-618852556 tempest-ImagesTestJSON-618852556-project-member] [instance: e0ef0a35-82a5-495b-9d5c-5805e8306390] instance snapshotting [ 794.811031] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b0ae5797-33fc-42df-8570-f40fc40756e3 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 794.829507] env[61806]: DEBUG nova.compute.manager [None req-3d5b4b69-10ec-4d1a-aa3d-4dddf4155c99 tempest-ServerShowV247Test-1218443059 tempest-ServerShowV247Test-1218443059-project-member] [instance: 7c3fd2ff-ebd5-454f-a743-7fbae1088941] Start building block device mappings for instance. {{(pid=61806) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 794.832650] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e0485466-e99b-41be-a523-919d239fd6d9 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 795.050033] env[61806]: DEBUG nova.network.neutron [None req-b3f33a6f-532a-458d-aec6-6ec54b0cadd5 tempest-ServerMetadataTestJSON-316746592 tempest-ServerMetadataTestJSON-316746592-project-member] [instance: 4184e263-6f56-4bc1-99b2-a2c460531516] Instance cache missing network info. {{(pid=61806) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 795.165616] env[61806]: DEBUG nova.scheduler.client.report [None req-35a8f95d-1ddd-492f-a93c-adb382770732 tempest-ServersV294TestFqdnHostnames-1709116086 tempest-ServersV294TestFqdnHostnames-1709116086-project-member] Inventory has not changed for provider a2858be1-fd22-4e08-979e-87ad25293407 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 139, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61806) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 795.248791] env[61806]: DEBUG oslo_vmware.api [None req-0b87073d-3de3-486a-876d-8ca5794993c1 tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] Task: {'id': task-1294530, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.060444} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 795.249731] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-0b87073d-3de3-486a-876d-8ca5794993c1 tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] [instance: 2a3f81e3-b84f-4370-a598-2f6a607c3d47] Extended root virtual disk {{(pid=61806) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 795.250561] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0c7c29b5-7bc9-4fc7-8acb-fe821295c607 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 795.276341] env[61806]: DEBUG nova.virt.vmwareapi.volumeops [None req-0b87073d-3de3-486a-876d-8ca5794993c1 tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] [instance: 2a3f81e3-b84f-4370-a598-2f6a607c3d47] Reconfiguring VM instance instance-00000036 to attach disk [datastore1] 2a3f81e3-b84f-4370-a598-2f6a607c3d47/2a3f81e3-b84f-4370-a598-2f6a607c3d47.vmdk or device None with type sparse {{(pid=61806) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 795.276647] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-b5807e9e-a9ce-48c7-b136-d456eeae15b5 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 795.302477] env[61806]: DEBUG oslo_vmware.api [None req-0b87073d-3de3-486a-876d-8ca5794993c1 tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] Waiting for the task: (returnval){ [ 795.302477] env[61806]: value = "task-1294531" [ 795.302477] env[61806]: _type = "Task" [ 795.302477] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 795.310347] env[61806]: DEBUG oslo_vmware.api [None req-0b87073d-3de3-486a-876d-8ca5794993c1 tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] Task: {'id': task-1294531, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 795.343653] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-9a79bd3c-e872-4196-bd5a-cef8933311f9 tempest-ImagesTestJSON-618852556 tempest-ImagesTestJSON-618852556-project-member] [instance: e0ef0a35-82a5-495b-9d5c-5805e8306390] Creating Snapshot of the VM instance {{(pid=61806) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 795.344037] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-a0bbb208-53ec-4709-b580-d66f28adffbc {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 795.353659] env[61806]: DEBUG oslo_vmware.api [None req-9a79bd3c-e872-4196-bd5a-cef8933311f9 tempest-ImagesTestJSON-618852556 tempest-ImagesTestJSON-618852556-project-member] Waiting for the task: (returnval){ [ 795.353659] env[61806]: value = "task-1294532" [ 795.353659] env[61806]: _type = "Task" [ 795.353659] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 795.366155] env[61806]: DEBUG oslo_vmware.api [None req-9a79bd3c-e872-4196-bd5a-cef8933311f9 tempest-ImagesTestJSON-618852556 tempest-ImagesTestJSON-618852556-project-member] Task: {'id': task-1294532, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 796.184137] env[61806]: DEBUG nova.network.neutron [None req-b3f33a6f-532a-458d-aec6-6ec54b0cadd5 tempest-ServerMetadataTestJSON-316746592 tempest-ServerMetadataTestJSON-316746592-project-member] [instance: 4184e263-6f56-4bc1-99b2-a2c460531516] Updating instance_info_cache with network_info: [{"id": "4427e5e3-c696-45c2-b1a8-c89329397198", "address": "fa:16:3e:e0:c6:6c", "network": {"id": "b567f10e-651f-4599-b465-0258697c1537", "bridge": "br-int", "label": "tempest-ServerMetadataTestJSON-1727546333-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "0dd7568d384e47b791d785730e401649", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ba4f6497-e2b4-43b5-9819-6927865ae974", "external-id": "nsx-vlan-transportzone-112", "segmentation_id": 112, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4427e5e3-c6", "ovs_interfaceid": "4427e5e3-c696-45c2-b1a8-c89329397198", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61806) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 796.186139] env[61806]: DEBUG oslo_concurrency.lockutils [None req-35a8f95d-1ddd-492f-a93c-adb382770732 tempest-ServersV294TestFqdnHostnames-1709116086 tempest-ServersV294TestFqdnHostnames-1709116086-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.380s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 796.186714] env[61806]: ERROR nova.compute.manager [None req-35a8f95d-1ddd-492f-a93c-adb382770732 tempest-ServersV294TestFqdnHostnames-1709116086 tempest-ServersV294TestFqdnHostnames-1709116086-project-member] [instance: c25ddf7c-d0cc-4b73-96da-1dcd6012f072] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 59bb93fe-b4ce-4348-bd86-42344407e007, please check neutron logs for more information. [ 796.186714] env[61806]: ERROR nova.compute.manager [instance: c25ddf7c-d0cc-4b73-96da-1dcd6012f072] Traceback (most recent call last): [ 796.186714] env[61806]: ERROR nova.compute.manager [instance: c25ddf7c-d0cc-4b73-96da-1dcd6012f072] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 796.186714] env[61806]: ERROR nova.compute.manager [instance: c25ddf7c-d0cc-4b73-96da-1dcd6012f072] self.driver.spawn(context, instance, image_meta, [ 796.186714] env[61806]: ERROR nova.compute.manager [instance: c25ddf7c-d0cc-4b73-96da-1dcd6012f072] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 796.186714] env[61806]: ERROR nova.compute.manager [instance: c25ddf7c-d0cc-4b73-96da-1dcd6012f072] self._vmops.spawn(context, instance, image_meta, injected_files, [ 796.186714] env[61806]: ERROR nova.compute.manager [instance: c25ddf7c-d0cc-4b73-96da-1dcd6012f072] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 796.186714] env[61806]: ERROR nova.compute.manager [instance: c25ddf7c-d0cc-4b73-96da-1dcd6012f072] vm_ref = self.build_virtual_machine(instance, [ 796.186714] env[61806]: ERROR nova.compute.manager [instance: c25ddf7c-d0cc-4b73-96da-1dcd6012f072] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 796.186714] env[61806]: ERROR nova.compute.manager [instance: c25ddf7c-d0cc-4b73-96da-1dcd6012f072] vif_infos = vmwarevif.get_vif_info(self._session, [ 796.186714] env[61806]: ERROR nova.compute.manager [instance: c25ddf7c-d0cc-4b73-96da-1dcd6012f072] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 796.187040] env[61806]: ERROR nova.compute.manager [instance: c25ddf7c-d0cc-4b73-96da-1dcd6012f072] for vif in network_info: [ 796.187040] env[61806]: ERROR nova.compute.manager [instance: c25ddf7c-d0cc-4b73-96da-1dcd6012f072] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 796.187040] env[61806]: ERROR nova.compute.manager [instance: c25ddf7c-d0cc-4b73-96da-1dcd6012f072] return self._sync_wrapper(fn, *args, **kwargs) [ 796.187040] env[61806]: ERROR nova.compute.manager [instance: c25ddf7c-d0cc-4b73-96da-1dcd6012f072] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 796.187040] env[61806]: ERROR nova.compute.manager [instance: c25ddf7c-d0cc-4b73-96da-1dcd6012f072] self.wait() [ 796.187040] env[61806]: ERROR nova.compute.manager [instance: c25ddf7c-d0cc-4b73-96da-1dcd6012f072] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 796.187040] env[61806]: ERROR nova.compute.manager [instance: c25ddf7c-d0cc-4b73-96da-1dcd6012f072] self[:] = self._gt.wait() [ 796.187040] env[61806]: ERROR nova.compute.manager [instance: c25ddf7c-d0cc-4b73-96da-1dcd6012f072] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 796.187040] env[61806]: ERROR nova.compute.manager [instance: c25ddf7c-d0cc-4b73-96da-1dcd6012f072] return self._exit_event.wait() [ 796.187040] env[61806]: ERROR nova.compute.manager [instance: c25ddf7c-d0cc-4b73-96da-1dcd6012f072] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 796.187040] env[61806]: ERROR nova.compute.manager [instance: c25ddf7c-d0cc-4b73-96da-1dcd6012f072] current.throw(*self._exc) [ 796.187040] env[61806]: ERROR nova.compute.manager [instance: c25ddf7c-d0cc-4b73-96da-1dcd6012f072] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 796.187040] env[61806]: ERROR nova.compute.manager [instance: c25ddf7c-d0cc-4b73-96da-1dcd6012f072] result = function(*args, **kwargs) [ 796.187421] env[61806]: ERROR nova.compute.manager [instance: c25ddf7c-d0cc-4b73-96da-1dcd6012f072] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 796.187421] env[61806]: ERROR nova.compute.manager [instance: c25ddf7c-d0cc-4b73-96da-1dcd6012f072] return func(*args, **kwargs) [ 796.187421] env[61806]: ERROR nova.compute.manager [instance: c25ddf7c-d0cc-4b73-96da-1dcd6012f072] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 796.187421] env[61806]: ERROR nova.compute.manager [instance: c25ddf7c-d0cc-4b73-96da-1dcd6012f072] raise e [ 796.187421] env[61806]: ERROR nova.compute.manager [instance: c25ddf7c-d0cc-4b73-96da-1dcd6012f072] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 796.187421] env[61806]: ERROR nova.compute.manager [instance: c25ddf7c-d0cc-4b73-96da-1dcd6012f072] nwinfo = self.network_api.allocate_for_instance( [ 796.187421] env[61806]: ERROR nova.compute.manager [instance: c25ddf7c-d0cc-4b73-96da-1dcd6012f072] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 796.187421] env[61806]: ERROR nova.compute.manager [instance: c25ddf7c-d0cc-4b73-96da-1dcd6012f072] created_port_ids = self._update_ports_for_instance( [ 796.187421] env[61806]: ERROR nova.compute.manager [instance: c25ddf7c-d0cc-4b73-96da-1dcd6012f072] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 796.187421] env[61806]: ERROR nova.compute.manager [instance: c25ddf7c-d0cc-4b73-96da-1dcd6012f072] with excutils.save_and_reraise_exception(): [ 796.187421] env[61806]: ERROR nova.compute.manager [instance: c25ddf7c-d0cc-4b73-96da-1dcd6012f072] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 796.187421] env[61806]: ERROR nova.compute.manager [instance: c25ddf7c-d0cc-4b73-96da-1dcd6012f072] self.force_reraise() [ 796.187421] env[61806]: ERROR nova.compute.manager [instance: c25ddf7c-d0cc-4b73-96da-1dcd6012f072] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 796.187770] env[61806]: ERROR nova.compute.manager [instance: c25ddf7c-d0cc-4b73-96da-1dcd6012f072] raise self.value [ 796.187770] env[61806]: ERROR nova.compute.manager [instance: c25ddf7c-d0cc-4b73-96da-1dcd6012f072] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 796.187770] env[61806]: ERROR nova.compute.manager [instance: c25ddf7c-d0cc-4b73-96da-1dcd6012f072] updated_port = self._update_port( [ 796.187770] env[61806]: ERROR nova.compute.manager [instance: c25ddf7c-d0cc-4b73-96da-1dcd6012f072] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 796.187770] env[61806]: ERROR nova.compute.manager [instance: c25ddf7c-d0cc-4b73-96da-1dcd6012f072] _ensure_no_port_binding_failure(port) [ 796.187770] env[61806]: ERROR nova.compute.manager [instance: c25ddf7c-d0cc-4b73-96da-1dcd6012f072] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 796.187770] env[61806]: ERROR nova.compute.manager [instance: c25ddf7c-d0cc-4b73-96da-1dcd6012f072] raise exception.PortBindingFailed(port_id=port['id']) [ 796.187770] env[61806]: ERROR nova.compute.manager [instance: c25ddf7c-d0cc-4b73-96da-1dcd6012f072] nova.exception.PortBindingFailed: Binding failed for port 59bb93fe-b4ce-4348-bd86-42344407e007, please check neutron logs for more information. [ 796.187770] env[61806]: ERROR nova.compute.manager [instance: c25ddf7c-d0cc-4b73-96da-1dcd6012f072] [ 796.187770] env[61806]: DEBUG nova.compute.utils [None req-35a8f95d-1ddd-492f-a93c-adb382770732 tempest-ServersV294TestFqdnHostnames-1709116086 tempest-ServersV294TestFqdnHostnames-1709116086-project-member] [instance: c25ddf7c-d0cc-4b73-96da-1dcd6012f072] Binding failed for port 59bb93fe-b4ce-4348-bd86-42344407e007, please check neutron logs for more information. {{(pid=61806) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 796.189195] env[61806]: DEBUG nova.compute.manager [None req-3d5b4b69-10ec-4d1a-aa3d-4dddf4155c99 tempest-ServerShowV247Test-1218443059 tempest-ServerShowV247Test-1218443059-project-member] [instance: 7c3fd2ff-ebd5-454f-a743-7fbae1088941] Start spawning the instance on the hypervisor. {{(pid=61806) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 796.192264] env[61806]: DEBUG nova.compute.manager [None req-35a8f95d-1ddd-492f-a93c-adb382770732 tempest-ServersV294TestFqdnHostnames-1709116086 tempest-ServersV294TestFqdnHostnames-1709116086-project-member] [instance: c25ddf7c-d0cc-4b73-96da-1dcd6012f072] Build of instance c25ddf7c-d0cc-4b73-96da-1dcd6012f072 was re-scheduled: Binding failed for port 59bb93fe-b4ce-4348-bd86-42344407e007, please check neutron logs for more information. {{(pid=61806) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 796.192264] env[61806]: DEBUG nova.compute.manager [None req-35a8f95d-1ddd-492f-a93c-adb382770732 tempest-ServersV294TestFqdnHostnames-1709116086 tempest-ServersV294TestFqdnHostnames-1709116086-project-member] [instance: c25ddf7c-d0cc-4b73-96da-1dcd6012f072] Unplugging VIFs for instance {{(pid=61806) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 796.192264] env[61806]: DEBUG oslo_concurrency.lockutils [None req-35a8f95d-1ddd-492f-a93c-adb382770732 tempest-ServersV294TestFqdnHostnames-1709116086 tempest-ServersV294TestFqdnHostnames-1709116086-project-member] Acquiring lock "refresh_cache-c25ddf7c-d0cc-4b73-96da-1dcd6012f072" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 796.192527] env[61806]: DEBUG oslo_concurrency.lockutils [None req-35a8f95d-1ddd-492f-a93c-adb382770732 tempest-ServersV294TestFqdnHostnames-1709116086 tempest-ServersV294TestFqdnHostnames-1709116086-project-member] Acquired lock "refresh_cache-c25ddf7c-d0cc-4b73-96da-1dcd6012f072" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 796.193259] env[61806]: DEBUG nova.network.neutron [None req-35a8f95d-1ddd-492f-a93c-adb382770732 tempest-ServersV294TestFqdnHostnames-1709116086 tempest-ServersV294TestFqdnHostnames-1709116086-project-member] [instance: c25ddf7c-d0cc-4b73-96da-1dcd6012f072] Building network info cache for instance {{(pid=61806) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 796.193463] env[61806]: DEBUG oslo_concurrency.lockutils [None req-6a30dd3c-3e08-4933-b4cc-6fae094056e7 tempest-ServerShowV247Test-1218443059 tempest-ServerShowV247Test-1218443059-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 15.082s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 796.194870] env[61806]: INFO nova.compute.claims [None req-6a30dd3c-3e08-4933-b4cc-6fae094056e7 tempest-ServerShowV247Test-1218443059 tempest-ServerShowV247Test-1218443059-project-member] [instance: 4febd093-9f2b-494e-b175-e4693b0e3e0d] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 796.213770] env[61806]: DEBUG oslo_vmware.api [None req-0b87073d-3de3-486a-876d-8ca5794993c1 tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] Task: {'id': task-1294531, 'name': ReconfigVM_Task, 'duration_secs': 0.323417} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 796.214354] env[61806]: DEBUG oslo_vmware.api [None req-9a79bd3c-e872-4196-bd5a-cef8933311f9 tempest-ImagesTestJSON-618852556 tempest-ImagesTestJSON-618852556-project-member] Task: {'id': task-1294532, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 796.214674] env[61806]: DEBUG nova.virt.vmwareapi.volumeops [None req-0b87073d-3de3-486a-876d-8ca5794993c1 tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] [instance: 2a3f81e3-b84f-4370-a598-2f6a607c3d47] Reconfigured VM instance instance-00000036 to attach disk [datastore1] 2a3f81e3-b84f-4370-a598-2f6a607c3d47/2a3f81e3-b84f-4370-a598-2f6a607c3d47.vmdk or device None with type sparse {{(pid=61806) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 796.215195] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-aa0c5a61-028a-4540-b471-babed7db0c2c {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 796.222643] env[61806]: DEBUG nova.virt.hardware [None req-3d5b4b69-10ec-4d1a-aa3d-4dddf4155c99 tempest-ServerShowV247Test-1218443059 tempest-ServerShowV247Test-1218443059-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-15T18:05:10Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-15T18:04:52Z,direct_url=,disk_format='vmdk',id=ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='84ba83fce2c349a6988173c5d6fc3b07',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-15T18:04:53Z,virtual_size=,visibility=), allow threads: False {{(pid=61806) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 796.222643] env[61806]: DEBUG nova.virt.hardware [None req-3d5b4b69-10ec-4d1a-aa3d-4dddf4155c99 tempest-ServerShowV247Test-1218443059 tempest-ServerShowV247Test-1218443059-project-member] Flavor limits 0:0:0 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 796.222643] env[61806]: DEBUG nova.virt.hardware [None req-3d5b4b69-10ec-4d1a-aa3d-4dddf4155c99 tempest-ServerShowV247Test-1218443059 tempest-ServerShowV247Test-1218443059-project-member] Image limits 0:0:0 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 796.222805] env[61806]: DEBUG nova.virt.hardware [None req-3d5b4b69-10ec-4d1a-aa3d-4dddf4155c99 tempest-ServerShowV247Test-1218443059 tempest-ServerShowV247Test-1218443059-project-member] Flavor pref 0:0:0 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 796.222965] env[61806]: DEBUG nova.virt.hardware [None req-3d5b4b69-10ec-4d1a-aa3d-4dddf4155c99 tempest-ServerShowV247Test-1218443059 tempest-ServerShowV247Test-1218443059-project-member] Image pref 0:0:0 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 796.223077] env[61806]: DEBUG nova.virt.hardware [None req-3d5b4b69-10ec-4d1a-aa3d-4dddf4155c99 tempest-ServerShowV247Test-1218443059 tempest-ServerShowV247Test-1218443059-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 796.223284] env[61806]: DEBUG nova.virt.hardware [None req-3d5b4b69-10ec-4d1a-aa3d-4dddf4155c99 tempest-ServerShowV247Test-1218443059 tempest-ServerShowV247Test-1218443059-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61806) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 796.223436] env[61806]: DEBUG nova.virt.hardware [None req-3d5b4b69-10ec-4d1a-aa3d-4dddf4155c99 tempest-ServerShowV247Test-1218443059 tempest-ServerShowV247Test-1218443059-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61806) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 796.225409] env[61806]: DEBUG nova.virt.hardware [None req-3d5b4b69-10ec-4d1a-aa3d-4dddf4155c99 tempest-ServerShowV247Test-1218443059 tempest-ServerShowV247Test-1218443059-project-member] Got 1 possible topologies {{(pid=61806) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 796.225409] env[61806]: DEBUG nova.virt.hardware [None req-3d5b4b69-10ec-4d1a-aa3d-4dddf4155c99 tempest-ServerShowV247Test-1218443059 tempest-ServerShowV247Test-1218443059-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61806) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 796.225409] env[61806]: DEBUG nova.virt.hardware [None req-3d5b4b69-10ec-4d1a-aa3d-4dddf4155c99 tempest-ServerShowV247Test-1218443059 tempest-ServerShowV247Test-1218443059-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61806) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 796.225409] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0c9081cc-5ec2-4470-84dc-256cc942227f {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 796.228360] env[61806]: DEBUG oslo_vmware.api [None req-0b87073d-3de3-486a-876d-8ca5794993c1 tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] Waiting for the task: (returnval){ [ 796.228360] env[61806]: value = "task-1294533" [ 796.228360] env[61806]: _type = "Task" [ 796.228360] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 796.235208] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-63a27012-9e82-41f2-92d5-8e5ce813fdaa {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 796.242413] env[61806]: DEBUG oslo_vmware.api [None req-0b87073d-3de3-486a-876d-8ca5794993c1 tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] Task: {'id': task-1294533, 'name': Rename_Task} progress is 10%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 796.252423] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-3d5b4b69-10ec-4d1a-aa3d-4dddf4155c99 tempest-ServerShowV247Test-1218443059 tempest-ServerShowV247Test-1218443059-project-member] [instance: 7c3fd2ff-ebd5-454f-a743-7fbae1088941] Instance VIF info [] {{(pid=61806) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 796.258049] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [None req-3d5b4b69-10ec-4d1a-aa3d-4dddf4155c99 tempest-ServerShowV247Test-1218443059 tempest-ServerShowV247Test-1218443059-project-member] Creating folder: Project (72590b62f3314d37ad62e0f54e315113). Parent ref: group-v277609. {{(pid=61806) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 796.258590] env[61806]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-c91adc78-8b21-4ad3-a512-573a1d469afb {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 796.269617] env[61806]: INFO nova.virt.vmwareapi.vm_util [None req-3d5b4b69-10ec-4d1a-aa3d-4dddf4155c99 tempest-ServerShowV247Test-1218443059 tempest-ServerShowV247Test-1218443059-project-member] Created folder: Project (72590b62f3314d37ad62e0f54e315113) in parent group-v277609. [ 796.269617] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [None req-3d5b4b69-10ec-4d1a-aa3d-4dddf4155c99 tempest-ServerShowV247Test-1218443059 tempest-ServerShowV247Test-1218443059-project-member] Creating folder: Instances. Parent ref: group-v277633. {{(pid=61806) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 796.269617] env[61806]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-b0a1e178-c906-4fe4-bab1-c46595bb3714 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 796.277902] env[61806]: INFO nova.virt.vmwareapi.vm_util [None req-3d5b4b69-10ec-4d1a-aa3d-4dddf4155c99 tempest-ServerShowV247Test-1218443059 tempest-ServerShowV247Test-1218443059-project-member] Created folder: Instances in parent group-v277633. [ 796.278146] env[61806]: DEBUG oslo.service.loopingcall [None req-3d5b4b69-10ec-4d1a-aa3d-4dddf4155c99 tempest-ServerShowV247Test-1218443059 tempest-ServerShowV247Test-1218443059-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61806) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 796.278328] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 7c3fd2ff-ebd5-454f-a743-7fbae1088941] Creating VM on the ESX host {{(pid=61806) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 796.278519] env[61806]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-80cd6bde-02ab-4669-91d3-306d270043b7 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 796.295042] env[61806]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 796.295042] env[61806]: value = "task-1294536" [ 796.295042] env[61806]: _type = "Task" [ 796.295042] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 796.302027] env[61806]: DEBUG oslo_vmware.api [-] Task: {'id': task-1294536, 'name': CreateVM_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 796.692880] env[61806]: DEBUG oslo_concurrency.lockutils [None req-b3f33a6f-532a-458d-aec6-6ec54b0cadd5 tempest-ServerMetadataTestJSON-316746592 tempest-ServerMetadataTestJSON-316746592-project-member] Releasing lock "refresh_cache-4184e263-6f56-4bc1-99b2-a2c460531516" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 796.693247] env[61806]: DEBUG nova.compute.manager [None req-b3f33a6f-532a-458d-aec6-6ec54b0cadd5 tempest-ServerMetadataTestJSON-316746592 tempest-ServerMetadataTestJSON-316746592-project-member] [instance: 4184e263-6f56-4bc1-99b2-a2c460531516] Instance network_info: |[{"id": "4427e5e3-c696-45c2-b1a8-c89329397198", "address": "fa:16:3e:e0:c6:6c", "network": {"id": "b567f10e-651f-4599-b465-0258697c1537", "bridge": "br-int", "label": "tempest-ServerMetadataTestJSON-1727546333-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "0dd7568d384e47b791d785730e401649", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ba4f6497-e2b4-43b5-9819-6927865ae974", "external-id": "nsx-vlan-transportzone-112", "segmentation_id": 112, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4427e5e3-c6", "ovs_interfaceid": "4427e5e3-c696-45c2-b1a8-c89329397198", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61806) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 796.696711] env[61806]: DEBUG oslo_concurrency.lockutils [req-84302afb-be07-48a8-8e9c-cf00600a752d req-cc329977-dc03-4c53-ae85-dc84bf7d15e9 service nova] Acquired lock "refresh_cache-4184e263-6f56-4bc1-99b2-a2c460531516" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 796.696711] env[61806]: DEBUG nova.network.neutron [req-84302afb-be07-48a8-8e9c-cf00600a752d req-cc329977-dc03-4c53-ae85-dc84bf7d15e9 service nova] [instance: 4184e263-6f56-4bc1-99b2-a2c460531516] Refreshing network info cache for port 4427e5e3-c696-45c2-b1a8-c89329397198 {{(pid=61806) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 796.697890] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-b3f33a6f-532a-458d-aec6-6ec54b0cadd5 tempest-ServerMetadataTestJSON-316746592 tempest-ServerMetadataTestJSON-316746592-project-member] [instance: 4184e263-6f56-4bc1-99b2-a2c460531516] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:e0:c6:6c', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'ba4f6497-e2b4-43b5-9819-6927865ae974', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '4427e5e3-c696-45c2-b1a8-c89329397198', 'vif_model': 'vmxnet3'}] {{(pid=61806) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 796.705481] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [None req-b3f33a6f-532a-458d-aec6-6ec54b0cadd5 tempest-ServerMetadataTestJSON-316746592 tempest-ServerMetadataTestJSON-316746592-project-member] Creating folder: Project (0dd7568d384e47b791d785730e401649). Parent ref: group-v277609. {{(pid=61806) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 796.709758] env[61806]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-5010fe1d-9e6b-4c40-bfdf-b3ca0924a3d7 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 796.719009] env[61806]: DEBUG oslo_vmware.api [None req-9a79bd3c-e872-4196-bd5a-cef8933311f9 tempest-ImagesTestJSON-618852556 tempest-ImagesTestJSON-618852556-project-member] Task: {'id': task-1294532, 'name': CreateSnapshot_Task, 'duration_secs': 1.083026} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 796.719288] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-9a79bd3c-e872-4196-bd5a-cef8933311f9 tempest-ImagesTestJSON-618852556 tempest-ImagesTestJSON-618852556-project-member] [instance: e0ef0a35-82a5-495b-9d5c-5805e8306390] Created Snapshot of the VM instance {{(pid=61806) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 796.720058] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-58f46c80-63f5-44ed-8c6e-b2f4528ed376 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 796.727608] env[61806]: INFO nova.virt.vmwareapi.vm_util [None req-b3f33a6f-532a-458d-aec6-6ec54b0cadd5 tempest-ServerMetadataTestJSON-316746592 tempest-ServerMetadataTestJSON-316746592-project-member] Created folder: Project (0dd7568d384e47b791d785730e401649) in parent group-v277609. [ 796.727608] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [None req-b3f33a6f-532a-458d-aec6-6ec54b0cadd5 tempest-ServerMetadataTestJSON-316746592 tempest-ServerMetadataTestJSON-316746592-project-member] Creating folder: Instances. Parent ref: group-v277637. {{(pid=61806) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 796.728247] env[61806]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-6039db0e-d39e-4ee8-ba15-50abba98ba4b {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 796.730551] env[61806]: DEBUG nova.network.neutron [None req-35a8f95d-1ddd-492f-a93c-adb382770732 tempest-ServersV294TestFqdnHostnames-1709116086 tempest-ServersV294TestFqdnHostnames-1709116086-project-member] [instance: c25ddf7c-d0cc-4b73-96da-1dcd6012f072] Instance cache missing network info. {{(pid=61806) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 796.747492] env[61806]: DEBUG oslo_vmware.api [None req-0b87073d-3de3-486a-876d-8ca5794993c1 tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] Task: {'id': task-1294533, 'name': Rename_Task, 'duration_secs': 0.153807} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 796.747605] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [None req-0b87073d-3de3-486a-876d-8ca5794993c1 tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] [instance: 2a3f81e3-b84f-4370-a598-2f6a607c3d47] Powering on the VM {{(pid=61806) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 796.748893] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-401109a5-14d6-4c73-b3ec-3077d5a307e9 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 796.750471] env[61806]: INFO nova.virt.vmwareapi.vm_util [None req-b3f33a6f-532a-458d-aec6-6ec54b0cadd5 tempest-ServerMetadataTestJSON-316746592 tempest-ServerMetadataTestJSON-316746592-project-member] Created folder: Instances in parent group-v277637. [ 796.750601] env[61806]: DEBUG oslo.service.loopingcall [None req-b3f33a6f-532a-458d-aec6-6ec54b0cadd5 tempest-ServerMetadataTestJSON-316746592 tempest-ServerMetadataTestJSON-316746592-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61806) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 796.750749] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 4184e263-6f56-4bc1-99b2-a2c460531516] Creating VM on the ESX host {{(pid=61806) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 796.751277] env[61806]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-8ae61cb8-f029-43e3-9fea-8f3885cd92f4 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 796.770737] env[61806]: DEBUG oslo_vmware.api [None req-0b87073d-3de3-486a-876d-8ca5794993c1 tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] Waiting for the task: (returnval){ [ 796.770737] env[61806]: value = "task-1294539" [ 796.770737] env[61806]: _type = "Task" [ 796.770737] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 796.776253] env[61806]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 796.776253] env[61806]: value = "task-1294540" [ 796.776253] env[61806]: _type = "Task" [ 796.776253] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 796.779760] env[61806]: DEBUG oslo_vmware.api [None req-0b87073d-3de3-486a-876d-8ca5794993c1 tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] Task: {'id': task-1294539, 'name': PowerOnVM_Task} progress is 33%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 796.789406] env[61806]: DEBUG oslo_vmware.api [-] Task: {'id': task-1294540, 'name': CreateVM_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 796.804495] env[61806]: DEBUG oslo_vmware.api [-] Task: {'id': task-1294536, 'name': CreateVM_Task, 'duration_secs': 0.297775} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 796.804694] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 7c3fd2ff-ebd5-454f-a743-7fbae1088941] Created VM on the ESX host {{(pid=61806) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 796.805084] env[61806]: DEBUG oslo_concurrency.lockutils [None req-3d5b4b69-10ec-4d1a-aa3d-4dddf4155c99 tempest-ServerShowV247Test-1218443059 tempest-ServerShowV247Test-1218443059-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 796.805252] env[61806]: DEBUG oslo_concurrency.lockutils [None req-3d5b4b69-10ec-4d1a-aa3d-4dddf4155c99 tempest-ServerShowV247Test-1218443059 tempest-ServerShowV247Test-1218443059-project-member] Acquired lock "[datastore2] devstack-image-cache_base/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 796.805615] env[61806]: DEBUG oslo_concurrency.lockutils [None req-3d5b4b69-10ec-4d1a-aa3d-4dddf4155c99 tempest-ServerShowV247Test-1218443059 tempest-ServerShowV247Test-1218443059-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 796.808034] env[61806]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f6428893-24c6-41af-9ce4-6631a1b63816 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 796.812901] env[61806]: DEBUG oslo_vmware.api [None req-3d5b4b69-10ec-4d1a-aa3d-4dddf4155c99 tempest-ServerShowV247Test-1218443059 tempest-ServerShowV247Test-1218443059-project-member] Waiting for the task: (returnval){ [ 796.812901] env[61806]: value = "session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]52ca2a93-1f76-6163-ea2d-2473dc5cef2b" [ 796.812901] env[61806]: _type = "Task" [ 796.812901] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 796.821497] env[61806]: DEBUG oslo_vmware.api [None req-3d5b4b69-10ec-4d1a-aa3d-4dddf4155c99 tempest-ServerShowV247Test-1218443059 tempest-ServerShowV247Test-1218443059-project-member] Task: {'id': session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]52ca2a93-1f76-6163-ea2d-2473dc5cef2b, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 796.847225] env[61806]: DEBUG nova.network.neutron [None req-35a8f95d-1ddd-492f-a93c-adb382770732 tempest-ServersV294TestFqdnHostnames-1709116086 tempest-ServersV294TestFqdnHostnames-1709116086-project-member] [instance: c25ddf7c-d0cc-4b73-96da-1dcd6012f072] Updating instance_info_cache with network_info: [] {{(pid=61806) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 797.252527] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-9a79bd3c-e872-4196-bd5a-cef8933311f9 tempest-ImagesTestJSON-618852556 tempest-ImagesTestJSON-618852556-project-member] [instance: e0ef0a35-82a5-495b-9d5c-5805e8306390] Creating linked-clone VM from snapshot {{(pid=61806) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 797.253586] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-c256b554-aeda-4455-b35a-d988c33b8906 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 797.263968] env[61806]: DEBUG oslo_vmware.api [None req-9a79bd3c-e872-4196-bd5a-cef8933311f9 tempest-ImagesTestJSON-618852556 tempest-ImagesTestJSON-618852556-project-member] Waiting for the task: (returnval){ [ 797.263968] env[61806]: value = "task-1294541" [ 797.263968] env[61806]: _type = "Task" [ 797.263968] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 797.271636] env[61806]: DEBUG oslo_vmware.api [None req-9a79bd3c-e872-4196-bd5a-cef8933311f9 tempest-ImagesTestJSON-618852556 tempest-ImagesTestJSON-618852556-project-member] Task: {'id': task-1294541, 'name': CloneVM_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 797.281754] env[61806]: DEBUG oslo_vmware.api [None req-0b87073d-3de3-486a-876d-8ca5794993c1 tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] Task: {'id': task-1294539, 'name': PowerOnVM_Task, 'duration_secs': 0.505476} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 797.284619] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [None req-0b87073d-3de3-486a-876d-8ca5794993c1 tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] [instance: 2a3f81e3-b84f-4370-a598-2f6a607c3d47] Powered on the VM {{(pid=61806) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 797.284831] env[61806]: INFO nova.compute.manager [None req-0b87073d-3de3-486a-876d-8ca5794993c1 tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] [instance: 2a3f81e3-b84f-4370-a598-2f6a607c3d47] Took 8.37 seconds to spawn the instance on the hypervisor. [ 797.285037] env[61806]: DEBUG nova.compute.manager [None req-0b87073d-3de3-486a-876d-8ca5794993c1 tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] [instance: 2a3f81e3-b84f-4370-a598-2f6a607c3d47] Checking state {{(pid=61806) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 797.287626] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-89fc016f-e74f-49fd-8a8c-622ad657d10b {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 797.296785] env[61806]: DEBUG oslo_vmware.api [-] Task: {'id': task-1294540, 'name': CreateVM_Task, 'duration_secs': 0.390799} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 797.298176] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 4184e263-6f56-4bc1-99b2-a2c460531516] Created VM on the ESX host {{(pid=61806) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 797.304654] env[61806]: DEBUG oslo_concurrency.lockutils [None req-b3f33a6f-532a-458d-aec6-6ec54b0cadd5 tempest-ServerMetadataTestJSON-316746592 tempest-ServerMetadataTestJSON-316746592-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 797.304654] env[61806]: DEBUG oslo_concurrency.lockutils [None req-b3f33a6f-532a-458d-aec6-6ec54b0cadd5 tempest-ServerMetadataTestJSON-316746592 tempest-ServerMetadataTestJSON-316746592-project-member] Acquired lock "[datastore1] devstack-image-cache_base/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 797.304871] env[61806]: DEBUG oslo_concurrency.lockutils [None req-b3f33a6f-532a-458d-aec6-6ec54b0cadd5 tempest-ServerMetadataTestJSON-316746592 tempest-ServerMetadataTestJSON-316746592-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 797.305490] env[61806]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d05475f9-3e87-40dd-bddf-cee98e2171c5 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 797.311686] env[61806]: DEBUG oslo_vmware.api [None req-b3f33a6f-532a-458d-aec6-6ec54b0cadd5 tempest-ServerMetadataTestJSON-316746592 tempest-ServerMetadataTestJSON-316746592-project-member] Waiting for the task: (returnval){ [ 797.311686] env[61806]: value = "session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]52bd4e70-9277-084e-976d-42777ea2a90d" [ 797.311686] env[61806]: _type = "Task" [ 797.311686] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 797.323709] env[61806]: DEBUG oslo_vmware.api [None req-b3f33a6f-532a-458d-aec6-6ec54b0cadd5 tempest-ServerMetadataTestJSON-316746592 tempest-ServerMetadataTestJSON-316746592-project-member] Task: {'id': session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]52bd4e70-9277-084e-976d-42777ea2a90d, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 797.328278] env[61806]: DEBUG oslo_vmware.api [None req-3d5b4b69-10ec-4d1a-aa3d-4dddf4155c99 tempest-ServerShowV247Test-1218443059 tempest-ServerShowV247Test-1218443059-project-member] Task: {'id': session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]52ca2a93-1f76-6163-ea2d-2473dc5cef2b, 'name': SearchDatastore_Task, 'duration_secs': 0.009871} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 797.328591] env[61806]: DEBUG oslo_concurrency.lockutils [None req-3d5b4b69-10ec-4d1a-aa3d-4dddf4155c99 tempest-ServerShowV247Test-1218443059 tempest-ServerShowV247Test-1218443059-project-member] Releasing lock "[datastore2] devstack-image-cache_base/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 797.328827] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-3d5b4b69-10ec-4d1a-aa3d-4dddf4155c99 tempest-ServerShowV247Test-1218443059 tempest-ServerShowV247Test-1218443059-project-member] [instance: 7c3fd2ff-ebd5-454f-a743-7fbae1088941] Processing image ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa {{(pid=61806) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 797.329123] env[61806]: DEBUG oslo_concurrency.lockutils [None req-3d5b4b69-10ec-4d1a-aa3d-4dddf4155c99 tempest-ServerShowV247Test-1218443059 tempest-ServerShowV247Test-1218443059-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa.vmdk" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 797.329310] env[61806]: DEBUG oslo_concurrency.lockutils [None req-3d5b4b69-10ec-4d1a-aa3d-4dddf4155c99 tempest-ServerShowV247Test-1218443059 tempest-ServerShowV247Test-1218443059-project-member] Acquired lock "[datastore2] devstack-image-cache_base/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa.vmdk" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 797.329474] env[61806]: DEBUG nova.virt.vmwareapi.ds_util [None req-3d5b4b69-10ec-4d1a-aa3d-4dddf4155c99 tempest-ServerShowV247Test-1218443059 tempest-ServerShowV247Test-1218443059-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61806) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 797.331976] env[61806]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-64bc7d82-e5d5-473c-87cb-716ede90971b {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 797.343953] env[61806]: DEBUG nova.virt.vmwareapi.ds_util [None req-3d5b4b69-10ec-4d1a-aa3d-4dddf4155c99 tempest-ServerShowV247Test-1218443059 tempest-ServerShowV247Test-1218443059-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61806) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 797.344280] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-3d5b4b69-10ec-4d1a-aa3d-4dddf4155c99 tempest-ServerShowV247Test-1218443059 tempest-ServerShowV247Test-1218443059-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61806) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 797.345204] env[61806]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-583df40d-c3fd-491f-a6d1-b89cbcf94c98 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 797.351642] env[61806]: DEBUG oslo_concurrency.lockutils [None req-35a8f95d-1ddd-492f-a93c-adb382770732 tempest-ServersV294TestFqdnHostnames-1709116086 tempest-ServersV294TestFqdnHostnames-1709116086-project-member] Releasing lock "refresh_cache-c25ddf7c-d0cc-4b73-96da-1dcd6012f072" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 797.351915] env[61806]: DEBUG nova.compute.manager [None req-35a8f95d-1ddd-492f-a93c-adb382770732 tempest-ServersV294TestFqdnHostnames-1709116086 tempest-ServersV294TestFqdnHostnames-1709116086-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61806) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 797.352245] env[61806]: DEBUG nova.compute.manager [None req-35a8f95d-1ddd-492f-a93c-adb382770732 tempest-ServersV294TestFqdnHostnames-1709116086 tempest-ServersV294TestFqdnHostnames-1709116086-project-member] [instance: c25ddf7c-d0cc-4b73-96da-1dcd6012f072] Deallocating network for instance {{(pid=61806) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 797.352434] env[61806]: DEBUG nova.network.neutron [None req-35a8f95d-1ddd-492f-a93c-adb382770732 tempest-ServersV294TestFqdnHostnames-1709116086 tempest-ServersV294TestFqdnHostnames-1709116086-project-member] [instance: c25ddf7c-d0cc-4b73-96da-1dcd6012f072] deallocate_for_instance() {{(pid=61806) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 797.358933] env[61806]: DEBUG oslo_vmware.api [None req-3d5b4b69-10ec-4d1a-aa3d-4dddf4155c99 tempest-ServerShowV247Test-1218443059 tempest-ServerShowV247Test-1218443059-project-member] Waiting for the task: (returnval){ [ 797.358933] env[61806]: value = "session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]52ff6be3-bdd3-82ba-a978-2ee5758d845f" [ 797.358933] env[61806]: _type = "Task" [ 797.358933] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 797.370253] env[61806]: DEBUG oslo_vmware.api [None req-3d5b4b69-10ec-4d1a-aa3d-4dddf4155c99 tempest-ServerShowV247Test-1218443059 tempest-ServerShowV247Test-1218443059-project-member] Task: {'id': session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]52ff6be3-bdd3-82ba-a978-2ee5758d845f, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 797.373868] env[61806]: DEBUG nova.network.neutron [None req-35a8f95d-1ddd-492f-a93c-adb382770732 tempest-ServersV294TestFqdnHostnames-1709116086 tempest-ServersV294TestFqdnHostnames-1709116086-project-member] [instance: c25ddf7c-d0cc-4b73-96da-1dcd6012f072] Instance cache missing network info. {{(pid=61806) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 797.473867] env[61806]: DEBUG nova.network.neutron [req-84302afb-be07-48a8-8e9c-cf00600a752d req-cc329977-dc03-4c53-ae85-dc84bf7d15e9 service nova] [instance: 4184e263-6f56-4bc1-99b2-a2c460531516] Updated VIF entry in instance network info cache for port 4427e5e3-c696-45c2-b1a8-c89329397198. {{(pid=61806) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 797.474253] env[61806]: DEBUG nova.network.neutron [req-84302afb-be07-48a8-8e9c-cf00600a752d req-cc329977-dc03-4c53-ae85-dc84bf7d15e9 service nova] [instance: 4184e263-6f56-4bc1-99b2-a2c460531516] Updating instance_info_cache with network_info: [{"id": "4427e5e3-c696-45c2-b1a8-c89329397198", "address": "fa:16:3e:e0:c6:6c", "network": {"id": "b567f10e-651f-4599-b465-0258697c1537", "bridge": "br-int", "label": "tempest-ServerMetadataTestJSON-1727546333-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "0dd7568d384e47b791d785730e401649", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ba4f6497-e2b4-43b5-9819-6927865ae974", "external-id": "nsx-vlan-transportzone-112", "segmentation_id": 112, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4427e5e3-c6", "ovs_interfaceid": "4427e5e3-c696-45c2-b1a8-c89329397198", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61806) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 797.494313] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bb5c7230-e5aa-4b30-bf2c-cf03896815ed {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 797.503064] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-583e1492-3adc-44af-afac-f4a1ef8d9e68 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 797.532132] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bb54049d-73bc-445d-b908-6ceec655793f {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 797.540532] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ca054d4e-3613-4061-bbe1-67a00f4323e8 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 797.555309] env[61806]: DEBUG nova.compute.provider_tree [None req-6a30dd3c-3e08-4933-b4cc-6fae094056e7 tempest-ServerShowV247Test-1218443059 tempest-ServerShowV247Test-1218443059-project-member] Updating inventory in ProviderTree for provider a2858be1-fd22-4e08-979e-87ad25293407 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 138, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61806) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 797.776393] env[61806]: DEBUG oslo_vmware.api [None req-9a79bd3c-e872-4196-bd5a-cef8933311f9 tempest-ImagesTestJSON-618852556 tempest-ImagesTestJSON-618852556-project-member] Task: {'id': task-1294541, 'name': CloneVM_Task} progress is 94%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 797.811666] env[61806]: INFO nova.compute.manager [None req-0b87073d-3de3-486a-876d-8ca5794993c1 tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] [instance: 2a3f81e3-b84f-4370-a598-2f6a607c3d47] Took 36.97 seconds to build instance. [ 797.827993] env[61806]: DEBUG oslo_vmware.api [None req-b3f33a6f-532a-458d-aec6-6ec54b0cadd5 tempest-ServerMetadataTestJSON-316746592 tempest-ServerMetadataTestJSON-316746592-project-member] Task: {'id': session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]52bd4e70-9277-084e-976d-42777ea2a90d, 'name': SearchDatastore_Task, 'duration_secs': 0.009883} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 797.828430] env[61806]: DEBUG oslo_concurrency.lockutils [None req-b3f33a6f-532a-458d-aec6-6ec54b0cadd5 tempest-ServerMetadataTestJSON-316746592 tempest-ServerMetadataTestJSON-316746592-project-member] Releasing lock "[datastore1] devstack-image-cache_base/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 797.828639] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-b3f33a6f-532a-458d-aec6-6ec54b0cadd5 tempest-ServerMetadataTestJSON-316746592 tempest-ServerMetadataTestJSON-316746592-project-member] [instance: 4184e263-6f56-4bc1-99b2-a2c460531516] Processing image ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa {{(pid=61806) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 797.828978] env[61806]: DEBUG oslo_concurrency.lockutils [None req-b3f33a6f-532a-458d-aec6-6ec54b0cadd5 tempest-ServerMetadataTestJSON-316746592 tempest-ServerMetadataTestJSON-316746592-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa.vmdk" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 797.829353] env[61806]: DEBUG oslo_concurrency.lockutils [None req-b3f33a6f-532a-458d-aec6-6ec54b0cadd5 tempest-ServerMetadataTestJSON-316746592 tempest-ServerMetadataTestJSON-316746592-project-member] Acquired lock "[datastore1] devstack-image-cache_base/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa.vmdk" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 797.829353] env[61806]: DEBUG nova.virt.vmwareapi.ds_util [None req-b3f33a6f-532a-458d-aec6-6ec54b0cadd5 tempest-ServerMetadataTestJSON-316746592 tempest-ServerMetadataTestJSON-316746592-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61806) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 797.829964] env[61806]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-71c2c395-fd02-4f48-b9e1-28c4066f68b0 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 797.838638] env[61806]: DEBUG nova.virt.vmwareapi.ds_util [None req-b3f33a6f-532a-458d-aec6-6ec54b0cadd5 tempest-ServerMetadataTestJSON-316746592 tempest-ServerMetadataTestJSON-316746592-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61806) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 797.838716] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-b3f33a6f-532a-458d-aec6-6ec54b0cadd5 tempest-ServerMetadataTestJSON-316746592 tempest-ServerMetadataTestJSON-316746592-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61806) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 797.839583] env[61806]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-1500c3fc-6d94-421d-8066-f35aa7737943 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 797.845251] env[61806]: DEBUG oslo_vmware.api [None req-b3f33a6f-532a-458d-aec6-6ec54b0cadd5 tempest-ServerMetadataTestJSON-316746592 tempest-ServerMetadataTestJSON-316746592-project-member] Waiting for the task: (returnval){ [ 797.845251] env[61806]: value = "session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]52c15527-b246-fcf7-e8e0-0dafe5e96d10" [ 797.845251] env[61806]: _type = "Task" [ 797.845251] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 797.855361] env[61806]: DEBUG oslo_vmware.api [None req-b3f33a6f-532a-458d-aec6-6ec54b0cadd5 tempest-ServerMetadataTestJSON-316746592 tempest-ServerMetadataTestJSON-316746592-project-member] Task: {'id': session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]52c15527-b246-fcf7-e8e0-0dafe5e96d10, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 797.870296] env[61806]: DEBUG oslo_vmware.api [None req-3d5b4b69-10ec-4d1a-aa3d-4dddf4155c99 tempest-ServerShowV247Test-1218443059 tempest-ServerShowV247Test-1218443059-project-member] Task: {'id': session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]52ff6be3-bdd3-82ba-a978-2ee5758d845f, 'name': SearchDatastore_Task, 'duration_secs': 0.010952} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 797.871339] env[61806]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-0595ded0-0ebb-41f0-9c74-d4ec1c936435 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 797.876754] env[61806]: DEBUG nova.network.neutron [None req-35a8f95d-1ddd-492f-a93c-adb382770732 tempest-ServersV294TestFqdnHostnames-1709116086 tempest-ServersV294TestFqdnHostnames-1709116086-project-member] [instance: c25ddf7c-d0cc-4b73-96da-1dcd6012f072] Updating instance_info_cache with network_info: [] {{(pid=61806) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 797.882179] env[61806]: DEBUG oslo_vmware.api [None req-3d5b4b69-10ec-4d1a-aa3d-4dddf4155c99 tempest-ServerShowV247Test-1218443059 tempest-ServerShowV247Test-1218443059-project-member] Waiting for the task: (returnval){ [ 797.882179] env[61806]: value = "session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]52e24e29-ff00-2992-3f13-2e6af66f93aa" [ 797.882179] env[61806]: _type = "Task" [ 797.882179] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 797.892404] env[61806]: DEBUG oslo_vmware.api [None req-3d5b4b69-10ec-4d1a-aa3d-4dddf4155c99 tempest-ServerShowV247Test-1218443059 tempest-ServerShowV247Test-1218443059-project-member] Task: {'id': session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]52e24e29-ff00-2992-3f13-2e6af66f93aa, 'name': SearchDatastore_Task, 'duration_secs': 0.010188} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 797.892404] env[61806]: DEBUG oslo_concurrency.lockutils [None req-3d5b4b69-10ec-4d1a-aa3d-4dddf4155c99 tempest-ServerShowV247Test-1218443059 tempest-ServerShowV247Test-1218443059-project-member] Releasing lock "[datastore2] devstack-image-cache_base/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa.vmdk" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 797.892564] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [None req-3d5b4b69-10ec-4d1a-aa3d-4dddf4155c99 tempest-ServerShowV247Test-1218443059 tempest-ServerShowV247Test-1218443059-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa.vmdk to [datastore2] 7c3fd2ff-ebd5-454f-a743-7fbae1088941/7c3fd2ff-ebd5-454f-a743-7fbae1088941.vmdk {{(pid=61806) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 797.893302] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-8226d469-fe38-4927-b7a5-1bd1782d780d {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 797.899119] env[61806]: DEBUG oslo_vmware.api [None req-3d5b4b69-10ec-4d1a-aa3d-4dddf4155c99 tempest-ServerShowV247Test-1218443059 tempest-ServerShowV247Test-1218443059-project-member] Waiting for the task: (returnval){ [ 797.899119] env[61806]: value = "task-1294542" [ 797.899119] env[61806]: _type = "Task" [ 797.899119] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 797.907336] env[61806]: DEBUG oslo_vmware.api [None req-3d5b4b69-10ec-4d1a-aa3d-4dddf4155c99 tempest-ServerShowV247Test-1218443059 tempest-ServerShowV247Test-1218443059-project-member] Task: {'id': task-1294542, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 797.977346] env[61806]: DEBUG oslo_concurrency.lockutils [req-84302afb-be07-48a8-8e9c-cf00600a752d req-cc329977-dc03-4c53-ae85-dc84bf7d15e9 service nova] Releasing lock "refresh_cache-4184e263-6f56-4bc1-99b2-a2c460531516" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 798.083092] env[61806]: ERROR nova.scheduler.client.report [None req-6a30dd3c-3e08-4933-b4cc-6fae094056e7 tempest-ServerShowV247Test-1218443059 tempest-ServerShowV247Test-1218443059-project-member] [req-a7a5671d-774e-4aa5-8522-f2ec0976a815] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 138, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID a2858be1-fd22-4e08-979e-87ad25293407. Got 409: {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-a7a5671d-774e-4aa5-8522-f2ec0976a815"}]} [ 798.103021] env[61806]: DEBUG nova.scheduler.client.report [None req-6a30dd3c-3e08-4933-b4cc-6fae094056e7 tempest-ServerShowV247Test-1218443059 tempest-ServerShowV247Test-1218443059-project-member] Refreshing inventories for resource provider a2858be1-fd22-4e08-979e-87ad25293407 {{(pid=61806) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:818}} [ 798.122226] env[61806]: DEBUG nova.scheduler.client.report [None req-6a30dd3c-3e08-4933-b4cc-6fae094056e7 tempest-ServerShowV247Test-1218443059 tempest-ServerShowV247Test-1218443059-project-member] Updating ProviderTree inventory for provider a2858be1-fd22-4e08-979e-87ad25293407 from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 139, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61806) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:782}} [ 798.122226] env[61806]: DEBUG nova.compute.provider_tree [None req-6a30dd3c-3e08-4933-b4cc-6fae094056e7 tempest-ServerShowV247Test-1218443059 tempest-ServerShowV247Test-1218443059-project-member] Updating inventory in ProviderTree for provider a2858be1-fd22-4e08-979e-87ad25293407 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 139, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61806) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 798.138025] env[61806]: DEBUG nova.scheduler.client.report [None req-6a30dd3c-3e08-4933-b4cc-6fae094056e7 tempest-ServerShowV247Test-1218443059 tempest-ServerShowV247Test-1218443059-project-member] Refreshing aggregate associations for resource provider a2858be1-fd22-4e08-979e-87ad25293407, aggregates: None {{(pid=61806) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:827}} [ 798.162120] env[61806]: DEBUG nova.scheduler.client.report [None req-6a30dd3c-3e08-4933-b4cc-6fae094056e7 tempest-ServerShowV247Test-1218443059 tempest-ServerShowV247Test-1218443059-project-member] Refreshing trait associations for resource provider a2858be1-fd22-4e08-979e-87ad25293407, traits: COMPUTE_IMAGE_TYPE_ISO,COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_SAME_HOST_COLD_MIGRATE,COMPUTE_NET_ATTACH_INTERFACE,COMPUTE_NODE {{(pid=61806) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:839}} [ 798.278488] env[61806]: DEBUG oslo_vmware.api [None req-9a79bd3c-e872-4196-bd5a-cef8933311f9 tempest-ImagesTestJSON-618852556 tempest-ImagesTestJSON-618852556-project-member] Task: {'id': task-1294541, 'name': CloneVM_Task} progress is 94%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 798.320625] env[61806]: DEBUG oslo_concurrency.lockutils [None req-0b87073d-3de3-486a-876d-8ca5794993c1 tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] Lock "2a3f81e3-b84f-4370-a598-2f6a607c3d47" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 145.695s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 798.358827] env[61806]: DEBUG oslo_vmware.api [None req-b3f33a6f-532a-458d-aec6-6ec54b0cadd5 tempest-ServerMetadataTestJSON-316746592 tempest-ServerMetadataTestJSON-316746592-project-member] Task: {'id': session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]52c15527-b246-fcf7-e8e0-0dafe5e96d10, 'name': SearchDatastore_Task, 'duration_secs': 0.01127} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 798.360290] env[61806]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-1766a134-2604-4289-adaa-419dc1052671 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 798.371015] env[61806]: DEBUG oslo_vmware.api [None req-b3f33a6f-532a-458d-aec6-6ec54b0cadd5 tempest-ServerMetadataTestJSON-316746592 tempest-ServerMetadataTestJSON-316746592-project-member] Waiting for the task: (returnval){ [ 798.371015] env[61806]: value = "session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]528601ed-343d-deba-ce63-36931099d9ab" [ 798.371015] env[61806]: _type = "Task" [ 798.371015] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 798.382642] env[61806]: DEBUG oslo_vmware.api [None req-b3f33a6f-532a-458d-aec6-6ec54b0cadd5 tempest-ServerMetadataTestJSON-316746592 tempest-ServerMetadataTestJSON-316746592-project-member] Task: {'id': session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]528601ed-343d-deba-ce63-36931099d9ab, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 798.386409] env[61806]: INFO nova.compute.manager [None req-35a8f95d-1ddd-492f-a93c-adb382770732 tempest-ServersV294TestFqdnHostnames-1709116086 tempest-ServersV294TestFqdnHostnames-1709116086-project-member] [instance: c25ddf7c-d0cc-4b73-96da-1dcd6012f072] Took 1.03 seconds to deallocate network for instance. [ 798.413917] env[61806]: DEBUG oslo_vmware.api [None req-3d5b4b69-10ec-4d1a-aa3d-4dddf4155c99 tempest-ServerShowV247Test-1218443059 tempest-ServerShowV247Test-1218443059-project-member] Task: {'id': task-1294542, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.496999} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 798.414075] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [None req-3d5b4b69-10ec-4d1a-aa3d-4dddf4155c99 tempest-ServerShowV247Test-1218443059 tempest-ServerShowV247Test-1218443059-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa.vmdk to [datastore2] 7c3fd2ff-ebd5-454f-a743-7fbae1088941/7c3fd2ff-ebd5-454f-a743-7fbae1088941.vmdk {{(pid=61806) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 798.414284] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-3d5b4b69-10ec-4d1a-aa3d-4dddf4155c99 tempest-ServerShowV247Test-1218443059 tempest-ServerShowV247Test-1218443059-project-member] [instance: 7c3fd2ff-ebd5-454f-a743-7fbae1088941] Extending root virtual disk to 1048576 {{(pid=61806) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 798.414529] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-aa701042-e9f3-4a77-8007-736868e51347 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 798.424231] env[61806]: DEBUG oslo_vmware.api [None req-3d5b4b69-10ec-4d1a-aa3d-4dddf4155c99 tempest-ServerShowV247Test-1218443059 tempest-ServerShowV247Test-1218443059-project-member] Waiting for the task: (returnval){ [ 798.424231] env[61806]: value = "task-1294543" [ 798.424231] env[61806]: _type = "Task" [ 798.424231] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 798.432274] env[61806]: DEBUG oslo_vmware.api [None req-3d5b4b69-10ec-4d1a-aa3d-4dddf4155c99 tempest-ServerShowV247Test-1218443059 tempest-ServerShowV247Test-1218443059-project-member] Task: {'id': task-1294543, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 798.461071] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ad8dc944-6023-4df0-97ea-5bb374c63780 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 798.468061] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2c751f6d-16f3-4e59-99bf-7e095adaf19a {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 798.499946] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-347177d5-9996-4076-a4fe-55406a627de0 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 798.508398] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1d038f35-9f2f-4b7d-ab57-1a48ea0c0324 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 798.522779] env[61806]: DEBUG nova.compute.provider_tree [None req-6a30dd3c-3e08-4933-b4cc-6fae094056e7 tempest-ServerShowV247Test-1218443059 tempest-ServerShowV247Test-1218443059-project-member] Updating inventory in ProviderTree for provider a2858be1-fd22-4e08-979e-87ad25293407 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 138, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61806) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 798.562861] env[61806]: DEBUG nova.compute.manager [req-a1cc378d-0b00-4410-b6b0-ca255d022aad req-b4f00449-742d-4c8c-b165-4e9c86ba13cf service nova] [instance: 2a3f81e3-b84f-4370-a598-2f6a607c3d47] Received event network-changed-a839bcaa-9990-4bad-be1f-1f06eb1c978e {{(pid=61806) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 798.562965] env[61806]: DEBUG nova.compute.manager [req-a1cc378d-0b00-4410-b6b0-ca255d022aad req-b4f00449-742d-4c8c-b165-4e9c86ba13cf service nova] [instance: 2a3f81e3-b84f-4370-a598-2f6a607c3d47] Refreshing instance network info cache due to event network-changed-a839bcaa-9990-4bad-be1f-1f06eb1c978e. {{(pid=61806) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 798.563190] env[61806]: DEBUG oslo_concurrency.lockutils [req-a1cc378d-0b00-4410-b6b0-ca255d022aad req-b4f00449-742d-4c8c-b165-4e9c86ba13cf service nova] Acquiring lock "refresh_cache-2a3f81e3-b84f-4370-a598-2f6a607c3d47" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 798.563347] env[61806]: DEBUG oslo_concurrency.lockutils [req-a1cc378d-0b00-4410-b6b0-ca255d022aad req-b4f00449-742d-4c8c-b165-4e9c86ba13cf service nova] Acquired lock "refresh_cache-2a3f81e3-b84f-4370-a598-2f6a607c3d47" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 798.563782] env[61806]: DEBUG nova.network.neutron [req-a1cc378d-0b00-4410-b6b0-ca255d022aad req-b4f00449-742d-4c8c-b165-4e9c86ba13cf service nova] [instance: 2a3f81e3-b84f-4370-a598-2f6a607c3d47] Refreshing network info cache for port a839bcaa-9990-4bad-be1f-1f06eb1c978e {{(pid=61806) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 798.778588] env[61806]: DEBUG oslo_vmware.api [None req-9a79bd3c-e872-4196-bd5a-cef8933311f9 tempest-ImagesTestJSON-618852556 tempest-ImagesTestJSON-618852556-project-member] Task: {'id': task-1294541, 'name': CloneVM_Task} progress is 95%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 798.823367] env[61806]: DEBUG nova.compute.manager [None req-8aa43bc8-f134-4682-8a54-ce9aa25e4ee5 tempest-AttachVolumeNegativeTest-1683033704 tempest-AttachVolumeNegativeTest-1683033704-project-member] [instance: 43dcce0d-4fdb-4ee1-8309-92e92d2331a9] Starting instance... {{(pid=61806) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 798.883454] env[61806]: DEBUG oslo_vmware.api [None req-b3f33a6f-532a-458d-aec6-6ec54b0cadd5 tempest-ServerMetadataTestJSON-316746592 tempest-ServerMetadataTestJSON-316746592-project-member] Task: {'id': session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]528601ed-343d-deba-ce63-36931099d9ab, 'name': SearchDatastore_Task, 'duration_secs': 0.013315} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 798.883770] env[61806]: DEBUG oslo_concurrency.lockutils [None req-b3f33a6f-532a-458d-aec6-6ec54b0cadd5 tempest-ServerMetadataTestJSON-316746592 tempest-ServerMetadataTestJSON-316746592-project-member] Releasing lock "[datastore1] devstack-image-cache_base/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa.vmdk" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 798.884051] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [None req-b3f33a6f-532a-458d-aec6-6ec54b0cadd5 tempest-ServerMetadataTestJSON-316746592 tempest-ServerMetadataTestJSON-316746592-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa.vmdk to [datastore1] 4184e263-6f56-4bc1-99b2-a2c460531516/4184e263-6f56-4bc1-99b2-a2c460531516.vmdk {{(pid=61806) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 798.884325] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-0763c667-8ef0-4349-95eb-4431d78bbf13 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 798.899021] env[61806]: DEBUG oslo_vmware.api [None req-b3f33a6f-532a-458d-aec6-6ec54b0cadd5 tempest-ServerMetadataTestJSON-316746592 tempest-ServerMetadataTestJSON-316746592-project-member] Waiting for the task: (returnval){ [ 798.899021] env[61806]: value = "task-1294544" [ 798.899021] env[61806]: _type = "Task" [ 798.899021] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 798.906064] env[61806]: DEBUG oslo_vmware.api [None req-b3f33a6f-532a-458d-aec6-6ec54b0cadd5 tempest-ServerMetadataTestJSON-316746592 tempest-ServerMetadataTestJSON-316746592-project-member] Task: {'id': task-1294544, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 798.936047] env[61806]: DEBUG oslo_vmware.api [None req-3d5b4b69-10ec-4d1a-aa3d-4dddf4155c99 tempest-ServerShowV247Test-1218443059 tempest-ServerShowV247Test-1218443059-project-member] Task: {'id': task-1294543, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.065642} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 798.936336] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-3d5b4b69-10ec-4d1a-aa3d-4dddf4155c99 tempest-ServerShowV247Test-1218443059 tempest-ServerShowV247Test-1218443059-project-member] [instance: 7c3fd2ff-ebd5-454f-a743-7fbae1088941] Extended root virtual disk {{(pid=61806) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 798.937169] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-05d0fc5c-90d1-45a8-a4a1-264a4b51c0a6 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 798.967967] env[61806]: DEBUG nova.virt.vmwareapi.volumeops [None req-3d5b4b69-10ec-4d1a-aa3d-4dddf4155c99 tempest-ServerShowV247Test-1218443059 tempest-ServerShowV247Test-1218443059-project-member] [instance: 7c3fd2ff-ebd5-454f-a743-7fbae1088941] Reconfiguring VM instance instance-00000038 to attach disk [datastore2] 7c3fd2ff-ebd5-454f-a743-7fbae1088941/7c3fd2ff-ebd5-454f-a743-7fbae1088941.vmdk or device None with type sparse {{(pid=61806) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 798.969783] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-32358a6e-adc9-46ec-8ae1-16b4e0da3058 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 798.990956] env[61806]: DEBUG oslo_vmware.api [None req-3d5b4b69-10ec-4d1a-aa3d-4dddf4155c99 tempest-ServerShowV247Test-1218443059 tempest-ServerShowV247Test-1218443059-project-member] Waiting for the task: (returnval){ [ 798.990956] env[61806]: value = "task-1294545" [ 798.990956] env[61806]: _type = "Task" [ 798.990956] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 799.003255] env[61806]: DEBUG oslo_vmware.api [None req-3d5b4b69-10ec-4d1a-aa3d-4dddf4155c99 tempest-ServerShowV247Test-1218443059 tempest-ServerShowV247Test-1218443059-project-member] Task: {'id': task-1294545, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 799.067253] env[61806]: DEBUG nova.scheduler.client.report [None req-6a30dd3c-3e08-4933-b4cc-6fae094056e7 tempest-ServerShowV247Test-1218443059 tempest-ServerShowV247Test-1218443059-project-member] Updated inventory for provider a2858be1-fd22-4e08-979e-87ad25293407 with generation 81 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 138, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61806) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:971}} [ 799.067573] env[61806]: DEBUG nova.compute.provider_tree [None req-6a30dd3c-3e08-4933-b4cc-6fae094056e7 tempest-ServerShowV247Test-1218443059 tempest-ServerShowV247Test-1218443059-project-member] Updating resource provider a2858be1-fd22-4e08-979e-87ad25293407 generation from 81 to 82 during operation: update_inventory {{(pid=61806) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 799.067754] env[61806]: DEBUG nova.compute.provider_tree [None req-6a30dd3c-3e08-4933-b4cc-6fae094056e7 tempest-ServerShowV247Test-1218443059 tempest-ServerShowV247Test-1218443059-project-member] Updating inventory in ProviderTree for provider a2858be1-fd22-4e08-979e-87ad25293407 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 138, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61806) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 799.281956] env[61806]: DEBUG oslo_vmware.api [None req-9a79bd3c-e872-4196-bd5a-cef8933311f9 tempest-ImagesTestJSON-618852556 tempest-ImagesTestJSON-618852556-project-member] Task: {'id': task-1294541, 'name': CloneVM_Task, 'duration_secs': 1.763866} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 799.282401] env[61806]: INFO nova.virt.vmwareapi.vmops [None req-9a79bd3c-e872-4196-bd5a-cef8933311f9 tempest-ImagesTestJSON-618852556 tempest-ImagesTestJSON-618852556-project-member] [instance: e0ef0a35-82a5-495b-9d5c-5805e8306390] Created linked-clone VM from snapshot [ 799.283063] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9ee2e918-2f46-4bde-851e-76d0f9f2b6ff {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 799.292626] env[61806]: DEBUG nova.virt.vmwareapi.images [None req-9a79bd3c-e872-4196-bd5a-cef8933311f9 tempest-ImagesTestJSON-618852556 tempest-ImagesTestJSON-618852556-project-member] [instance: e0ef0a35-82a5-495b-9d5c-5805e8306390] Uploading image e1cfdcc0-a515-4a37-bf5d-903e61ffb7f2 {{(pid=61806) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:442}} [ 799.323052] env[61806]: DEBUG oslo_vmware.rw_handles [None req-9a79bd3c-e872-4196-bd5a-cef8933311f9 tempest-ImagesTestJSON-618852556 tempest-ImagesTestJSON-618852556-project-member] Creating HttpNfcLease lease for exporting VM: (result){ [ 799.323052] env[61806]: value = "vm-277640" [ 799.323052] env[61806]: _type = "VirtualMachine" [ 799.323052] env[61806]: }. {{(pid=61806) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:478}} [ 799.323052] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ExportVm with opID=oslo.vmware-602167af-f03c-4484-a83c-0466e238a2cf {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 799.332277] env[61806]: DEBUG oslo_vmware.rw_handles [None req-9a79bd3c-e872-4196-bd5a-cef8933311f9 tempest-ImagesTestJSON-618852556 tempest-ImagesTestJSON-618852556-project-member] Lease: (returnval){ [ 799.332277] env[61806]: value = "session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]520d6d96-a88b-8f91-81ac-7f1b88bc506a" [ 799.332277] env[61806]: _type = "HttpNfcLease" [ 799.332277] env[61806]: } obtained for exporting VM: (result){ [ 799.332277] env[61806]: value = "vm-277640" [ 799.332277] env[61806]: _type = "VirtualMachine" [ 799.332277] env[61806]: }. {{(pid=61806) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:481}} [ 799.332666] env[61806]: DEBUG oslo_vmware.api [None req-9a79bd3c-e872-4196-bd5a-cef8933311f9 tempest-ImagesTestJSON-618852556 tempest-ImagesTestJSON-618852556-project-member] Waiting for the lease: (returnval){ [ 799.332666] env[61806]: value = "session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]520d6d96-a88b-8f91-81ac-7f1b88bc506a" [ 799.332666] env[61806]: _type = "HttpNfcLease" [ 799.332666] env[61806]: } to be ready. {{(pid=61806) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 799.347874] env[61806]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 799.347874] env[61806]: value = "session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]520d6d96-a88b-8f91-81ac-7f1b88bc506a" [ 799.347874] env[61806]: _type = "HttpNfcLease" [ 799.347874] env[61806]: } is initializing. {{(pid=61806) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 799.352844] env[61806]: DEBUG oslo_concurrency.lockutils [None req-8aa43bc8-f134-4682-8a54-ce9aa25e4ee5 tempest-AttachVolumeNegativeTest-1683033704 tempest-AttachVolumeNegativeTest-1683033704-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 799.413214] env[61806]: DEBUG oslo_vmware.api [None req-b3f33a6f-532a-458d-aec6-6ec54b0cadd5 tempest-ServerMetadataTestJSON-316746592 tempest-ServerMetadataTestJSON-316746592-project-member] Task: {'id': task-1294544, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.508245} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 799.413333] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [None req-b3f33a6f-532a-458d-aec6-6ec54b0cadd5 tempest-ServerMetadataTestJSON-316746592 tempest-ServerMetadataTestJSON-316746592-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa.vmdk to [datastore1] 4184e263-6f56-4bc1-99b2-a2c460531516/4184e263-6f56-4bc1-99b2-a2c460531516.vmdk {{(pid=61806) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 799.413624] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-b3f33a6f-532a-458d-aec6-6ec54b0cadd5 tempest-ServerMetadataTestJSON-316746592 tempest-ServerMetadataTestJSON-316746592-project-member] [instance: 4184e263-6f56-4bc1-99b2-a2c460531516] Extending root virtual disk to 1048576 {{(pid=61806) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 799.413947] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-1fbd401a-5258-4f0a-a236-7d7b800eb805 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 799.423061] env[61806]: DEBUG nova.network.neutron [req-a1cc378d-0b00-4410-b6b0-ca255d022aad req-b4f00449-742d-4c8c-b165-4e9c86ba13cf service nova] [instance: 2a3f81e3-b84f-4370-a598-2f6a607c3d47] Updated VIF entry in instance network info cache for port a839bcaa-9990-4bad-be1f-1f06eb1c978e. {{(pid=61806) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 799.423426] env[61806]: DEBUG nova.network.neutron [req-a1cc378d-0b00-4410-b6b0-ca255d022aad req-b4f00449-742d-4c8c-b165-4e9c86ba13cf service nova] [instance: 2a3f81e3-b84f-4370-a598-2f6a607c3d47] Updating instance_info_cache with network_info: [{"id": "a839bcaa-9990-4bad-be1f-1f06eb1c978e", "address": "fa:16:3e:0b:8a:a6", "network": {"id": "a7ffda72-e57c-41fa-a7ed-d7b12084bc9d", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-62794865-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.193", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "6e391e1267014fafbb8f5a3211299819", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8b3cfeb1-f262-4fd9-b506-8e9c0733e2d8", "external-id": "nsx-vlan-transportzone-119", "segmentation_id": 119, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa839bcaa-99", "ovs_interfaceid": "a839bcaa-9990-4bad-be1f-1f06eb1c978e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61806) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 799.424782] env[61806]: DEBUG oslo_vmware.api [None req-b3f33a6f-532a-458d-aec6-6ec54b0cadd5 tempest-ServerMetadataTestJSON-316746592 tempest-ServerMetadataTestJSON-316746592-project-member] Waiting for the task: (returnval){ [ 799.424782] env[61806]: value = "task-1294547" [ 799.424782] env[61806]: _type = "Task" [ 799.424782] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 799.434272] env[61806]: DEBUG oslo_vmware.api [None req-b3f33a6f-532a-458d-aec6-6ec54b0cadd5 tempest-ServerMetadataTestJSON-316746592 tempest-ServerMetadataTestJSON-316746592-project-member] Task: {'id': task-1294547, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 799.435082] env[61806]: INFO nova.scheduler.client.report [None req-35a8f95d-1ddd-492f-a93c-adb382770732 tempest-ServersV294TestFqdnHostnames-1709116086 tempest-ServersV294TestFqdnHostnames-1709116086-project-member] Deleted allocations for instance c25ddf7c-d0cc-4b73-96da-1dcd6012f072 [ 799.500365] env[61806]: DEBUG oslo_vmware.api [None req-3d5b4b69-10ec-4d1a-aa3d-4dddf4155c99 tempest-ServerShowV247Test-1218443059 tempest-ServerShowV247Test-1218443059-project-member] Task: {'id': task-1294545, 'name': ReconfigVM_Task, 'duration_secs': 0.462157} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 799.500925] env[61806]: DEBUG nova.virt.vmwareapi.volumeops [None req-3d5b4b69-10ec-4d1a-aa3d-4dddf4155c99 tempest-ServerShowV247Test-1218443059 tempest-ServerShowV247Test-1218443059-project-member] [instance: 7c3fd2ff-ebd5-454f-a743-7fbae1088941] Reconfigured VM instance instance-00000038 to attach disk [datastore2] 7c3fd2ff-ebd5-454f-a743-7fbae1088941/7c3fd2ff-ebd5-454f-a743-7fbae1088941.vmdk or device None with type sparse {{(pid=61806) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 799.502403] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-693c9ec2-66a4-4c6c-993c-d04376b57ba0 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 799.508805] env[61806]: DEBUG oslo_vmware.api [None req-3d5b4b69-10ec-4d1a-aa3d-4dddf4155c99 tempest-ServerShowV247Test-1218443059 tempest-ServerShowV247Test-1218443059-project-member] Waiting for the task: (returnval){ [ 799.508805] env[61806]: value = "task-1294548" [ 799.508805] env[61806]: _type = "Task" [ 799.508805] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 799.515676] env[61806]: DEBUG oslo_vmware.api [None req-3d5b4b69-10ec-4d1a-aa3d-4dddf4155c99 tempest-ServerShowV247Test-1218443059 tempest-ServerShowV247Test-1218443059-project-member] Task: {'id': task-1294548, 'name': Rename_Task} progress is 5%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 799.574549] env[61806]: DEBUG oslo_concurrency.lockutils [None req-6a30dd3c-3e08-4933-b4cc-6fae094056e7 tempest-ServerShowV247Test-1218443059 tempest-ServerShowV247Test-1218443059-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 3.381s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 799.575082] env[61806]: DEBUG nova.compute.manager [None req-6a30dd3c-3e08-4933-b4cc-6fae094056e7 tempest-ServerShowV247Test-1218443059 tempest-ServerShowV247Test-1218443059-project-member] [instance: 4febd093-9f2b-494e-b175-e4693b0e3e0d] Start building networks asynchronously for instance. {{(pid=61806) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 799.580184] env[61806]: DEBUG oslo_concurrency.lockutils [None req-f7f6aee3-3d33-4e4b-b451-fa3af2e320d4 tempest-ServerDiagnosticsV248Test-463927155 tempest-ServerDiagnosticsV248Test-463927155-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 16.665s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 799.580184] env[61806]: INFO nova.compute.claims [None req-f7f6aee3-3d33-4e4b-b451-fa3af2e320d4 tempest-ServerDiagnosticsV248Test-463927155 tempest-ServerDiagnosticsV248Test-463927155-project-member] [instance: a934d02d-26aa-4900-b473-a58489e5629e] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 799.845500] env[61806]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 799.845500] env[61806]: value = "session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]520d6d96-a88b-8f91-81ac-7f1b88bc506a" [ 799.845500] env[61806]: _type = "HttpNfcLease" [ 799.845500] env[61806]: } is ready. {{(pid=61806) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 799.845816] env[61806]: DEBUG oslo_vmware.rw_handles [None req-9a79bd3c-e872-4196-bd5a-cef8933311f9 tempest-ImagesTestJSON-618852556 tempest-ImagesTestJSON-618852556-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 799.845816] env[61806]: value = "session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]520d6d96-a88b-8f91-81ac-7f1b88bc506a" [ 799.845816] env[61806]: _type = "HttpNfcLease" [ 799.845816] env[61806]: }. {{(pid=61806) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:486}} [ 799.847214] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e3c2876d-75ad-41a7-8dd5-b15b0177921b {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 799.855805] env[61806]: DEBUG oslo_vmware.rw_handles [None req-9a79bd3c-e872-4196-bd5a-cef8933311f9 tempest-ImagesTestJSON-618852556 tempest-ImagesTestJSON-618852556-project-member] Found VMDK URL: https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/529481d3-96c1-2bc3-572e-202b8bc69f4d/disk-0.vmdk from lease info. {{(pid=61806) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 799.856449] env[61806]: DEBUG oslo_vmware.rw_handles [None req-9a79bd3c-e872-4196-bd5a-cef8933311f9 tempest-ImagesTestJSON-618852556 tempest-ImagesTestJSON-618852556-project-member] Opening URL: https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/529481d3-96c1-2bc3-572e-202b8bc69f4d/disk-0.vmdk for reading. {{(pid=61806) _create_read_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:99}} [ 799.929958] env[61806]: DEBUG oslo_concurrency.lockutils [req-a1cc378d-0b00-4410-b6b0-ca255d022aad req-b4f00449-742d-4c8c-b165-4e9c86ba13cf service nova] Releasing lock "refresh_cache-2a3f81e3-b84f-4370-a598-2f6a607c3d47" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 799.936361] env[61806]: DEBUG oslo_vmware.api [None req-b3f33a6f-532a-458d-aec6-6ec54b0cadd5 tempest-ServerMetadataTestJSON-316746592 tempest-ServerMetadataTestJSON-316746592-project-member] Task: {'id': task-1294547, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.068996} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 799.939044] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-b3f33a6f-532a-458d-aec6-6ec54b0cadd5 tempest-ServerMetadataTestJSON-316746592 tempest-ServerMetadataTestJSON-316746592-project-member] [instance: 4184e263-6f56-4bc1-99b2-a2c460531516] Extended root virtual disk {{(pid=61806) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 799.939044] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c30835ef-9c28-4c0f-aad8-27f853f562ba {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 799.953272] env[61806]: DEBUG oslo_concurrency.lockutils [None req-35a8f95d-1ddd-492f-a93c-adb382770732 tempest-ServersV294TestFqdnHostnames-1709116086 tempest-ServersV294TestFqdnHostnames-1709116086-project-member] Lock "c25ddf7c-d0cc-4b73-96da-1dcd6012f072" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 162.872s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 799.961827] env[61806]: DEBUG nova.virt.vmwareapi.volumeops [None req-b3f33a6f-532a-458d-aec6-6ec54b0cadd5 tempest-ServerMetadataTestJSON-316746592 tempest-ServerMetadataTestJSON-316746592-project-member] [instance: 4184e263-6f56-4bc1-99b2-a2c460531516] Reconfiguring VM instance instance-00000037 to attach disk [datastore1] 4184e263-6f56-4bc1-99b2-a2c460531516/4184e263-6f56-4bc1-99b2-a2c460531516.vmdk or device None with type sparse {{(pid=61806) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 799.963190] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-c2787163-4950-4e2b-9d5d-b8dadce21368 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 799.983626] env[61806]: DEBUG oslo_vmware.api [None req-b3f33a6f-532a-458d-aec6-6ec54b0cadd5 tempest-ServerMetadataTestJSON-316746592 tempest-ServerMetadataTestJSON-316746592-project-member] Waiting for the task: (returnval){ [ 799.983626] env[61806]: value = "task-1294549" [ 799.983626] env[61806]: _type = "Task" [ 799.983626] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 799.991684] env[61806]: DEBUG oslo_vmware.api [None req-b3f33a6f-532a-458d-aec6-6ec54b0cadd5 tempest-ServerMetadataTestJSON-316746592 tempest-ServerMetadataTestJSON-316746592-project-member] Task: {'id': task-1294549, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 800.018601] env[61806]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-b0d26081-d628-42d2-b8f8-bc01db35b441 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 800.020453] env[61806]: DEBUG oslo_vmware.api [None req-3d5b4b69-10ec-4d1a-aa3d-4dddf4155c99 tempest-ServerShowV247Test-1218443059 tempest-ServerShowV247Test-1218443059-project-member] Task: {'id': task-1294548, 'name': Rename_Task, 'duration_secs': 0.131245} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 800.020453] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [None req-3d5b4b69-10ec-4d1a-aa3d-4dddf4155c99 tempest-ServerShowV247Test-1218443059 tempest-ServerShowV247Test-1218443059-project-member] [instance: 7c3fd2ff-ebd5-454f-a743-7fbae1088941] Powering on the VM {{(pid=61806) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 800.021102] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-b0a5dff1-e859-494b-890d-207ca36df48a {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 800.028020] env[61806]: DEBUG oslo_vmware.api [None req-3d5b4b69-10ec-4d1a-aa3d-4dddf4155c99 tempest-ServerShowV247Test-1218443059 tempest-ServerShowV247Test-1218443059-project-member] Waiting for the task: (returnval){ [ 800.028020] env[61806]: value = "task-1294550" [ 800.028020] env[61806]: _type = "Task" [ 800.028020] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 800.034095] env[61806]: DEBUG oslo_vmware.api [None req-3d5b4b69-10ec-4d1a-aa3d-4dddf4155c99 tempest-ServerShowV247Test-1218443059 tempest-ServerShowV247Test-1218443059-project-member] Task: {'id': task-1294550, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 800.085911] env[61806]: DEBUG nova.compute.utils [None req-6a30dd3c-3e08-4933-b4cc-6fae094056e7 tempest-ServerShowV247Test-1218443059 tempest-ServerShowV247Test-1218443059-project-member] Using /dev/sd instead of None {{(pid=61806) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 800.091270] env[61806]: DEBUG nova.compute.manager [None req-6a30dd3c-3e08-4933-b4cc-6fae094056e7 tempest-ServerShowV247Test-1218443059 tempest-ServerShowV247Test-1218443059-project-member] [instance: 4febd093-9f2b-494e-b175-e4693b0e3e0d] Not allocating networking since 'none' was specified. {{(pid=61806) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1977}} [ 800.237634] env[61806]: DEBUG oslo_concurrency.lockutils [None req-c9a1ae7a-8313-4bdf-8de4-77929ce2d9bc tempest-ServerMetadataNegativeTestJSON-456453902 tempest-ServerMetadataNegativeTestJSON-456453902-project-member] Acquiring lock "7f4907d0-178d-452a-8149-030becde8779" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 800.238256] env[61806]: DEBUG oslo_concurrency.lockutils [None req-c9a1ae7a-8313-4bdf-8de4-77929ce2d9bc tempest-ServerMetadataNegativeTestJSON-456453902 tempest-ServerMetadataNegativeTestJSON-456453902-project-member] Lock "7f4907d0-178d-452a-8149-030becde8779" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 800.478363] env[61806]: DEBUG nova.compute.manager [None req-9c52c9e9-816e-4638-b13a-a63a9756140c tempest-SecurityGroupsTestJSON-1213485581 tempest-SecurityGroupsTestJSON-1213485581-project-member] [instance: b41531f2-f28c-4d82-9682-0b557bbaa491] Starting instance... {{(pid=61806) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 800.493415] env[61806]: DEBUG oslo_vmware.api [None req-b3f33a6f-532a-458d-aec6-6ec54b0cadd5 tempest-ServerMetadataTestJSON-316746592 tempest-ServerMetadataTestJSON-316746592-project-member] Task: {'id': task-1294549, 'name': ReconfigVM_Task, 'duration_secs': 0.446282} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 800.493840] env[61806]: DEBUG nova.virt.vmwareapi.volumeops [None req-b3f33a6f-532a-458d-aec6-6ec54b0cadd5 tempest-ServerMetadataTestJSON-316746592 tempest-ServerMetadataTestJSON-316746592-project-member] [instance: 4184e263-6f56-4bc1-99b2-a2c460531516] Reconfigured VM instance instance-00000037 to attach disk [datastore1] 4184e263-6f56-4bc1-99b2-a2c460531516/4184e263-6f56-4bc1-99b2-a2c460531516.vmdk or device None with type sparse {{(pid=61806) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 800.494564] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-696d59b8-8c0c-4532-9b42-871500c57817 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 800.501179] env[61806]: DEBUG oslo_vmware.api [None req-b3f33a6f-532a-458d-aec6-6ec54b0cadd5 tempest-ServerMetadataTestJSON-316746592 tempest-ServerMetadataTestJSON-316746592-project-member] Waiting for the task: (returnval){ [ 800.501179] env[61806]: value = "task-1294551" [ 800.501179] env[61806]: _type = "Task" [ 800.501179] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 800.509144] env[61806]: DEBUG oslo_vmware.api [None req-b3f33a6f-532a-458d-aec6-6ec54b0cadd5 tempest-ServerMetadataTestJSON-316746592 tempest-ServerMetadataTestJSON-316746592-project-member] Task: {'id': task-1294551, 'name': Rename_Task} progress is 5%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 800.537975] env[61806]: DEBUG oslo_vmware.api [None req-3d5b4b69-10ec-4d1a-aa3d-4dddf4155c99 tempest-ServerShowV247Test-1218443059 tempest-ServerShowV247Test-1218443059-project-member] Task: {'id': task-1294550, 'name': PowerOnVM_Task, 'duration_secs': 0.449349} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 800.538409] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [None req-3d5b4b69-10ec-4d1a-aa3d-4dddf4155c99 tempest-ServerShowV247Test-1218443059 tempest-ServerShowV247Test-1218443059-project-member] [instance: 7c3fd2ff-ebd5-454f-a743-7fbae1088941] Powered on the VM {{(pid=61806) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 800.538711] env[61806]: INFO nova.compute.manager [None req-3d5b4b69-10ec-4d1a-aa3d-4dddf4155c99 tempest-ServerShowV247Test-1218443059 tempest-ServerShowV247Test-1218443059-project-member] [instance: 7c3fd2ff-ebd5-454f-a743-7fbae1088941] Took 4.35 seconds to spawn the instance on the hypervisor. [ 800.539222] env[61806]: DEBUG nova.compute.manager [None req-3d5b4b69-10ec-4d1a-aa3d-4dddf4155c99 tempest-ServerShowV247Test-1218443059 tempest-ServerShowV247Test-1218443059-project-member] [instance: 7c3fd2ff-ebd5-454f-a743-7fbae1088941] Checking state {{(pid=61806) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 800.540636] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c37c48f8-8e9c-4289-ac83-47d441d8c900 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 800.591432] env[61806]: DEBUG nova.compute.manager [None req-6a30dd3c-3e08-4933-b4cc-6fae094056e7 tempest-ServerShowV247Test-1218443059 tempest-ServerShowV247Test-1218443059-project-member] [instance: 4febd093-9f2b-494e-b175-e4693b0e3e0d] Start building block device mappings for instance. {{(pid=61806) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 800.923827] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f9d6269a-6b25-4928-834a-dcac160f0ad4 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 800.931954] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-08e8f7a4-6f7a-4e04-9a62-4c5504f22fa4 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 800.965634] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5d3e6d50-1413-4bf8-aa1d-752a8c4bffca {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 800.974155] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0811f39c-74a3-4b76-abcd-9212f18018e3 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 800.992379] env[61806]: DEBUG nova.compute.provider_tree [None req-f7f6aee3-3d33-4e4b-b451-fa3af2e320d4 tempest-ServerDiagnosticsV248Test-463927155 tempest-ServerDiagnosticsV248Test-463927155-project-member] Inventory has not changed in ProviderTree for provider: a2858be1-fd22-4e08-979e-87ad25293407 {{(pid=61806) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 801.012539] env[61806]: DEBUG oslo_concurrency.lockutils [None req-9c52c9e9-816e-4638-b13a-a63a9756140c tempest-SecurityGroupsTestJSON-1213485581 tempest-SecurityGroupsTestJSON-1213485581-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 801.019035] env[61806]: DEBUG oslo_vmware.api [None req-b3f33a6f-532a-458d-aec6-6ec54b0cadd5 tempest-ServerMetadataTestJSON-316746592 tempest-ServerMetadataTestJSON-316746592-project-member] Task: {'id': task-1294551, 'name': Rename_Task, 'duration_secs': 0.245065} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 801.020037] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [None req-b3f33a6f-532a-458d-aec6-6ec54b0cadd5 tempest-ServerMetadataTestJSON-316746592 tempest-ServerMetadataTestJSON-316746592-project-member] [instance: 4184e263-6f56-4bc1-99b2-a2c460531516] Powering on the VM {{(pid=61806) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 801.020178] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-75bbcb58-ad85-46d2-b1bb-e0e49e5734fc {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 801.027671] env[61806]: DEBUG oslo_vmware.api [None req-b3f33a6f-532a-458d-aec6-6ec54b0cadd5 tempest-ServerMetadataTestJSON-316746592 tempest-ServerMetadataTestJSON-316746592-project-member] Waiting for the task: (returnval){ [ 801.027671] env[61806]: value = "task-1294552" [ 801.027671] env[61806]: _type = "Task" [ 801.027671] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 801.037700] env[61806]: DEBUG oslo_vmware.api [None req-b3f33a6f-532a-458d-aec6-6ec54b0cadd5 tempest-ServerMetadataTestJSON-316746592 tempest-ServerMetadataTestJSON-316746592-project-member] Task: {'id': task-1294552, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 801.060807] env[61806]: INFO nova.compute.manager [None req-3d5b4b69-10ec-4d1a-aa3d-4dddf4155c99 tempest-ServerShowV247Test-1218443059 tempest-ServerShowV247Test-1218443059-project-member] [instance: 7c3fd2ff-ebd5-454f-a743-7fbae1088941] Took 34.07 seconds to build instance. [ 801.499316] env[61806]: DEBUG nova.scheduler.client.report [None req-f7f6aee3-3d33-4e4b-b451-fa3af2e320d4 tempest-ServerDiagnosticsV248Test-463927155 tempest-ServerDiagnosticsV248Test-463927155-project-member] Inventory has not changed for provider a2858be1-fd22-4e08-979e-87ad25293407 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 138, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61806) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 801.539968] env[61806]: DEBUG oslo_vmware.api [None req-b3f33a6f-532a-458d-aec6-6ec54b0cadd5 tempest-ServerMetadataTestJSON-316746592 tempest-ServerMetadataTestJSON-316746592-project-member] Task: {'id': task-1294552, 'name': PowerOnVM_Task} progress is 89%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 801.564193] env[61806]: DEBUG oslo_concurrency.lockutils [None req-3d5b4b69-10ec-4d1a-aa3d-4dddf4155c99 tempest-ServerShowV247Test-1218443059 tempest-ServerShowV247Test-1218443059-project-member] Lock "7c3fd2ff-ebd5-454f-a743-7fbae1088941" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 140.456s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 801.601940] env[61806]: DEBUG nova.compute.manager [None req-6a30dd3c-3e08-4933-b4cc-6fae094056e7 tempest-ServerShowV247Test-1218443059 tempest-ServerShowV247Test-1218443059-project-member] [instance: 4febd093-9f2b-494e-b175-e4693b0e3e0d] Start spawning the instance on the hypervisor. {{(pid=61806) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 801.635949] env[61806]: DEBUG nova.virt.hardware [None req-6a30dd3c-3e08-4933-b4cc-6fae094056e7 tempest-ServerShowV247Test-1218443059 tempest-ServerShowV247Test-1218443059-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-15T18:05:10Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-15T18:04:52Z,direct_url=,disk_format='vmdk',id=ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='84ba83fce2c349a6988173c5d6fc3b07',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-15T18:04:53Z,virtual_size=,visibility=), allow threads: False {{(pid=61806) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 801.635949] env[61806]: DEBUG nova.virt.hardware [None req-6a30dd3c-3e08-4933-b4cc-6fae094056e7 tempest-ServerShowV247Test-1218443059 tempest-ServerShowV247Test-1218443059-project-member] Flavor limits 0:0:0 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 801.635949] env[61806]: DEBUG nova.virt.hardware [None req-6a30dd3c-3e08-4933-b4cc-6fae094056e7 tempest-ServerShowV247Test-1218443059 tempest-ServerShowV247Test-1218443059-project-member] Image limits 0:0:0 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 801.635949] env[61806]: DEBUG nova.virt.hardware [None req-6a30dd3c-3e08-4933-b4cc-6fae094056e7 tempest-ServerShowV247Test-1218443059 tempest-ServerShowV247Test-1218443059-project-member] Flavor pref 0:0:0 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 801.636141] env[61806]: DEBUG nova.virt.hardware [None req-6a30dd3c-3e08-4933-b4cc-6fae094056e7 tempest-ServerShowV247Test-1218443059 tempest-ServerShowV247Test-1218443059-project-member] Image pref 0:0:0 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 801.636287] env[61806]: DEBUG nova.virt.hardware [None req-6a30dd3c-3e08-4933-b4cc-6fae094056e7 tempest-ServerShowV247Test-1218443059 tempest-ServerShowV247Test-1218443059-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 801.636537] env[61806]: DEBUG nova.virt.hardware [None req-6a30dd3c-3e08-4933-b4cc-6fae094056e7 tempest-ServerShowV247Test-1218443059 tempest-ServerShowV247Test-1218443059-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61806) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 801.636706] env[61806]: DEBUG nova.virt.hardware [None req-6a30dd3c-3e08-4933-b4cc-6fae094056e7 tempest-ServerShowV247Test-1218443059 tempest-ServerShowV247Test-1218443059-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61806) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 801.636880] env[61806]: DEBUG nova.virt.hardware [None req-6a30dd3c-3e08-4933-b4cc-6fae094056e7 tempest-ServerShowV247Test-1218443059 tempest-ServerShowV247Test-1218443059-project-member] Got 1 possible topologies {{(pid=61806) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 801.637060] env[61806]: DEBUG nova.virt.hardware [None req-6a30dd3c-3e08-4933-b4cc-6fae094056e7 tempest-ServerShowV247Test-1218443059 tempest-ServerShowV247Test-1218443059-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61806) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 801.637241] env[61806]: DEBUG nova.virt.hardware [None req-6a30dd3c-3e08-4933-b4cc-6fae094056e7 tempest-ServerShowV247Test-1218443059 tempest-ServerShowV247Test-1218443059-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61806) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 801.638294] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-87b2022d-8d76-4e70-834d-3273fe548da7 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 801.646635] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8ec0ae37-0116-4287-a13e-2d518970b5e1 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 801.666503] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-6a30dd3c-3e08-4933-b4cc-6fae094056e7 tempest-ServerShowV247Test-1218443059 tempest-ServerShowV247Test-1218443059-project-member] [instance: 4febd093-9f2b-494e-b175-e4693b0e3e0d] Instance VIF info [] {{(pid=61806) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 801.673087] env[61806]: DEBUG oslo.service.loopingcall [None req-6a30dd3c-3e08-4933-b4cc-6fae094056e7 tempest-ServerShowV247Test-1218443059 tempest-ServerShowV247Test-1218443059-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61806) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 801.673373] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 4febd093-9f2b-494e-b175-e4693b0e3e0d] Creating VM on the ESX host {{(pid=61806) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 801.673608] env[61806]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-ee66e66a-c10e-4cb5-b45c-8af433c942e6 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 801.690754] env[61806]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 801.690754] env[61806]: value = "task-1294553" [ 801.690754] env[61806]: _type = "Task" [ 801.690754] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 801.701705] env[61806]: DEBUG oslo_vmware.api [-] Task: {'id': task-1294553, 'name': CreateVM_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 802.004424] env[61806]: DEBUG oslo_concurrency.lockutils [None req-f7f6aee3-3d33-4e4b-b451-fa3af2e320d4 tempest-ServerDiagnosticsV248Test-463927155 tempest-ServerDiagnosticsV248Test-463927155-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.426s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 802.005162] env[61806]: DEBUG nova.compute.manager [None req-f7f6aee3-3d33-4e4b-b451-fa3af2e320d4 tempest-ServerDiagnosticsV248Test-463927155 tempest-ServerDiagnosticsV248Test-463927155-project-member] [instance: a934d02d-26aa-4900-b473-a58489e5629e] Start building networks asynchronously for instance. {{(pid=61806) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 802.009171] env[61806]: DEBUG oslo_concurrency.lockutils [None req-38045444-a393-40c6-8f1e-16b0707aea48 tempest-ServerShowV257Test-1159790492 tempest-ServerShowV257Test-1159790492-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 17.074s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 802.013020] env[61806]: INFO nova.compute.claims [None req-38045444-a393-40c6-8f1e-16b0707aea48 tempest-ServerShowV257Test-1159790492 tempest-ServerShowV257Test-1159790492-project-member] [instance: 616ec206-9804-469e-ab5c-41aea7f048aa] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 802.039804] env[61806]: DEBUG oslo_vmware.api [None req-b3f33a6f-532a-458d-aec6-6ec54b0cadd5 tempest-ServerMetadataTestJSON-316746592 tempest-ServerMetadataTestJSON-316746592-project-member] Task: {'id': task-1294552, 'name': PowerOnVM_Task, 'duration_secs': 0.640372} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 802.040562] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [None req-b3f33a6f-532a-458d-aec6-6ec54b0cadd5 tempest-ServerMetadataTestJSON-316746592 tempest-ServerMetadataTestJSON-316746592-project-member] [instance: 4184e263-6f56-4bc1-99b2-a2c460531516] Powered on the VM {{(pid=61806) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 802.041445] env[61806]: INFO nova.compute.manager [None req-b3f33a6f-532a-458d-aec6-6ec54b0cadd5 tempest-ServerMetadataTestJSON-316746592 tempest-ServerMetadataTestJSON-316746592-project-member] [instance: 4184e263-6f56-4bc1-99b2-a2c460531516] Took 8.61 seconds to spawn the instance on the hypervisor. [ 802.041445] env[61806]: DEBUG nova.compute.manager [None req-b3f33a6f-532a-458d-aec6-6ec54b0cadd5 tempest-ServerMetadataTestJSON-316746592 tempest-ServerMetadataTestJSON-316746592-project-member] [instance: 4184e263-6f56-4bc1-99b2-a2c460531516] Checking state {{(pid=61806) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 802.042361] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-528fb24e-134a-446e-8091-930b987f701f {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 802.068276] env[61806]: DEBUG nova.compute.manager [None req-7fa92f02-ad2b-4004-a9a7-cb08f83ec0f1 tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] [instance: 67d1bdd8-cf8f-4fb5-bd3c-42008e4d60e3] Starting instance... {{(pid=61806) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 802.207018] env[61806]: DEBUG oslo_vmware.api [-] Task: {'id': task-1294553, 'name': CreateVM_Task} progress is 99%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 802.517782] env[61806]: DEBUG nova.compute.utils [None req-f7f6aee3-3d33-4e4b-b451-fa3af2e320d4 tempest-ServerDiagnosticsV248Test-463927155 tempest-ServerDiagnosticsV248Test-463927155-project-member] Using /dev/sd instead of None {{(pid=61806) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 802.521490] env[61806]: DEBUG nova.compute.manager [None req-f7f6aee3-3d33-4e4b-b451-fa3af2e320d4 tempest-ServerDiagnosticsV248Test-463927155 tempest-ServerDiagnosticsV248Test-463927155-project-member] [instance: a934d02d-26aa-4900-b473-a58489e5629e] Not allocating networking since 'none' was specified. {{(pid=61806) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1977}} [ 802.563618] env[61806]: INFO nova.compute.manager [None req-b3f33a6f-532a-458d-aec6-6ec54b0cadd5 tempest-ServerMetadataTestJSON-316746592 tempest-ServerMetadataTestJSON-316746592-project-member] [instance: 4184e263-6f56-4bc1-99b2-a2c460531516] Took 37.41 seconds to build instance. [ 802.593438] env[61806]: DEBUG oslo_concurrency.lockutils [None req-7fa92f02-ad2b-4004-a9a7-cb08f83ec0f1 tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 802.703446] env[61806]: DEBUG oslo_vmware.api [-] Task: {'id': task-1294553, 'name': CreateVM_Task, 'duration_secs': 0.599453} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 802.703621] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 4febd093-9f2b-494e-b175-e4693b0e3e0d] Created VM on the ESX host {{(pid=61806) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 802.704073] env[61806]: DEBUG oslo_concurrency.lockutils [None req-6a30dd3c-3e08-4933-b4cc-6fae094056e7 tempest-ServerShowV247Test-1218443059 tempest-ServerShowV247Test-1218443059-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 802.704256] env[61806]: DEBUG oslo_concurrency.lockutils [None req-6a30dd3c-3e08-4933-b4cc-6fae094056e7 tempest-ServerShowV247Test-1218443059 tempest-ServerShowV247Test-1218443059-project-member] Acquired lock "[datastore2] devstack-image-cache_base/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 802.704689] env[61806]: DEBUG oslo_concurrency.lockutils [None req-6a30dd3c-3e08-4933-b4cc-6fae094056e7 tempest-ServerShowV247Test-1218443059 tempest-ServerShowV247Test-1218443059-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 802.704983] env[61806]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-7a4331a0-5e49-43cd-8a21-45651e20d7b4 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 802.712206] env[61806]: DEBUG oslo_vmware.api [None req-6a30dd3c-3e08-4933-b4cc-6fae094056e7 tempest-ServerShowV247Test-1218443059 tempest-ServerShowV247Test-1218443059-project-member] Waiting for the task: (returnval){ [ 802.712206] env[61806]: value = "session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]5207362f-56ef-b48c-0a7e-b7db61dd950a" [ 802.712206] env[61806]: _type = "Task" [ 802.712206] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 802.722729] env[61806]: DEBUG oslo_vmware.api [None req-6a30dd3c-3e08-4933-b4cc-6fae094056e7 tempest-ServerShowV247Test-1218443059 tempest-ServerShowV247Test-1218443059-project-member] Task: {'id': session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]5207362f-56ef-b48c-0a7e-b7db61dd950a, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 803.022704] env[61806]: DEBUG nova.compute.manager [None req-f7f6aee3-3d33-4e4b-b451-fa3af2e320d4 tempest-ServerDiagnosticsV248Test-463927155 tempest-ServerDiagnosticsV248Test-463927155-project-member] [instance: a934d02d-26aa-4900-b473-a58489e5629e] Start building block device mappings for instance. {{(pid=61806) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 803.066022] env[61806]: DEBUG oslo_concurrency.lockutils [None req-b3f33a6f-532a-458d-aec6-6ec54b0cadd5 tempest-ServerMetadataTestJSON-316746592 tempest-ServerMetadataTestJSON-316746592-project-member] Lock "4184e263-6f56-4bc1-99b2-a2c460531516" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 148.195s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 803.225913] env[61806]: DEBUG oslo_vmware.api [None req-6a30dd3c-3e08-4933-b4cc-6fae094056e7 tempest-ServerShowV247Test-1218443059 tempest-ServerShowV247Test-1218443059-project-member] Task: {'id': session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]5207362f-56ef-b48c-0a7e-b7db61dd950a, 'name': SearchDatastore_Task, 'duration_secs': 0.016216} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 803.227016] env[61806]: DEBUG oslo_concurrency.lockutils [None req-6a30dd3c-3e08-4933-b4cc-6fae094056e7 tempest-ServerShowV247Test-1218443059 tempest-ServerShowV247Test-1218443059-project-member] Releasing lock "[datastore2] devstack-image-cache_base/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 803.227016] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-6a30dd3c-3e08-4933-b4cc-6fae094056e7 tempest-ServerShowV247Test-1218443059 tempest-ServerShowV247Test-1218443059-project-member] [instance: 4febd093-9f2b-494e-b175-e4693b0e3e0d] Processing image ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa {{(pid=61806) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 803.227016] env[61806]: DEBUG oslo_concurrency.lockutils [None req-6a30dd3c-3e08-4933-b4cc-6fae094056e7 tempest-ServerShowV247Test-1218443059 tempest-ServerShowV247Test-1218443059-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa.vmdk" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 803.227016] env[61806]: DEBUG oslo_concurrency.lockutils [None req-6a30dd3c-3e08-4933-b4cc-6fae094056e7 tempest-ServerShowV247Test-1218443059 tempest-ServerShowV247Test-1218443059-project-member] Acquired lock "[datastore2] devstack-image-cache_base/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa.vmdk" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 803.227220] env[61806]: DEBUG nova.virt.vmwareapi.ds_util [None req-6a30dd3c-3e08-4933-b4cc-6fae094056e7 tempest-ServerShowV247Test-1218443059 tempest-ServerShowV247Test-1218443059-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61806) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 803.227463] env[61806]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-08b25074-c581-4368-927a-ccc692e347b6 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 803.238683] env[61806]: DEBUG nova.virt.vmwareapi.ds_util [None req-6a30dd3c-3e08-4933-b4cc-6fae094056e7 tempest-ServerShowV247Test-1218443059 tempest-ServerShowV247Test-1218443059-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61806) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 803.238903] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-6a30dd3c-3e08-4933-b4cc-6fae094056e7 tempest-ServerShowV247Test-1218443059 tempest-ServerShowV247Test-1218443059-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61806) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 803.239770] env[61806]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-a9e27023-45b6-48a5-8394-d09eadc8f19d {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 803.245709] env[61806]: DEBUG oslo_vmware.api [None req-6a30dd3c-3e08-4933-b4cc-6fae094056e7 tempest-ServerShowV247Test-1218443059 tempest-ServerShowV247Test-1218443059-project-member] Waiting for the task: (returnval){ [ 803.245709] env[61806]: value = "session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]52db1c76-d5d3-2129-4c88-4a763eae4612" [ 803.245709] env[61806]: _type = "Task" [ 803.245709] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 803.256123] env[61806]: DEBUG oslo_vmware.api [None req-6a30dd3c-3e08-4933-b4cc-6fae094056e7 tempest-ServerShowV247Test-1218443059 tempest-ServerShowV247Test-1218443059-project-member] Task: {'id': session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]52db1c76-d5d3-2129-4c88-4a763eae4612, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 803.316037] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9ac7de0e-6c71-4d66-84e8-9e680b9019fd {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 803.323059] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f09e6c8e-c259-4d82-a1c4-d4a758ac8eae {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 803.353454] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5a547666-9382-4816-bfcb-a0891715cae4 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 803.361188] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3f54c095-efe3-489d-b2c9-cbb60db185ca {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 803.376112] env[61806]: DEBUG nova.compute.provider_tree [None req-38045444-a393-40c6-8f1e-16b0707aea48 tempest-ServerShowV257Test-1159790492 tempest-ServerShowV257Test-1159790492-project-member] Inventory has not changed in ProviderTree for provider: a2858be1-fd22-4e08-979e-87ad25293407 {{(pid=61806) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 803.569856] env[61806]: DEBUG nova.compute.manager [None req-cbb90e6d-7dac-4c30-b860-1f0db7dd091e tempest-MultipleCreateTestJSON-1589015462 tempest-MultipleCreateTestJSON-1589015462-project-member] [instance: 25218e83-c1ab-4b97-ade1-7c32b6f99305] Starting instance... {{(pid=61806) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 803.759199] env[61806]: DEBUG oslo_vmware.api [None req-6a30dd3c-3e08-4933-b4cc-6fae094056e7 tempest-ServerShowV247Test-1218443059 tempest-ServerShowV247Test-1218443059-project-member] Task: {'id': session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]52db1c76-d5d3-2129-4c88-4a763eae4612, 'name': SearchDatastore_Task, 'duration_secs': 0.01316} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 803.760013] env[61806]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-79d2d1dd-4785-4a17-b966-37de033e4322 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 803.765851] env[61806]: DEBUG oslo_vmware.api [None req-6a30dd3c-3e08-4933-b4cc-6fae094056e7 tempest-ServerShowV247Test-1218443059 tempest-ServerShowV247Test-1218443059-project-member] Waiting for the task: (returnval){ [ 803.765851] env[61806]: value = "session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]5265dce1-683e-9780-f399-b14f5aec2184" [ 803.765851] env[61806]: _type = "Task" [ 803.765851] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 803.773863] env[61806]: DEBUG oslo_vmware.api [None req-6a30dd3c-3e08-4933-b4cc-6fae094056e7 tempest-ServerShowV247Test-1218443059 tempest-ServerShowV247Test-1218443059-project-member] Task: {'id': session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]5265dce1-683e-9780-f399-b14f5aec2184, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 803.879686] env[61806]: DEBUG nova.scheduler.client.report [None req-38045444-a393-40c6-8f1e-16b0707aea48 tempest-ServerShowV257Test-1159790492 tempest-ServerShowV257Test-1159790492-project-member] Inventory has not changed for provider a2858be1-fd22-4e08-979e-87ad25293407 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 138, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61806) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 804.034408] env[61806]: DEBUG nova.compute.manager [None req-f7f6aee3-3d33-4e4b-b451-fa3af2e320d4 tempest-ServerDiagnosticsV248Test-463927155 tempest-ServerDiagnosticsV248Test-463927155-project-member] [instance: a934d02d-26aa-4900-b473-a58489e5629e] Start spawning the instance on the hypervisor. {{(pid=61806) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 804.062330] env[61806]: DEBUG nova.virt.hardware [None req-f7f6aee3-3d33-4e4b-b451-fa3af2e320d4 tempest-ServerDiagnosticsV248Test-463927155 tempest-ServerDiagnosticsV248Test-463927155-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-15T18:05:10Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-15T18:04:52Z,direct_url=,disk_format='vmdk',id=ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='84ba83fce2c349a6988173c5d6fc3b07',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-15T18:04:53Z,virtual_size=,visibility=), allow threads: False {{(pid=61806) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 804.062628] env[61806]: DEBUG nova.virt.hardware [None req-f7f6aee3-3d33-4e4b-b451-fa3af2e320d4 tempest-ServerDiagnosticsV248Test-463927155 tempest-ServerDiagnosticsV248Test-463927155-project-member] Flavor limits 0:0:0 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 804.062794] env[61806]: DEBUG nova.virt.hardware [None req-f7f6aee3-3d33-4e4b-b451-fa3af2e320d4 tempest-ServerDiagnosticsV248Test-463927155 tempest-ServerDiagnosticsV248Test-463927155-project-member] Image limits 0:0:0 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 804.062983] env[61806]: DEBUG nova.virt.hardware [None req-f7f6aee3-3d33-4e4b-b451-fa3af2e320d4 tempest-ServerDiagnosticsV248Test-463927155 tempest-ServerDiagnosticsV248Test-463927155-project-member] Flavor pref 0:0:0 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 804.063193] env[61806]: DEBUG nova.virt.hardware [None req-f7f6aee3-3d33-4e4b-b451-fa3af2e320d4 tempest-ServerDiagnosticsV248Test-463927155 tempest-ServerDiagnosticsV248Test-463927155-project-member] Image pref 0:0:0 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 804.063335] env[61806]: DEBUG nova.virt.hardware [None req-f7f6aee3-3d33-4e4b-b451-fa3af2e320d4 tempest-ServerDiagnosticsV248Test-463927155 tempest-ServerDiagnosticsV248Test-463927155-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 804.063601] env[61806]: DEBUG nova.virt.hardware [None req-f7f6aee3-3d33-4e4b-b451-fa3af2e320d4 tempest-ServerDiagnosticsV248Test-463927155 tempest-ServerDiagnosticsV248Test-463927155-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61806) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 804.063796] env[61806]: DEBUG nova.virt.hardware [None req-f7f6aee3-3d33-4e4b-b451-fa3af2e320d4 tempest-ServerDiagnosticsV248Test-463927155 tempest-ServerDiagnosticsV248Test-463927155-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61806) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 804.063976] env[61806]: DEBUG nova.virt.hardware [None req-f7f6aee3-3d33-4e4b-b451-fa3af2e320d4 tempest-ServerDiagnosticsV248Test-463927155 tempest-ServerDiagnosticsV248Test-463927155-project-member] Got 1 possible topologies {{(pid=61806) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 804.064158] env[61806]: DEBUG nova.virt.hardware [None req-f7f6aee3-3d33-4e4b-b451-fa3af2e320d4 tempest-ServerDiagnosticsV248Test-463927155 tempest-ServerDiagnosticsV248Test-463927155-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61806) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 804.064358] env[61806]: DEBUG nova.virt.hardware [None req-f7f6aee3-3d33-4e4b-b451-fa3af2e320d4 tempest-ServerDiagnosticsV248Test-463927155 tempest-ServerDiagnosticsV248Test-463927155-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61806) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 804.065285] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3ac665be-aef9-438c-9a5e-bdb345af45ca {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 804.077092] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d4da796c-4f2d-4b94-9c11-d02b90df20c5 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 804.093724] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-f7f6aee3-3d33-4e4b-b451-fa3af2e320d4 tempest-ServerDiagnosticsV248Test-463927155 tempest-ServerDiagnosticsV248Test-463927155-project-member] [instance: a934d02d-26aa-4900-b473-a58489e5629e] Instance VIF info [] {{(pid=61806) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 804.099507] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [None req-f7f6aee3-3d33-4e4b-b451-fa3af2e320d4 tempest-ServerDiagnosticsV248Test-463927155 tempest-ServerDiagnosticsV248Test-463927155-project-member] Creating folder: Project (ea57bc170e3742b98e68a32fa177f352). Parent ref: group-v277609. {{(pid=61806) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 804.100753] env[61806]: DEBUG oslo_concurrency.lockutils [None req-cbb90e6d-7dac-4c30-b860-1f0db7dd091e tempest-MultipleCreateTestJSON-1589015462 tempest-MultipleCreateTestJSON-1589015462-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 804.101049] env[61806]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-48d2ed1f-edde-47e8-9339-5cbbfc247d05 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 804.111278] env[61806]: INFO nova.virt.vmwareapi.vm_util [None req-f7f6aee3-3d33-4e4b-b451-fa3af2e320d4 tempest-ServerDiagnosticsV248Test-463927155 tempest-ServerDiagnosticsV248Test-463927155-project-member] Created folder: Project (ea57bc170e3742b98e68a32fa177f352) in parent group-v277609. [ 804.111484] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [None req-f7f6aee3-3d33-4e4b-b451-fa3af2e320d4 tempest-ServerDiagnosticsV248Test-463927155 tempest-ServerDiagnosticsV248Test-463927155-project-member] Creating folder: Instances. Parent ref: group-v277642. {{(pid=61806) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 804.111832] env[61806]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-cc243e48-7142-419e-ba0b-e6c7a7663c7f {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 804.122974] env[61806]: INFO nova.virt.vmwareapi.vm_util [None req-f7f6aee3-3d33-4e4b-b451-fa3af2e320d4 tempest-ServerDiagnosticsV248Test-463927155 tempest-ServerDiagnosticsV248Test-463927155-project-member] Created folder: Instances in parent group-v277642. [ 804.122974] env[61806]: DEBUG oslo.service.loopingcall [None req-f7f6aee3-3d33-4e4b-b451-fa3af2e320d4 tempest-ServerDiagnosticsV248Test-463927155 tempest-ServerDiagnosticsV248Test-463927155-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61806) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 804.122974] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: a934d02d-26aa-4900-b473-a58489e5629e] Creating VM on the ESX host {{(pid=61806) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 804.122974] env[61806]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-1b486844-24ae-4a73-b061-d9d0bd560777 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 804.144535] env[61806]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 804.144535] env[61806]: value = "task-1294556" [ 804.144535] env[61806]: _type = "Task" [ 804.144535] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 804.157248] env[61806]: DEBUG oslo_vmware.api [-] Task: {'id': task-1294556, 'name': CreateVM_Task} progress is 6%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 804.284814] env[61806]: DEBUG oslo_vmware.api [None req-6a30dd3c-3e08-4933-b4cc-6fae094056e7 tempest-ServerShowV247Test-1218443059 tempest-ServerShowV247Test-1218443059-project-member] Task: {'id': session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]5265dce1-683e-9780-f399-b14f5aec2184, 'name': SearchDatastore_Task, 'duration_secs': 0.016509} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 804.284814] env[61806]: DEBUG oslo_concurrency.lockutils [None req-6a30dd3c-3e08-4933-b4cc-6fae094056e7 tempest-ServerShowV247Test-1218443059 tempest-ServerShowV247Test-1218443059-project-member] Releasing lock "[datastore2] devstack-image-cache_base/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa.vmdk" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 804.284814] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [None req-6a30dd3c-3e08-4933-b4cc-6fae094056e7 tempest-ServerShowV247Test-1218443059 tempest-ServerShowV247Test-1218443059-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa.vmdk to [datastore2] 4febd093-9f2b-494e-b175-e4693b0e3e0d/4febd093-9f2b-494e-b175-e4693b0e3e0d.vmdk {{(pid=61806) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 804.284814] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-1813c6d0-69d7-404f-ae6e-38d58db99df6 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 804.292887] env[61806]: DEBUG oslo_vmware.api [None req-6a30dd3c-3e08-4933-b4cc-6fae094056e7 tempest-ServerShowV247Test-1218443059 tempest-ServerShowV247Test-1218443059-project-member] Waiting for the task: (returnval){ [ 804.292887] env[61806]: value = "task-1294557" [ 804.292887] env[61806]: _type = "Task" [ 804.292887] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 804.298678] env[61806]: DEBUG oslo_vmware.api [None req-6a30dd3c-3e08-4933-b4cc-6fae094056e7 tempest-ServerShowV247Test-1218443059 tempest-ServerShowV247Test-1218443059-project-member] Task: {'id': task-1294557, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 804.384823] env[61806]: DEBUG oslo_concurrency.lockutils [None req-38045444-a393-40c6-8f1e-16b0707aea48 tempest-ServerShowV257Test-1159790492 tempest-ServerShowV257Test-1159790492-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.375s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 804.385399] env[61806]: DEBUG nova.compute.manager [None req-38045444-a393-40c6-8f1e-16b0707aea48 tempest-ServerShowV257Test-1159790492 tempest-ServerShowV257Test-1159790492-project-member] [instance: 616ec206-9804-469e-ab5c-41aea7f048aa] Start building networks asynchronously for instance. {{(pid=61806) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 804.388147] env[61806]: DEBUG oslo_concurrency.lockutils [None req-dbeadda9-7825-4678-af18-5f977e49124f tempest-ServerActionsV293TestJSON-906151111 tempest-ServerActionsV293TestJSON-906151111-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 15.171s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 804.389923] env[61806]: INFO nova.compute.claims [None req-dbeadda9-7825-4678-af18-5f977e49124f tempest-ServerActionsV293TestJSON-906151111 tempest-ServerActionsV293TestJSON-906151111-project-member] [instance: 0758901a-7093-41d3-b0e2-5c519333abdd] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 804.659218] env[61806]: DEBUG oslo_vmware.api [-] Task: {'id': task-1294556, 'name': CreateVM_Task} progress is 99%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 804.804277] env[61806]: DEBUG oslo_vmware.api [None req-6a30dd3c-3e08-4933-b4cc-6fae094056e7 tempest-ServerShowV247Test-1218443059 tempest-ServerShowV247Test-1218443059-project-member] Task: {'id': task-1294557, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 804.894954] env[61806]: DEBUG nova.compute.utils [None req-38045444-a393-40c6-8f1e-16b0707aea48 tempest-ServerShowV257Test-1159790492 tempest-ServerShowV257Test-1159790492-project-member] Using /dev/sd instead of None {{(pid=61806) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 804.899032] env[61806]: DEBUG nova.compute.manager [None req-38045444-a393-40c6-8f1e-16b0707aea48 tempest-ServerShowV257Test-1159790492 tempest-ServerShowV257Test-1159790492-project-member] [instance: 616ec206-9804-469e-ab5c-41aea7f048aa] Not allocating networking since 'none' was specified. {{(pid=61806) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1977}} [ 805.001372] env[61806]: DEBUG oslo_concurrency.lockutils [None req-fbc15506-7910-4d6f-bfbc-c3a1ce19fdcc tempest-ServerMetadataTestJSON-316746592 tempest-ServerMetadataTestJSON-316746592-project-member] Acquiring lock "4184e263-6f56-4bc1-99b2-a2c460531516" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 805.001663] env[61806]: DEBUG oslo_concurrency.lockutils [None req-fbc15506-7910-4d6f-bfbc-c3a1ce19fdcc tempest-ServerMetadataTestJSON-316746592 tempest-ServerMetadataTestJSON-316746592-project-member] Lock "4184e263-6f56-4bc1-99b2-a2c460531516" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 805.001879] env[61806]: DEBUG oslo_concurrency.lockutils [None req-fbc15506-7910-4d6f-bfbc-c3a1ce19fdcc tempest-ServerMetadataTestJSON-316746592 tempest-ServerMetadataTestJSON-316746592-project-member] Acquiring lock "4184e263-6f56-4bc1-99b2-a2c460531516-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 805.002081] env[61806]: DEBUG oslo_concurrency.lockutils [None req-fbc15506-7910-4d6f-bfbc-c3a1ce19fdcc tempest-ServerMetadataTestJSON-316746592 tempest-ServerMetadataTestJSON-316746592-project-member] Lock "4184e263-6f56-4bc1-99b2-a2c460531516-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 805.002727] env[61806]: DEBUG oslo_concurrency.lockutils [None req-fbc15506-7910-4d6f-bfbc-c3a1ce19fdcc tempest-ServerMetadataTestJSON-316746592 tempest-ServerMetadataTestJSON-316746592-project-member] Lock "4184e263-6f56-4bc1-99b2-a2c460531516-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 805.006705] env[61806]: INFO nova.compute.manager [None req-fbc15506-7910-4d6f-bfbc-c3a1ce19fdcc tempest-ServerMetadataTestJSON-316746592 tempest-ServerMetadataTestJSON-316746592-project-member] [instance: 4184e263-6f56-4bc1-99b2-a2c460531516] Terminating instance [ 805.009651] env[61806]: DEBUG nova.compute.manager [None req-fbc15506-7910-4d6f-bfbc-c3a1ce19fdcc tempest-ServerMetadataTestJSON-316746592 tempest-ServerMetadataTestJSON-316746592-project-member] [instance: 4184e263-6f56-4bc1-99b2-a2c460531516] Start destroying the instance on the hypervisor. {{(pid=61806) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 805.010016] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-fbc15506-7910-4d6f-bfbc-c3a1ce19fdcc tempest-ServerMetadataTestJSON-316746592 tempest-ServerMetadataTestJSON-316746592-project-member] [instance: 4184e263-6f56-4bc1-99b2-a2c460531516] Destroying instance {{(pid=61806) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 805.010827] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-86e393e2-d259-4411-8a4e-c5ea0db052ed {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 805.021163] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [None req-fbc15506-7910-4d6f-bfbc-c3a1ce19fdcc tempest-ServerMetadataTestJSON-316746592 tempest-ServerMetadataTestJSON-316746592-project-member] [instance: 4184e263-6f56-4bc1-99b2-a2c460531516] Powering off the VM {{(pid=61806) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 805.021457] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-56dcd899-5f1b-4ec2-baa9-31d45c5e4ba1 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 805.028055] env[61806]: DEBUG oslo_vmware.api [None req-fbc15506-7910-4d6f-bfbc-c3a1ce19fdcc tempest-ServerMetadataTestJSON-316746592 tempest-ServerMetadataTestJSON-316746592-project-member] Waiting for the task: (returnval){ [ 805.028055] env[61806]: value = "task-1294558" [ 805.028055] env[61806]: _type = "Task" [ 805.028055] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 805.036425] env[61806]: DEBUG oslo_vmware.api [None req-fbc15506-7910-4d6f-bfbc-c3a1ce19fdcc tempest-ServerMetadataTestJSON-316746592 tempest-ServerMetadataTestJSON-316746592-project-member] Task: {'id': task-1294558, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 805.155939] env[61806]: DEBUG oslo_vmware.api [-] Task: {'id': task-1294556, 'name': CreateVM_Task} progress is 99%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 805.302705] env[61806]: DEBUG oslo_vmware.api [None req-6a30dd3c-3e08-4933-b4cc-6fae094056e7 tempest-ServerShowV247Test-1218443059 tempest-ServerShowV247Test-1218443059-project-member] Task: {'id': task-1294557, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.595614} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 805.303040] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [None req-6a30dd3c-3e08-4933-b4cc-6fae094056e7 tempest-ServerShowV247Test-1218443059 tempest-ServerShowV247Test-1218443059-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa.vmdk to [datastore2] 4febd093-9f2b-494e-b175-e4693b0e3e0d/4febd093-9f2b-494e-b175-e4693b0e3e0d.vmdk {{(pid=61806) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 805.303304] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-6a30dd3c-3e08-4933-b4cc-6fae094056e7 tempest-ServerShowV247Test-1218443059 tempest-ServerShowV247Test-1218443059-project-member] [instance: 4febd093-9f2b-494e-b175-e4693b0e3e0d] Extending root virtual disk to 1048576 {{(pid=61806) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 805.304199] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-0553beeb-b11e-4925-b63a-a56814407071 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 805.310795] env[61806]: DEBUG oslo_vmware.api [None req-6a30dd3c-3e08-4933-b4cc-6fae094056e7 tempest-ServerShowV247Test-1218443059 tempest-ServerShowV247Test-1218443059-project-member] Waiting for the task: (returnval){ [ 805.310795] env[61806]: value = "task-1294559" [ 805.310795] env[61806]: _type = "Task" [ 805.310795] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 805.319464] env[61806]: DEBUG oslo_vmware.api [None req-6a30dd3c-3e08-4933-b4cc-6fae094056e7 tempest-ServerShowV247Test-1218443059 tempest-ServerShowV247Test-1218443059-project-member] Task: {'id': task-1294559, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 805.400793] env[61806]: DEBUG nova.compute.manager [None req-38045444-a393-40c6-8f1e-16b0707aea48 tempest-ServerShowV257Test-1159790492 tempest-ServerShowV257Test-1159790492-project-member] [instance: 616ec206-9804-469e-ab5c-41aea7f048aa] Start building block device mappings for instance. {{(pid=61806) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 805.542256] env[61806]: DEBUG oslo_vmware.api [None req-fbc15506-7910-4d6f-bfbc-c3a1ce19fdcc tempest-ServerMetadataTestJSON-316746592 tempest-ServerMetadataTestJSON-316746592-project-member] Task: {'id': task-1294558, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 805.660770] env[61806]: DEBUG oslo_vmware.api [-] Task: {'id': task-1294556, 'name': CreateVM_Task, 'duration_secs': 1.382646} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 805.663782] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: a934d02d-26aa-4900-b473-a58489e5629e] Created VM on the ESX host {{(pid=61806) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 805.664487] env[61806]: DEBUG oslo_concurrency.lockutils [None req-f7f6aee3-3d33-4e4b-b451-fa3af2e320d4 tempest-ServerDiagnosticsV248Test-463927155 tempest-ServerDiagnosticsV248Test-463927155-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 805.664700] env[61806]: DEBUG oslo_concurrency.lockutils [None req-f7f6aee3-3d33-4e4b-b451-fa3af2e320d4 tempest-ServerDiagnosticsV248Test-463927155 tempest-ServerDiagnosticsV248Test-463927155-project-member] Acquired lock "[datastore2] devstack-image-cache_base/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 805.665046] env[61806]: DEBUG oslo_concurrency.lockutils [None req-f7f6aee3-3d33-4e4b-b451-fa3af2e320d4 tempest-ServerDiagnosticsV248Test-463927155 tempest-ServerDiagnosticsV248Test-463927155-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 805.665331] env[61806]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-9aefd578-c3a8-45df-90e7-6a3069108872 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 805.670426] env[61806]: DEBUG oslo_vmware.api [None req-f7f6aee3-3d33-4e4b-b451-fa3af2e320d4 tempest-ServerDiagnosticsV248Test-463927155 tempest-ServerDiagnosticsV248Test-463927155-project-member] Waiting for the task: (returnval){ [ 805.670426] env[61806]: value = "session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]5200254a-cd60-3118-6eff-1e2eeccd42d5" [ 805.670426] env[61806]: _type = "Task" [ 805.670426] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 805.680479] env[61806]: DEBUG oslo_vmware.api [None req-f7f6aee3-3d33-4e4b-b451-fa3af2e320d4 tempest-ServerDiagnosticsV248Test-463927155 tempest-ServerDiagnosticsV248Test-463927155-project-member] Task: {'id': session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]5200254a-cd60-3118-6eff-1e2eeccd42d5, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 805.716089] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-896a5a98-fb7a-4375-97c3-ff502077227d {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 805.724151] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0ba559f7-1427-4bc1-b149-3c5801373d04 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 805.757676] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2f1fbd68-8744-4e39-976c-cd3fad7c1cab {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 805.766878] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-edcdc4e6-6036-4366-b471-661a441017d0 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 805.784743] env[61806]: DEBUG nova.compute.provider_tree [None req-dbeadda9-7825-4678-af18-5f977e49124f tempest-ServerActionsV293TestJSON-906151111 tempest-ServerActionsV293TestJSON-906151111-project-member] Inventory has not changed in ProviderTree for provider: a2858be1-fd22-4e08-979e-87ad25293407 {{(pid=61806) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 805.821219] env[61806]: DEBUG oslo_vmware.api [None req-6a30dd3c-3e08-4933-b4cc-6fae094056e7 tempest-ServerShowV247Test-1218443059 tempest-ServerShowV247Test-1218443059-project-member] Task: {'id': task-1294559, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.104149} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 805.821527] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-6a30dd3c-3e08-4933-b4cc-6fae094056e7 tempest-ServerShowV247Test-1218443059 tempest-ServerShowV247Test-1218443059-project-member] [instance: 4febd093-9f2b-494e-b175-e4693b0e3e0d] Extended root virtual disk {{(pid=61806) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 805.822392] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e968e97b-b0fe-4701-bdd0-598ae1388981 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 805.843830] env[61806]: DEBUG nova.virt.vmwareapi.volumeops [None req-6a30dd3c-3e08-4933-b4cc-6fae094056e7 tempest-ServerShowV247Test-1218443059 tempest-ServerShowV247Test-1218443059-project-member] [instance: 4febd093-9f2b-494e-b175-e4693b0e3e0d] Reconfiguring VM instance instance-00000039 to attach disk [datastore2] 4febd093-9f2b-494e-b175-e4693b0e3e0d/4febd093-9f2b-494e-b175-e4693b0e3e0d.vmdk or device None with type sparse {{(pid=61806) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 805.844299] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-c2b8e4a1-b9c0-4a15-9131-e0ac48106dc3 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 805.864930] env[61806]: DEBUG oslo_vmware.api [None req-6a30dd3c-3e08-4933-b4cc-6fae094056e7 tempest-ServerShowV247Test-1218443059 tempest-ServerShowV247Test-1218443059-project-member] Waiting for the task: (returnval){ [ 805.864930] env[61806]: value = "task-1294560" [ 805.864930] env[61806]: _type = "Task" [ 805.864930] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 805.875277] env[61806]: DEBUG oslo_vmware.api [None req-6a30dd3c-3e08-4933-b4cc-6fae094056e7 tempest-ServerShowV247Test-1218443059 tempest-ServerShowV247Test-1218443059-project-member] Task: {'id': task-1294560, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 806.039119] env[61806]: DEBUG oslo_vmware.api [None req-fbc15506-7910-4d6f-bfbc-c3a1ce19fdcc tempest-ServerMetadataTestJSON-316746592 tempest-ServerMetadataTestJSON-316746592-project-member] Task: {'id': task-1294558, 'name': PowerOffVM_Task, 'duration_secs': 0.623083} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 806.039462] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [None req-fbc15506-7910-4d6f-bfbc-c3a1ce19fdcc tempest-ServerMetadataTestJSON-316746592 tempest-ServerMetadataTestJSON-316746592-project-member] [instance: 4184e263-6f56-4bc1-99b2-a2c460531516] Powered off the VM {{(pid=61806) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 806.039655] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-fbc15506-7910-4d6f-bfbc-c3a1ce19fdcc tempest-ServerMetadataTestJSON-316746592 tempest-ServerMetadataTestJSON-316746592-project-member] [instance: 4184e263-6f56-4bc1-99b2-a2c460531516] Unregistering the VM {{(pid=61806) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 806.039922] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-127f5097-ffb6-4fcb-ba88-684b9a0c10d6 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 806.102584] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-fbc15506-7910-4d6f-bfbc-c3a1ce19fdcc tempest-ServerMetadataTestJSON-316746592 tempest-ServerMetadataTestJSON-316746592-project-member] [instance: 4184e263-6f56-4bc1-99b2-a2c460531516] Unregistered the VM {{(pid=61806) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 806.102818] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-fbc15506-7910-4d6f-bfbc-c3a1ce19fdcc tempest-ServerMetadataTestJSON-316746592 tempest-ServerMetadataTestJSON-316746592-project-member] [instance: 4184e263-6f56-4bc1-99b2-a2c460531516] Deleting contents of the VM from datastore datastore1 {{(pid=61806) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 806.102999] env[61806]: DEBUG nova.virt.vmwareapi.ds_util [None req-fbc15506-7910-4d6f-bfbc-c3a1ce19fdcc tempest-ServerMetadataTestJSON-316746592 tempest-ServerMetadataTestJSON-316746592-project-member] Deleting the datastore file [datastore1] 4184e263-6f56-4bc1-99b2-a2c460531516 {{(pid=61806) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 806.103291] env[61806]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-5f76f980-072e-4a69-94ba-d23d9836b59b {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 806.109362] env[61806]: DEBUG oslo_vmware.api [None req-fbc15506-7910-4d6f-bfbc-c3a1ce19fdcc tempest-ServerMetadataTestJSON-316746592 tempest-ServerMetadataTestJSON-316746592-project-member] Waiting for the task: (returnval){ [ 806.109362] env[61806]: value = "task-1294562" [ 806.109362] env[61806]: _type = "Task" [ 806.109362] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 806.117531] env[61806]: DEBUG oslo_vmware.api [None req-fbc15506-7910-4d6f-bfbc-c3a1ce19fdcc tempest-ServerMetadataTestJSON-316746592 tempest-ServerMetadataTestJSON-316746592-project-member] Task: {'id': task-1294562, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 806.180942] env[61806]: DEBUG oslo_vmware.api [None req-f7f6aee3-3d33-4e4b-b451-fa3af2e320d4 tempest-ServerDiagnosticsV248Test-463927155 tempest-ServerDiagnosticsV248Test-463927155-project-member] Task: {'id': session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]5200254a-cd60-3118-6eff-1e2eeccd42d5, 'name': SearchDatastore_Task, 'duration_secs': 0.019102} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 806.181288] env[61806]: DEBUG oslo_concurrency.lockutils [None req-f7f6aee3-3d33-4e4b-b451-fa3af2e320d4 tempest-ServerDiagnosticsV248Test-463927155 tempest-ServerDiagnosticsV248Test-463927155-project-member] Releasing lock "[datastore2] devstack-image-cache_base/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 806.181577] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-f7f6aee3-3d33-4e4b-b451-fa3af2e320d4 tempest-ServerDiagnosticsV248Test-463927155 tempest-ServerDiagnosticsV248Test-463927155-project-member] [instance: a934d02d-26aa-4900-b473-a58489e5629e] Processing image ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa {{(pid=61806) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 806.181825] env[61806]: DEBUG oslo_concurrency.lockutils [None req-f7f6aee3-3d33-4e4b-b451-fa3af2e320d4 tempest-ServerDiagnosticsV248Test-463927155 tempest-ServerDiagnosticsV248Test-463927155-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa.vmdk" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 806.181991] env[61806]: DEBUG oslo_concurrency.lockutils [None req-f7f6aee3-3d33-4e4b-b451-fa3af2e320d4 tempest-ServerDiagnosticsV248Test-463927155 tempest-ServerDiagnosticsV248Test-463927155-project-member] Acquired lock "[datastore2] devstack-image-cache_base/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa.vmdk" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 806.182239] env[61806]: DEBUG nova.virt.vmwareapi.ds_util [None req-f7f6aee3-3d33-4e4b-b451-fa3af2e320d4 tempest-ServerDiagnosticsV248Test-463927155 tempest-ServerDiagnosticsV248Test-463927155-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61806) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 806.182547] env[61806]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-2c1bad11-03f2-4eb8-9803-6992f2a22a6d {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 806.202503] env[61806]: DEBUG nova.virt.vmwareapi.ds_util [None req-f7f6aee3-3d33-4e4b-b451-fa3af2e320d4 tempest-ServerDiagnosticsV248Test-463927155 tempest-ServerDiagnosticsV248Test-463927155-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61806) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 806.202846] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-f7f6aee3-3d33-4e4b-b451-fa3af2e320d4 tempest-ServerDiagnosticsV248Test-463927155 tempest-ServerDiagnosticsV248Test-463927155-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61806) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 806.203747] env[61806]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-22da9c6b-f3f8-401a-a232-1f4f22a76638 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 806.210907] env[61806]: DEBUG oslo_vmware.api [None req-f7f6aee3-3d33-4e4b-b451-fa3af2e320d4 tempest-ServerDiagnosticsV248Test-463927155 tempest-ServerDiagnosticsV248Test-463927155-project-member] Waiting for the task: (returnval){ [ 806.210907] env[61806]: value = "session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]52d11ef2-bcdd-6065-03ef-c50ce8e0ceb8" [ 806.210907] env[61806]: _type = "Task" [ 806.210907] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 806.223221] env[61806]: DEBUG oslo_vmware.api [None req-f7f6aee3-3d33-4e4b-b451-fa3af2e320d4 tempest-ServerDiagnosticsV248Test-463927155 tempest-ServerDiagnosticsV248Test-463927155-project-member] Task: {'id': session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]52d11ef2-bcdd-6065-03ef-c50ce8e0ceb8, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 806.288081] env[61806]: DEBUG nova.scheduler.client.report [None req-dbeadda9-7825-4678-af18-5f977e49124f tempest-ServerActionsV293TestJSON-906151111 tempest-ServerActionsV293TestJSON-906151111-project-member] Inventory has not changed for provider a2858be1-fd22-4e08-979e-87ad25293407 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 138, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61806) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 806.376499] env[61806]: DEBUG oslo_vmware.api [None req-6a30dd3c-3e08-4933-b4cc-6fae094056e7 tempest-ServerShowV247Test-1218443059 tempest-ServerShowV247Test-1218443059-project-member] Task: {'id': task-1294560, 'name': ReconfigVM_Task, 'duration_secs': 0.383648} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 806.376753] env[61806]: DEBUG nova.virt.vmwareapi.volumeops [None req-6a30dd3c-3e08-4933-b4cc-6fae094056e7 tempest-ServerShowV247Test-1218443059 tempest-ServerShowV247Test-1218443059-project-member] [instance: 4febd093-9f2b-494e-b175-e4693b0e3e0d] Reconfigured VM instance instance-00000039 to attach disk [datastore2] 4febd093-9f2b-494e-b175-e4693b0e3e0d/4febd093-9f2b-494e-b175-e4693b0e3e0d.vmdk or device None with type sparse {{(pid=61806) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 806.377985] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-2425e30b-24bd-4aaa-8bed-cf5109292684 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 806.384985] env[61806]: DEBUG oslo_vmware.api [None req-6a30dd3c-3e08-4933-b4cc-6fae094056e7 tempest-ServerShowV247Test-1218443059 tempest-ServerShowV247Test-1218443059-project-member] Waiting for the task: (returnval){ [ 806.384985] env[61806]: value = "task-1294563" [ 806.384985] env[61806]: _type = "Task" [ 806.384985] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 806.397325] env[61806]: DEBUG oslo_vmware.api [None req-6a30dd3c-3e08-4933-b4cc-6fae094056e7 tempest-ServerShowV247Test-1218443059 tempest-ServerShowV247Test-1218443059-project-member] Task: {'id': task-1294563, 'name': Rename_Task} progress is 5%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 806.423148] env[61806]: DEBUG nova.compute.manager [None req-38045444-a393-40c6-8f1e-16b0707aea48 tempest-ServerShowV257Test-1159790492 tempest-ServerShowV257Test-1159790492-project-member] [instance: 616ec206-9804-469e-ab5c-41aea7f048aa] Start spawning the instance on the hypervisor. {{(pid=61806) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 806.446488] env[61806]: DEBUG nova.virt.hardware [None req-38045444-a393-40c6-8f1e-16b0707aea48 tempest-ServerShowV257Test-1159790492 tempest-ServerShowV257Test-1159790492-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-15T18:05:10Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-15T18:04:52Z,direct_url=,disk_format='vmdk',id=ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='84ba83fce2c349a6988173c5d6fc3b07',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-15T18:04:53Z,virtual_size=,visibility=), allow threads: False {{(pid=61806) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 806.446799] env[61806]: DEBUG nova.virt.hardware [None req-38045444-a393-40c6-8f1e-16b0707aea48 tempest-ServerShowV257Test-1159790492 tempest-ServerShowV257Test-1159790492-project-member] Flavor limits 0:0:0 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 806.446935] env[61806]: DEBUG nova.virt.hardware [None req-38045444-a393-40c6-8f1e-16b0707aea48 tempest-ServerShowV257Test-1159790492 tempest-ServerShowV257Test-1159790492-project-member] Image limits 0:0:0 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 806.447149] env[61806]: DEBUG nova.virt.hardware [None req-38045444-a393-40c6-8f1e-16b0707aea48 tempest-ServerShowV257Test-1159790492 tempest-ServerShowV257Test-1159790492-project-member] Flavor pref 0:0:0 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 806.447302] env[61806]: DEBUG nova.virt.hardware [None req-38045444-a393-40c6-8f1e-16b0707aea48 tempest-ServerShowV257Test-1159790492 tempest-ServerShowV257Test-1159790492-project-member] Image pref 0:0:0 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 806.447550] env[61806]: DEBUG nova.virt.hardware [None req-38045444-a393-40c6-8f1e-16b0707aea48 tempest-ServerShowV257Test-1159790492 tempest-ServerShowV257Test-1159790492-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 806.447665] env[61806]: DEBUG nova.virt.hardware [None req-38045444-a393-40c6-8f1e-16b0707aea48 tempest-ServerShowV257Test-1159790492 tempest-ServerShowV257Test-1159790492-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61806) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 806.447826] env[61806]: DEBUG nova.virt.hardware [None req-38045444-a393-40c6-8f1e-16b0707aea48 tempest-ServerShowV257Test-1159790492 tempest-ServerShowV257Test-1159790492-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61806) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 806.447994] env[61806]: DEBUG nova.virt.hardware [None req-38045444-a393-40c6-8f1e-16b0707aea48 tempest-ServerShowV257Test-1159790492 tempest-ServerShowV257Test-1159790492-project-member] Got 1 possible topologies {{(pid=61806) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 806.448189] env[61806]: DEBUG nova.virt.hardware [None req-38045444-a393-40c6-8f1e-16b0707aea48 tempest-ServerShowV257Test-1159790492 tempest-ServerShowV257Test-1159790492-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61806) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 806.448405] env[61806]: DEBUG nova.virt.hardware [None req-38045444-a393-40c6-8f1e-16b0707aea48 tempest-ServerShowV257Test-1159790492 tempest-ServerShowV257Test-1159790492-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61806) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 806.449279] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c100f2ac-e3a2-4ae6-8b41-3be1024b7fd2 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 806.461602] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d87e957e-55de-400f-b0a7-96e7ccfe2f75 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 806.478141] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-38045444-a393-40c6-8f1e-16b0707aea48 tempest-ServerShowV257Test-1159790492 tempest-ServerShowV257Test-1159790492-project-member] [instance: 616ec206-9804-469e-ab5c-41aea7f048aa] Instance VIF info [] {{(pid=61806) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 806.484189] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [None req-38045444-a393-40c6-8f1e-16b0707aea48 tempest-ServerShowV257Test-1159790492 tempest-ServerShowV257Test-1159790492-project-member] Creating folder: Project (245eb8e77ece4262b2eb47212d9739bd). Parent ref: group-v277609. {{(pid=61806) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 806.484540] env[61806]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-da0d17a2-4217-4cd3-87a3-f2d0cbbe280a {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 806.493915] env[61806]: INFO nova.virt.vmwareapi.vm_util [None req-38045444-a393-40c6-8f1e-16b0707aea48 tempest-ServerShowV257Test-1159790492 tempest-ServerShowV257Test-1159790492-project-member] Created folder: Project (245eb8e77ece4262b2eb47212d9739bd) in parent group-v277609. [ 806.494130] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [None req-38045444-a393-40c6-8f1e-16b0707aea48 tempest-ServerShowV257Test-1159790492 tempest-ServerShowV257Test-1159790492-project-member] Creating folder: Instances. Parent ref: group-v277645. {{(pid=61806) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 806.494380] env[61806]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-d3f2d6bb-35a4-4b14-90d4-6cfe4cf48648 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 806.503468] env[61806]: INFO nova.virt.vmwareapi.vm_util [None req-38045444-a393-40c6-8f1e-16b0707aea48 tempest-ServerShowV257Test-1159790492 tempest-ServerShowV257Test-1159790492-project-member] Created folder: Instances in parent group-v277645. [ 806.503722] env[61806]: DEBUG oslo.service.loopingcall [None req-38045444-a393-40c6-8f1e-16b0707aea48 tempest-ServerShowV257Test-1159790492 tempest-ServerShowV257Test-1159790492-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61806) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 806.503922] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 616ec206-9804-469e-ab5c-41aea7f048aa] Creating VM on the ESX host {{(pid=61806) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 806.504152] env[61806]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-1a0ff23e-81e1-42db-972a-51a4c063c51b {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 806.520960] env[61806]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 806.520960] env[61806]: value = "task-1294566" [ 806.520960] env[61806]: _type = "Task" [ 806.520960] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 806.528348] env[61806]: DEBUG oslo_vmware.api [-] Task: {'id': task-1294566, 'name': CreateVM_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 806.620135] env[61806]: DEBUG oslo_vmware.api [None req-fbc15506-7910-4d6f-bfbc-c3a1ce19fdcc tempest-ServerMetadataTestJSON-316746592 tempest-ServerMetadataTestJSON-316746592-project-member] Task: {'id': task-1294562, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.37088} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 806.620404] env[61806]: DEBUG nova.virt.vmwareapi.ds_util [None req-fbc15506-7910-4d6f-bfbc-c3a1ce19fdcc tempest-ServerMetadataTestJSON-316746592 tempest-ServerMetadataTestJSON-316746592-project-member] Deleted the datastore file {{(pid=61806) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 806.620595] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-fbc15506-7910-4d6f-bfbc-c3a1ce19fdcc tempest-ServerMetadataTestJSON-316746592 tempest-ServerMetadataTestJSON-316746592-project-member] [instance: 4184e263-6f56-4bc1-99b2-a2c460531516] Deleted contents of the VM from datastore datastore1 {{(pid=61806) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 806.620779] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-fbc15506-7910-4d6f-bfbc-c3a1ce19fdcc tempest-ServerMetadataTestJSON-316746592 tempest-ServerMetadataTestJSON-316746592-project-member] [instance: 4184e263-6f56-4bc1-99b2-a2c460531516] Instance destroyed {{(pid=61806) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 806.620979] env[61806]: INFO nova.compute.manager [None req-fbc15506-7910-4d6f-bfbc-c3a1ce19fdcc tempest-ServerMetadataTestJSON-316746592 tempest-ServerMetadataTestJSON-316746592-project-member] [instance: 4184e263-6f56-4bc1-99b2-a2c460531516] Took 1.61 seconds to destroy the instance on the hypervisor. [ 806.621300] env[61806]: DEBUG oslo.service.loopingcall [None req-fbc15506-7910-4d6f-bfbc-c3a1ce19fdcc tempest-ServerMetadataTestJSON-316746592 tempest-ServerMetadataTestJSON-316746592-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61806) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 806.621508] env[61806]: DEBUG nova.compute.manager [-] [instance: 4184e263-6f56-4bc1-99b2-a2c460531516] Deallocating network for instance {{(pid=61806) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 806.621616] env[61806]: DEBUG nova.network.neutron [-] [instance: 4184e263-6f56-4bc1-99b2-a2c460531516] deallocate_for_instance() {{(pid=61806) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 806.721309] env[61806]: DEBUG oslo_vmware.api [None req-f7f6aee3-3d33-4e4b-b451-fa3af2e320d4 tempest-ServerDiagnosticsV248Test-463927155 tempest-ServerDiagnosticsV248Test-463927155-project-member] Task: {'id': session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]52d11ef2-bcdd-6065-03ef-c50ce8e0ceb8, 'name': SearchDatastore_Task, 'duration_secs': 0.02054} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 806.722476] env[61806]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-cdd4e9b7-7798-4b96-94da-5b5390b3c3e8 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 806.727782] env[61806]: DEBUG oslo_vmware.api [None req-f7f6aee3-3d33-4e4b-b451-fa3af2e320d4 tempest-ServerDiagnosticsV248Test-463927155 tempest-ServerDiagnosticsV248Test-463927155-project-member] Waiting for the task: (returnval){ [ 806.727782] env[61806]: value = "session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]52c9fd34-c5a9-2d16-b882-82ec17c81fd1" [ 806.727782] env[61806]: _type = "Task" [ 806.727782] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 806.738334] env[61806]: DEBUG oslo_vmware.api [None req-f7f6aee3-3d33-4e4b-b451-fa3af2e320d4 tempest-ServerDiagnosticsV248Test-463927155 tempest-ServerDiagnosticsV248Test-463927155-project-member] Task: {'id': session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]52c9fd34-c5a9-2d16-b882-82ec17c81fd1, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 806.793798] env[61806]: DEBUG oslo_concurrency.lockutils [None req-dbeadda9-7825-4678-af18-5f977e49124f tempest-ServerActionsV293TestJSON-906151111 tempest-ServerActionsV293TestJSON-906151111-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.405s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 806.794070] env[61806]: DEBUG nova.compute.manager [None req-dbeadda9-7825-4678-af18-5f977e49124f tempest-ServerActionsV293TestJSON-906151111 tempest-ServerActionsV293TestJSON-906151111-project-member] [instance: 0758901a-7093-41d3-b0e2-5c519333abdd] Start building networks asynchronously for instance. {{(pid=61806) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 806.797615] env[61806]: DEBUG oslo_concurrency.lockutils [None req-5af76806-1cdf-4002-962d-34b8992ee9dc tempest-VolumesAdminNegativeTest-1387177911 tempest-VolumesAdminNegativeTest-1387177911-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 13.068s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 806.799068] env[61806]: INFO nova.compute.claims [None req-5af76806-1cdf-4002-962d-34b8992ee9dc tempest-VolumesAdminNegativeTest-1387177911 tempest-VolumesAdminNegativeTest-1387177911-project-member] [instance: 4373b735-31cf-4b53-b655-38555cf212a5] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 806.900428] env[61806]: DEBUG oslo_vmware.api [None req-6a30dd3c-3e08-4933-b4cc-6fae094056e7 tempest-ServerShowV247Test-1218443059 tempest-ServerShowV247Test-1218443059-project-member] Task: {'id': task-1294563, 'name': Rename_Task, 'duration_secs': 0.155712} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 806.900428] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [None req-6a30dd3c-3e08-4933-b4cc-6fae094056e7 tempest-ServerShowV247Test-1218443059 tempest-ServerShowV247Test-1218443059-project-member] [instance: 4febd093-9f2b-494e-b175-e4693b0e3e0d] Powering on the VM {{(pid=61806) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 806.900428] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-99ea71ec-315f-4059-8729-54875982896d {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 806.906661] env[61806]: DEBUG oslo_vmware.api [None req-6a30dd3c-3e08-4933-b4cc-6fae094056e7 tempest-ServerShowV247Test-1218443059 tempest-ServerShowV247Test-1218443059-project-member] Waiting for the task: (returnval){ [ 806.906661] env[61806]: value = "task-1294567" [ 806.906661] env[61806]: _type = "Task" [ 806.906661] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 806.917461] env[61806]: DEBUG oslo_vmware.api [None req-6a30dd3c-3e08-4933-b4cc-6fae094056e7 tempest-ServerShowV247Test-1218443059 tempest-ServerShowV247Test-1218443059-project-member] Task: {'id': task-1294567, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 806.957547] env[61806]: DEBUG nova.compute.manager [req-16d605ea-5464-403a-b5c9-d5b4eb04c70a req-cc8a1d8c-c282-4e96-99c3-6dad7972e711 service nova] [instance: 4184e263-6f56-4bc1-99b2-a2c460531516] Received event network-vif-deleted-4427e5e3-c696-45c2-b1a8-c89329397198 {{(pid=61806) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 806.957785] env[61806]: INFO nova.compute.manager [req-16d605ea-5464-403a-b5c9-d5b4eb04c70a req-cc8a1d8c-c282-4e96-99c3-6dad7972e711 service nova] [instance: 4184e263-6f56-4bc1-99b2-a2c460531516] Neutron deleted interface 4427e5e3-c696-45c2-b1a8-c89329397198; detaching it from the instance and deleting it from the info cache [ 806.957957] env[61806]: DEBUG nova.network.neutron [req-16d605ea-5464-403a-b5c9-d5b4eb04c70a req-cc8a1d8c-c282-4e96-99c3-6dad7972e711 service nova] [instance: 4184e263-6f56-4bc1-99b2-a2c460531516] Updating instance_info_cache with network_info: [] {{(pid=61806) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 806.970553] env[61806]: DEBUG oslo_concurrency.lockutils [None req-334c6338-dc9d-4c32-ad77-bd3fe007216c tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Acquiring lock "be19fc44-ecbe-489c-9b6e-be8957bb1dd0" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 806.971157] env[61806]: DEBUG oslo_concurrency.lockutils [None req-334c6338-dc9d-4c32-ad77-bd3fe007216c tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Lock "be19fc44-ecbe-489c-9b6e-be8957bb1dd0" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 807.032349] env[61806]: DEBUG oslo_vmware.api [-] Task: {'id': task-1294566, 'name': CreateVM_Task, 'duration_secs': 0.415766} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 807.032542] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 616ec206-9804-469e-ab5c-41aea7f048aa] Created VM on the ESX host {{(pid=61806) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 807.033123] env[61806]: DEBUG oslo_concurrency.lockutils [None req-38045444-a393-40c6-8f1e-16b0707aea48 tempest-ServerShowV257Test-1159790492 tempest-ServerShowV257Test-1159790492-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 807.034037] env[61806]: DEBUG oslo_concurrency.lockutils [None req-38045444-a393-40c6-8f1e-16b0707aea48 tempest-ServerShowV257Test-1159790492 tempest-ServerShowV257Test-1159790492-project-member] Acquired lock "[datastore1] devstack-image-cache_base/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 807.034037] env[61806]: DEBUG oslo_concurrency.lockutils [None req-38045444-a393-40c6-8f1e-16b0707aea48 tempest-ServerShowV257Test-1159790492 tempest-ServerShowV257Test-1159790492-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 807.034037] env[61806]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-6adff0b2-1d1a-4944-acf1-c0cc8b0b580d {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 807.039089] env[61806]: DEBUG oslo_vmware.api [None req-38045444-a393-40c6-8f1e-16b0707aea48 tempest-ServerShowV257Test-1159790492 tempest-ServerShowV257Test-1159790492-project-member] Waiting for the task: (returnval){ [ 807.039089] env[61806]: value = "session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]52082708-4a9e-eab6-ec41-a74c90a1d4ba" [ 807.039089] env[61806]: _type = "Task" [ 807.039089] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 807.046942] env[61806]: DEBUG oslo_vmware.api [None req-38045444-a393-40c6-8f1e-16b0707aea48 tempest-ServerShowV257Test-1159790492 tempest-ServerShowV257Test-1159790492-project-member] Task: {'id': session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]52082708-4a9e-eab6-ec41-a74c90a1d4ba, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 807.238788] env[61806]: DEBUG oslo_vmware.api [None req-f7f6aee3-3d33-4e4b-b451-fa3af2e320d4 tempest-ServerDiagnosticsV248Test-463927155 tempest-ServerDiagnosticsV248Test-463927155-project-member] Task: {'id': session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]52c9fd34-c5a9-2d16-b882-82ec17c81fd1, 'name': SearchDatastore_Task, 'duration_secs': 0.012983} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 807.239150] env[61806]: DEBUG oslo_concurrency.lockutils [None req-f7f6aee3-3d33-4e4b-b451-fa3af2e320d4 tempest-ServerDiagnosticsV248Test-463927155 tempest-ServerDiagnosticsV248Test-463927155-project-member] Releasing lock "[datastore2] devstack-image-cache_base/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa.vmdk" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 807.239451] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [None req-f7f6aee3-3d33-4e4b-b451-fa3af2e320d4 tempest-ServerDiagnosticsV248Test-463927155 tempest-ServerDiagnosticsV248Test-463927155-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa.vmdk to [datastore2] a934d02d-26aa-4900-b473-a58489e5629e/a934d02d-26aa-4900-b473-a58489e5629e.vmdk {{(pid=61806) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 807.239752] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-eff45577-be97-4530-9010-fd13c908dc92 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 807.246629] env[61806]: DEBUG oslo_vmware.api [None req-f7f6aee3-3d33-4e4b-b451-fa3af2e320d4 tempest-ServerDiagnosticsV248Test-463927155 tempest-ServerDiagnosticsV248Test-463927155-project-member] Waiting for the task: (returnval){ [ 807.246629] env[61806]: value = "task-1294568" [ 807.246629] env[61806]: _type = "Task" [ 807.246629] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 807.254906] env[61806]: DEBUG oslo_vmware.api [None req-f7f6aee3-3d33-4e4b-b451-fa3af2e320d4 tempest-ServerDiagnosticsV248Test-463927155 tempest-ServerDiagnosticsV248Test-463927155-project-member] Task: {'id': task-1294568, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 807.303776] env[61806]: DEBUG nova.compute.utils [None req-dbeadda9-7825-4678-af18-5f977e49124f tempest-ServerActionsV293TestJSON-906151111 tempest-ServerActionsV293TestJSON-906151111-project-member] Using /dev/sd instead of None {{(pid=61806) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 807.308299] env[61806]: DEBUG nova.compute.manager [None req-dbeadda9-7825-4678-af18-5f977e49124f tempest-ServerActionsV293TestJSON-906151111 tempest-ServerActionsV293TestJSON-906151111-project-member] [instance: 0758901a-7093-41d3-b0e2-5c519333abdd] Allocating IP information in the background. {{(pid=61806) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 807.308427] env[61806]: DEBUG nova.network.neutron [None req-dbeadda9-7825-4678-af18-5f977e49124f tempest-ServerActionsV293TestJSON-906151111 tempest-ServerActionsV293TestJSON-906151111-project-member] [instance: 0758901a-7093-41d3-b0e2-5c519333abdd] allocate_for_instance() {{(pid=61806) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 807.348697] env[61806]: DEBUG nova.policy [None req-dbeadda9-7825-4678-af18-5f977e49124f tempest-ServerActionsV293TestJSON-906151111 tempest-ServerActionsV293TestJSON-906151111-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'b89b5574639a43b3bc2936b3c00d7b32', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '7e8bbf4274b240619d03f7903cc57ff5', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61806) authorize /opt/stack/nova/nova/policy.py:201}} [ 807.417160] env[61806]: DEBUG oslo_vmware.api [None req-6a30dd3c-3e08-4933-b4cc-6fae094056e7 tempest-ServerShowV247Test-1218443059 tempest-ServerShowV247Test-1218443059-project-member] Task: {'id': task-1294567, 'name': PowerOnVM_Task, 'duration_secs': 0.493487} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 807.417160] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [None req-6a30dd3c-3e08-4933-b4cc-6fae094056e7 tempest-ServerShowV247Test-1218443059 tempest-ServerShowV247Test-1218443059-project-member] [instance: 4febd093-9f2b-494e-b175-e4693b0e3e0d] Powered on the VM {{(pid=61806) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 807.417268] env[61806]: INFO nova.compute.manager [None req-6a30dd3c-3e08-4933-b4cc-6fae094056e7 tempest-ServerShowV247Test-1218443059 tempest-ServerShowV247Test-1218443059-project-member] [instance: 4febd093-9f2b-494e-b175-e4693b0e3e0d] Took 5.82 seconds to spawn the instance on the hypervisor. [ 807.417378] env[61806]: DEBUG nova.compute.manager [None req-6a30dd3c-3e08-4933-b4cc-6fae094056e7 tempest-ServerShowV247Test-1218443059 tempest-ServerShowV247Test-1218443059-project-member] [instance: 4febd093-9f2b-494e-b175-e4693b0e3e0d] Checking state {{(pid=61806) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 807.418222] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c6706f87-f615-4767-91a3-b1b75b1010c3 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 807.434315] env[61806]: DEBUG nova.network.neutron [-] [instance: 4184e263-6f56-4bc1-99b2-a2c460531516] Updating instance_info_cache with network_info: [] {{(pid=61806) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 807.461157] env[61806]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-544036b9-f28d-42a3-b1b5-8040f2c461a3 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 807.473415] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5fac5610-d8de-41fa-9899-006ae1737919 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 807.509019] env[61806]: DEBUG nova.compute.manager [req-16d605ea-5464-403a-b5c9-d5b4eb04c70a req-cc8a1d8c-c282-4e96-99c3-6dad7972e711 service nova] [instance: 4184e263-6f56-4bc1-99b2-a2c460531516] Detach interface failed, port_id=4427e5e3-c696-45c2-b1a8-c89329397198, reason: Instance 4184e263-6f56-4bc1-99b2-a2c460531516 could not be found. {{(pid=61806) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10963}} [ 807.551877] env[61806]: DEBUG oslo_vmware.api [None req-38045444-a393-40c6-8f1e-16b0707aea48 tempest-ServerShowV257Test-1159790492 tempest-ServerShowV257Test-1159790492-project-member] Task: {'id': session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]52082708-4a9e-eab6-ec41-a74c90a1d4ba, 'name': SearchDatastore_Task, 'duration_secs': 0.019039} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 807.552281] env[61806]: DEBUG oslo_concurrency.lockutils [None req-38045444-a393-40c6-8f1e-16b0707aea48 tempest-ServerShowV257Test-1159790492 tempest-ServerShowV257Test-1159790492-project-member] Releasing lock "[datastore1] devstack-image-cache_base/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 807.552538] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-38045444-a393-40c6-8f1e-16b0707aea48 tempest-ServerShowV257Test-1159790492 tempest-ServerShowV257Test-1159790492-project-member] [instance: 616ec206-9804-469e-ab5c-41aea7f048aa] Processing image ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa {{(pid=61806) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 807.552793] env[61806]: DEBUG oslo_concurrency.lockutils [None req-38045444-a393-40c6-8f1e-16b0707aea48 tempest-ServerShowV257Test-1159790492 tempest-ServerShowV257Test-1159790492-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa.vmdk" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 807.553338] env[61806]: DEBUG oslo_concurrency.lockutils [None req-38045444-a393-40c6-8f1e-16b0707aea48 tempest-ServerShowV257Test-1159790492 tempest-ServerShowV257Test-1159790492-project-member] Acquired lock "[datastore1] devstack-image-cache_base/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa.vmdk" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 807.554474] env[61806]: DEBUG nova.virt.vmwareapi.ds_util [None req-38045444-a393-40c6-8f1e-16b0707aea48 tempest-ServerShowV257Test-1159790492 tempest-ServerShowV257Test-1159790492-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61806) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 807.554474] env[61806]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-96af46d3-0f7b-4de8-b5c6-0e7daa48989b {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 807.565330] env[61806]: DEBUG nova.virt.vmwareapi.ds_util [None req-38045444-a393-40c6-8f1e-16b0707aea48 tempest-ServerShowV257Test-1159790492 tempest-ServerShowV257Test-1159790492-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61806) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 807.565594] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-38045444-a393-40c6-8f1e-16b0707aea48 tempest-ServerShowV257Test-1159790492 tempest-ServerShowV257Test-1159790492-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61806) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 807.566343] env[61806]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ee85c511-2744-47cc-abe6-5b0aab7dd5a9 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 807.572529] env[61806]: DEBUG oslo_vmware.api [None req-38045444-a393-40c6-8f1e-16b0707aea48 tempest-ServerShowV257Test-1159790492 tempest-ServerShowV257Test-1159790492-project-member] Waiting for the task: (returnval){ [ 807.572529] env[61806]: value = "session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]52b2744d-9e99-6a91-4024-76d2f91e79f2" [ 807.572529] env[61806]: _type = "Task" [ 807.572529] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 807.581131] env[61806]: DEBUG oslo_vmware.api [None req-38045444-a393-40c6-8f1e-16b0707aea48 tempest-ServerShowV257Test-1159790492 tempest-ServerShowV257Test-1159790492-project-member] Task: {'id': session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]52b2744d-9e99-6a91-4024-76d2f91e79f2, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 807.725841] env[61806]: DEBUG nova.network.neutron [None req-dbeadda9-7825-4678-af18-5f977e49124f tempest-ServerActionsV293TestJSON-906151111 tempest-ServerActionsV293TestJSON-906151111-project-member] [instance: 0758901a-7093-41d3-b0e2-5c519333abdd] Successfully created port: 5ed8d8cc-fa88-438d-80e3-22e0c009373d {{(pid=61806) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 807.759394] env[61806]: DEBUG oslo_vmware.api [None req-f7f6aee3-3d33-4e4b-b451-fa3af2e320d4 tempest-ServerDiagnosticsV248Test-463927155 tempest-ServerDiagnosticsV248Test-463927155-project-member] Task: {'id': task-1294568, 'name': CopyVirtualDisk_Task} progress is 25%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 807.813953] env[61806]: DEBUG nova.compute.manager [None req-dbeadda9-7825-4678-af18-5f977e49124f tempest-ServerActionsV293TestJSON-906151111 tempest-ServerActionsV293TestJSON-906151111-project-member] [instance: 0758901a-7093-41d3-b0e2-5c519333abdd] Start building block device mappings for instance. {{(pid=61806) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 807.943082] env[61806]: INFO nova.compute.manager [-] [instance: 4184e263-6f56-4bc1-99b2-a2c460531516] Took 1.32 seconds to deallocate network for instance. [ 807.952125] env[61806]: INFO nova.compute.manager [None req-6a30dd3c-3e08-4933-b4cc-6fae094056e7 tempest-ServerShowV247Test-1218443059 tempest-ServerShowV247Test-1218443059-project-member] [instance: 4febd093-9f2b-494e-b175-e4693b0e3e0d] Took 26.86 seconds to build instance. [ 808.086804] env[61806]: DEBUG oslo_vmware.api [None req-38045444-a393-40c6-8f1e-16b0707aea48 tempest-ServerShowV257Test-1159790492 tempest-ServerShowV257Test-1159790492-project-member] Task: {'id': session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]52b2744d-9e99-6a91-4024-76d2f91e79f2, 'name': SearchDatastore_Task, 'duration_secs': 0.023113} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 808.092074] env[61806]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-bc01fd2b-be66-429e-a64e-904eea51cc69 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 808.099507] env[61806]: DEBUG oslo_vmware.api [None req-38045444-a393-40c6-8f1e-16b0707aea48 tempest-ServerShowV257Test-1159790492 tempest-ServerShowV257Test-1159790492-project-member] Waiting for the task: (returnval){ [ 808.099507] env[61806]: value = "session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]52c95ba7-08ce-71d8-b3f0-1962dbeba6cf" [ 808.099507] env[61806]: _type = "Task" [ 808.099507] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 808.111511] env[61806]: DEBUG oslo_vmware.api [None req-38045444-a393-40c6-8f1e-16b0707aea48 tempest-ServerShowV257Test-1159790492 tempest-ServerShowV257Test-1159790492-project-member] Task: {'id': session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]52c95ba7-08ce-71d8-b3f0-1962dbeba6cf, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 808.165992] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-857e54d4-d586-4e5b-843c-16dfb1d3c6e5 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 808.170744] env[61806]: DEBUG oslo_vmware.rw_handles [None req-9a79bd3c-e872-4196-bd5a-cef8933311f9 tempest-ImagesTestJSON-618852556 tempest-ImagesTestJSON-618852556-project-member] Getting lease state for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/529481d3-96c1-2bc3-572e-202b8bc69f4d/disk-0.vmdk. {{(pid=61806) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 808.171669] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-85abd70b-efe8-41d1-903a-873196fe2721 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 808.181560] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-788922d5-1c8d-4469-b4fd-abd6bda085e8 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 808.184834] env[61806]: DEBUG oslo_vmware.rw_handles [None req-9a79bd3c-e872-4196-bd5a-cef8933311f9 tempest-ImagesTestJSON-618852556 tempest-ImagesTestJSON-618852556-project-member] Lease for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/529481d3-96c1-2bc3-572e-202b8bc69f4d/disk-0.vmdk is in state: ready. {{(pid=61806) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 808.186181] env[61806]: ERROR oslo_vmware.rw_handles [None req-9a79bd3c-e872-4196-bd5a-cef8933311f9 tempest-ImagesTestJSON-618852556 tempest-ImagesTestJSON-618852556-project-member] Aborting lease for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/529481d3-96c1-2bc3-572e-202b8bc69f4d/disk-0.vmdk due to incomplete transfer. [ 808.186181] env[61806]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseAbort with opID=oslo.vmware-a3823bcd-cd4a-4a0b-a08c-3b07cc7387ba {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 808.218973] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-19be17f9-c220-4c77-ab77-4bd347c5430e {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 808.222199] env[61806]: DEBUG oslo_vmware.rw_handles [None req-9a79bd3c-e872-4196-bd5a-cef8933311f9 tempest-ImagesTestJSON-618852556 tempest-ImagesTestJSON-618852556-project-member] Closed VMDK read handle for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/529481d3-96c1-2bc3-572e-202b8bc69f4d/disk-0.vmdk. {{(pid=61806) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:735}} [ 808.222417] env[61806]: DEBUG nova.virt.vmwareapi.images [None req-9a79bd3c-e872-4196-bd5a-cef8933311f9 tempest-ImagesTestJSON-618852556 tempest-ImagesTestJSON-618852556-project-member] [instance: e0ef0a35-82a5-495b-9d5c-5805e8306390] Uploaded image e1cfdcc0-a515-4a37-bf5d-903e61ffb7f2 to the Glance image server {{(pid=61806) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:473}} [ 808.225021] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [None req-9a79bd3c-e872-4196-bd5a-cef8933311f9 tempest-ImagesTestJSON-618852556 tempest-ImagesTestJSON-618852556-project-member] [instance: e0ef0a35-82a5-495b-9d5c-5805e8306390] Destroying the VM {{(pid=61806) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1358}} [ 808.225314] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-e9c1bae0-786a-4611-a532-45aa094899bf {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 808.233245] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b22baa07-a7ed-4dd6-b636-6c9d1115992d {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 808.237295] env[61806]: DEBUG oslo_vmware.api [None req-9a79bd3c-e872-4196-bd5a-cef8933311f9 tempest-ImagesTestJSON-618852556 tempest-ImagesTestJSON-618852556-project-member] Waiting for the task: (returnval){ [ 808.237295] env[61806]: value = "task-1294569" [ 808.237295] env[61806]: _type = "Task" [ 808.237295] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 808.250366] env[61806]: DEBUG nova.compute.provider_tree [None req-5af76806-1cdf-4002-962d-34b8992ee9dc tempest-VolumesAdminNegativeTest-1387177911 tempest-VolumesAdminNegativeTest-1387177911-project-member] Inventory has not changed in ProviderTree for provider: a2858be1-fd22-4e08-979e-87ad25293407 {{(pid=61806) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 808.260761] env[61806]: DEBUG oslo_vmware.api [None req-9a79bd3c-e872-4196-bd5a-cef8933311f9 tempest-ImagesTestJSON-618852556 tempest-ImagesTestJSON-618852556-project-member] Task: {'id': task-1294569, 'name': Destroy_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 808.266027] env[61806]: DEBUG oslo_vmware.api [None req-f7f6aee3-3d33-4e4b-b451-fa3af2e320d4 tempest-ServerDiagnosticsV248Test-463927155 tempest-ServerDiagnosticsV248Test-463927155-project-member] Task: {'id': task-1294568, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.71629} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 808.266272] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [None req-f7f6aee3-3d33-4e4b-b451-fa3af2e320d4 tempest-ServerDiagnosticsV248Test-463927155 tempest-ServerDiagnosticsV248Test-463927155-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa.vmdk to [datastore2] a934d02d-26aa-4900-b473-a58489e5629e/a934d02d-26aa-4900-b473-a58489e5629e.vmdk {{(pid=61806) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 808.266490] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-f7f6aee3-3d33-4e4b-b451-fa3af2e320d4 tempest-ServerDiagnosticsV248Test-463927155 tempest-ServerDiagnosticsV248Test-463927155-project-member] [instance: a934d02d-26aa-4900-b473-a58489e5629e] Extending root virtual disk to 1048576 {{(pid=61806) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 808.266738] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-c8588887-9c86-4ef5-9f1e-feaa47bf1c93 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 808.272770] env[61806]: DEBUG oslo_vmware.api [None req-f7f6aee3-3d33-4e4b-b451-fa3af2e320d4 tempest-ServerDiagnosticsV248Test-463927155 tempest-ServerDiagnosticsV248Test-463927155-project-member] Waiting for the task: (returnval){ [ 808.272770] env[61806]: value = "task-1294570" [ 808.272770] env[61806]: _type = "Task" [ 808.272770] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 808.280259] env[61806]: DEBUG oslo_vmware.api [None req-f7f6aee3-3d33-4e4b-b451-fa3af2e320d4 tempest-ServerDiagnosticsV248Test-463927155 tempest-ServerDiagnosticsV248Test-463927155-project-member] Task: {'id': task-1294570, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 808.323634] env[61806]: INFO nova.virt.block_device [None req-dbeadda9-7825-4678-af18-5f977e49124f tempest-ServerActionsV293TestJSON-906151111 tempest-ServerActionsV293TestJSON-906151111-project-member] [instance: 0758901a-7093-41d3-b0e2-5c519333abdd] Booting with volume 1b19f6ff-b694-445e-9bc1-d581e9268324 at /dev/sda [ 808.362278] env[61806]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-7cf2b833-0790-4cf7-83cd-2e02c34cf7ca {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 808.372286] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-871d12b1-9d31-4a43-8c22-6cd92064d22f {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 808.399250] env[61806]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-86ccc142-d068-4b63-b3ce-e8d6e90c8adf {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 808.407261] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cf38002e-e476-406c-8c5a-479a8ed374c5 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 808.435540] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aeb0ee5a-56bf-489a-ac74-cebe0ec4f0f6 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 808.441826] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f9a6959d-230e-435c-aba7-6e21348e01dc {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 808.452548] env[61806]: DEBUG oslo_concurrency.lockutils [None req-fbc15506-7910-4d6f-bfbc-c3a1ce19fdcc tempest-ServerMetadataTestJSON-316746592 tempest-ServerMetadataTestJSON-316746592-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 808.461994] env[61806]: DEBUG oslo_concurrency.lockutils [None req-6a30dd3c-3e08-4933-b4cc-6fae094056e7 tempest-ServerShowV247Test-1218443059 tempest-ServerShowV247Test-1218443059-project-member] Lock "4febd093-9f2b-494e-b175-e4693b0e3e0d" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 147.008s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 808.461994] env[61806]: DEBUG nova.virt.block_device [None req-dbeadda9-7825-4678-af18-5f977e49124f tempest-ServerActionsV293TestJSON-906151111 tempest-ServerActionsV293TestJSON-906151111-project-member] [instance: 0758901a-7093-41d3-b0e2-5c519333abdd] Updating existing volume attachment record: 43e7030f-8dfb-4113-9e7c-2eabaf17a86f {{(pid=61806) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 808.610999] env[61806]: DEBUG oslo_vmware.api [None req-38045444-a393-40c6-8f1e-16b0707aea48 tempest-ServerShowV257Test-1159790492 tempest-ServerShowV257Test-1159790492-project-member] Task: {'id': session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]52c95ba7-08ce-71d8-b3f0-1962dbeba6cf, 'name': SearchDatastore_Task, 'duration_secs': 0.027657} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 808.612471] env[61806]: DEBUG oslo_concurrency.lockutils [None req-38045444-a393-40c6-8f1e-16b0707aea48 tempest-ServerShowV257Test-1159790492 tempest-ServerShowV257Test-1159790492-project-member] Releasing lock "[datastore1] devstack-image-cache_base/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa.vmdk" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 808.612750] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [None req-38045444-a393-40c6-8f1e-16b0707aea48 tempest-ServerShowV257Test-1159790492 tempest-ServerShowV257Test-1159790492-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa.vmdk to [datastore1] 616ec206-9804-469e-ab5c-41aea7f048aa/616ec206-9804-469e-ab5c-41aea7f048aa.vmdk {{(pid=61806) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 808.615905] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-db36f27d-43c7-4280-8e09-9e9ac9475612 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 808.626526] env[61806]: DEBUG oslo_vmware.api [None req-38045444-a393-40c6-8f1e-16b0707aea48 tempest-ServerShowV257Test-1159790492 tempest-ServerShowV257Test-1159790492-project-member] Waiting for the task: (returnval){ [ 808.626526] env[61806]: value = "task-1294571" [ 808.626526] env[61806]: _type = "Task" [ 808.626526] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 808.636368] env[61806]: DEBUG oslo_vmware.api [None req-38045444-a393-40c6-8f1e-16b0707aea48 tempest-ServerShowV257Test-1159790492 tempest-ServerShowV257Test-1159790492-project-member] Task: {'id': task-1294571, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 808.749234] env[61806]: DEBUG oslo_vmware.api [None req-9a79bd3c-e872-4196-bd5a-cef8933311f9 tempest-ImagesTestJSON-618852556 tempest-ImagesTestJSON-618852556-project-member] Task: {'id': task-1294569, 'name': Destroy_Task} progress is 33%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 808.761655] env[61806]: DEBUG nova.scheduler.client.report [None req-5af76806-1cdf-4002-962d-34b8992ee9dc tempest-VolumesAdminNegativeTest-1387177911 tempest-VolumesAdminNegativeTest-1387177911-project-member] Inventory has not changed for provider a2858be1-fd22-4e08-979e-87ad25293407 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 138, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61806) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 808.784492] env[61806]: DEBUG oslo_vmware.api [None req-f7f6aee3-3d33-4e4b-b451-fa3af2e320d4 tempest-ServerDiagnosticsV248Test-463927155 tempest-ServerDiagnosticsV248Test-463927155-project-member] Task: {'id': task-1294570, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.147516} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 808.784855] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-f7f6aee3-3d33-4e4b-b451-fa3af2e320d4 tempest-ServerDiagnosticsV248Test-463927155 tempest-ServerDiagnosticsV248Test-463927155-project-member] [instance: a934d02d-26aa-4900-b473-a58489e5629e] Extended root virtual disk {{(pid=61806) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 808.785711] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8312e3be-212b-413a-b354-2bb1823141d4 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 808.810790] env[61806]: DEBUG nova.virt.vmwareapi.volumeops [None req-f7f6aee3-3d33-4e4b-b451-fa3af2e320d4 tempest-ServerDiagnosticsV248Test-463927155 tempest-ServerDiagnosticsV248Test-463927155-project-member] [instance: a934d02d-26aa-4900-b473-a58489e5629e] Reconfiguring VM instance instance-0000003a to attach disk [datastore2] a934d02d-26aa-4900-b473-a58489e5629e/a934d02d-26aa-4900-b473-a58489e5629e.vmdk or device None with type sparse {{(pid=61806) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 808.812271] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-129b50a5-e4db-494b-ab7a-caed9f30e400 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 808.837034] env[61806]: DEBUG oslo_vmware.api [None req-f7f6aee3-3d33-4e4b-b451-fa3af2e320d4 tempest-ServerDiagnosticsV248Test-463927155 tempest-ServerDiagnosticsV248Test-463927155-project-member] Waiting for the task: (returnval){ [ 808.837034] env[61806]: value = "task-1294572" [ 808.837034] env[61806]: _type = "Task" [ 808.837034] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 808.848348] env[61806]: DEBUG oslo_vmware.api [None req-f7f6aee3-3d33-4e4b-b451-fa3af2e320d4 tempest-ServerDiagnosticsV248Test-463927155 tempest-ServerDiagnosticsV248Test-463927155-project-member] Task: {'id': task-1294572, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 808.968455] env[61806]: DEBUG nova.compute.manager [None req-cbb90e6d-7dac-4c30-b860-1f0db7dd091e tempest-MultipleCreateTestJSON-1589015462 tempest-MultipleCreateTestJSON-1589015462-project-member] [instance: ef502a7d-7dc0-4319-8ce3-6a8701288237] Starting instance... {{(pid=61806) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 809.121776] env[61806]: INFO nova.compute.manager [None req-7cf484b6-b263-4f99-a8c0-b55d49de0809 tempest-ServerShowV247Test-1218443059 tempest-ServerShowV247Test-1218443059-project-member] [instance: 4febd093-9f2b-494e-b175-e4693b0e3e0d] Rebuilding instance [ 809.139827] env[61806]: DEBUG oslo_vmware.api [None req-38045444-a393-40c6-8f1e-16b0707aea48 tempest-ServerShowV257Test-1159790492 tempest-ServerShowV257Test-1159790492-project-member] Task: {'id': task-1294571, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 809.175437] env[61806]: DEBUG nova.compute.manager [None req-7cf484b6-b263-4f99-a8c0-b55d49de0809 tempest-ServerShowV247Test-1218443059 tempest-ServerShowV247Test-1218443059-project-member] [instance: 4febd093-9f2b-494e-b175-e4693b0e3e0d] Checking state {{(pid=61806) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 809.176670] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2a4d2ee1-8bdd-4a84-81cc-d0203ba87b5c {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 809.235680] env[61806]: DEBUG nova.compute.manager [req-992270c0-eff1-443c-9908-293983606445 req-1a4da547-2160-49b6-8bc5-231fe6d9f39e service nova] [instance: 0758901a-7093-41d3-b0e2-5c519333abdd] Received event network-vif-plugged-5ed8d8cc-fa88-438d-80e3-22e0c009373d {{(pid=61806) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 809.235944] env[61806]: DEBUG oslo_concurrency.lockutils [req-992270c0-eff1-443c-9908-293983606445 req-1a4da547-2160-49b6-8bc5-231fe6d9f39e service nova] Acquiring lock "0758901a-7093-41d3-b0e2-5c519333abdd-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 809.236211] env[61806]: DEBUG oslo_concurrency.lockutils [req-992270c0-eff1-443c-9908-293983606445 req-1a4da547-2160-49b6-8bc5-231fe6d9f39e service nova] Lock "0758901a-7093-41d3-b0e2-5c519333abdd-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 809.236340] env[61806]: DEBUG oslo_concurrency.lockutils [req-992270c0-eff1-443c-9908-293983606445 req-1a4da547-2160-49b6-8bc5-231fe6d9f39e service nova] Lock "0758901a-7093-41d3-b0e2-5c519333abdd-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 809.236518] env[61806]: DEBUG nova.compute.manager [req-992270c0-eff1-443c-9908-293983606445 req-1a4da547-2160-49b6-8bc5-231fe6d9f39e service nova] [instance: 0758901a-7093-41d3-b0e2-5c519333abdd] No waiting events found dispatching network-vif-plugged-5ed8d8cc-fa88-438d-80e3-22e0c009373d {{(pid=61806) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 809.236694] env[61806]: WARNING nova.compute.manager [req-992270c0-eff1-443c-9908-293983606445 req-1a4da547-2160-49b6-8bc5-231fe6d9f39e service nova] [instance: 0758901a-7093-41d3-b0e2-5c519333abdd] Received unexpected event network-vif-plugged-5ed8d8cc-fa88-438d-80e3-22e0c009373d for instance with vm_state building and task_state block_device_mapping. [ 809.248981] env[61806]: DEBUG oslo_vmware.api [None req-9a79bd3c-e872-4196-bd5a-cef8933311f9 tempest-ImagesTestJSON-618852556 tempest-ImagesTestJSON-618852556-project-member] Task: {'id': task-1294569, 'name': Destroy_Task, 'duration_secs': 0.926827} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 809.249304] env[61806]: INFO nova.virt.vmwareapi.vm_util [None req-9a79bd3c-e872-4196-bd5a-cef8933311f9 tempest-ImagesTestJSON-618852556 tempest-ImagesTestJSON-618852556-project-member] [instance: e0ef0a35-82a5-495b-9d5c-5805e8306390] Destroyed the VM [ 809.249628] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-9a79bd3c-e872-4196-bd5a-cef8933311f9 tempest-ImagesTestJSON-618852556 tempest-ImagesTestJSON-618852556-project-member] [instance: e0ef0a35-82a5-495b-9d5c-5805e8306390] Deleting Snapshot of the VM instance {{(pid=61806) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 809.249879] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-da901814-82d3-4963-9d89-15e6a5912d73 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 809.257795] env[61806]: DEBUG oslo_vmware.api [None req-9a79bd3c-e872-4196-bd5a-cef8933311f9 tempest-ImagesTestJSON-618852556 tempest-ImagesTestJSON-618852556-project-member] Waiting for the task: (returnval){ [ 809.257795] env[61806]: value = "task-1294573" [ 809.257795] env[61806]: _type = "Task" [ 809.257795] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 809.267468] env[61806]: DEBUG oslo_concurrency.lockutils [None req-5af76806-1cdf-4002-962d-34b8992ee9dc tempest-VolumesAdminNegativeTest-1387177911 tempest-VolumesAdminNegativeTest-1387177911-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.470s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 809.267945] env[61806]: DEBUG nova.compute.manager [None req-5af76806-1cdf-4002-962d-34b8992ee9dc tempest-VolumesAdminNegativeTest-1387177911 tempest-VolumesAdminNegativeTest-1387177911-project-member] [instance: 4373b735-31cf-4b53-b655-38555cf212a5] Start building networks asynchronously for instance. {{(pid=61806) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 809.270416] env[61806]: DEBUG oslo_vmware.api [None req-9a79bd3c-e872-4196-bd5a-cef8933311f9 tempest-ImagesTestJSON-618852556 tempest-ImagesTestJSON-618852556-project-member] Task: {'id': task-1294573, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 809.270691] env[61806]: DEBUG oslo_concurrency.lockutils [None req-61b80e06-de2f-412b-be27-b9318eeb5c25 tempest-ServerAddressesTestJSON-661730579 tempest-ServerAddressesTestJSON-661730579-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 14.607s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 809.272117] env[61806]: INFO nova.compute.claims [None req-61b80e06-de2f-412b-be27-b9318eeb5c25 tempest-ServerAddressesTestJSON-661730579 tempest-ServerAddressesTestJSON-661730579-project-member] [instance: a842df44-d8a9-4376-b9fc-5ca19a68a4b7] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 809.342291] env[61806]: DEBUG nova.network.neutron [None req-dbeadda9-7825-4678-af18-5f977e49124f tempest-ServerActionsV293TestJSON-906151111 tempest-ServerActionsV293TestJSON-906151111-project-member] [instance: 0758901a-7093-41d3-b0e2-5c519333abdd] Successfully updated port: 5ed8d8cc-fa88-438d-80e3-22e0c009373d {{(pid=61806) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 809.350565] env[61806]: DEBUG oslo_vmware.api [None req-f7f6aee3-3d33-4e4b-b451-fa3af2e320d4 tempest-ServerDiagnosticsV248Test-463927155 tempest-ServerDiagnosticsV248Test-463927155-project-member] Task: {'id': task-1294572, 'name': ReconfigVM_Task, 'duration_secs': 0.370265} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 809.352649] env[61806]: DEBUG nova.virt.vmwareapi.volumeops [None req-f7f6aee3-3d33-4e4b-b451-fa3af2e320d4 tempest-ServerDiagnosticsV248Test-463927155 tempest-ServerDiagnosticsV248Test-463927155-project-member] [instance: a934d02d-26aa-4900-b473-a58489e5629e] Reconfigured VM instance instance-0000003a to attach disk [datastore2] a934d02d-26aa-4900-b473-a58489e5629e/a934d02d-26aa-4900-b473-a58489e5629e.vmdk or device None with type sparse {{(pid=61806) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 809.353592] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-fc12ff39-be57-48c8-8668-1eb381684a9a {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 809.364070] env[61806]: DEBUG oslo_vmware.api [None req-f7f6aee3-3d33-4e4b-b451-fa3af2e320d4 tempest-ServerDiagnosticsV248Test-463927155 tempest-ServerDiagnosticsV248Test-463927155-project-member] Waiting for the task: (returnval){ [ 809.364070] env[61806]: value = "task-1294574" [ 809.364070] env[61806]: _type = "Task" [ 809.364070] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 809.375210] env[61806]: DEBUG oslo_vmware.api [None req-f7f6aee3-3d33-4e4b-b451-fa3af2e320d4 tempest-ServerDiagnosticsV248Test-463927155 tempest-ServerDiagnosticsV248Test-463927155-project-member] Task: {'id': task-1294574, 'name': Rename_Task} progress is 5%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 809.415568] env[61806]: DEBUG oslo_concurrency.lockutils [None req-4b919e4f-1f0b-40e0-b482-0547e22d55ff tempest-ImagesTestJSON-618852556 tempest-ImagesTestJSON-618852556-project-member] Acquiring lock "28e0baab-8516-42e3-acc2-9b8eb5192f57" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 809.415885] env[61806]: DEBUG oslo_concurrency.lockutils [None req-4b919e4f-1f0b-40e0-b482-0547e22d55ff tempest-ImagesTestJSON-618852556 tempest-ImagesTestJSON-618852556-project-member] Lock "28e0baab-8516-42e3-acc2-9b8eb5192f57" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 809.497151] env[61806]: DEBUG oslo_concurrency.lockutils [None req-cbb90e6d-7dac-4c30-b860-1f0db7dd091e tempest-MultipleCreateTestJSON-1589015462 tempest-MultipleCreateTestJSON-1589015462-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 809.638647] env[61806]: DEBUG oslo_vmware.api [None req-38045444-a393-40c6-8f1e-16b0707aea48 tempest-ServerShowV257Test-1159790492 tempest-ServerShowV257Test-1159790492-project-member] Task: {'id': task-1294571, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.554232} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 809.638965] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [None req-38045444-a393-40c6-8f1e-16b0707aea48 tempest-ServerShowV257Test-1159790492 tempest-ServerShowV257Test-1159790492-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa.vmdk to [datastore1] 616ec206-9804-469e-ab5c-41aea7f048aa/616ec206-9804-469e-ab5c-41aea7f048aa.vmdk {{(pid=61806) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 809.639206] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-38045444-a393-40c6-8f1e-16b0707aea48 tempest-ServerShowV257Test-1159790492 tempest-ServerShowV257Test-1159790492-project-member] [instance: 616ec206-9804-469e-ab5c-41aea7f048aa] Extending root virtual disk to 1048576 {{(pid=61806) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 809.639484] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-fd904a3e-39f3-4ddc-9e21-5789ddd09822 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 809.647479] env[61806]: DEBUG oslo_vmware.api [None req-38045444-a393-40c6-8f1e-16b0707aea48 tempest-ServerShowV257Test-1159790492 tempest-ServerShowV257Test-1159790492-project-member] Waiting for the task: (returnval){ [ 809.647479] env[61806]: value = "task-1294575" [ 809.647479] env[61806]: _type = "Task" [ 809.647479] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 809.656391] env[61806]: DEBUG oslo_vmware.api [None req-38045444-a393-40c6-8f1e-16b0707aea48 tempest-ServerShowV257Test-1159790492 tempest-ServerShowV257Test-1159790492-project-member] Task: {'id': task-1294575, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 809.689347] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [None req-7cf484b6-b263-4f99-a8c0-b55d49de0809 tempest-ServerShowV247Test-1218443059 tempest-ServerShowV247Test-1218443059-project-member] [instance: 4febd093-9f2b-494e-b175-e4693b0e3e0d] Powering off the VM {{(pid=61806) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 809.689692] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-5cf19d88-8b9d-4935-872a-f2a921f22e70 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 809.696784] env[61806]: DEBUG oslo_vmware.api [None req-7cf484b6-b263-4f99-a8c0-b55d49de0809 tempest-ServerShowV247Test-1218443059 tempest-ServerShowV247Test-1218443059-project-member] Waiting for the task: (returnval){ [ 809.696784] env[61806]: value = "task-1294576" [ 809.696784] env[61806]: _type = "Task" [ 809.696784] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 809.705598] env[61806]: DEBUG oslo_vmware.api [None req-7cf484b6-b263-4f99-a8c0-b55d49de0809 tempest-ServerShowV247Test-1218443059 tempest-ServerShowV247Test-1218443059-project-member] Task: {'id': task-1294576, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 809.768636] env[61806]: DEBUG oslo_vmware.api [None req-9a79bd3c-e872-4196-bd5a-cef8933311f9 tempest-ImagesTestJSON-618852556 tempest-ImagesTestJSON-618852556-project-member] Task: {'id': task-1294573, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 809.779063] env[61806]: DEBUG nova.compute.utils [None req-5af76806-1cdf-4002-962d-34b8992ee9dc tempest-VolumesAdminNegativeTest-1387177911 tempest-VolumesAdminNegativeTest-1387177911-project-member] Using /dev/sd instead of None {{(pid=61806) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 809.780627] env[61806]: DEBUG nova.compute.manager [None req-5af76806-1cdf-4002-962d-34b8992ee9dc tempest-VolumesAdminNegativeTest-1387177911 tempest-VolumesAdminNegativeTest-1387177911-project-member] [instance: 4373b735-31cf-4b53-b655-38555cf212a5] Allocating IP information in the background. {{(pid=61806) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 809.780809] env[61806]: DEBUG nova.network.neutron [None req-5af76806-1cdf-4002-962d-34b8992ee9dc tempest-VolumesAdminNegativeTest-1387177911 tempest-VolumesAdminNegativeTest-1387177911-project-member] [instance: 4373b735-31cf-4b53-b655-38555cf212a5] allocate_for_instance() {{(pid=61806) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 809.845226] env[61806]: DEBUG nova.policy [None req-5af76806-1cdf-4002-962d-34b8992ee9dc tempest-VolumesAdminNegativeTest-1387177911 tempest-VolumesAdminNegativeTest-1387177911-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '5b2be9100e8f4a1387600ab9b76961c6', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '44b0a78c155d47d98223d2543ac7673b', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61806) authorize /opt/stack/nova/nova/policy.py:201}} [ 809.847300] env[61806]: DEBUG oslo_concurrency.lockutils [None req-dbeadda9-7825-4678-af18-5f977e49124f tempest-ServerActionsV293TestJSON-906151111 tempest-ServerActionsV293TestJSON-906151111-project-member] Acquiring lock "refresh_cache-0758901a-7093-41d3-b0e2-5c519333abdd" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 809.847479] env[61806]: DEBUG oslo_concurrency.lockutils [None req-dbeadda9-7825-4678-af18-5f977e49124f tempest-ServerActionsV293TestJSON-906151111 tempest-ServerActionsV293TestJSON-906151111-project-member] Acquired lock "refresh_cache-0758901a-7093-41d3-b0e2-5c519333abdd" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 809.847653] env[61806]: DEBUG nova.network.neutron [None req-dbeadda9-7825-4678-af18-5f977e49124f tempest-ServerActionsV293TestJSON-906151111 tempest-ServerActionsV293TestJSON-906151111-project-member] [instance: 0758901a-7093-41d3-b0e2-5c519333abdd] Building network info cache for instance {{(pid=61806) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 809.876534] env[61806]: DEBUG oslo_vmware.api [None req-f7f6aee3-3d33-4e4b-b451-fa3af2e320d4 tempest-ServerDiagnosticsV248Test-463927155 tempest-ServerDiagnosticsV248Test-463927155-project-member] Task: {'id': task-1294574, 'name': Rename_Task, 'duration_secs': 0.155937} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 809.876687] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [None req-f7f6aee3-3d33-4e4b-b451-fa3af2e320d4 tempest-ServerDiagnosticsV248Test-463927155 tempest-ServerDiagnosticsV248Test-463927155-project-member] [instance: a934d02d-26aa-4900-b473-a58489e5629e] Powering on the VM {{(pid=61806) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 809.876898] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-b792ee37-b9d7-46a9-9f48-238ff102d151 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 809.886138] env[61806]: DEBUG oslo_vmware.api [None req-f7f6aee3-3d33-4e4b-b451-fa3af2e320d4 tempest-ServerDiagnosticsV248Test-463927155 tempest-ServerDiagnosticsV248Test-463927155-project-member] Waiting for the task: (returnval){ [ 809.886138] env[61806]: value = "task-1294577" [ 809.886138] env[61806]: _type = "Task" [ 809.886138] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 809.897130] env[61806]: DEBUG oslo_vmware.api [None req-f7f6aee3-3d33-4e4b-b451-fa3af2e320d4 tempest-ServerDiagnosticsV248Test-463927155 tempest-ServerDiagnosticsV248Test-463927155-project-member] Task: {'id': task-1294577, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 810.163991] env[61806]: DEBUG oslo_vmware.api [None req-38045444-a393-40c6-8f1e-16b0707aea48 tempest-ServerShowV257Test-1159790492 tempest-ServerShowV257Test-1159790492-project-member] Task: {'id': task-1294575, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.07549} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 810.164317] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-38045444-a393-40c6-8f1e-16b0707aea48 tempest-ServerShowV257Test-1159790492 tempest-ServerShowV257Test-1159790492-project-member] [instance: 616ec206-9804-469e-ab5c-41aea7f048aa] Extended root virtual disk {{(pid=61806) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 810.165121] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9f5e2955-308f-4267-a6a2-c6e3ad3a360d {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 810.189795] env[61806]: DEBUG nova.virt.vmwareapi.volumeops [None req-38045444-a393-40c6-8f1e-16b0707aea48 tempest-ServerShowV257Test-1159790492 tempest-ServerShowV257Test-1159790492-project-member] [instance: 616ec206-9804-469e-ab5c-41aea7f048aa] Reconfiguring VM instance instance-0000003b to attach disk [datastore1] 616ec206-9804-469e-ab5c-41aea7f048aa/616ec206-9804-469e-ab5c-41aea7f048aa.vmdk or device None with type sparse {{(pid=61806) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 810.190291] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-1c57f405-ff8a-4698-a123-3387b4e4d502 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 810.206902] env[61806]: DEBUG nova.network.neutron [None req-5af76806-1cdf-4002-962d-34b8992ee9dc tempest-VolumesAdminNegativeTest-1387177911 tempest-VolumesAdminNegativeTest-1387177911-project-member] [instance: 4373b735-31cf-4b53-b655-38555cf212a5] Successfully created port: 2df33c67-a8f5-4af4-ae74-29dd266d5054 {{(pid=61806) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 810.219201] env[61806]: DEBUG oslo_vmware.api [None req-7cf484b6-b263-4f99-a8c0-b55d49de0809 tempest-ServerShowV247Test-1218443059 tempest-ServerShowV247Test-1218443059-project-member] Task: {'id': task-1294576, 'name': PowerOffVM_Task, 'duration_secs': 0.127571} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 810.221034] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [None req-7cf484b6-b263-4f99-a8c0-b55d49de0809 tempest-ServerShowV247Test-1218443059 tempest-ServerShowV247Test-1218443059-project-member] [instance: 4febd093-9f2b-494e-b175-e4693b0e3e0d] Powered off the VM {{(pid=61806) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 810.221447] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-7cf484b6-b263-4f99-a8c0-b55d49de0809 tempest-ServerShowV247Test-1218443059 tempest-ServerShowV247Test-1218443059-project-member] [instance: 4febd093-9f2b-494e-b175-e4693b0e3e0d] Destroying instance {{(pid=61806) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 810.221789] env[61806]: DEBUG oslo_vmware.api [None req-38045444-a393-40c6-8f1e-16b0707aea48 tempest-ServerShowV257Test-1159790492 tempest-ServerShowV257Test-1159790492-project-member] Waiting for the task: (returnval){ [ 810.221789] env[61806]: value = "task-1294578" [ 810.221789] env[61806]: _type = "Task" [ 810.221789] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 810.222958] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f0bb343f-87dd-4fe5-b4e8-626132880ec7 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 810.233979] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-7cf484b6-b263-4f99-a8c0-b55d49de0809 tempest-ServerShowV247Test-1218443059 tempest-ServerShowV247Test-1218443059-project-member] [instance: 4febd093-9f2b-494e-b175-e4693b0e3e0d] Unregistering the VM {{(pid=61806) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 810.237460] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-62412d57-bea0-4445-893b-9c40ef49a41f {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 810.239303] env[61806]: DEBUG oslo_vmware.api [None req-38045444-a393-40c6-8f1e-16b0707aea48 tempest-ServerShowV257Test-1159790492 tempest-ServerShowV257Test-1159790492-project-member] Task: {'id': task-1294578, 'name': ReconfigVM_Task} progress is 10%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 810.266976] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-7cf484b6-b263-4f99-a8c0-b55d49de0809 tempest-ServerShowV247Test-1218443059 tempest-ServerShowV247Test-1218443059-project-member] [instance: 4febd093-9f2b-494e-b175-e4693b0e3e0d] Unregistered the VM {{(pid=61806) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 810.268157] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-7cf484b6-b263-4f99-a8c0-b55d49de0809 tempest-ServerShowV247Test-1218443059 tempest-ServerShowV247Test-1218443059-project-member] [instance: 4febd093-9f2b-494e-b175-e4693b0e3e0d] Deleting contents of the VM from datastore datastore2 {{(pid=61806) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 810.268157] env[61806]: DEBUG nova.virt.vmwareapi.ds_util [None req-7cf484b6-b263-4f99-a8c0-b55d49de0809 tempest-ServerShowV247Test-1218443059 tempest-ServerShowV247Test-1218443059-project-member] Deleting the datastore file [datastore2] 4febd093-9f2b-494e-b175-e4693b0e3e0d {{(pid=61806) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 810.268157] env[61806]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-f6b6822c-1131-41e8-8e1b-db88fd9ffa00 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 810.274391] env[61806]: DEBUG oslo_vmware.api [None req-9a79bd3c-e872-4196-bd5a-cef8933311f9 tempest-ImagesTestJSON-618852556 tempest-ImagesTestJSON-618852556-project-member] Task: {'id': task-1294573, 'name': RemoveSnapshot_Task, 'duration_secs': 0.98627} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 810.274712] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-9a79bd3c-e872-4196-bd5a-cef8933311f9 tempest-ImagesTestJSON-618852556 tempest-ImagesTestJSON-618852556-project-member] [instance: e0ef0a35-82a5-495b-9d5c-5805e8306390] Deleted Snapshot of the VM instance {{(pid=61806) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 810.274815] env[61806]: INFO nova.compute.manager [None req-9a79bd3c-e872-4196-bd5a-cef8933311f9 tempest-ImagesTestJSON-618852556 tempest-ImagesTestJSON-618852556-project-member] [instance: e0ef0a35-82a5-495b-9d5c-5805e8306390] Took 15.46 seconds to snapshot the instance on the hypervisor. [ 810.288891] env[61806]: DEBUG oslo_vmware.api [None req-7cf484b6-b263-4f99-a8c0-b55d49de0809 tempest-ServerShowV247Test-1218443059 tempest-ServerShowV247Test-1218443059-project-member] Waiting for the task: (returnval){ [ 810.288891] env[61806]: value = "task-1294580" [ 810.288891] env[61806]: _type = "Task" [ 810.288891] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 810.289764] env[61806]: DEBUG nova.compute.manager [None req-5af76806-1cdf-4002-962d-34b8992ee9dc tempest-VolumesAdminNegativeTest-1387177911 tempest-VolumesAdminNegativeTest-1387177911-project-member] [instance: 4373b735-31cf-4b53-b655-38555cf212a5] Start building block device mappings for instance. {{(pid=61806) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 810.303348] env[61806]: DEBUG oslo_vmware.api [None req-7cf484b6-b263-4f99-a8c0-b55d49de0809 tempest-ServerShowV247Test-1218443059 tempest-ServerShowV247Test-1218443059-project-member] Task: {'id': task-1294580, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 810.399741] env[61806]: DEBUG oslo_vmware.api [None req-f7f6aee3-3d33-4e4b-b451-fa3af2e320d4 tempest-ServerDiagnosticsV248Test-463927155 tempest-ServerDiagnosticsV248Test-463927155-project-member] Task: {'id': task-1294577, 'name': PowerOnVM_Task, 'duration_secs': 0.509197} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 810.400080] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [None req-f7f6aee3-3d33-4e4b-b451-fa3af2e320d4 tempest-ServerDiagnosticsV248Test-463927155 tempest-ServerDiagnosticsV248Test-463927155-project-member] [instance: a934d02d-26aa-4900-b473-a58489e5629e] Powered on the VM {{(pid=61806) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 810.400572] env[61806]: INFO nova.compute.manager [None req-f7f6aee3-3d33-4e4b-b451-fa3af2e320d4 tempest-ServerDiagnosticsV248Test-463927155 tempest-ServerDiagnosticsV248Test-463927155-project-member] [instance: a934d02d-26aa-4900-b473-a58489e5629e] Took 6.37 seconds to spawn the instance on the hypervisor. [ 810.400791] env[61806]: DEBUG nova.compute.manager [None req-f7f6aee3-3d33-4e4b-b451-fa3af2e320d4 tempest-ServerDiagnosticsV248Test-463927155 tempest-ServerDiagnosticsV248Test-463927155-project-member] [instance: a934d02d-26aa-4900-b473-a58489e5629e] Checking state {{(pid=61806) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 810.402016] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5ffe498c-3c58-4b97-a437-fba6e02e2fc1 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 810.414381] env[61806]: DEBUG nova.network.neutron [None req-dbeadda9-7825-4678-af18-5f977e49124f tempest-ServerActionsV293TestJSON-906151111 tempest-ServerActionsV293TestJSON-906151111-project-member] [instance: 0758901a-7093-41d3-b0e2-5c519333abdd] Instance cache missing network info. {{(pid=61806) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 810.573052] env[61806]: DEBUG nova.compute.manager [None req-dbeadda9-7825-4678-af18-5f977e49124f tempest-ServerActionsV293TestJSON-906151111 tempest-ServerActionsV293TestJSON-906151111-project-member] [instance: 0758901a-7093-41d3-b0e2-5c519333abdd] Start spawning the instance on the hypervisor. {{(pid=61806) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 810.573052] env[61806]: DEBUG nova.virt.hardware [None req-dbeadda9-7825-4678-af18-5f977e49124f tempest-ServerActionsV293TestJSON-906151111 tempest-ServerActionsV293TestJSON-906151111-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-15T18:05:10Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format=,created_at=,direct_url=,disk_format=,id=,min_disk=0,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=1073741824,status='active',tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=61806) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 810.573052] env[61806]: DEBUG nova.virt.hardware [None req-dbeadda9-7825-4678-af18-5f977e49124f tempest-ServerActionsV293TestJSON-906151111 tempest-ServerActionsV293TestJSON-906151111-project-member] Flavor limits 0:0:0 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 810.573052] env[61806]: DEBUG nova.virt.hardware [None req-dbeadda9-7825-4678-af18-5f977e49124f tempest-ServerActionsV293TestJSON-906151111 tempest-ServerActionsV293TestJSON-906151111-project-member] Image limits 0:0:0 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 810.573317] env[61806]: DEBUG nova.virt.hardware [None req-dbeadda9-7825-4678-af18-5f977e49124f tempest-ServerActionsV293TestJSON-906151111 tempest-ServerActionsV293TestJSON-906151111-project-member] Flavor pref 0:0:0 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 810.573317] env[61806]: DEBUG nova.virt.hardware [None req-dbeadda9-7825-4678-af18-5f977e49124f tempest-ServerActionsV293TestJSON-906151111 tempest-ServerActionsV293TestJSON-906151111-project-member] Image pref 0:0:0 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 810.573317] env[61806]: DEBUG nova.virt.hardware [None req-dbeadda9-7825-4678-af18-5f977e49124f tempest-ServerActionsV293TestJSON-906151111 tempest-ServerActionsV293TestJSON-906151111-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 810.573485] env[61806]: DEBUG nova.virt.hardware [None req-dbeadda9-7825-4678-af18-5f977e49124f tempest-ServerActionsV293TestJSON-906151111 tempest-ServerActionsV293TestJSON-906151111-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61806) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 810.574042] env[61806]: DEBUG nova.virt.hardware [None req-dbeadda9-7825-4678-af18-5f977e49124f tempest-ServerActionsV293TestJSON-906151111 tempest-ServerActionsV293TestJSON-906151111-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61806) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 810.574042] env[61806]: DEBUG nova.virt.hardware [None req-dbeadda9-7825-4678-af18-5f977e49124f tempest-ServerActionsV293TestJSON-906151111 tempest-ServerActionsV293TestJSON-906151111-project-member] Got 1 possible topologies {{(pid=61806) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 810.574042] env[61806]: DEBUG nova.virt.hardware [None req-dbeadda9-7825-4678-af18-5f977e49124f tempest-ServerActionsV293TestJSON-906151111 tempest-ServerActionsV293TestJSON-906151111-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61806) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 810.574246] env[61806]: DEBUG nova.virt.hardware [None req-dbeadda9-7825-4678-af18-5f977e49124f tempest-ServerActionsV293TestJSON-906151111 tempest-ServerActionsV293TestJSON-906151111-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61806) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 810.575453] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a63926ed-9191-4629-b24e-292538eca0a2 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 810.587647] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c311fb29-e2ee-405b-8aa6-b1b8a0771359 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 810.604964] env[61806]: DEBUG nova.network.neutron [None req-dbeadda9-7825-4678-af18-5f977e49124f tempest-ServerActionsV293TestJSON-906151111 tempest-ServerActionsV293TestJSON-906151111-project-member] [instance: 0758901a-7093-41d3-b0e2-5c519333abdd] Updating instance_info_cache with network_info: [{"id": "5ed8d8cc-fa88-438d-80e3-22e0c009373d", "address": "fa:16:3e:4b:68:4e", "network": {"id": "b087ee9c-cf27-42df-9c9a-a3652a50fbca", "bridge": "br-int", "label": "tempest-ServerActionsV293TestJSON-1492533501-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7e8bbf4274b240619d03f7903cc57ff5", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "92e4d027-e755-417b-8eea-9a8f24b85140", "external-id": "nsx-vlan-transportzone-756", "segmentation_id": 756, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap5ed8d8cc-fa", "ovs_interfaceid": "5ed8d8cc-fa88-438d-80e3-22e0c009373d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61806) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 810.623641] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-470ac5fd-1b38-4bf3-9e2d-d94f4280ac4e {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 810.631515] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f53b8822-1db7-424a-ab07-afffd9699638 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 810.666160] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-70d88259-a754-40a0-b771-a1ab0d3334a9 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 810.674372] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7bf1eced-7099-4af1-9c16-a1d45e673f91 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 810.687650] env[61806]: DEBUG nova.compute.provider_tree [None req-61b80e06-de2f-412b-be27-b9318eeb5c25 tempest-ServerAddressesTestJSON-661730579 tempest-ServerAddressesTestJSON-661730579-project-member] Inventory has not changed in ProviderTree for provider: a2858be1-fd22-4e08-979e-87ad25293407 {{(pid=61806) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 810.735802] env[61806]: DEBUG oslo_vmware.api [None req-38045444-a393-40c6-8f1e-16b0707aea48 tempest-ServerShowV257Test-1159790492 tempest-ServerShowV257Test-1159790492-project-member] Task: {'id': task-1294578, 'name': ReconfigVM_Task, 'duration_secs': 0.314463} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 810.735802] env[61806]: DEBUG nova.virt.vmwareapi.volumeops [None req-38045444-a393-40c6-8f1e-16b0707aea48 tempest-ServerShowV257Test-1159790492 tempest-ServerShowV257Test-1159790492-project-member] [instance: 616ec206-9804-469e-ab5c-41aea7f048aa] Reconfigured VM instance instance-0000003b to attach disk [datastore1] 616ec206-9804-469e-ab5c-41aea7f048aa/616ec206-9804-469e-ab5c-41aea7f048aa.vmdk or device None with type sparse {{(pid=61806) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 810.736279] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-77f63ff7-2199-42f5-b833-7dc77601e658 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 810.743076] env[61806]: DEBUG oslo_vmware.api [None req-38045444-a393-40c6-8f1e-16b0707aea48 tempest-ServerShowV257Test-1159790492 tempest-ServerShowV257Test-1159790492-project-member] Waiting for the task: (returnval){ [ 810.743076] env[61806]: value = "task-1294581" [ 810.743076] env[61806]: _type = "Task" [ 810.743076] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 810.750894] env[61806]: DEBUG oslo_vmware.api [None req-38045444-a393-40c6-8f1e-16b0707aea48 tempest-ServerShowV257Test-1159790492 tempest-ServerShowV257Test-1159790492-project-member] Task: {'id': task-1294581, 'name': Rename_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 810.804029] env[61806]: DEBUG oslo_vmware.api [None req-7cf484b6-b263-4f99-a8c0-b55d49de0809 tempest-ServerShowV247Test-1218443059 tempest-ServerShowV247Test-1218443059-project-member] Task: {'id': task-1294580, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.124526} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 810.804814] env[61806]: DEBUG nova.virt.vmwareapi.ds_util [None req-7cf484b6-b263-4f99-a8c0-b55d49de0809 tempest-ServerShowV247Test-1218443059 tempest-ServerShowV247Test-1218443059-project-member] Deleted the datastore file {{(pid=61806) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 810.804814] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-7cf484b6-b263-4f99-a8c0-b55d49de0809 tempest-ServerShowV247Test-1218443059 tempest-ServerShowV247Test-1218443059-project-member] [instance: 4febd093-9f2b-494e-b175-e4693b0e3e0d] Deleted contents of the VM from datastore datastore2 {{(pid=61806) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 810.804814] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-7cf484b6-b263-4f99-a8c0-b55d49de0809 tempest-ServerShowV247Test-1218443059 tempest-ServerShowV247Test-1218443059-project-member] [instance: 4febd093-9f2b-494e-b175-e4693b0e3e0d] Instance destroyed {{(pid=61806) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 810.925657] env[61806]: INFO nova.compute.manager [None req-f7f6aee3-3d33-4e4b-b451-fa3af2e320d4 tempest-ServerDiagnosticsV248Test-463927155 tempest-ServerDiagnosticsV248Test-463927155-project-member] [instance: a934d02d-26aa-4900-b473-a58489e5629e] Took 28.03 seconds to build instance. [ 810.973958] env[61806]: DEBUG nova.compute.manager [None req-ebc0d068-cd04-41d9-938d-5e857a99d145 tempest-ServerDiagnosticsV248Test-1367482897 tempest-ServerDiagnosticsV248Test-1367482897-project-admin] [instance: a934d02d-26aa-4900-b473-a58489e5629e] Checking state {{(pid=61806) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 810.975441] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-76fb7d2f-9c0a-4734-9e13-e25e68c5d413 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 810.982994] env[61806]: INFO nova.compute.manager [None req-ebc0d068-cd04-41d9-938d-5e857a99d145 tempest-ServerDiagnosticsV248Test-1367482897 tempest-ServerDiagnosticsV248Test-1367482897-project-admin] [instance: a934d02d-26aa-4900-b473-a58489e5629e] Retrieving diagnostics [ 810.983865] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7f03ea98-1217-4ba6-9622-59ae725297bf {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 811.107713] env[61806]: DEBUG oslo_concurrency.lockutils [None req-dbeadda9-7825-4678-af18-5f977e49124f tempest-ServerActionsV293TestJSON-906151111 tempest-ServerActionsV293TestJSON-906151111-project-member] Releasing lock "refresh_cache-0758901a-7093-41d3-b0e2-5c519333abdd" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 811.107970] env[61806]: DEBUG nova.compute.manager [None req-dbeadda9-7825-4678-af18-5f977e49124f tempest-ServerActionsV293TestJSON-906151111 tempest-ServerActionsV293TestJSON-906151111-project-member] [instance: 0758901a-7093-41d3-b0e2-5c519333abdd] Instance network_info: |[{"id": "5ed8d8cc-fa88-438d-80e3-22e0c009373d", "address": "fa:16:3e:4b:68:4e", "network": {"id": "b087ee9c-cf27-42df-9c9a-a3652a50fbca", "bridge": "br-int", "label": "tempest-ServerActionsV293TestJSON-1492533501-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7e8bbf4274b240619d03f7903cc57ff5", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "92e4d027-e755-417b-8eea-9a8f24b85140", "external-id": "nsx-vlan-transportzone-756", "segmentation_id": 756, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap5ed8d8cc-fa", "ovs_interfaceid": "5ed8d8cc-fa88-438d-80e3-22e0c009373d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61806) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 811.108445] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-dbeadda9-7825-4678-af18-5f977e49124f tempest-ServerActionsV293TestJSON-906151111 tempest-ServerActionsV293TestJSON-906151111-project-member] [instance: 0758901a-7093-41d3-b0e2-5c519333abdd] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:4b:68:4e', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '92e4d027-e755-417b-8eea-9a8f24b85140', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '5ed8d8cc-fa88-438d-80e3-22e0c009373d', 'vif_model': 'vmxnet3'}] {{(pid=61806) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 811.115957] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [None req-dbeadda9-7825-4678-af18-5f977e49124f tempest-ServerActionsV293TestJSON-906151111 tempest-ServerActionsV293TestJSON-906151111-project-member] Creating folder: Project (7e8bbf4274b240619d03f7903cc57ff5). Parent ref: group-v277609. {{(pid=61806) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 811.116582] env[61806]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-a929d3a6-7e86-488f-843f-6be69101517c {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 811.130971] env[61806]: WARNING suds.client [-] Web service reported a SOAP processing fault using an unexpected HTTP status code 200. Reporting as an internal server error. [ 811.131146] env[61806]: DEBUG oslo_vmware.api [-] Fault list: [DuplicateName] {{(pid=61806) _invoke_api /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:337}} [ 811.131527] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [None req-dbeadda9-7825-4678-af18-5f977e49124f tempest-ServerActionsV293TestJSON-906151111 tempest-ServerActionsV293TestJSON-906151111-project-member] Folder already exists: Project (7e8bbf4274b240619d03f7903cc57ff5). Parent ref: group-v277609. {{(pid=61806) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 811.131765] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [None req-dbeadda9-7825-4678-af18-5f977e49124f tempest-ServerActionsV293TestJSON-906151111 tempest-ServerActionsV293TestJSON-906151111-project-member] Creating folder: Instances. Parent ref: group-v277623. {{(pid=61806) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 811.132015] env[61806]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-34e0aff9-c84e-4e7f-a3d6-7ffef76b8ee8 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 811.141822] env[61806]: INFO nova.virt.vmwareapi.vm_util [None req-dbeadda9-7825-4678-af18-5f977e49124f tempest-ServerActionsV293TestJSON-906151111 tempest-ServerActionsV293TestJSON-906151111-project-member] Created folder: Instances in parent group-v277623. [ 811.142083] env[61806]: DEBUG oslo.service.loopingcall [None req-dbeadda9-7825-4678-af18-5f977e49124f tempest-ServerActionsV293TestJSON-906151111 tempest-ServerActionsV293TestJSON-906151111-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61806) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 811.142316] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 0758901a-7093-41d3-b0e2-5c519333abdd] Creating VM on the ESX host {{(pid=61806) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 811.142520] env[61806]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-2637652c-33ba-417f-983a-8bbe53377b35 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 811.163556] env[61806]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 811.163556] env[61806]: value = "task-1294584" [ 811.163556] env[61806]: _type = "Task" [ 811.163556] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 811.171965] env[61806]: DEBUG oslo_vmware.api [-] Task: {'id': task-1294584, 'name': CreateVM_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 811.191505] env[61806]: DEBUG nova.scheduler.client.report [None req-61b80e06-de2f-412b-be27-b9318eeb5c25 tempest-ServerAddressesTestJSON-661730579 tempest-ServerAddressesTestJSON-661730579-project-member] Inventory has not changed for provider a2858be1-fd22-4e08-979e-87ad25293407 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 138, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61806) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 811.253987] env[61806]: DEBUG oslo_vmware.api [None req-38045444-a393-40c6-8f1e-16b0707aea48 tempest-ServerShowV257Test-1159790492 tempest-ServerShowV257Test-1159790492-project-member] Task: {'id': task-1294581, 'name': Rename_Task, 'duration_secs': 0.145323} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 811.254278] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [None req-38045444-a393-40c6-8f1e-16b0707aea48 tempest-ServerShowV257Test-1159790492 tempest-ServerShowV257Test-1159790492-project-member] [instance: 616ec206-9804-469e-ab5c-41aea7f048aa] Powering on the VM {{(pid=61806) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 811.254537] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-4e2fabac-db27-471b-b590-241326add9f1 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 811.262423] env[61806]: DEBUG oslo_vmware.api [None req-38045444-a393-40c6-8f1e-16b0707aea48 tempest-ServerShowV257Test-1159790492 tempest-ServerShowV257Test-1159790492-project-member] Waiting for the task: (returnval){ [ 811.262423] env[61806]: value = "task-1294585" [ 811.262423] env[61806]: _type = "Task" [ 811.262423] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 811.270798] env[61806]: DEBUG oslo_vmware.api [None req-38045444-a393-40c6-8f1e-16b0707aea48 tempest-ServerShowV257Test-1159790492 tempest-ServerShowV257Test-1159790492-project-member] Task: {'id': task-1294585, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 811.280833] env[61806]: DEBUG nova.compute.manager [req-e32edd4a-0391-485b-b792-2538e19e7f70 req-c20d0d66-aa77-4fa3-8f33-d39f5d325462 service nova] [instance: 0758901a-7093-41d3-b0e2-5c519333abdd] Received event network-changed-5ed8d8cc-fa88-438d-80e3-22e0c009373d {{(pid=61806) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 811.281050] env[61806]: DEBUG nova.compute.manager [req-e32edd4a-0391-485b-b792-2538e19e7f70 req-c20d0d66-aa77-4fa3-8f33-d39f5d325462 service nova] [instance: 0758901a-7093-41d3-b0e2-5c519333abdd] Refreshing instance network info cache due to event network-changed-5ed8d8cc-fa88-438d-80e3-22e0c009373d. {{(pid=61806) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 811.281270] env[61806]: DEBUG oslo_concurrency.lockutils [req-e32edd4a-0391-485b-b792-2538e19e7f70 req-c20d0d66-aa77-4fa3-8f33-d39f5d325462 service nova] Acquiring lock "refresh_cache-0758901a-7093-41d3-b0e2-5c519333abdd" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 811.281472] env[61806]: DEBUG oslo_concurrency.lockutils [req-e32edd4a-0391-485b-b792-2538e19e7f70 req-c20d0d66-aa77-4fa3-8f33-d39f5d325462 service nova] Acquired lock "refresh_cache-0758901a-7093-41d3-b0e2-5c519333abdd" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 811.281585] env[61806]: DEBUG nova.network.neutron [req-e32edd4a-0391-485b-b792-2538e19e7f70 req-c20d0d66-aa77-4fa3-8f33-d39f5d325462 service nova] [instance: 0758901a-7093-41d3-b0e2-5c519333abdd] Refreshing network info cache for port 5ed8d8cc-fa88-438d-80e3-22e0c009373d {{(pid=61806) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 811.307058] env[61806]: DEBUG nova.compute.manager [None req-5af76806-1cdf-4002-962d-34b8992ee9dc tempest-VolumesAdminNegativeTest-1387177911 tempest-VolumesAdminNegativeTest-1387177911-project-member] [instance: 4373b735-31cf-4b53-b655-38555cf212a5] Start spawning the instance on the hypervisor. {{(pid=61806) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 811.339037] env[61806]: DEBUG nova.virt.hardware [None req-5af76806-1cdf-4002-962d-34b8992ee9dc tempest-VolumesAdminNegativeTest-1387177911 tempest-VolumesAdminNegativeTest-1387177911-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-15T18:05:10Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-15T18:04:52Z,direct_url=,disk_format='vmdk',id=ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='84ba83fce2c349a6988173c5d6fc3b07',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-15T18:04:53Z,virtual_size=,visibility=), allow threads: False {{(pid=61806) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 811.339037] env[61806]: DEBUG nova.virt.hardware [None req-5af76806-1cdf-4002-962d-34b8992ee9dc tempest-VolumesAdminNegativeTest-1387177911 tempest-VolumesAdminNegativeTest-1387177911-project-member] Flavor limits 0:0:0 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 811.339037] env[61806]: DEBUG nova.virt.hardware [None req-5af76806-1cdf-4002-962d-34b8992ee9dc tempest-VolumesAdminNegativeTest-1387177911 tempest-VolumesAdminNegativeTest-1387177911-project-member] Image limits 0:0:0 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 811.339223] env[61806]: DEBUG nova.virt.hardware [None req-5af76806-1cdf-4002-962d-34b8992ee9dc tempest-VolumesAdminNegativeTest-1387177911 tempest-VolumesAdminNegativeTest-1387177911-project-member] Flavor pref 0:0:0 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 811.339223] env[61806]: DEBUG nova.virt.hardware [None req-5af76806-1cdf-4002-962d-34b8992ee9dc tempest-VolumesAdminNegativeTest-1387177911 tempest-VolumesAdminNegativeTest-1387177911-project-member] Image pref 0:0:0 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 811.339370] env[61806]: DEBUG nova.virt.hardware [None req-5af76806-1cdf-4002-962d-34b8992ee9dc tempest-VolumesAdminNegativeTest-1387177911 tempest-VolumesAdminNegativeTest-1387177911-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 811.339616] env[61806]: DEBUG nova.virt.hardware [None req-5af76806-1cdf-4002-962d-34b8992ee9dc tempest-VolumesAdminNegativeTest-1387177911 tempest-VolumesAdminNegativeTest-1387177911-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61806) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 811.339791] env[61806]: DEBUG nova.virt.hardware [None req-5af76806-1cdf-4002-962d-34b8992ee9dc tempest-VolumesAdminNegativeTest-1387177911 tempest-VolumesAdminNegativeTest-1387177911-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61806) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 811.339989] env[61806]: DEBUG nova.virt.hardware [None req-5af76806-1cdf-4002-962d-34b8992ee9dc tempest-VolumesAdminNegativeTest-1387177911 tempest-VolumesAdminNegativeTest-1387177911-project-member] Got 1 possible topologies {{(pid=61806) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 811.340203] env[61806]: DEBUG nova.virt.hardware [None req-5af76806-1cdf-4002-962d-34b8992ee9dc tempest-VolumesAdminNegativeTest-1387177911 tempest-VolumesAdminNegativeTest-1387177911-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61806) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 811.340386] env[61806]: DEBUG nova.virt.hardware [None req-5af76806-1cdf-4002-962d-34b8992ee9dc tempest-VolumesAdminNegativeTest-1387177911 tempest-VolumesAdminNegativeTest-1387177911-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61806) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 811.341410] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a750db31-53aa-466e-8111-a574cda5f1ad {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 811.351174] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a55913b6-c1d8-4650-963d-1398cb49b5e6 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 811.427696] env[61806]: DEBUG oslo_concurrency.lockutils [None req-f7f6aee3-3d33-4e4b-b451-fa3af2e320d4 tempest-ServerDiagnosticsV248Test-463927155 tempest-ServerDiagnosticsV248Test-463927155-project-member] Lock "a934d02d-26aa-4900-b473-a58489e5629e" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 128.956s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 811.680024] env[61806]: DEBUG oslo_vmware.api [-] Task: {'id': task-1294584, 'name': CreateVM_Task, 'duration_secs': 0.384446} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 811.680024] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 0758901a-7093-41d3-b0e2-5c519333abdd] Created VM on the ESX host {{(pid=61806) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 811.680024] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-dbeadda9-7825-4678-af18-5f977e49124f tempest-ServerActionsV293TestJSON-906151111 tempest-ServerActionsV293TestJSON-906151111-project-member] [instance: 0758901a-7093-41d3-b0e2-5c519333abdd] Block device information present: {'root_device_name': '/dev/sda', 'image': [], 'ephemerals': [], 'block_device_mapping': [{'delete_on_termination': True, 'device_type': None, 'attachment_id': '43e7030f-8dfb-4113-9e7c-2eabaf17a86f', 'connection_info': {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-277626', 'volume_id': '1b19f6ff-b694-445e-9bc1-d581e9268324', 'name': 'volume-1b19f6ff-b694-445e-9bc1-d581e9268324', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '0758901a-7093-41d3-b0e2-5c519333abdd', 'attached_at': '', 'detached_at': '', 'volume_id': '1b19f6ff-b694-445e-9bc1-d581e9268324', 'serial': '1b19f6ff-b694-445e-9bc1-d581e9268324'}, 'guest_format': None, 'mount_device': '/dev/sda', 'boot_index': 0, 'disk_bus': None, 'volume_type': None}], 'swap': None} {{(pid=61806) spawn /opt/stack/nova/nova/virt/vmwareapi/vmops.py:799}} [ 811.680408] env[61806]: DEBUG nova.virt.vmwareapi.volumeops [None req-dbeadda9-7825-4678-af18-5f977e49124f tempest-ServerActionsV293TestJSON-906151111 tempest-ServerActionsV293TestJSON-906151111-project-member] [instance: 0758901a-7093-41d3-b0e2-5c519333abdd] Root volume attach. Driver type: vmdk {{(pid=61806) attach_root_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:661}} [ 811.681818] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1b00b5b2-20ab-44a2-bfb6-db8a2068c3b6 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 811.694756] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7744de2a-3df8-47b5-bf75-0c694986dd85 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 811.699678] env[61806]: DEBUG oslo_concurrency.lockutils [None req-61b80e06-de2f-412b-be27-b9318eeb5c25 tempest-ServerAddressesTestJSON-661730579 tempest-ServerAddressesTestJSON-661730579-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.429s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 811.700630] env[61806]: DEBUG nova.compute.manager [None req-61b80e06-de2f-412b-be27-b9318eeb5c25 tempest-ServerAddressesTestJSON-661730579 tempest-ServerAddressesTestJSON-661730579-project-member] [instance: a842df44-d8a9-4376-b9fc-5ca19a68a4b7] Start building networks asynchronously for instance. {{(pid=61806) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 811.704890] env[61806]: DEBUG oslo_concurrency.lockutils [None req-8aa43bc8-f134-4682-8a54-ce9aa25e4ee5 tempest-AttachVolumeNegativeTest-1683033704 tempest-AttachVolumeNegativeTest-1683033704-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 12.354s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 811.707275] env[61806]: INFO nova.compute.claims [None req-8aa43bc8-f134-4682-8a54-ce9aa25e4ee5 tempest-AttachVolumeNegativeTest-1683033704 tempest-AttachVolumeNegativeTest-1683033704-project-member] [instance: 43dcce0d-4fdb-4ee1-8309-92e92d2331a9] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 811.718701] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a10a94f6-f0d7-4a47-9145-78f555d06694 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 811.728285] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.RelocateVM_Task with opID=oslo.vmware-8c588bc5-bc9d-415e-9291-e936f766da36 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 811.739423] env[61806]: DEBUG oslo_vmware.api [None req-dbeadda9-7825-4678-af18-5f977e49124f tempest-ServerActionsV293TestJSON-906151111 tempest-ServerActionsV293TestJSON-906151111-project-member] Waiting for the task: (returnval){ [ 811.739423] env[61806]: value = "task-1294586" [ 811.739423] env[61806]: _type = "Task" [ 811.739423] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 811.752732] env[61806]: DEBUG oslo_vmware.api [None req-dbeadda9-7825-4678-af18-5f977e49124f tempest-ServerActionsV293TestJSON-906151111 tempest-ServerActionsV293TestJSON-906151111-project-member] Task: {'id': task-1294586, 'name': RelocateVM_Task} progress is 5%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 811.775275] env[61806]: DEBUG oslo_vmware.api [None req-38045444-a393-40c6-8f1e-16b0707aea48 tempest-ServerShowV257Test-1159790492 tempest-ServerShowV257Test-1159790492-project-member] Task: {'id': task-1294585, 'name': PowerOnVM_Task} progress is 100%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 811.853944] env[61806]: DEBUG nova.virt.hardware [None req-7cf484b6-b263-4f99-a8c0-b55d49de0809 tempest-ServerShowV247Test-1218443059 tempest-ServerShowV247Test-1218443059-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-15T18:05:10Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-15T18:04:52Z,direct_url=,disk_format='vmdk',id=ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='84ba83fce2c349a6988173c5d6fc3b07',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-15T18:04:53Z,virtual_size=,visibility=), allow threads: False {{(pid=61806) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 811.854224] env[61806]: DEBUG nova.virt.hardware [None req-7cf484b6-b263-4f99-a8c0-b55d49de0809 tempest-ServerShowV247Test-1218443059 tempest-ServerShowV247Test-1218443059-project-member] Flavor limits 0:0:0 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 811.854403] env[61806]: DEBUG nova.virt.hardware [None req-7cf484b6-b263-4f99-a8c0-b55d49de0809 tempest-ServerShowV247Test-1218443059 tempest-ServerShowV247Test-1218443059-project-member] Image limits 0:0:0 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 811.854616] env[61806]: DEBUG nova.virt.hardware [None req-7cf484b6-b263-4f99-a8c0-b55d49de0809 tempest-ServerShowV247Test-1218443059 tempest-ServerShowV247Test-1218443059-project-member] Flavor pref 0:0:0 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 811.854771] env[61806]: DEBUG nova.virt.hardware [None req-7cf484b6-b263-4f99-a8c0-b55d49de0809 tempest-ServerShowV247Test-1218443059 tempest-ServerShowV247Test-1218443059-project-member] Image pref 0:0:0 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 811.854946] env[61806]: DEBUG nova.virt.hardware [None req-7cf484b6-b263-4f99-a8c0-b55d49de0809 tempest-ServerShowV247Test-1218443059 tempest-ServerShowV247Test-1218443059-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 811.855178] env[61806]: DEBUG nova.virt.hardware [None req-7cf484b6-b263-4f99-a8c0-b55d49de0809 tempest-ServerShowV247Test-1218443059 tempest-ServerShowV247Test-1218443059-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61806) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 811.855382] env[61806]: DEBUG nova.virt.hardware [None req-7cf484b6-b263-4f99-a8c0-b55d49de0809 tempest-ServerShowV247Test-1218443059 tempest-ServerShowV247Test-1218443059-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61806) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 811.855515] env[61806]: DEBUG nova.virt.hardware [None req-7cf484b6-b263-4f99-a8c0-b55d49de0809 tempest-ServerShowV247Test-1218443059 tempest-ServerShowV247Test-1218443059-project-member] Got 1 possible topologies {{(pid=61806) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 811.855844] env[61806]: DEBUG nova.virt.hardware [None req-7cf484b6-b263-4f99-a8c0-b55d49de0809 tempest-ServerShowV247Test-1218443059 tempest-ServerShowV247Test-1218443059-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61806) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 811.855950] env[61806]: DEBUG nova.virt.hardware [None req-7cf484b6-b263-4f99-a8c0-b55d49de0809 tempest-ServerShowV247Test-1218443059 tempest-ServerShowV247Test-1218443059-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61806) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 811.856823] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0f18e356-c671-447e-9974-3a356f088cba {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 811.859850] env[61806]: DEBUG nova.network.neutron [None req-5af76806-1cdf-4002-962d-34b8992ee9dc tempest-VolumesAdminNegativeTest-1387177911 tempest-VolumesAdminNegativeTest-1387177911-project-member] [instance: 4373b735-31cf-4b53-b655-38555cf212a5] Successfully updated port: 2df33c67-a8f5-4af4-ae74-29dd266d5054 {{(pid=61806) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 811.868853] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-539f5cac-5e13-4b6a-963b-479c42854ca4 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 811.884343] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-7cf484b6-b263-4f99-a8c0-b55d49de0809 tempest-ServerShowV247Test-1218443059 tempest-ServerShowV247Test-1218443059-project-member] [instance: 4febd093-9f2b-494e-b175-e4693b0e3e0d] Instance VIF info [] {{(pid=61806) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 811.889808] env[61806]: DEBUG oslo.service.loopingcall [None req-7cf484b6-b263-4f99-a8c0-b55d49de0809 tempest-ServerShowV247Test-1218443059 tempest-ServerShowV247Test-1218443059-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61806) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 811.892226] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 4febd093-9f2b-494e-b175-e4693b0e3e0d] Creating VM on the ESX host {{(pid=61806) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 811.892480] env[61806]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-49915e97-00ac-4b21-9b0c-d995f0b42c94 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 811.910624] env[61806]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 811.910624] env[61806]: value = "task-1294587" [ 811.910624] env[61806]: _type = "Task" [ 811.910624] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 811.918663] env[61806]: DEBUG oslo_vmware.api [-] Task: {'id': task-1294587, 'name': CreateVM_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 811.930832] env[61806]: DEBUG nova.compute.manager [None req-0ea3a38f-3871-4412-a6bc-d21c2086d262 tempest-AttachInterfacesV270Test-1974652806 tempest-AttachInterfacesV270Test-1974652806-project-member] [instance: 117139cd-4ed4-4bea-8ad3-f4ac32c692e4] Starting instance... {{(pid=61806) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 812.033763] env[61806]: DEBUG nova.network.neutron [req-e32edd4a-0391-485b-b792-2538e19e7f70 req-c20d0d66-aa77-4fa3-8f33-d39f5d325462 service nova] [instance: 0758901a-7093-41d3-b0e2-5c519333abdd] Updated VIF entry in instance network info cache for port 5ed8d8cc-fa88-438d-80e3-22e0c009373d. {{(pid=61806) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 812.034174] env[61806]: DEBUG nova.network.neutron [req-e32edd4a-0391-485b-b792-2538e19e7f70 req-c20d0d66-aa77-4fa3-8f33-d39f5d325462 service nova] [instance: 0758901a-7093-41d3-b0e2-5c519333abdd] Updating instance_info_cache with network_info: [{"id": "5ed8d8cc-fa88-438d-80e3-22e0c009373d", "address": "fa:16:3e:4b:68:4e", "network": {"id": "b087ee9c-cf27-42df-9c9a-a3652a50fbca", "bridge": "br-int", "label": "tempest-ServerActionsV293TestJSON-1492533501-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7e8bbf4274b240619d03f7903cc57ff5", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "92e4d027-e755-417b-8eea-9a8f24b85140", "external-id": "nsx-vlan-transportzone-756", "segmentation_id": 756, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap5ed8d8cc-fa", "ovs_interfaceid": "5ed8d8cc-fa88-438d-80e3-22e0c009373d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61806) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 812.215038] env[61806]: DEBUG nova.compute.utils [None req-61b80e06-de2f-412b-be27-b9318eeb5c25 tempest-ServerAddressesTestJSON-661730579 tempest-ServerAddressesTestJSON-661730579-project-member] Using /dev/sd instead of None {{(pid=61806) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 812.217914] env[61806]: DEBUG nova.compute.manager [None req-61b80e06-de2f-412b-be27-b9318eeb5c25 tempest-ServerAddressesTestJSON-661730579 tempest-ServerAddressesTestJSON-661730579-project-member] [instance: a842df44-d8a9-4376-b9fc-5ca19a68a4b7] Allocating IP information in the background. {{(pid=61806) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 812.218100] env[61806]: DEBUG nova.network.neutron [None req-61b80e06-de2f-412b-be27-b9318eeb5c25 tempest-ServerAddressesTestJSON-661730579 tempest-ServerAddressesTestJSON-661730579-project-member] [instance: a842df44-d8a9-4376-b9fc-5ca19a68a4b7] allocate_for_instance() {{(pid=61806) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 812.253343] env[61806]: DEBUG oslo_vmware.api [None req-dbeadda9-7825-4678-af18-5f977e49124f tempest-ServerActionsV293TestJSON-906151111 tempest-ServerActionsV293TestJSON-906151111-project-member] Task: {'id': task-1294586, 'name': RelocateVM_Task} progress is 20%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 812.273779] env[61806]: DEBUG oslo_vmware.api [None req-38045444-a393-40c6-8f1e-16b0707aea48 tempest-ServerShowV257Test-1159790492 tempest-ServerShowV257Test-1159790492-project-member] Task: {'id': task-1294585, 'name': PowerOnVM_Task, 'duration_secs': 0.531005} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 812.275387] env[61806]: DEBUG nova.policy [None req-61b80e06-de2f-412b-be27-b9318eeb5c25 tempest-ServerAddressesTestJSON-661730579 tempest-ServerAddressesTestJSON-661730579-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'a340b454c2b54350ad3756efbb425624', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '52cc8936b48b44c3b4f1172b1344cba6', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61806) authorize /opt/stack/nova/nova/policy.py:201}} [ 812.276855] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [None req-38045444-a393-40c6-8f1e-16b0707aea48 tempest-ServerShowV257Test-1159790492 tempest-ServerShowV257Test-1159790492-project-member] [instance: 616ec206-9804-469e-ab5c-41aea7f048aa] Powered on the VM {{(pid=61806) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 812.277108] env[61806]: INFO nova.compute.manager [None req-38045444-a393-40c6-8f1e-16b0707aea48 tempest-ServerShowV257Test-1159790492 tempest-ServerShowV257Test-1159790492-project-member] [instance: 616ec206-9804-469e-ab5c-41aea7f048aa] Took 5.86 seconds to spawn the instance on the hypervisor. [ 812.277277] env[61806]: DEBUG nova.compute.manager [None req-38045444-a393-40c6-8f1e-16b0707aea48 tempest-ServerShowV257Test-1159790492 tempest-ServerShowV257Test-1159790492-project-member] [instance: 616ec206-9804-469e-ab5c-41aea7f048aa] Checking state {{(pid=61806) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 812.278303] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-85e28114-565a-420c-978f-e87652dbcafa {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 812.362604] env[61806]: DEBUG oslo_concurrency.lockutils [None req-5af76806-1cdf-4002-962d-34b8992ee9dc tempest-VolumesAdminNegativeTest-1387177911 tempest-VolumesAdminNegativeTest-1387177911-project-member] Acquiring lock "refresh_cache-4373b735-31cf-4b53-b655-38555cf212a5" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 812.362756] env[61806]: DEBUG oslo_concurrency.lockutils [None req-5af76806-1cdf-4002-962d-34b8992ee9dc tempest-VolumesAdminNegativeTest-1387177911 tempest-VolumesAdminNegativeTest-1387177911-project-member] Acquired lock "refresh_cache-4373b735-31cf-4b53-b655-38555cf212a5" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 812.362910] env[61806]: DEBUG nova.network.neutron [None req-5af76806-1cdf-4002-962d-34b8992ee9dc tempest-VolumesAdminNegativeTest-1387177911 tempest-VolumesAdminNegativeTest-1387177911-project-member] [instance: 4373b735-31cf-4b53-b655-38555cf212a5] Building network info cache for instance {{(pid=61806) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 812.420815] env[61806]: DEBUG oslo_vmware.api [-] Task: {'id': task-1294587, 'name': CreateVM_Task, 'duration_secs': 0.296418} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 812.420943] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 4febd093-9f2b-494e-b175-e4693b0e3e0d] Created VM on the ESX host {{(pid=61806) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 812.421358] env[61806]: DEBUG oslo_concurrency.lockutils [None req-7cf484b6-b263-4f99-a8c0-b55d49de0809 tempest-ServerShowV247Test-1218443059 tempest-ServerShowV247Test-1218443059-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 812.422052] env[61806]: DEBUG oslo_concurrency.lockutils [None req-7cf484b6-b263-4f99-a8c0-b55d49de0809 tempest-ServerShowV247Test-1218443059 tempest-ServerShowV247Test-1218443059-project-member] Acquired lock "[datastore1] devstack-image-cache_base/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 812.422052] env[61806]: DEBUG oslo_concurrency.lockutils [None req-7cf484b6-b263-4f99-a8c0-b55d49de0809 tempest-ServerShowV247Test-1218443059 tempest-ServerShowV247Test-1218443059-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 812.422201] env[61806]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-548d5086-0c85-4cbf-9fc9-b7d92b5347d2 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 812.427192] env[61806]: DEBUG oslo_vmware.api [None req-7cf484b6-b263-4f99-a8c0-b55d49de0809 tempest-ServerShowV247Test-1218443059 tempest-ServerShowV247Test-1218443059-project-member] Waiting for the task: (returnval){ [ 812.427192] env[61806]: value = "session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]52e1c675-de02-a322-7578-c450463f5499" [ 812.427192] env[61806]: _type = "Task" [ 812.427192] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 812.438151] env[61806]: DEBUG oslo_vmware.api [None req-7cf484b6-b263-4f99-a8c0-b55d49de0809 tempest-ServerShowV247Test-1218443059 tempest-ServerShowV247Test-1218443059-project-member] Task: {'id': session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]52e1c675-de02-a322-7578-c450463f5499, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 812.460361] env[61806]: DEBUG oslo_concurrency.lockutils [None req-0ea3a38f-3871-4412-a6bc-d21c2086d262 tempest-AttachInterfacesV270Test-1974652806 tempest-AttachInterfacesV270Test-1974652806-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 812.536757] env[61806]: DEBUG oslo_concurrency.lockutils [req-e32edd4a-0391-485b-b792-2538e19e7f70 req-c20d0d66-aa77-4fa3-8f33-d39f5d325462 service nova] Releasing lock "refresh_cache-0758901a-7093-41d3-b0e2-5c519333abdd" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 812.587127] env[61806]: DEBUG nova.network.neutron [None req-61b80e06-de2f-412b-be27-b9318eeb5c25 tempest-ServerAddressesTestJSON-661730579 tempest-ServerAddressesTestJSON-661730579-project-member] [instance: a842df44-d8a9-4376-b9fc-5ca19a68a4b7] Successfully created port: cb3cf8c0-2bc5-4308-899d-d6d2734f7922 {{(pid=61806) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 812.719334] env[61806]: DEBUG nova.compute.manager [None req-61b80e06-de2f-412b-be27-b9318eeb5c25 tempest-ServerAddressesTestJSON-661730579 tempest-ServerAddressesTestJSON-661730579-project-member] [instance: a842df44-d8a9-4376-b9fc-5ca19a68a4b7] Start building block device mappings for instance. {{(pid=61806) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 812.755022] env[61806]: DEBUG oslo_vmware.api [None req-dbeadda9-7825-4678-af18-5f977e49124f tempest-ServerActionsV293TestJSON-906151111 tempest-ServerActionsV293TestJSON-906151111-project-member] Task: {'id': task-1294586, 'name': RelocateVM_Task, 'duration_secs': 0.546808} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 812.755022] env[61806]: DEBUG nova.virt.vmwareapi.volumeops [None req-dbeadda9-7825-4678-af18-5f977e49124f tempest-ServerActionsV293TestJSON-906151111 tempest-ServerActionsV293TestJSON-906151111-project-member] [instance: 0758901a-7093-41d3-b0e2-5c519333abdd] Volume attach. Driver type: vmdk {{(pid=61806) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 812.755022] env[61806]: DEBUG nova.virt.vmwareapi.volumeops [None req-dbeadda9-7825-4678-af18-5f977e49124f tempest-ServerActionsV293TestJSON-906151111 tempest-ServerActionsV293TestJSON-906151111-project-member] [instance: 0758901a-7093-41d3-b0e2-5c519333abdd] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-277626', 'volume_id': '1b19f6ff-b694-445e-9bc1-d581e9268324', 'name': 'volume-1b19f6ff-b694-445e-9bc1-d581e9268324', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '0758901a-7093-41d3-b0e2-5c519333abdd', 'attached_at': '', 'detached_at': '', 'volume_id': '1b19f6ff-b694-445e-9bc1-d581e9268324', 'serial': '1b19f6ff-b694-445e-9bc1-d581e9268324'} {{(pid=61806) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 812.755613] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ee079693-000b-4701-b955-c16f4bb10daa {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 812.778886] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-24d797e6-7522-4431-a76b-db7eff22f1fd {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 812.813187] env[61806]: DEBUG nova.virt.vmwareapi.volumeops [None req-dbeadda9-7825-4678-af18-5f977e49124f tempest-ServerActionsV293TestJSON-906151111 tempest-ServerActionsV293TestJSON-906151111-project-member] [instance: 0758901a-7093-41d3-b0e2-5c519333abdd] Reconfiguring VM instance instance-0000003c to attach disk [datastore1] volume-1b19f6ff-b694-445e-9bc1-d581e9268324/volume-1b19f6ff-b694-445e-9bc1-d581e9268324.vmdk or device None with type thin {{(pid=61806) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 812.818167] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-557d4541-d4f8-4388-844d-9fb1085899c3 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 812.835505] env[61806]: INFO nova.compute.manager [None req-38045444-a393-40c6-8f1e-16b0707aea48 tempest-ServerShowV257Test-1159790492 tempest-ServerShowV257Test-1159790492-project-member] [instance: 616ec206-9804-469e-ab5c-41aea7f048aa] Took 27.92 seconds to build instance. [ 812.844238] env[61806]: DEBUG oslo_vmware.api [None req-dbeadda9-7825-4678-af18-5f977e49124f tempest-ServerActionsV293TestJSON-906151111 tempest-ServerActionsV293TestJSON-906151111-project-member] Waiting for the task: (returnval){ [ 812.844238] env[61806]: value = "task-1294588" [ 812.844238] env[61806]: _type = "Task" [ 812.844238] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 812.857921] env[61806]: DEBUG oslo_vmware.api [None req-dbeadda9-7825-4678-af18-5f977e49124f tempest-ServerActionsV293TestJSON-906151111 tempest-ServerActionsV293TestJSON-906151111-project-member] Task: {'id': task-1294588, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 812.903418] env[61806]: DEBUG nova.network.neutron [None req-5af76806-1cdf-4002-962d-34b8992ee9dc tempest-VolumesAdminNegativeTest-1387177911 tempest-VolumesAdminNegativeTest-1387177911-project-member] [instance: 4373b735-31cf-4b53-b655-38555cf212a5] Instance cache missing network info. {{(pid=61806) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 812.939271] env[61806]: DEBUG oslo_vmware.api [None req-7cf484b6-b263-4f99-a8c0-b55d49de0809 tempest-ServerShowV247Test-1218443059 tempest-ServerShowV247Test-1218443059-project-member] Task: {'id': session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]52e1c675-de02-a322-7578-c450463f5499, 'name': SearchDatastore_Task, 'duration_secs': 0.019844} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 812.942514] env[61806]: DEBUG oslo_concurrency.lockutils [None req-7cf484b6-b263-4f99-a8c0-b55d49de0809 tempest-ServerShowV247Test-1218443059 tempest-ServerShowV247Test-1218443059-project-member] Releasing lock "[datastore1] devstack-image-cache_base/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 812.942824] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-7cf484b6-b263-4f99-a8c0-b55d49de0809 tempest-ServerShowV247Test-1218443059 tempest-ServerShowV247Test-1218443059-project-member] [instance: 4febd093-9f2b-494e-b175-e4693b0e3e0d] Processing image ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa {{(pid=61806) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 812.943115] env[61806]: DEBUG oslo_concurrency.lockutils [None req-7cf484b6-b263-4f99-a8c0-b55d49de0809 tempest-ServerShowV247Test-1218443059 tempest-ServerShowV247Test-1218443059-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa.vmdk" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 812.943302] env[61806]: DEBUG oslo_concurrency.lockutils [None req-7cf484b6-b263-4f99-a8c0-b55d49de0809 tempest-ServerShowV247Test-1218443059 tempest-ServerShowV247Test-1218443059-project-member] Acquired lock "[datastore1] devstack-image-cache_base/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa.vmdk" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 812.943512] env[61806]: DEBUG nova.virt.vmwareapi.ds_util [None req-7cf484b6-b263-4f99-a8c0-b55d49de0809 tempest-ServerShowV247Test-1218443059 tempest-ServerShowV247Test-1218443059-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61806) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 812.946870] env[61806]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-6efc60f4-e4ca-4e37-b654-62b42c65b51d {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 812.959773] env[61806]: DEBUG nova.virt.vmwareapi.ds_util [None req-7cf484b6-b263-4f99-a8c0-b55d49de0809 tempest-ServerShowV247Test-1218443059 tempest-ServerShowV247Test-1218443059-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61806) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 812.960018] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-7cf484b6-b263-4f99-a8c0-b55d49de0809 tempest-ServerShowV247Test-1218443059 tempest-ServerShowV247Test-1218443059-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61806) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 812.960876] env[61806]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-76069ad7-5831-41fd-aa7e-2f2dfa21cfb0 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 812.968197] env[61806]: DEBUG oslo_vmware.api [None req-7cf484b6-b263-4f99-a8c0-b55d49de0809 tempest-ServerShowV247Test-1218443059 tempest-ServerShowV247Test-1218443059-project-member] Waiting for the task: (returnval){ [ 812.968197] env[61806]: value = "session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]52409612-bfec-1690-fd02-11bada93516e" [ 812.968197] env[61806]: _type = "Task" [ 812.968197] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 812.981181] env[61806]: DEBUG oslo_vmware.api [None req-7cf484b6-b263-4f99-a8c0-b55d49de0809 tempest-ServerShowV247Test-1218443059 tempest-ServerShowV247Test-1218443059-project-member] Task: {'id': session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]52409612-bfec-1690-fd02-11bada93516e, 'name': SearchDatastore_Task, 'duration_secs': 0.010396} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 812.986873] env[61806]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-695d09ea-aa11-4161-b9e3-41275397871d {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 812.994416] env[61806]: DEBUG oslo_vmware.api [None req-7cf484b6-b263-4f99-a8c0-b55d49de0809 tempest-ServerShowV247Test-1218443059 tempest-ServerShowV247Test-1218443059-project-member] Waiting for the task: (returnval){ [ 812.994416] env[61806]: value = "session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]5294e416-6c3a-a7e2-e551-96e4f6d525f7" [ 812.994416] env[61806]: _type = "Task" [ 812.994416] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 813.003372] env[61806]: DEBUG oslo_vmware.api [None req-7cf484b6-b263-4f99-a8c0-b55d49de0809 tempest-ServerShowV247Test-1218443059 tempest-ServerShowV247Test-1218443059-project-member] Task: {'id': session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]5294e416-6c3a-a7e2-e551-96e4f6d525f7, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 813.077659] env[61806]: DEBUG nova.network.neutron [None req-5af76806-1cdf-4002-962d-34b8992ee9dc tempest-VolumesAdminNegativeTest-1387177911 tempest-VolumesAdminNegativeTest-1387177911-project-member] [instance: 4373b735-31cf-4b53-b655-38555cf212a5] Updating instance_info_cache with network_info: [{"id": "2df33c67-a8f5-4af4-ae74-29dd266d5054", "address": "fa:16:3e:fe:14:20", "network": {"id": "2767c0cb-36c3-4a3a-9755-278a7c2296b5", "bridge": "br-int", "label": "tempest-VolumesAdminNegativeTest-745798855-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "44b0a78c155d47d98223d2543ac7673b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "abcf0d10-3f3f-45dc-923e-1c78766e2dad", "external-id": "nsx-vlan-transportzone-405", "segmentation_id": 405, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2df33c67-a8", "ovs_interfaceid": "2df33c67-a8f5-4af4-ae74-29dd266d5054", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61806) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 813.098367] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7a494362-9802-4fd8-8755-ca43350d1516 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 813.109544] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1aef3e62-a407-4ffa-a8ec-a0b0a9250e52 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 813.142807] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9d977816-98fe-4e0b-a22e-3b2d1d57b48d {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 813.152094] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4ca016a3-6ba4-40c1-ad86-8ca2aa16874f {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 813.170395] env[61806]: DEBUG nova.compute.provider_tree [None req-8aa43bc8-f134-4682-8a54-ce9aa25e4ee5 tempest-AttachVolumeNegativeTest-1683033704 tempest-AttachVolumeNegativeTest-1683033704-project-member] Inventory has not changed in ProviderTree for provider: a2858be1-fd22-4e08-979e-87ad25293407 {{(pid=61806) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 813.177023] env[61806]: INFO nova.compute.manager [None req-4e219b6d-f08a-4088-99ff-2fcd194df36a tempest-ServerShowV257Test-1159790492 tempest-ServerShowV257Test-1159790492-project-member] [instance: 616ec206-9804-469e-ab5c-41aea7f048aa] Rebuilding instance [ 813.216374] env[61806]: DEBUG nova.compute.manager [None req-4e219b6d-f08a-4088-99ff-2fcd194df36a tempest-ServerShowV257Test-1159790492 tempest-ServerShowV257Test-1159790492-project-member] [instance: 616ec206-9804-469e-ab5c-41aea7f048aa] Checking state {{(pid=61806) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 813.217266] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-068d236e-a988-4199-8e6e-2bdb1b836f33 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 813.338346] env[61806]: DEBUG oslo_concurrency.lockutils [None req-38045444-a393-40c6-8f1e-16b0707aea48 tempest-ServerShowV257Test-1159790492 tempest-ServerShowV257Test-1159790492-project-member] Lock "616ec206-9804-469e-ab5c-41aea7f048aa" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 118.092s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 813.345754] env[61806]: DEBUG nova.compute.manager [req-73ea1082-8136-4be2-8518-198322c844e8 req-b5b52310-ce7c-4f5a-96a9-1c3d550bb745 service nova] [instance: 4373b735-31cf-4b53-b655-38555cf212a5] Received event network-vif-plugged-2df33c67-a8f5-4af4-ae74-29dd266d5054 {{(pid=61806) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 813.346042] env[61806]: DEBUG oslo_concurrency.lockutils [req-73ea1082-8136-4be2-8518-198322c844e8 req-b5b52310-ce7c-4f5a-96a9-1c3d550bb745 service nova] Acquiring lock "4373b735-31cf-4b53-b655-38555cf212a5-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 813.346289] env[61806]: DEBUG oslo_concurrency.lockutils [req-73ea1082-8136-4be2-8518-198322c844e8 req-b5b52310-ce7c-4f5a-96a9-1c3d550bb745 service nova] Lock "4373b735-31cf-4b53-b655-38555cf212a5-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 813.346355] env[61806]: DEBUG oslo_concurrency.lockutils [req-73ea1082-8136-4be2-8518-198322c844e8 req-b5b52310-ce7c-4f5a-96a9-1c3d550bb745 service nova] Lock "4373b735-31cf-4b53-b655-38555cf212a5-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 813.346526] env[61806]: DEBUG nova.compute.manager [req-73ea1082-8136-4be2-8518-198322c844e8 req-b5b52310-ce7c-4f5a-96a9-1c3d550bb745 service nova] [instance: 4373b735-31cf-4b53-b655-38555cf212a5] No waiting events found dispatching network-vif-plugged-2df33c67-a8f5-4af4-ae74-29dd266d5054 {{(pid=61806) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 813.346704] env[61806]: WARNING nova.compute.manager [req-73ea1082-8136-4be2-8518-198322c844e8 req-b5b52310-ce7c-4f5a-96a9-1c3d550bb745 service nova] [instance: 4373b735-31cf-4b53-b655-38555cf212a5] Received unexpected event network-vif-plugged-2df33c67-a8f5-4af4-ae74-29dd266d5054 for instance with vm_state building and task_state spawning. [ 813.346874] env[61806]: DEBUG nova.compute.manager [req-73ea1082-8136-4be2-8518-198322c844e8 req-b5b52310-ce7c-4f5a-96a9-1c3d550bb745 service nova] [instance: 4373b735-31cf-4b53-b655-38555cf212a5] Received event network-changed-2df33c67-a8f5-4af4-ae74-29dd266d5054 {{(pid=61806) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 813.347078] env[61806]: DEBUG nova.compute.manager [req-73ea1082-8136-4be2-8518-198322c844e8 req-b5b52310-ce7c-4f5a-96a9-1c3d550bb745 service nova] [instance: 4373b735-31cf-4b53-b655-38555cf212a5] Refreshing instance network info cache due to event network-changed-2df33c67-a8f5-4af4-ae74-29dd266d5054. {{(pid=61806) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 813.347241] env[61806]: DEBUG oslo_concurrency.lockutils [req-73ea1082-8136-4be2-8518-198322c844e8 req-b5b52310-ce7c-4f5a-96a9-1c3d550bb745 service nova] Acquiring lock "refresh_cache-4373b735-31cf-4b53-b655-38555cf212a5" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 813.358753] env[61806]: DEBUG oslo_vmware.api [None req-dbeadda9-7825-4678-af18-5f977e49124f tempest-ServerActionsV293TestJSON-906151111 tempest-ServerActionsV293TestJSON-906151111-project-member] Task: {'id': task-1294588, 'name': ReconfigVM_Task, 'duration_secs': 0.359462} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 813.359033] env[61806]: DEBUG nova.virt.vmwareapi.volumeops [None req-dbeadda9-7825-4678-af18-5f977e49124f tempest-ServerActionsV293TestJSON-906151111 tempest-ServerActionsV293TestJSON-906151111-project-member] [instance: 0758901a-7093-41d3-b0e2-5c519333abdd] Reconfigured VM instance instance-0000003c to attach disk [datastore1] volume-1b19f6ff-b694-445e-9bc1-d581e9268324/volume-1b19f6ff-b694-445e-9bc1-d581e9268324.vmdk or device None with type thin {{(pid=61806) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 813.364329] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-638ca69d-d14a-4dfc-9346-6bd108c17def {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 813.383855] env[61806]: DEBUG oslo_vmware.api [None req-dbeadda9-7825-4678-af18-5f977e49124f tempest-ServerActionsV293TestJSON-906151111 tempest-ServerActionsV293TestJSON-906151111-project-member] Waiting for the task: (returnval){ [ 813.383855] env[61806]: value = "task-1294589" [ 813.383855] env[61806]: _type = "Task" [ 813.383855] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 813.395899] env[61806]: DEBUG oslo_vmware.api [None req-dbeadda9-7825-4678-af18-5f977e49124f tempest-ServerActionsV293TestJSON-906151111 tempest-ServerActionsV293TestJSON-906151111-project-member] Task: {'id': task-1294589, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 813.505458] env[61806]: DEBUG oslo_vmware.api [None req-7cf484b6-b263-4f99-a8c0-b55d49de0809 tempest-ServerShowV247Test-1218443059 tempest-ServerShowV247Test-1218443059-project-member] Task: {'id': session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]5294e416-6c3a-a7e2-e551-96e4f6d525f7, 'name': SearchDatastore_Task, 'duration_secs': 0.01075} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 813.505753] env[61806]: DEBUG oslo_concurrency.lockutils [None req-7cf484b6-b263-4f99-a8c0-b55d49de0809 tempest-ServerShowV247Test-1218443059 tempest-ServerShowV247Test-1218443059-project-member] Releasing lock "[datastore1] devstack-image-cache_base/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa.vmdk" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 813.506031] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [None req-7cf484b6-b263-4f99-a8c0-b55d49de0809 tempest-ServerShowV247Test-1218443059 tempest-ServerShowV247Test-1218443059-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa.vmdk to [datastore1] 4febd093-9f2b-494e-b175-e4693b0e3e0d/4febd093-9f2b-494e-b175-e4693b0e3e0d.vmdk {{(pid=61806) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 813.506321] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-8599134f-e01b-4f2e-a87b-f13e90d3e758 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 813.515300] env[61806]: DEBUG oslo_vmware.api [None req-7cf484b6-b263-4f99-a8c0-b55d49de0809 tempest-ServerShowV247Test-1218443059 tempest-ServerShowV247Test-1218443059-project-member] Waiting for the task: (returnval){ [ 813.515300] env[61806]: value = "task-1294590" [ 813.515300] env[61806]: _type = "Task" [ 813.515300] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 813.523853] env[61806]: DEBUG oslo_vmware.api [None req-7cf484b6-b263-4f99-a8c0-b55d49de0809 tempest-ServerShowV247Test-1218443059 tempest-ServerShowV247Test-1218443059-project-member] Task: {'id': task-1294590, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 813.580864] env[61806]: DEBUG oslo_concurrency.lockutils [None req-5af76806-1cdf-4002-962d-34b8992ee9dc tempest-VolumesAdminNegativeTest-1387177911 tempest-VolumesAdminNegativeTest-1387177911-project-member] Releasing lock "refresh_cache-4373b735-31cf-4b53-b655-38555cf212a5" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 813.581249] env[61806]: DEBUG nova.compute.manager [None req-5af76806-1cdf-4002-962d-34b8992ee9dc tempest-VolumesAdminNegativeTest-1387177911 tempest-VolumesAdminNegativeTest-1387177911-project-member] [instance: 4373b735-31cf-4b53-b655-38555cf212a5] Instance network_info: |[{"id": "2df33c67-a8f5-4af4-ae74-29dd266d5054", "address": "fa:16:3e:fe:14:20", "network": {"id": "2767c0cb-36c3-4a3a-9755-278a7c2296b5", "bridge": "br-int", "label": "tempest-VolumesAdminNegativeTest-745798855-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "44b0a78c155d47d98223d2543ac7673b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "abcf0d10-3f3f-45dc-923e-1c78766e2dad", "external-id": "nsx-vlan-transportzone-405", "segmentation_id": 405, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2df33c67-a8", "ovs_interfaceid": "2df33c67-a8f5-4af4-ae74-29dd266d5054", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61806) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 813.581710] env[61806]: DEBUG oslo_concurrency.lockutils [req-73ea1082-8136-4be2-8518-198322c844e8 req-b5b52310-ce7c-4f5a-96a9-1c3d550bb745 service nova] Acquired lock "refresh_cache-4373b735-31cf-4b53-b655-38555cf212a5" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 813.582104] env[61806]: DEBUG nova.network.neutron [req-73ea1082-8136-4be2-8518-198322c844e8 req-b5b52310-ce7c-4f5a-96a9-1c3d550bb745 service nova] [instance: 4373b735-31cf-4b53-b655-38555cf212a5] Refreshing network info cache for port 2df33c67-a8f5-4af4-ae74-29dd266d5054 {{(pid=61806) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 813.583685] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-5af76806-1cdf-4002-962d-34b8992ee9dc tempest-VolumesAdminNegativeTest-1387177911 tempest-VolumesAdminNegativeTest-1387177911-project-member] [instance: 4373b735-31cf-4b53-b655-38555cf212a5] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:fe:14:20', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'abcf0d10-3f3f-45dc-923e-1c78766e2dad', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '2df33c67-a8f5-4af4-ae74-29dd266d5054', 'vif_model': 'vmxnet3'}] {{(pid=61806) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 813.592709] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [None req-5af76806-1cdf-4002-962d-34b8992ee9dc tempest-VolumesAdminNegativeTest-1387177911 tempest-VolumesAdminNegativeTest-1387177911-project-member] Creating folder: Project (44b0a78c155d47d98223d2543ac7673b). Parent ref: group-v277609. {{(pid=61806) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 813.595831] env[61806]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-7f8b4407-5c1d-415c-9037-ae7b29232ae3 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 813.609622] env[61806]: INFO nova.virt.vmwareapi.vm_util [None req-5af76806-1cdf-4002-962d-34b8992ee9dc tempest-VolumesAdminNegativeTest-1387177911 tempest-VolumesAdminNegativeTest-1387177911-project-member] Created folder: Project (44b0a78c155d47d98223d2543ac7673b) in parent group-v277609. [ 813.609794] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [None req-5af76806-1cdf-4002-962d-34b8992ee9dc tempest-VolumesAdminNegativeTest-1387177911 tempest-VolumesAdminNegativeTest-1387177911-project-member] Creating folder: Instances. Parent ref: group-v277651. {{(pid=61806) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 813.610168] env[61806]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-4d7cf147-77ed-4afa-8d84-7d8301dd1815 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 813.624169] env[61806]: INFO nova.virt.vmwareapi.vm_util [None req-5af76806-1cdf-4002-962d-34b8992ee9dc tempest-VolumesAdminNegativeTest-1387177911 tempest-VolumesAdminNegativeTest-1387177911-project-member] Created folder: Instances in parent group-v277651. [ 813.624546] env[61806]: DEBUG oslo.service.loopingcall [None req-5af76806-1cdf-4002-962d-34b8992ee9dc tempest-VolumesAdminNegativeTest-1387177911 tempest-VolumesAdminNegativeTest-1387177911-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61806) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 813.624844] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 4373b735-31cf-4b53-b655-38555cf212a5] Creating VM on the ESX host {{(pid=61806) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 813.625835] env[61806]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-bf08f7af-b3ae-47f0-b710-f8f06533ee57 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 813.656360] env[61806]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 813.656360] env[61806]: value = "task-1294593" [ 813.656360] env[61806]: _type = "Task" [ 813.656360] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 813.666731] env[61806]: DEBUG oslo_vmware.api [-] Task: {'id': task-1294593, 'name': CreateVM_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 813.673289] env[61806]: DEBUG nova.scheduler.client.report [None req-8aa43bc8-f134-4682-8a54-ce9aa25e4ee5 tempest-AttachVolumeNegativeTest-1683033704 tempest-AttachVolumeNegativeTest-1683033704-project-member] Inventory has not changed for provider a2858be1-fd22-4e08-979e-87ad25293407 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 138, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61806) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 813.729304] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [None req-4e219b6d-f08a-4088-99ff-2fcd194df36a tempest-ServerShowV257Test-1159790492 tempest-ServerShowV257Test-1159790492-project-member] [instance: 616ec206-9804-469e-ab5c-41aea7f048aa] Powering off the VM {{(pid=61806) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 813.729701] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-1286adb1-096e-4730-b047-1ca471861ef8 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 813.733626] env[61806]: DEBUG nova.compute.manager [None req-61b80e06-de2f-412b-be27-b9318eeb5c25 tempest-ServerAddressesTestJSON-661730579 tempest-ServerAddressesTestJSON-661730579-project-member] [instance: a842df44-d8a9-4376-b9fc-5ca19a68a4b7] Start spawning the instance on the hypervisor. {{(pid=61806) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 813.745673] env[61806]: DEBUG oslo_vmware.api [None req-4e219b6d-f08a-4088-99ff-2fcd194df36a tempest-ServerShowV257Test-1159790492 tempest-ServerShowV257Test-1159790492-project-member] Waiting for the task: (returnval){ [ 813.745673] env[61806]: value = "task-1294594" [ 813.745673] env[61806]: _type = "Task" [ 813.745673] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 813.755982] env[61806]: DEBUG oslo_vmware.api [None req-4e219b6d-f08a-4088-99ff-2fcd194df36a tempest-ServerShowV257Test-1159790492 tempest-ServerShowV257Test-1159790492-project-member] Task: {'id': task-1294594, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 813.769998] env[61806]: DEBUG nova.virt.hardware [None req-61b80e06-de2f-412b-be27-b9318eeb5c25 tempest-ServerAddressesTestJSON-661730579 tempest-ServerAddressesTestJSON-661730579-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-15T18:05:10Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-15T18:04:52Z,direct_url=,disk_format='vmdk',id=ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='84ba83fce2c349a6988173c5d6fc3b07',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-15T18:04:53Z,virtual_size=,visibility=), allow threads: False {{(pid=61806) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 813.770284] env[61806]: DEBUG nova.virt.hardware [None req-61b80e06-de2f-412b-be27-b9318eeb5c25 tempest-ServerAddressesTestJSON-661730579 tempest-ServerAddressesTestJSON-661730579-project-member] Flavor limits 0:0:0 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 813.770449] env[61806]: DEBUG nova.virt.hardware [None req-61b80e06-de2f-412b-be27-b9318eeb5c25 tempest-ServerAddressesTestJSON-661730579 tempest-ServerAddressesTestJSON-661730579-project-member] Image limits 0:0:0 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 813.770639] env[61806]: DEBUG nova.virt.hardware [None req-61b80e06-de2f-412b-be27-b9318eeb5c25 tempest-ServerAddressesTestJSON-661730579 tempest-ServerAddressesTestJSON-661730579-project-member] Flavor pref 0:0:0 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 813.770792] env[61806]: DEBUG nova.virt.hardware [None req-61b80e06-de2f-412b-be27-b9318eeb5c25 tempest-ServerAddressesTestJSON-661730579 tempest-ServerAddressesTestJSON-661730579-project-member] Image pref 0:0:0 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 813.770947] env[61806]: DEBUG nova.virt.hardware [None req-61b80e06-de2f-412b-be27-b9318eeb5c25 tempest-ServerAddressesTestJSON-661730579 tempest-ServerAddressesTestJSON-661730579-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 813.771182] env[61806]: DEBUG nova.virt.hardware [None req-61b80e06-de2f-412b-be27-b9318eeb5c25 tempest-ServerAddressesTestJSON-661730579 tempest-ServerAddressesTestJSON-661730579-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61806) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 813.771349] env[61806]: DEBUG nova.virt.hardware [None req-61b80e06-de2f-412b-be27-b9318eeb5c25 tempest-ServerAddressesTestJSON-661730579 tempest-ServerAddressesTestJSON-661730579-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61806) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 813.771524] env[61806]: DEBUG nova.virt.hardware [None req-61b80e06-de2f-412b-be27-b9318eeb5c25 tempest-ServerAddressesTestJSON-661730579 tempest-ServerAddressesTestJSON-661730579-project-member] Got 1 possible topologies {{(pid=61806) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 813.771692] env[61806]: DEBUG nova.virt.hardware [None req-61b80e06-de2f-412b-be27-b9318eeb5c25 tempest-ServerAddressesTestJSON-661730579 tempest-ServerAddressesTestJSON-661730579-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61806) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 813.771870] env[61806]: DEBUG nova.virt.hardware [None req-61b80e06-de2f-412b-be27-b9318eeb5c25 tempest-ServerAddressesTestJSON-661730579 tempest-ServerAddressesTestJSON-661730579-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61806) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 813.772754] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0b38c9d7-a0ed-4065-8345-b6290f46be1c {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 813.780860] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e4ce0dd3-c381-4099-b0fa-5730e60cc5b2 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 813.844280] env[61806]: DEBUG nova.compute.manager [None req-4e0c8443-0fd4-459b-a4c5-7cf373aa1089 tempest-ServerAddressesNegativeTestJSON-1526246390 tempest-ServerAddressesNegativeTestJSON-1526246390-project-member] [instance: c8033551-6591-4e37-ae78-4efe7145b10b] Starting instance... {{(pid=61806) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 813.891643] env[61806]: DEBUG nova.network.neutron [req-73ea1082-8136-4be2-8518-198322c844e8 req-b5b52310-ce7c-4f5a-96a9-1c3d550bb745 service nova] [instance: 4373b735-31cf-4b53-b655-38555cf212a5] Updated VIF entry in instance network info cache for port 2df33c67-a8f5-4af4-ae74-29dd266d5054. {{(pid=61806) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 813.892056] env[61806]: DEBUG nova.network.neutron [req-73ea1082-8136-4be2-8518-198322c844e8 req-b5b52310-ce7c-4f5a-96a9-1c3d550bb745 service nova] [instance: 4373b735-31cf-4b53-b655-38555cf212a5] Updating instance_info_cache with network_info: [{"id": "2df33c67-a8f5-4af4-ae74-29dd266d5054", "address": "fa:16:3e:fe:14:20", "network": {"id": "2767c0cb-36c3-4a3a-9755-278a7c2296b5", "bridge": "br-int", "label": "tempest-VolumesAdminNegativeTest-745798855-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "44b0a78c155d47d98223d2543ac7673b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "abcf0d10-3f3f-45dc-923e-1c78766e2dad", "external-id": "nsx-vlan-transportzone-405", "segmentation_id": 405, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2df33c67-a8", "ovs_interfaceid": "2df33c67-a8f5-4af4-ae74-29dd266d5054", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61806) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 813.897053] env[61806]: DEBUG oslo_vmware.api [None req-dbeadda9-7825-4678-af18-5f977e49124f tempest-ServerActionsV293TestJSON-906151111 tempest-ServerActionsV293TestJSON-906151111-project-member] Task: {'id': task-1294589, 'name': ReconfigVM_Task, 'duration_secs': 0.252573} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 813.897931] env[61806]: DEBUG nova.virt.vmwareapi.volumeops [None req-dbeadda9-7825-4678-af18-5f977e49124f tempest-ServerActionsV293TestJSON-906151111 tempest-ServerActionsV293TestJSON-906151111-project-member] [instance: 0758901a-7093-41d3-b0e2-5c519333abdd] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-277626', 'volume_id': '1b19f6ff-b694-445e-9bc1-d581e9268324', 'name': 'volume-1b19f6ff-b694-445e-9bc1-d581e9268324', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '0758901a-7093-41d3-b0e2-5c519333abdd', 'attached_at': '', 'detached_at': '', 'volume_id': '1b19f6ff-b694-445e-9bc1-d581e9268324', 'serial': '1b19f6ff-b694-445e-9bc1-d581e9268324'} {{(pid=61806) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 813.898362] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-4fc719ac-989a-41a3-9bfc-9124d945c97b {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 813.905492] env[61806]: DEBUG oslo_vmware.api [None req-dbeadda9-7825-4678-af18-5f977e49124f tempest-ServerActionsV293TestJSON-906151111 tempest-ServerActionsV293TestJSON-906151111-project-member] Waiting for the task: (returnval){ [ 813.905492] env[61806]: value = "task-1294595" [ 813.905492] env[61806]: _type = "Task" [ 813.905492] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 813.914513] env[61806]: DEBUG oslo_vmware.api [None req-dbeadda9-7825-4678-af18-5f977e49124f tempest-ServerActionsV293TestJSON-906151111 tempest-ServerActionsV293TestJSON-906151111-project-member] Task: {'id': task-1294595, 'name': Rename_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 814.026389] env[61806]: DEBUG oslo_vmware.api [None req-7cf484b6-b263-4f99-a8c0-b55d49de0809 tempest-ServerShowV247Test-1218443059 tempest-ServerShowV247Test-1218443059-project-member] Task: {'id': task-1294590, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 814.167932] env[61806]: DEBUG oslo_vmware.api [-] Task: {'id': task-1294593, 'name': CreateVM_Task} progress is 25%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 814.180820] env[61806]: DEBUG oslo_concurrency.lockutils [None req-8aa43bc8-f134-4682-8a54-ce9aa25e4ee5 tempest-AttachVolumeNegativeTest-1683033704 tempest-AttachVolumeNegativeTest-1683033704-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.473s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 814.180820] env[61806]: DEBUG nova.compute.manager [None req-8aa43bc8-f134-4682-8a54-ce9aa25e4ee5 tempest-AttachVolumeNegativeTest-1683033704 tempest-AttachVolumeNegativeTest-1683033704-project-member] [instance: 43dcce0d-4fdb-4ee1-8309-92e92d2331a9] Start building networks asynchronously for instance. {{(pid=61806) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 814.181723] env[61806]: DEBUG oslo_concurrency.lockutils [None req-9c52c9e9-816e-4638-b13a-a63a9756140c tempest-SecurityGroupsTestJSON-1213485581 tempest-SecurityGroupsTestJSON-1213485581-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 13.169s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 814.183833] env[61806]: INFO nova.compute.claims [None req-9c52c9e9-816e-4638-b13a-a63a9756140c tempest-SecurityGroupsTestJSON-1213485581 tempest-SecurityGroupsTestJSON-1213485581-project-member] [instance: b41531f2-f28c-4d82-9682-0b557bbaa491] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 814.256245] env[61806]: DEBUG oslo_vmware.api [None req-4e219b6d-f08a-4088-99ff-2fcd194df36a tempest-ServerShowV257Test-1159790492 tempest-ServerShowV257Test-1159790492-project-member] Task: {'id': task-1294594, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 814.272359] env[61806]: DEBUG nova.compute.manager [req-4b5ce4bd-2207-40b8-aa77-d79ee03f71eb req-c03f5830-3d3d-4e8b-9a58-56e39e555932 service nova] [instance: a842df44-d8a9-4376-b9fc-5ca19a68a4b7] Received event network-vif-plugged-cb3cf8c0-2bc5-4308-899d-d6d2734f7922 {{(pid=61806) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 814.272605] env[61806]: DEBUG oslo_concurrency.lockutils [req-4b5ce4bd-2207-40b8-aa77-d79ee03f71eb req-c03f5830-3d3d-4e8b-9a58-56e39e555932 service nova] Acquiring lock "a842df44-d8a9-4376-b9fc-5ca19a68a4b7-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 814.272874] env[61806]: DEBUG oslo_concurrency.lockutils [req-4b5ce4bd-2207-40b8-aa77-d79ee03f71eb req-c03f5830-3d3d-4e8b-9a58-56e39e555932 service nova] Lock "a842df44-d8a9-4376-b9fc-5ca19a68a4b7-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 814.273067] env[61806]: DEBUG oslo_concurrency.lockutils [req-4b5ce4bd-2207-40b8-aa77-d79ee03f71eb req-c03f5830-3d3d-4e8b-9a58-56e39e555932 service nova] Lock "a842df44-d8a9-4376-b9fc-5ca19a68a4b7-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 814.273273] env[61806]: DEBUG nova.compute.manager [req-4b5ce4bd-2207-40b8-aa77-d79ee03f71eb req-c03f5830-3d3d-4e8b-9a58-56e39e555932 service nova] [instance: a842df44-d8a9-4376-b9fc-5ca19a68a4b7] No waiting events found dispatching network-vif-plugged-cb3cf8c0-2bc5-4308-899d-d6d2734f7922 {{(pid=61806) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 814.273687] env[61806]: WARNING nova.compute.manager [req-4b5ce4bd-2207-40b8-aa77-d79ee03f71eb req-c03f5830-3d3d-4e8b-9a58-56e39e555932 service nova] [instance: a842df44-d8a9-4376-b9fc-5ca19a68a4b7] Received unexpected event network-vif-plugged-cb3cf8c0-2bc5-4308-899d-d6d2734f7922 for instance with vm_state building and task_state spawning. [ 814.358520] env[61806]: DEBUG nova.network.neutron [None req-61b80e06-de2f-412b-be27-b9318eeb5c25 tempest-ServerAddressesTestJSON-661730579 tempest-ServerAddressesTestJSON-661730579-project-member] [instance: a842df44-d8a9-4376-b9fc-5ca19a68a4b7] Successfully updated port: cb3cf8c0-2bc5-4308-899d-d6d2734f7922 {{(pid=61806) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 814.368361] env[61806]: DEBUG oslo_concurrency.lockutils [None req-4e0c8443-0fd4-459b-a4c5-7cf373aa1089 tempest-ServerAddressesNegativeTestJSON-1526246390 tempest-ServerAddressesNegativeTestJSON-1526246390-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 814.398986] env[61806]: DEBUG oslo_concurrency.lockutils [req-73ea1082-8136-4be2-8518-198322c844e8 req-b5b52310-ce7c-4f5a-96a9-1c3d550bb745 service nova] Releasing lock "refresh_cache-4373b735-31cf-4b53-b655-38555cf212a5" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 814.416325] env[61806]: DEBUG oslo_vmware.api [None req-dbeadda9-7825-4678-af18-5f977e49124f tempest-ServerActionsV293TestJSON-906151111 tempest-ServerActionsV293TestJSON-906151111-project-member] Task: {'id': task-1294595, 'name': Rename_Task} progress is 14%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 814.526288] env[61806]: DEBUG oslo_vmware.api [None req-7cf484b6-b263-4f99-a8c0-b55d49de0809 tempest-ServerShowV247Test-1218443059 tempest-ServerShowV247Test-1218443059-project-member] Task: {'id': task-1294590, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 814.667511] env[61806]: DEBUG oslo_vmware.api [-] Task: {'id': task-1294593, 'name': CreateVM_Task, 'duration_secs': 0.601965} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 814.667745] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 4373b735-31cf-4b53-b655-38555cf212a5] Created VM on the ESX host {{(pid=61806) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 814.668507] env[61806]: DEBUG oslo_concurrency.lockutils [None req-5af76806-1cdf-4002-962d-34b8992ee9dc tempest-VolumesAdminNegativeTest-1387177911 tempest-VolumesAdminNegativeTest-1387177911-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 814.668688] env[61806]: DEBUG oslo_concurrency.lockutils [None req-5af76806-1cdf-4002-962d-34b8992ee9dc tempest-VolumesAdminNegativeTest-1387177911 tempest-VolumesAdminNegativeTest-1387177911-project-member] Acquired lock "[datastore1] devstack-image-cache_base/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 814.669043] env[61806]: DEBUG oslo_concurrency.lockutils [None req-5af76806-1cdf-4002-962d-34b8992ee9dc tempest-VolumesAdminNegativeTest-1387177911 tempest-VolumesAdminNegativeTest-1387177911-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 814.669313] env[61806]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-629dd8fa-874d-4a46-9f16-7f52490b9dff {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 814.674643] env[61806]: DEBUG oslo_vmware.api [None req-5af76806-1cdf-4002-962d-34b8992ee9dc tempest-VolumesAdminNegativeTest-1387177911 tempest-VolumesAdminNegativeTest-1387177911-project-member] Waiting for the task: (returnval){ [ 814.674643] env[61806]: value = "session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]527f989c-63d9-0ec9-bcee-80b32acf965e" [ 814.674643] env[61806]: _type = "Task" [ 814.674643] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 814.685298] env[61806]: DEBUG oslo_vmware.api [None req-5af76806-1cdf-4002-962d-34b8992ee9dc tempest-VolumesAdminNegativeTest-1387177911 tempest-VolumesAdminNegativeTest-1387177911-project-member] Task: {'id': session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]527f989c-63d9-0ec9-bcee-80b32acf965e, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 814.692732] env[61806]: DEBUG nova.compute.utils [None req-8aa43bc8-f134-4682-8a54-ce9aa25e4ee5 tempest-AttachVolumeNegativeTest-1683033704 tempest-AttachVolumeNegativeTest-1683033704-project-member] Using /dev/sd instead of None {{(pid=61806) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 814.694590] env[61806]: DEBUG nova.compute.manager [None req-8aa43bc8-f134-4682-8a54-ce9aa25e4ee5 tempest-AttachVolumeNegativeTest-1683033704 tempest-AttachVolumeNegativeTest-1683033704-project-member] [instance: 43dcce0d-4fdb-4ee1-8309-92e92d2331a9] Allocating IP information in the background. {{(pid=61806) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 814.694913] env[61806]: DEBUG nova.network.neutron [None req-8aa43bc8-f134-4682-8a54-ce9aa25e4ee5 tempest-AttachVolumeNegativeTest-1683033704 tempest-AttachVolumeNegativeTest-1683033704-project-member] [instance: 43dcce0d-4fdb-4ee1-8309-92e92d2331a9] allocate_for_instance() {{(pid=61806) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 814.737403] env[61806]: DEBUG nova.policy [None req-8aa43bc8-f134-4682-8a54-ce9aa25e4ee5 tempest-AttachVolumeNegativeTest-1683033704 tempest-AttachVolumeNegativeTest-1683033704-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '62b5e014cfeb4403a1edf89108c55e11', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '22e47882c2c44ce492dff6b1c5d782e2', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61806) authorize /opt/stack/nova/nova/policy.py:201}} [ 814.755507] env[61806]: DEBUG oslo_vmware.api [None req-4e219b6d-f08a-4088-99ff-2fcd194df36a tempest-ServerShowV257Test-1159790492 tempest-ServerShowV257Test-1159790492-project-member] Task: {'id': task-1294594, 'name': PowerOffVM_Task} progress is 100%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 814.861954] env[61806]: DEBUG oslo_concurrency.lockutils [None req-61b80e06-de2f-412b-be27-b9318eeb5c25 tempest-ServerAddressesTestJSON-661730579 tempest-ServerAddressesTestJSON-661730579-project-member] Acquiring lock "refresh_cache-a842df44-d8a9-4376-b9fc-5ca19a68a4b7" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 814.862366] env[61806]: DEBUG oslo_concurrency.lockutils [None req-61b80e06-de2f-412b-be27-b9318eeb5c25 tempest-ServerAddressesTestJSON-661730579 tempest-ServerAddressesTestJSON-661730579-project-member] Acquired lock "refresh_cache-a842df44-d8a9-4376-b9fc-5ca19a68a4b7" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 814.862366] env[61806]: DEBUG nova.network.neutron [None req-61b80e06-de2f-412b-be27-b9318eeb5c25 tempest-ServerAddressesTestJSON-661730579 tempest-ServerAddressesTestJSON-661730579-project-member] [instance: a842df44-d8a9-4376-b9fc-5ca19a68a4b7] Building network info cache for instance {{(pid=61806) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 814.917861] env[61806]: DEBUG oslo_vmware.api [None req-dbeadda9-7825-4678-af18-5f977e49124f tempest-ServerActionsV293TestJSON-906151111 tempest-ServerActionsV293TestJSON-906151111-project-member] Task: {'id': task-1294595, 'name': Rename_Task, 'duration_secs': 0.842221} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 814.918260] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [None req-dbeadda9-7825-4678-af18-5f977e49124f tempest-ServerActionsV293TestJSON-906151111 tempest-ServerActionsV293TestJSON-906151111-project-member] [instance: 0758901a-7093-41d3-b0e2-5c519333abdd] Powering on the VM {{(pid=61806) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 814.918653] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-6ed6c56f-8d7f-41b9-95f4-47bb9c7a6719 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 814.929576] env[61806]: DEBUG oslo_vmware.api [None req-dbeadda9-7825-4678-af18-5f977e49124f tempest-ServerActionsV293TestJSON-906151111 tempest-ServerActionsV293TestJSON-906151111-project-member] Waiting for the task: (returnval){ [ 814.929576] env[61806]: value = "task-1294596" [ 814.929576] env[61806]: _type = "Task" [ 814.929576] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 814.942250] env[61806]: DEBUG oslo_vmware.api [None req-dbeadda9-7825-4678-af18-5f977e49124f tempest-ServerActionsV293TestJSON-906151111 tempest-ServerActionsV293TestJSON-906151111-project-member] Task: {'id': task-1294596, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 815.028919] env[61806]: DEBUG oslo_vmware.api [None req-7cf484b6-b263-4f99-a8c0-b55d49de0809 tempest-ServerShowV247Test-1218443059 tempest-ServerShowV247Test-1218443059-project-member] Task: {'id': task-1294590, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 815.104221] env[61806]: DEBUG nova.network.neutron [None req-8aa43bc8-f134-4682-8a54-ce9aa25e4ee5 tempest-AttachVolumeNegativeTest-1683033704 tempest-AttachVolumeNegativeTest-1683033704-project-member] [instance: 43dcce0d-4fdb-4ee1-8309-92e92d2331a9] Successfully created port: 6361e800-fbf2-48e0-a11d-b00d9a9955b4 {{(pid=61806) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 815.186222] env[61806]: DEBUG oslo_vmware.api [None req-5af76806-1cdf-4002-962d-34b8992ee9dc tempest-VolumesAdminNegativeTest-1387177911 tempest-VolumesAdminNegativeTest-1387177911-project-member] Task: {'id': session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]527f989c-63d9-0ec9-bcee-80b32acf965e, 'name': SearchDatastore_Task, 'duration_secs': 0.01271} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 815.186550] env[61806]: DEBUG oslo_concurrency.lockutils [None req-5af76806-1cdf-4002-962d-34b8992ee9dc tempest-VolumesAdminNegativeTest-1387177911 tempest-VolumesAdminNegativeTest-1387177911-project-member] Releasing lock "[datastore1] devstack-image-cache_base/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 815.186795] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-5af76806-1cdf-4002-962d-34b8992ee9dc tempest-VolumesAdminNegativeTest-1387177911 tempest-VolumesAdminNegativeTest-1387177911-project-member] [instance: 4373b735-31cf-4b53-b655-38555cf212a5] Processing image ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa {{(pid=61806) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 815.187086] env[61806]: DEBUG oslo_concurrency.lockutils [None req-5af76806-1cdf-4002-962d-34b8992ee9dc tempest-VolumesAdminNegativeTest-1387177911 tempest-VolumesAdminNegativeTest-1387177911-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa.vmdk" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 815.187199] env[61806]: DEBUG oslo_concurrency.lockutils [None req-5af76806-1cdf-4002-962d-34b8992ee9dc tempest-VolumesAdminNegativeTest-1387177911 tempest-VolumesAdminNegativeTest-1387177911-project-member] Acquired lock "[datastore1] devstack-image-cache_base/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa.vmdk" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 815.187373] env[61806]: DEBUG nova.virt.vmwareapi.ds_util [None req-5af76806-1cdf-4002-962d-34b8992ee9dc tempest-VolumesAdminNegativeTest-1387177911 tempest-VolumesAdminNegativeTest-1387177911-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61806) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 815.187637] env[61806]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-e1b61d35-e967-442c-a46d-4cdec9a2f684 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 815.195158] env[61806]: DEBUG nova.compute.manager [None req-8aa43bc8-f134-4682-8a54-ce9aa25e4ee5 tempest-AttachVolumeNegativeTest-1683033704 tempest-AttachVolumeNegativeTest-1683033704-project-member] [instance: 43dcce0d-4fdb-4ee1-8309-92e92d2331a9] Start building block device mappings for instance. {{(pid=61806) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 815.203330] env[61806]: DEBUG nova.virt.vmwareapi.ds_util [None req-5af76806-1cdf-4002-962d-34b8992ee9dc tempest-VolumesAdminNegativeTest-1387177911 tempest-VolumesAdminNegativeTest-1387177911-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61806) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 815.203529] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-5af76806-1cdf-4002-962d-34b8992ee9dc tempest-VolumesAdminNegativeTest-1387177911 tempest-VolumesAdminNegativeTest-1387177911-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61806) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 815.204518] env[61806]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-3a279136-92f9-46cf-8d5d-221183296be1 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 815.222020] env[61806]: DEBUG oslo_vmware.api [None req-5af76806-1cdf-4002-962d-34b8992ee9dc tempest-VolumesAdminNegativeTest-1387177911 tempest-VolumesAdminNegativeTest-1387177911-project-member] Waiting for the task: (returnval){ [ 815.222020] env[61806]: value = "session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]524e3103-4659-e4b5-6ac1-af940b9a30b6" [ 815.222020] env[61806]: _type = "Task" [ 815.222020] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 815.231875] env[61806]: DEBUG oslo_vmware.api [None req-5af76806-1cdf-4002-962d-34b8992ee9dc tempest-VolumesAdminNegativeTest-1387177911 tempest-VolumesAdminNegativeTest-1387177911-project-member] Task: {'id': session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]524e3103-4659-e4b5-6ac1-af940b9a30b6, 'name': SearchDatastore_Task, 'duration_secs': 0.014082} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 815.232891] env[61806]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-94bc06ee-693c-403e-a176-c5a0baec21bd {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 815.242068] env[61806]: DEBUG oslo_vmware.api [None req-5af76806-1cdf-4002-962d-34b8992ee9dc tempest-VolumesAdminNegativeTest-1387177911 tempest-VolumesAdminNegativeTest-1387177911-project-member] Waiting for the task: (returnval){ [ 815.242068] env[61806]: value = "session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]52d36a72-118c-fc2f-4dbb-95ca0e119540" [ 815.242068] env[61806]: _type = "Task" [ 815.242068] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 815.252343] env[61806]: DEBUG oslo_vmware.api [None req-5af76806-1cdf-4002-962d-34b8992ee9dc tempest-VolumesAdminNegativeTest-1387177911 tempest-VolumesAdminNegativeTest-1387177911-project-member] Task: {'id': session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]52d36a72-118c-fc2f-4dbb-95ca0e119540, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 815.261258] env[61806]: DEBUG oslo_vmware.api [None req-4e219b6d-f08a-4088-99ff-2fcd194df36a tempest-ServerShowV257Test-1159790492 tempest-ServerShowV257Test-1159790492-project-member] Task: {'id': task-1294594, 'name': PowerOffVM_Task, 'duration_secs': 1.014658} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 815.261593] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [None req-4e219b6d-f08a-4088-99ff-2fcd194df36a tempest-ServerShowV257Test-1159790492 tempest-ServerShowV257Test-1159790492-project-member] [instance: 616ec206-9804-469e-ab5c-41aea7f048aa] Powered off the VM {{(pid=61806) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 815.261789] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-4e219b6d-f08a-4088-99ff-2fcd194df36a tempest-ServerShowV257Test-1159790492 tempest-ServerShowV257Test-1159790492-project-member] [instance: 616ec206-9804-469e-ab5c-41aea7f048aa] Destroying instance {{(pid=61806) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 815.265223] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2fc16fe0-3f72-4110-94b1-aedf8b8f8d6b {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 815.271459] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-4e219b6d-f08a-4088-99ff-2fcd194df36a tempest-ServerShowV257Test-1159790492 tempest-ServerShowV257Test-1159790492-project-member] [instance: 616ec206-9804-469e-ab5c-41aea7f048aa] Unregistering the VM {{(pid=61806) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 815.274160] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-03a196ca-2331-4634-9da1-56b9a2c46eba {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 815.308018] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-4e219b6d-f08a-4088-99ff-2fcd194df36a tempest-ServerShowV257Test-1159790492 tempest-ServerShowV257Test-1159790492-project-member] [instance: 616ec206-9804-469e-ab5c-41aea7f048aa] Unregistered the VM {{(pid=61806) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 815.308383] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-4e219b6d-f08a-4088-99ff-2fcd194df36a tempest-ServerShowV257Test-1159790492 tempest-ServerShowV257Test-1159790492-project-member] [instance: 616ec206-9804-469e-ab5c-41aea7f048aa] Deleting contents of the VM from datastore datastore1 {{(pid=61806) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 815.308652] env[61806]: DEBUG nova.virt.vmwareapi.ds_util [None req-4e219b6d-f08a-4088-99ff-2fcd194df36a tempest-ServerShowV257Test-1159790492 tempest-ServerShowV257Test-1159790492-project-member] Deleting the datastore file [datastore1] 616ec206-9804-469e-ab5c-41aea7f048aa {{(pid=61806) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 815.309021] env[61806]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-1ffb12ba-71ec-41c5-9a80-9808f27cb727 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 815.320583] env[61806]: DEBUG oslo_vmware.api [None req-4e219b6d-f08a-4088-99ff-2fcd194df36a tempest-ServerShowV257Test-1159790492 tempest-ServerShowV257Test-1159790492-project-member] Waiting for the task: (returnval){ [ 815.320583] env[61806]: value = "task-1294598" [ 815.320583] env[61806]: _type = "Task" [ 815.320583] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 815.330595] env[61806]: DEBUG oslo_vmware.api [None req-4e219b6d-f08a-4088-99ff-2fcd194df36a tempest-ServerShowV257Test-1159790492 tempest-ServerShowV257Test-1159790492-project-member] Task: {'id': task-1294598, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 815.417542] env[61806]: DEBUG nova.network.neutron [None req-61b80e06-de2f-412b-be27-b9318eeb5c25 tempest-ServerAddressesTestJSON-661730579 tempest-ServerAddressesTestJSON-661730579-project-member] [instance: a842df44-d8a9-4376-b9fc-5ca19a68a4b7] Instance cache missing network info. {{(pid=61806) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 815.447182] env[61806]: DEBUG oslo_vmware.api [None req-dbeadda9-7825-4678-af18-5f977e49124f tempest-ServerActionsV293TestJSON-906151111 tempest-ServerActionsV293TestJSON-906151111-project-member] Task: {'id': task-1294596, 'name': PowerOnVM_Task} progress is 89%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 815.500814] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bc7ab779-183a-41ae-8d21-4a1b67cc0087 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 815.509037] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f65770a9-46a8-4fb1-a3c3-69f9e60fd9e6 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 815.544362] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6274c38e-5810-46e7-a187-abf3d04d7a7d {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 815.556502] env[61806]: DEBUG oslo_vmware.api [None req-7cf484b6-b263-4f99-a8c0-b55d49de0809 tempest-ServerShowV247Test-1218443059 tempest-ServerShowV247Test-1218443059-project-member] Task: {'id': task-1294590, 'name': CopyVirtualDisk_Task, 'duration_secs': 1.575624} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 815.557705] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9c35dfbe-6ac8-432d-af15-78fd510c19ff {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 815.561380] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [None req-7cf484b6-b263-4f99-a8c0-b55d49de0809 tempest-ServerShowV247Test-1218443059 tempest-ServerShowV247Test-1218443059-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa.vmdk to [datastore1] 4febd093-9f2b-494e-b175-e4693b0e3e0d/4febd093-9f2b-494e-b175-e4693b0e3e0d.vmdk {{(pid=61806) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 815.561613] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-7cf484b6-b263-4f99-a8c0-b55d49de0809 tempest-ServerShowV247Test-1218443059 tempest-ServerShowV247Test-1218443059-project-member] [instance: 4febd093-9f2b-494e-b175-e4693b0e3e0d] Extending root virtual disk to 1048576 {{(pid=61806) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 815.561869] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-1aa909ad-74b5-4c8a-99de-7a4d32c43572 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 815.574860] env[61806]: DEBUG nova.compute.provider_tree [None req-9c52c9e9-816e-4638-b13a-a63a9756140c tempest-SecurityGroupsTestJSON-1213485581 tempest-SecurityGroupsTestJSON-1213485581-project-member] Inventory has not changed in ProviderTree for provider: a2858be1-fd22-4e08-979e-87ad25293407 {{(pid=61806) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 815.577697] env[61806]: DEBUG oslo_vmware.api [None req-7cf484b6-b263-4f99-a8c0-b55d49de0809 tempest-ServerShowV247Test-1218443059 tempest-ServerShowV247Test-1218443059-project-member] Waiting for the task: (returnval){ [ 815.577697] env[61806]: value = "task-1294599" [ 815.577697] env[61806]: _type = "Task" [ 815.577697] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 815.587064] env[61806]: DEBUG oslo_vmware.api [None req-7cf484b6-b263-4f99-a8c0-b55d49de0809 tempest-ServerShowV247Test-1218443059 tempest-ServerShowV247Test-1218443059-project-member] Task: {'id': task-1294599, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 815.616218] env[61806]: DEBUG nova.network.neutron [None req-61b80e06-de2f-412b-be27-b9318eeb5c25 tempest-ServerAddressesTestJSON-661730579 tempest-ServerAddressesTestJSON-661730579-project-member] [instance: a842df44-d8a9-4376-b9fc-5ca19a68a4b7] Updating instance_info_cache with network_info: [{"id": "cb3cf8c0-2bc5-4308-899d-d6d2734f7922", "address": "fa:16:3e:69:44:e9", "network": {"id": "43aa1d97-f97a-4a31-8939-51c56101b18e", "bridge": "br-int", "label": "tempest-ServerAddressesTestJSON-1770170618-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "52cc8936b48b44c3b4f1172b1344cba6", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "496faa4d-d874-449b-905e-328ddd60b31b", "external-id": "nsx-vlan-transportzone-391", "segmentation_id": 391, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapcb3cf8c0-2b", "ovs_interfaceid": "cb3cf8c0-2bc5-4308-899d-d6d2734f7922", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61806) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 815.752339] env[61806]: DEBUG oslo_vmware.api [None req-5af76806-1cdf-4002-962d-34b8992ee9dc tempest-VolumesAdminNegativeTest-1387177911 tempest-VolumesAdminNegativeTest-1387177911-project-member] Task: {'id': session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]52d36a72-118c-fc2f-4dbb-95ca0e119540, 'name': SearchDatastore_Task, 'duration_secs': 0.023086} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 815.752637] env[61806]: DEBUG oslo_concurrency.lockutils [None req-5af76806-1cdf-4002-962d-34b8992ee9dc tempest-VolumesAdminNegativeTest-1387177911 tempest-VolumesAdminNegativeTest-1387177911-project-member] Releasing lock "[datastore1] devstack-image-cache_base/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa.vmdk" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 815.752898] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [None req-5af76806-1cdf-4002-962d-34b8992ee9dc tempest-VolumesAdminNegativeTest-1387177911 tempest-VolumesAdminNegativeTest-1387177911-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa.vmdk to [datastore1] 4373b735-31cf-4b53-b655-38555cf212a5/4373b735-31cf-4b53-b655-38555cf212a5.vmdk {{(pid=61806) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 815.753177] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-2a04e095-546f-4aee-beff-c0194912d5d1 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 815.759780] env[61806]: DEBUG oslo_vmware.api [None req-5af76806-1cdf-4002-962d-34b8992ee9dc tempest-VolumesAdminNegativeTest-1387177911 tempest-VolumesAdminNegativeTest-1387177911-project-member] Waiting for the task: (returnval){ [ 815.759780] env[61806]: value = "task-1294600" [ 815.759780] env[61806]: _type = "Task" [ 815.759780] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 815.767431] env[61806]: DEBUG oslo_vmware.api [None req-5af76806-1cdf-4002-962d-34b8992ee9dc tempest-VolumesAdminNegativeTest-1387177911 tempest-VolumesAdminNegativeTest-1387177911-project-member] Task: {'id': task-1294600, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 815.832235] env[61806]: DEBUG oslo_vmware.api [None req-4e219b6d-f08a-4088-99ff-2fcd194df36a tempest-ServerShowV257Test-1159790492 tempest-ServerShowV257Test-1159790492-project-member] Task: {'id': task-1294598, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.190517} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 815.832561] env[61806]: DEBUG nova.virt.vmwareapi.ds_util [None req-4e219b6d-f08a-4088-99ff-2fcd194df36a tempest-ServerShowV257Test-1159790492 tempest-ServerShowV257Test-1159790492-project-member] Deleted the datastore file {{(pid=61806) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 815.832749] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-4e219b6d-f08a-4088-99ff-2fcd194df36a tempest-ServerShowV257Test-1159790492 tempest-ServerShowV257Test-1159790492-project-member] [instance: 616ec206-9804-469e-ab5c-41aea7f048aa] Deleted contents of the VM from datastore datastore1 {{(pid=61806) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 815.832928] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-4e219b6d-f08a-4088-99ff-2fcd194df36a tempest-ServerShowV257Test-1159790492 tempest-ServerShowV257Test-1159790492-project-member] [instance: 616ec206-9804-469e-ab5c-41aea7f048aa] Instance destroyed {{(pid=61806) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 815.941707] env[61806]: DEBUG oslo_vmware.api [None req-dbeadda9-7825-4678-af18-5f977e49124f tempest-ServerActionsV293TestJSON-906151111 tempest-ServerActionsV293TestJSON-906151111-project-member] Task: {'id': task-1294596, 'name': PowerOnVM_Task, 'duration_secs': 0.605598} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 815.942023] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [None req-dbeadda9-7825-4678-af18-5f977e49124f tempest-ServerActionsV293TestJSON-906151111 tempest-ServerActionsV293TestJSON-906151111-project-member] [instance: 0758901a-7093-41d3-b0e2-5c519333abdd] Powered on the VM {{(pid=61806) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 815.942231] env[61806]: INFO nova.compute.manager [None req-dbeadda9-7825-4678-af18-5f977e49124f tempest-ServerActionsV293TestJSON-906151111 tempest-ServerActionsV293TestJSON-906151111-project-member] [instance: 0758901a-7093-41d3-b0e2-5c519333abdd] Took 5.37 seconds to spawn the instance on the hypervisor. [ 815.942472] env[61806]: DEBUG nova.compute.manager [None req-dbeadda9-7825-4678-af18-5f977e49124f tempest-ServerActionsV293TestJSON-906151111 tempest-ServerActionsV293TestJSON-906151111-project-member] [instance: 0758901a-7093-41d3-b0e2-5c519333abdd] Checking state {{(pid=61806) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 815.943292] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-58627c9e-6b2f-4373-9c45-1ca46dce76ad {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 816.079887] env[61806]: DEBUG nova.scheduler.client.report [None req-9c52c9e9-816e-4638-b13a-a63a9756140c tempest-SecurityGroupsTestJSON-1213485581 tempest-SecurityGroupsTestJSON-1213485581-project-member] Inventory has not changed for provider a2858be1-fd22-4e08-979e-87ad25293407 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 138, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61806) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 816.092061] env[61806]: DEBUG oslo_vmware.api [None req-7cf484b6-b263-4f99-a8c0-b55d49de0809 tempest-ServerShowV247Test-1218443059 tempest-ServerShowV247Test-1218443059-project-member] Task: {'id': task-1294599, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.071606} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 816.092363] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-7cf484b6-b263-4f99-a8c0-b55d49de0809 tempest-ServerShowV247Test-1218443059 tempest-ServerShowV247Test-1218443059-project-member] [instance: 4febd093-9f2b-494e-b175-e4693b0e3e0d] Extended root virtual disk {{(pid=61806) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 816.093139] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5d92555f-af5e-44c9-b15f-e8d1d625b4e3 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 816.114924] env[61806]: DEBUG nova.virt.vmwareapi.volumeops [None req-7cf484b6-b263-4f99-a8c0-b55d49de0809 tempest-ServerShowV247Test-1218443059 tempest-ServerShowV247Test-1218443059-project-member] [instance: 4febd093-9f2b-494e-b175-e4693b0e3e0d] Reconfiguring VM instance instance-00000039 to attach disk [datastore1] 4febd093-9f2b-494e-b175-e4693b0e3e0d/4febd093-9f2b-494e-b175-e4693b0e3e0d.vmdk or device None with type sparse {{(pid=61806) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 816.115480] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-3f8c96c9-5ea2-438c-8301-966aad8c1636 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 816.129785] env[61806]: DEBUG oslo_concurrency.lockutils [None req-61b80e06-de2f-412b-be27-b9318eeb5c25 tempest-ServerAddressesTestJSON-661730579 tempest-ServerAddressesTestJSON-661730579-project-member] Releasing lock "refresh_cache-a842df44-d8a9-4376-b9fc-5ca19a68a4b7" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 816.130091] env[61806]: DEBUG nova.compute.manager [None req-61b80e06-de2f-412b-be27-b9318eeb5c25 tempest-ServerAddressesTestJSON-661730579 tempest-ServerAddressesTestJSON-661730579-project-member] [instance: a842df44-d8a9-4376-b9fc-5ca19a68a4b7] Instance network_info: |[{"id": "cb3cf8c0-2bc5-4308-899d-d6d2734f7922", "address": "fa:16:3e:69:44:e9", "network": {"id": "43aa1d97-f97a-4a31-8939-51c56101b18e", "bridge": "br-int", "label": "tempest-ServerAddressesTestJSON-1770170618-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "52cc8936b48b44c3b4f1172b1344cba6", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "496faa4d-d874-449b-905e-328ddd60b31b", "external-id": "nsx-vlan-transportzone-391", "segmentation_id": 391, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapcb3cf8c0-2b", "ovs_interfaceid": "cb3cf8c0-2bc5-4308-899d-d6d2734f7922", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61806) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 816.130743] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-61b80e06-de2f-412b-be27-b9318eeb5c25 tempest-ServerAddressesTestJSON-661730579 tempest-ServerAddressesTestJSON-661730579-project-member] [instance: a842df44-d8a9-4376-b9fc-5ca19a68a4b7] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:69:44:e9', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '496faa4d-d874-449b-905e-328ddd60b31b', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'cb3cf8c0-2bc5-4308-899d-d6d2734f7922', 'vif_model': 'vmxnet3'}] {{(pid=61806) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 816.138391] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [None req-61b80e06-de2f-412b-be27-b9318eeb5c25 tempest-ServerAddressesTestJSON-661730579 tempest-ServerAddressesTestJSON-661730579-project-member] Creating folder: Project (52cc8936b48b44c3b4f1172b1344cba6). Parent ref: group-v277609. {{(pid=61806) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 816.139209] env[61806]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-180dcf3b-0728-4765-942d-cfe3ca4b74fb {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 816.144448] env[61806]: DEBUG oslo_vmware.api [None req-7cf484b6-b263-4f99-a8c0-b55d49de0809 tempest-ServerShowV247Test-1218443059 tempest-ServerShowV247Test-1218443059-project-member] Waiting for the task: (returnval){ [ 816.144448] env[61806]: value = "task-1294601" [ 816.144448] env[61806]: _type = "Task" [ 816.144448] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 816.150506] env[61806]: INFO nova.virt.vmwareapi.vm_util [None req-61b80e06-de2f-412b-be27-b9318eeb5c25 tempest-ServerAddressesTestJSON-661730579 tempest-ServerAddressesTestJSON-661730579-project-member] Created folder: Project (52cc8936b48b44c3b4f1172b1344cba6) in parent group-v277609. [ 816.150706] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [None req-61b80e06-de2f-412b-be27-b9318eeb5c25 tempest-ServerAddressesTestJSON-661730579 tempest-ServerAddressesTestJSON-661730579-project-member] Creating folder: Instances. Parent ref: group-v277654. {{(pid=61806) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 816.151363] env[61806]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-a47d8c05-e647-4187-92e5-2bdb8c0c9c29 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 816.155379] env[61806]: DEBUG oslo_vmware.api [None req-7cf484b6-b263-4f99-a8c0-b55d49de0809 tempest-ServerShowV247Test-1218443059 tempest-ServerShowV247Test-1218443059-project-member] Task: {'id': task-1294601, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 816.164284] env[61806]: INFO nova.virt.vmwareapi.vm_util [None req-61b80e06-de2f-412b-be27-b9318eeb5c25 tempest-ServerAddressesTestJSON-661730579 tempest-ServerAddressesTestJSON-661730579-project-member] Created folder: Instances in parent group-v277654. [ 816.164571] env[61806]: DEBUG oslo.service.loopingcall [None req-61b80e06-de2f-412b-be27-b9318eeb5c25 tempest-ServerAddressesTestJSON-661730579 tempest-ServerAddressesTestJSON-661730579-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61806) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 816.164779] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: a842df44-d8a9-4376-b9fc-5ca19a68a4b7] Creating VM on the ESX host {{(pid=61806) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 816.164990] env[61806]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-424d31a1-7aed-4c68-b8a2-4f098a806587 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 816.186815] env[61806]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 816.186815] env[61806]: value = "task-1294604" [ 816.186815] env[61806]: _type = "Task" [ 816.186815] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 816.194816] env[61806]: DEBUG oslo_vmware.api [-] Task: {'id': task-1294604, 'name': CreateVM_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 816.210772] env[61806]: DEBUG nova.compute.manager [None req-8aa43bc8-f134-4682-8a54-ce9aa25e4ee5 tempest-AttachVolumeNegativeTest-1683033704 tempest-AttachVolumeNegativeTest-1683033704-project-member] [instance: 43dcce0d-4fdb-4ee1-8309-92e92d2331a9] Start spawning the instance on the hypervisor. {{(pid=61806) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 816.245644] env[61806]: DEBUG nova.virt.hardware [None req-8aa43bc8-f134-4682-8a54-ce9aa25e4ee5 tempest-AttachVolumeNegativeTest-1683033704 tempest-AttachVolumeNegativeTest-1683033704-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-15T18:05:10Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-15T18:04:52Z,direct_url=,disk_format='vmdk',id=ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='84ba83fce2c349a6988173c5d6fc3b07',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-15T18:04:53Z,virtual_size=,visibility=), allow threads: False {{(pid=61806) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 816.245996] env[61806]: DEBUG nova.virt.hardware [None req-8aa43bc8-f134-4682-8a54-ce9aa25e4ee5 tempest-AttachVolumeNegativeTest-1683033704 tempest-AttachVolumeNegativeTest-1683033704-project-member] Flavor limits 0:0:0 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 816.246203] env[61806]: DEBUG nova.virt.hardware [None req-8aa43bc8-f134-4682-8a54-ce9aa25e4ee5 tempest-AttachVolumeNegativeTest-1683033704 tempest-AttachVolumeNegativeTest-1683033704-project-member] Image limits 0:0:0 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 816.246401] env[61806]: DEBUG nova.virt.hardware [None req-8aa43bc8-f134-4682-8a54-ce9aa25e4ee5 tempest-AttachVolumeNegativeTest-1683033704 tempest-AttachVolumeNegativeTest-1683033704-project-member] Flavor pref 0:0:0 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 816.246555] env[61806]: DEBUG nova.virt.hardware [None req-8aa43bc8-f134-4682-8a54-ce9aa25e4ee5 tempest-AttachVolumeNegativeTest-1683033704 tempest-AttachVolumeNegativeTest-1683033704-project-member] Image pref 0:0:0 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 816.246708] env[61806]: DEBUG nova.virt.hardware [None req-8aa43bc8-f134-4682-8a54-ce9aa25e4ee5 tempest-AttachVolumeNegativeTest-1683033704 tempest-AttachVolumeNegativeTest-1683033704-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 816.246945] env[61806]: DEBUG nova.virt.hardware [None req-8aa43bc8-f134-4682-8a54-ce9aa25e4ee5 tempest-AttachVolumeNegativeTest-1683033704 tempest-AttachVolumeNegativeTest-1683033704-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61806) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 816.247159] env[61806]: DEBUG nova.virt.hardware [None req-8aa43bc8-f134-4682-8a54-ce9aa25e4ee5 tempest-AttachVolumeNegativeTest-1683033704 tempest-AttachVolumeNegativeTest-1683033704-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61806) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 816.247345] env[61806]: DEBUG nova.virt.hardware [None req-8aa43bc8-f134-4682-8a54-ce9aa25e4ee5 tempest-AttachVolumeNegativeTest-1683033704 tempest-AttachVolumeNegativeTest-1683033704-project-member] Got 1 possible topologies {{(pid=61806) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 816.247518] env[61806]: DEBUG nova.virt.hardware [None req-8aa43bc8-f134-4682-8a54-ce9aa25e4ee5 tempest-AttachVolumeNegativeTest-1683033704 tempest-AttachVolumeNegativeTest-1683033704-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61806) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 816.247695] env[61806]: DEBUG nova.virt.hardware [None req-8aa43bc8-f134-4682-8a54-ce9aa25e4ee5 tempest-AttachVolumeNegativeTest-1683033704 tempest-AttachVolumeNegativeTest-1683033704-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61806) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 816.249035] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-52178d90-c32e-463d-93de-509b1394b33a {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 816.258777] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-867d12bf-0fd0-4db3-a3ff-9f494d1cae6b {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 816.283377] env[61806]: DEBUG oslo_vmware.api [None req-5af76806-1cdf-4002-962d-34b8992ee9dc tempest-VolumesAdminNegativeTest-1387177911 tempest-VolumesAdminNegativeTest-1387177911-project-member] Task: {'id': task-1294600, 'name': CopyVirtualDisk_Task} progress is 4%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 816.302048] env[61806]: DEBUG nova.compute.manager [req-020e7328-be26-46fa-9c10-01f8b970b777 req-4f9c072d-5eb9-4f48-8b80-a05cfda99308 service nova] [instance: a842df44-d8a9-4376-b9fc-5ca19a68a4b7] Received event network-changed-cb3cf8c0-2bc5-4308-899d-d6d2734f7922 {{(pid=61806) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 816.302274] env[61806]: DEBUG nova.compute.manager [req-020e7328-be26-46fa-9c10-01f8b970b777 req-4f9c072d-5eb9-4f48-8b80-a05cfda99308 service nova] [instance: a842df44-d8a9-4376-b9fc-5ca19a68a4b7] Refreshing instance network info cache due to event network-changed-cb3cf8c0-2bc5-4308-899d-d6d2734f7922. {{(pid=61806) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 816.302515] env[61806]: DEBUG oslo_concurrency.lockutils [req-020e7328-be26-46fa-9c10-01f8b970b777 req-4f9c072d-5eb9-4f48-8b80-a05cfda99308 service nova] Acquiring lock "refresh_cache-a842df44-d8a9-4376-b9fc-5ca19a68a4b7" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 816.302675] env[61806]: DEBUG oslo_concurrency.lockutils [req-020e7328-be26-46fa-9c10-01f8b970b777 req-4f9c072d-5eb9-4f48-8b80-a05cfda99308 service nova] Acquired lock "refresh_cache-a842df44-d8a9-4376-b9fc-5ca19a68a4b7" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 816.302825] env[61806]: DEBUG nova.network.neutron [req-020e7328-be26-46fa-9c10-01f8b970b777 req-4f9c072d-5eb9-4f48-8b80-a05cfda99308 service nova] [instance: a842df44-d8a9-4376-b9fc-5ca19a68a4b7] Refreshing network info cache for port cb3cf8c0-2bc5-4308-899d-d6d2734f7922 {{(pid=61806) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 816.464165] env[61806]: INFO nova.compute.manager [None req-dbeadda9-7825-4678-af18-5f977e49124f tempest-ServerActionsV293TestJSON-906151111 tempest-ServerActionsV293TestJSON-906151111-project-member] [instance: 0758901a-7093-41d3-b0e2-5c519333abdd] Took 27.26 seconds to build instance. [ 816.588024] env[61806]: DEBUG oslo_concurrency.lockutils [None req-9c52c9e9-816e-4638-b13a-a63a9756140c tempest-SecurityGroupsTestJSON-1213485581 tempest-SecurityGroupsTestJSON-1213485581-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.406s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 816.588668] env[61806]: DEBUG nova.compute.manager [None req-9c52c9e9-816e-4638-b13a-a63a9756140c tempest-SecurityGroupsTestJSON-1213485581 tempest-SecurityGroupsTestJSON-1213485581-project-member] [instance: b41531f2-f28c-4d82-9682-0b557bbaa491] Start building networks asynchronously for instance. {{(pid=61806) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 816.591635] env[61806]: DEBUG oslo_concurrency.lockutils [None req-7fa92f02-ad2b-4004-a9a7-cb08f83ec0f1 tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 13.998s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 816.593320] env[61806]: INFO nova.compute.claims [None req-7fa92f02-ad2b-4004-a9a7-cb08f83ec0f1 tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] [instance: 67d1bdd8-cf8f-4fb5-bd3c-42008e4d60e3] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 816.655506] env[61806]: DEBUG oslo_vmware.api [None req-7cf484b6-b263-4f99-a8c0-b55d49de0809 tempest-ServerShowV247Test-1218443059 tempest-ServerShowV247Test-1218443059-project-member] Task: {'id': task-1294601, 'name': ReconfigVM_Task} progress is 99%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 816.697614] env[61806]: DEBUG oslo_vmware.api [-] Task: {'id': task-1294604, 'name': CreateVM_Task} progress is 99%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 816.717642] env[61806]: DEBUG nova.network.neutron [None req-8aa43bc8-f134-4682-8a54-ce9aa25e4ee5 tempest-AttachVolumeNegativeTest-1683033704 tempest-AttachVolumeNegativeTest-1683033704-project-member] [instance: 43dcce0d-4fdb-4ee1-8309-92e92d2331a9] Successfully updated port: 6361e800-fbf2-48e0-a11d-b00d9a9955b4 {{(pid=61806) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 816.773951] env[61806]: DEBUG oslo_vmware.api [None req-5af76806-1cdf-4002-962d-34b8992ee9dc tempest-VolumesAdminNegativeTest-1387177911 tempest-VolumesAdminNegativeTest-1387177911-project-member] Task: {'id': task-1294600, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.746944} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 816.774294] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [None req-5af76806-1cdf-4002-962d-34b8992ee9dc tempest-VolumesAdminNegativeTest-1387177911 tempest-VolumesAdminNegativeTest-1387177911-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa.vmdk to [datastore1] 4373b735-31cf-4b53-b655-38555cf212a5/4373b735-31cf-4b53-b655-38555cf212a5.vmdk {{(pid=61806) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 816.774589] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-5af76806-1cdf-4002-962d-34b8992ee9dc tempest-VolumesAdminNegativeTest-1387177911 tempest-VolumesAdminNegativeTest-1387177911-project-member] [instance: 4373b735-31cf-4b53-b655-38555cf212a5] Extending root virtual disk to 1048576 {{(pid=61806) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 816.774891] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-58665c7c-bb94-4ae4-bd37-719785d5bd6b {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 816.782057] env[61806]: DEBUG oslo_vmware.api [None req-5af76806-1cdf-4002-962d-34b8992ee9dc tempest-VolumesAdminNegativeTest-1387177911 tempest-VolumesAdminNegativeTest-1387177911-project-member] Waiting for the task: (returnval){ [ 816.782057] env[61806]: value = "task-1294605" [ 816.782057] env[61806]: _type = "Task" [ 816.782057] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 816.790885] env[61806]: DEBUG oslo_vmware.api [None req-5af76806-1cdf-4002-962d-34b8992ee9dc tempest-VolumesAdminNegativeTest-1387177911 tempest-VolumesAdminNegativeTest-1387177911-project-member] Task: {'id': task-1294605, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 816.869693] env[61806]: DEBUG nova.virt.hardware [None req-4e219b6d-f08a-4088-99ff-2fcd194df36a tempest-ServerShowV257Test-1159790492 tempest-ServerShowV257Test-1159790492-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-15T18:05:10Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-15T18:04:52Z,direct_url=,disk_format='vmdk',id=ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='84ba83fce2c349a6988173c5d6fc3b07',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-15T18:04:53Z,virtual_size=,visibility=), allow threads: False {{(pid=61806) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 816.869918] env[61806]: DEBUG nova.virt.hardware [None req-4e219b6d-f08a-4088-99ff-2fcd194df36a tempest-ServerShowV257Test-1159790492 tempest-ServerShowV257Test-1159790492-project-member] Flavor limits 0:0:0 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 816.870096] env[61806]: DEBUG nova.virt.hardware [None req-4e219b6d-f08a-4088-99ff-2fcd194df36a tempest-ServerShowV257Test-1159790492 tempest-ServerShowV257Test-1159790492-project-member] Image limits 0:0:0 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 816.870290] env[61806]: DEBUG nova.virt.hardware [None req-4e219b6d-f08a-4088-99ff-2fcd194df36a tempest-ServerShowV257Test-1159790492 tempest-ServerShowV257Test-1159790492-project-member] Flavor pref 0:0:0 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 816.870448] env[61806]: DEBUG nova.virt.hardware [None req-4e219b6d-f08a-4088-99ff-2fcd194df36a tempest-ServerShowV257Test-1159790492 tempest-ServerShowV257Test-1159790492-project-member] Image pref 0:0:0 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 816.870602] env[61806]: DEBUG nova.virt.hardware [None req-4e219b6d-f08a-4088-99ff-2fcd194df36a tempest-ServerShowV257Test-1159790492 tempest-ServerShowV257Test-1159790492-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 816.870874] env[61806]: DEBUG nova.virt.hardware [None req-4e219b6d-f08a-4088-99ff-2fcd194df36a tempest-ServerShowV257Test-1159790492 tempest-ServerShowV257Test-1159790492-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61806) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 816.870978] env[61806]: DEBUG nova.virt.hardware [None req-4e219b6d-f08a-4088-99ff-2fcd194df36a tempest-ServerShowV257Test-1159790492 tempest-ServerShowV257Test-1159790492-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61806) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 816.871179] env[61806]: DEBUG nova.virt.hardware [None req-4e219b6d-f08a-4088-99ff-2fcd194df36a tempest-ServerShowV257Test-1159790492 tempest-ServerShowV257Test-1159790492-project-member] Got 1 possible topologies {{(pid=61806) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 816.871326] env[61806]: DEBUG nova.virt.hardware [None req-4e219b6d-f08a-4088-99ff-2fcd194df36a tempest-ServerShowV257Test-1159790492 tempest-ServerShowV257Test-1159790492-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61806) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 816.871517] env[61806]: DEBUG nova.virt.hardware [None req-4e219b6d-f08a-4088-99ff-2fcd194df36a tempest-ServerShowV257Test-1159790492 tempest-ServerShowV257Test-1159790492-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61806) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 816.872474] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a7ec5755-bc89-4a9e-a9af-b1b05c7318a6 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 816.884181] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c32ac8fd-d7bd-4e2a-87b1-51f9519ae3e1 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 816.897875] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-4e219b6d-f08a-4088-99ff-2fcd194df36a tempest-ServerShowV257Test-1159790492 tempest-ServerShowV257Test-1159790492-project-member] [instance: 616ec206-9804-469e-ab5c-41aea7f048aa] Instance VIF info [] {{(pid=61806) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 816.903630] env[61806]: DEBUG oslo.service.loopingcall [None req-4e219b6d-f08a-4088-99ff-2fcd194df36a tempest-ServerShowV257Test-1159790492 tempest-ServerShowV257Test-1159790492-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61806) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 816.905762] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 616ec206-9804-469e-ab5c-41aea7f048aa] Creating VM on the ESX host {{(pid=61806) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 816.906052] env[61806]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-fb4c534c-e109-47e2-916e-f072afaf6c8e {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 816.923971] env[61806]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 816.923971] env[61806]: value = "task-1294606" [ 816.923971] env[61806]: _type = "Task" [ 816.923971] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 816.932355] env[61806]: DEBUG oslo_vmware.api [-] Task: {'id': task-1294606, 'name': CreateVM_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 816.966386] env[61806]: DEBUG oslo_concurrency.lockutils [None req-dbeadda9-7825-4678-af18-5f977e49124f tempest-ServerActionsV293TestJSON-906151111 tempest-ServerActionsV293TestJSON-906151111-project-member] Lock "0758901a-7093-41d3-b0e2-5c519333abdd" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 107.660s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 817.066681] env[61806]: DEBUG nova.network.neutron [req-020e7328-be26-46fa-9c10-01f8b970b777 req-4f9c072d-5eb9-4f48-8b80-a05cfda99308 service nova] [instance: a842df44-d8a9-4376-b9fc-5ca19a68a4b7] Updated VIF entry in instance network info cache for port cb3cf8c0-2bc5-4308-899d-d6d2734f7922. {{(pid=61806) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 817.066770] env[61806]: DEBUG nova.network.neutron [req-020e7328-be26-46fa-9c10-01f8b970b777 req-4f9c072d-5eb9-4f48-8b80-a05cfda99308 service nova] [instance: a842df44-d8a9-4376-b9fc-5ca19a68a4b7] Updating instance_info_cache with network_info: [{"id": "cb3cf8c0-2bc5-4308-899d-d6d2734f7922", "address": "fa:16:3e:69:44:e9", "network": {"id": "43aa1d97-f97a-4a31-8939-51c56101b18e", "bridge": "br-int", "label": "tempest-ServerAddressesTestJSON-1770170618-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "52cc8936b48b44c3b4f1172b1344cba6", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "496faa4d-d874-449b-905e-328ddd60b31b", "external-id": "nsx-vlan-transportzone-391", "segmentation_id": 391, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapcb3cf8c0-2b", "ovs_interfaceid": "cb3cf8c0-2bc5-4308-899d-d6d2734f7922", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61806) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 817.098262] env[61806]: DEBUG nova.compute.utils [None req-9c52c9e9-816e-4638-b13a-a63a9756140c tempest-SecurityGroupsTestJSON-1213485581 tempest-SecurityGroupsTestJSON-1213485581-project-member] Using /dev/sd instead of None {{(pid=61806) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 817.102178] env[61806]: DEBUG nova.compute.manager [None req-9c52c9e9-816e-4638-b13a-a63a9756140c tempest-SecurityGroupsTestJSON-1213485581 tempest-SecurityGroupsTestJSON-1213485581-project-member] [instance: b41531f2-f28c-4d82-9682-0b557bbaa491] Allocating IP information in the background. {{(pid=61806) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 817.102386] env[61806]: DEBUG nova.network.neutron [None req-9c52c9e9-816e-4638-b13a-a63a9756140c tempest-SecurityGroupsTestJSON-1213485581 tempest-SecurityGroupsTestJSON-1213485581-project-member] [instance: b41531f2-f28c-4d82-9682-0b557bbaa491] allocate_for_instance() {{(pid=61806) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 817.157355] env[61806]: DEBUG oslo_vmware.api [None req-7cf484b6-b263-4f99-a8c0-b55d49de0809 tempest-ServerShowV247Test-1218443059 tempest-ServerShowV247Test-1218443059-project-member] Task: {'id': task-1294601, 'name': ReconfigVM_Task, 'duration_secs': 0.599433} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 817.158061] env[61806]: DEBUG nova.virt.vmwareapi.volumeops [None req-7cf484b6-b263-4f99-a8c0-b55d49de0809 tempest-ServerShowV247Test-1218443059 tempest-ServerShowV247Test-1218443059-project-member] [instance: 4febd093-9f2b-494e-b175-e4693b0e3e0d] Reconfigured VM instance instance-00000039 to attach disk [datastore1] 4febd093-9f2b-494e-b175-e4693b0e3e0d/4febd093-9f2b-494e-b175-e4693b0e3e0d.vmdk or device None with type sparse {{(pid=61806) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 817.158367] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-fc0904be-18d5-402d-bf1f-fceae0a3ab2c {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 817.161113] env[61806]: DEBUG nova.policy [None req-9c52c9e9-816e-4638-b13a-a63a9756140c tempest-SecurityGroupsTestJSON-1213485581 tempest-SecurityGroupsTestJSON-1213485581-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'e32051e5482847ce85b4f9fa6e89f9e1', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'cfce93add38d40d79f294017f52e13d5', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61806) authorize /opt/stack/nova/nova/policy.py:201}} [ 817.168458] env[61806]: DEBUG oslo_vmware.api [None req-7cf484b6-b263-4f99-a8c0-b55d49de0809 tempest-ServerShowV247Test-1218443059 tempest-ServerShowV247Test-1218443059-project-member] Waiting for the task: (returnval){ [ 817.168458] env[61806]: value = "task-1294607" [ 817.168458] env[61806]: _type = "Task" [ 817.168458] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 817.177682] env[61806]: DEBUG oslo_vmware.api [None req-7cf484b6-b263-4f99-a8c0-b55d49de0809 tempest-ServerShowV247Test-1218443059 tempest-ServerShowV247Test-1218443059-project-member] Task: {'id': task-1294607, 'name': Rename_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 817.198318] env[61806]: DEBUG oslo_vmware.api [-] Task: {'id': task-1294604, 'name': CreateVM_Task, 'duration_secs': 0.5652} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 817.198532] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: a842df44-d8a9-4376-b9fc-5ca19a68a4b7] Created VM on the ESX host {{(pid=61806) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 817.199294] env[61806]: DEBUG oslo_concurrency.lockutils [None req-61b80e06-de2f-412b-be27-b9318eeb5c25 tempest-ServerAddressesTestJSON-661730579 tempest-ServerAddressesTestJSON-661730579-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 817.199495] env[61806]: DEBUG oslo_concurrency.lockutils [None req-61b80e06-de2f-412b-be27-b9318eeb5c25 tempest-ServerAddressesTestJSON-661730579 tempest-ServerAddressesTestJSON-661730579-project-member] Acquired lock "[datastore1] devstack-image-cache_base/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 817.200043] env[61806]: DEBUG oslo_concurrency.lockutils [None req-61b80e06-de2f-412b-be27-b9318eeb5c25 tempest-ServerAddressesTestJSON-661730579 tempest-ServerAddressesTestJSON-661730579-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 817.200318] env[61806]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d4433b76-1787-44bc-93fc-dd3e10582189 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 817.206543] env[61806]: DEBUG oslo_vmware.api [None req-61b80e06-de2f-412b-be27-b9318eeb5c25 tempest-ServerAddressesTestJSON-661730579 tempest-ServerAddressesTestJSON-661730579-project-member] Waiting for the task: (returnval){ [ 817.206543] env[61806]: value = "session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]526718bc-4f54-ec13-9fef-0facbcdaf2ad" [ 817.206543] env[61806]: _type = "Task" [ 817.206543] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 817.219527] env[61806]: DEBUG oslo_vmware.api [None req-61b80e06-de2f-412b-be27-b9318eeb5c25 tempest-ServerAddressesTestJSON-661730579 tempest-ServerAddressesTestJSON-661730579-project-member] Task: {'id': session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]526718bc-4f54-ec13-9fef-0facbcdaf2ad, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 817.220262] env[61806]: DEBUG oslo_concurrency.lockutils [None req-8aa43bc8-f134-4682-8a54-ce9aa25e4ee5 tempest-AttachVolumeNegativeTest-1683033704 tempest-AttachVolumeNegativeTest-1683033704-project-member] Acquiring lock "refresh_cache-43dcce0d-4fdb-4ee1-8309-92e92d2331a9" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 817.220480] env[61806]: DEBUG oslo_concurrency.lockutils [None req-8aa43bc8-f134-4682-8a54-ce9aa25e4ee5 tempest-AttachVolumeNegativeTest-1683033704 tempest-AttachVolumeNegativeTest-1683033704-project-member] Acquired lock "refresh_cache-43dcce0d-4fdb-4ee1-8309-92e92d2331a9" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 817.221362] env[61806]: DEBUG nova.network.neutron [None req-8aa43bc8-f134-4682-8a54-ce9aa25e4ee5 tempest-AttachVolumeNegativeTest-1683033704 tempest-AttachVolumeNegativeTest-1683033704-project-member] [instance: 43dcce0d-4fdb-4ee1-8309-92e92d2331a9] Building network info cache for instance {{(pid=61806) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 817.294552] env[61806]: DEBUG oslo_vmware.api [None req-5af76806-1cdf-4002-962d-34b8992ee9dc tempest-VolumesAdminNegativeTest-1387177911 tempest-VolumesAdminNegativeTest-1387177911-project-member] Task: {'id': task-1294605, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.082552} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 817.294915] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-5af76806-1cdf-4002-962d-34b8992ee9dc tempest-VolumesAdminNegativeTest-1387177911 tempest-VolumesAdminNegativeTest-1387177911-project-member] [instance: 4373b735-31cf-4b53-b655-38555cf212a5] Extended root virtual disk {{(pid=61806) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 817.295716] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e022c774-a714-43a1-84ae-188e2de906b2 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 817.320994] env[61806]: DEBUG nova.virt.vmwareapi.volumeops [None req-5af76806-1cdf-4002-962d-34b8992ee9dc tempest-VolumesAdminNegativeTest-1387177911 tempest-VolumesAdminNegativeTest-1387177911-project-member] [instance: 4373b735-31cf-4b53-b655-38555cf212a5] Reconfiguring VM instance instance-0000003d to attach disk [datastore1] 4373b735-31cf-4b53-b655-38555cf212a5/4373b735-31cf-4b53-b655-38555cf212a5.vmdk or device None with type sparse {{(pid=61806) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 817.321353] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-7bfa67bf-2402-4fec-9152-f41c71c9d3e7 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 817.342700] env[61806]: DEBUG oslo_vmware.api [None req-5af76806-1cdf-4002-962d-34b8992ee9dc tempest-VolumesAdminNegativeTest-1387177911 tempest-VolumesAdminNegativeTest-1387177911-project-member] Waiting for the task: (returnval){ [ 817.342700] env[61806]: value = "task-1294608" [ 817.342700] env[61806]: _type = "Task" [ 817.342700] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 817.352235] env[61806]: DEBUG oslo_vmware.api [None req-5af76806-1cdf-4002-962d-34b8992ee9dc tempest-VolumesAdminNegativeTest-1387177911 tempest-VolumesAdminNegativeTest-1387177911-project-member] Task: {'id': task-1294608, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 817.434354] env[61806]: DEBUG oslo_vmware.api [-] Task: {'id': task-1294606, 'name': CreateVM_Task, 'duration_secs': 0.354309} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 817.434566] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 616ec206-9804-469e-ab5c-41aea7f048aa] Created VM on the ESX host {{(pid=61806) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 817.435055] env[61806]: DEBUG oslo_concurrency.lockutils [None req-4e219b6d-f08a-4088-99ff-2fcd194df36a tempest-ServerShowV257Test-1159790492 tempest-ServerShowV257Test-1159790492-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 817.435232] env[61806]: DEBUG oslo_concurrency.lockutils [None req-4e219b6d-f08a-4088-99ff-2fcd194df36a tempest-ServerShowV257Test-1159790492 tempest-ServerShowV257Test-1159790492-project-member] Acquired lock "[datastore2] devstack-image-cache_base/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 817.435559] env[61806]: DEBUG oslo_concurrency.lockutils [None req-4e219b6d-f08a-4088-99ff-2fcd194df36a tempest-ServerShowV257Test-1159790492 tempest-ServerShowV257Test-1159790492-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 817.435822] env[61806]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-4012de57-927f-4f40-a8d5-6058217ba9e9 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 817.440954] env[61806]: DEBUG oslo_vmware.api [None req-4e219b6d-f08a-4088-99ff-2fcd194df36a tempest-ServerShowV257Test-1159790492 tempest-ServerShowV257Test-1159790492-project-member] Waiting for the task: (returnval){ [ 817.440954] env[61806]: value = "session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]5237496a-8fba-1722-57e3-29f09d1f98a7" [ 817.440954] env[61806]: _type = "Task" [ 817.440954] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 817.449678] env[61806]: DEBUG oslo_vmware.api [None req-4e219b6d-f08a-4088-99ff-2fcd194df36a tempest-ServerShowV257Test-1159790492 tempest-ServerShowV257Test-1159790492-project-member] Task: {'id': session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]5237496a-8fba-1722-57e3-29f09d1f98a7, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 817.469376] env[61806]: DEBUG nova.compute.manager [None req-c9a1ae7a-8313-4bdf-8de4-77929ce2d9bc tempest-ServerMetadataNegativeTestJSON-456453902 tempest-ServerMetadataNegativeTestJSON-456453902-project-member] [instance: 7f4907d0-178d-452a-8149-030becde8779] Starting instance... {{(pid=61806) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 817.569659] env[61806]: DEBUG oslo_concurrency.lockutils [req-020e7328-be26-46fa-9c10-01f8b970b777 req-4f9c072d-5eb9-4f48-8b80-a05cfda99308 service nova] Releasing lock "refresh_cache-a842df44-d8a9-4376-b9fc-5ca19a68a4b7" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 817.603301] env[61806]: DEBUG nova.compute.manager [None req-9c52c9e9-816e-4638-b13a-a63a9756140c tempest-SecurityGroupsTestJSON-1213485581 tempest-SecurityGroupsTestJSON-1213485581-project-member] [instance: b41531f2-f28c-4d82-9682-0b557bbaa491] Start building block device mappings for instance. {{(pid=61806) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 817.685314] env[61806]: DEBUG oslo_vmware.api [None req-7cf484b6-b263-4f99-a8c0-b55d49de0809 tempest-ServerShowV247Test-1218443059 tempest-ServerShowV247Test-1218443059-project-member] Task: {'id': task-1294607, 'name': Rename_Task, 'duration_secs': 0.338832} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 817.685314] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [None req-7cf484b6-b263-4f99-a8c0-b55d49de0809 tempest-ServerShowV247Test-1218443059 tempest-ServerShowV247Test-1218443059-project-member] [instance: 4febd093-9f2b-494e-b175-e4693b0e3e0d] Powering on the VM {{(pid=61806) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 817.685314] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-db4062f4-982a-406b-8245-f7a92ab1e1f8 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 817.692609] env[61806]: DEBUG oslo_vmware.api [None req-7cf484b6-b263-4f99-a8c0-b55d49de0809 tempest-ServerShowV247Test-1218443059 tempest-ServerShowV247Test-1218443059-project-member] Waiting for the task: (returnval){ [ 817.692609] env[61806]: value = "task-1294609" [ 817.692609] env[61806]: _type = "Task" [ 817.692609] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 817.708984] env[61806]: DEBUG oslo_vmware.api [None req-7cf484b6-b263-4f99-a8c0-b55d49de0809 tempest-ServerShowV247Test-1218443059 tempest-ServerShowV247Test-1218443059-project-member] Task: {'id': task-1294609, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 817.721084] env[61806]: DEBUG oslo_vmware.api [None req-61b80e06-de2f-412b-be27-b9318eeb5c25 tempest-ServerAddressesTestJSON-661730579 tempest-ServerAddressesTestJSON-661730579-project-member] Task: {'id': session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]526718bc-4f54-ec13-9fef-0facbcdaf2ad, 'name': SearchDatastore_Task, 'duration_secs': 0.024022} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 817.722211] env[61806]: DEBUG oslo_concurrency.lockutils [None req-61b80e06-de2f-412b-be27-b9318eeb5c25 tempest-ServerAddressesTestJSON-661730579 tempest-ServerAddressesTestJSON-661730579-project-member] Releasing lock "[datastore1] devstack-image-cache_base/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 817.722211] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-61b80e06-de2f-412b-be27-b9318eeb5c25 tempest-ServerAddressesTestJSON-661730579 tempest-ServerAddressesTestJSON-661730579-project-member] [instance: a842df44-d8a9-4376-b9fc-5ca19a68a4b7] Processing image ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa {{(pid=61806) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 817.722211] env[61806]: DEBUG oslo_concurrency.lockutils [None req-61b80e06-de2f-412b-be27-b9318eeb5c25 tempest-ServerAddressesTestJSON-661730579 tempest-ServerAddressesTestJSON-661730579-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa.vmdk" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 817.722211] env[61806]: DEBUG oslo_concurrency.lockutils [None req-61b80e06-de2f-412b-be27-b9318eeb5c25 tempest-ServerAddressesTestJSON-661730579 tempest-ServerAddressesTestJSON-661730579-project-member] Acquired lock "[datastore1] devstack-image-cache_base/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa.vmdk" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 817.722448] env[61806]: DEBUG nova.virt.vmwareapi.ds_util [None req-61b80e06-de2f-412b-be27-b9318eeb5c25 tempest-ServerAddressesTestJSON-661730579 tempest-ServerAddressesTestJSON-661730579-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61806) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 817.722517] env[61806]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-e6b49c37-bbfb-4c20-845f-736ee71d687b {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 817.739328] env[61806]: DEBUG nova.virt.vmwareapi.ds_util [None req-61b80e06-de2f-412b-be27-b9318eeb5c25 tempest-ServerAddressesTestJSON-661730579 tempest-ServerAddressesTestJSON-661730579-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61806) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 817.739328] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-61b80e06-de2f-412b-be27-b9318eeb5c25 tempest-ServerAddressesTestJSON-661730579 tempest-ServerAddressesTestJSON-661730579-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61806) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 817.739328] env[61806]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-621db5f4-dcd6-4b9a-8f42-7edbae9460f2 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 817.748021] env[61806]: DEBUG oslo_vmware.api [None req-61b80e06-de2f-412b-be27-b9318eeb5c25 tempest-ServerAddressesTestJSON-661730579 tempest-ServerAddressesTestJSON-661730579-project-member] Waiting for the task: (returnval){ [ 817.748021] env[61806]: value = "session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]52fcdef8-7f17-3add-f2d4-9b38c12a25ae" [ 817.748021] env[61806]: _type = "Task" [ 817.748021] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 817.761871] env[61806]: DEBUG oslo_vmware.api [None req-61b80e06-de2f-412b-be27-b9318eeb5c25 tempest-ServerAddressesTestJSON-661730579 tempest-ServerAddressesTestJSON-661730579-project-member] Task: {'id': session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]52fcdef8-7f17-3add-f2d4-9b38c12a25ae, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 817.784377] env[61806]: DEBUG nova.network.neutron [None req-8aa43bc8-f134-4682-8a54-ce9aa25e4ee5 tempest-AttachVolumeNegativeTest-1683033704 tempest-AttachVolumeNegativeTest-1683033704-project-member] [instance: 43dcce0d-4fdb-4ee1-8309-92e92d2331a9] Instance cache missing network info. {{(pid=61806) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 817.790201] env[61806]: DEBUG nova.network.neutron [None req-9c52c9e9-816e-4638-b13a-a63a9756140c tempest-SecurityGroupsTestJSON-1213485581 tempest-SecurityGroupsTestJSON-1213485581-project-member] [instance: b41531f2-f28c-4d82-9682-0b557bbaa491] Successfully created port: 513ca4b5-b62c-486a-9860-cf267a15c3e5 {{(pid=61806) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 817.855612] env[61806]: DEBUG oslo_vmware.api [None req-5af76806-1cdf-4002-962d-34b8992ee9dc tempest-VolumesAdminNegativeTest-1387177911 tempest-VolumesAdminNegativeTest-1387177911-project-member] Task: {'id': task-1294608, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 817.949573] env[61806]: DEBUG nova.network.neutron [None req-8aa43bc8-f134-4682-8a54-ce9aa25e4ee5 tempest-AttachVolumeNegativeTest-1683033704 tempest-AttachVolumeNegativeTest-1683033704-project-member] [instance: 43dcce0d-4fdb-4ee1-8309-92e92d2331a9] Updating instance_info_cache with network_info: [{"id": "6361e800-fbf2-48e0-a11d-b00d9a9955b4", "address": "fa:16:3e:db:ef:a2", "network": {"id": "d636ca55-0737-48e4-8d9d-7fe0f797f49e", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-2090334349-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "22e47882c2c44ce492dff6b1c5d782e2", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6f493cd8-1cb4-42a1-8d56-bfa6ac7cf563", "external-id": "nsx-vlan-transportzone-931", "segmentation_id": 931, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap6361e800-fb", "ovs_interfaceid": "6361e800-fbf2-48e0-a11d-b00d9a9955b4", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61806) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 817.958183] env[61806]: DEBUG oslo_vmware.api [None req-4e219b6d-f08a-4088-99ff-2fcd194df36a tempest-ServerShowV257Test-1159790492 tempest-ServerShowV257Test-1159790492-project-member] Task: {'id': session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]5237496a-8fba-1722-57e3-29f09d1f98a7, 'name': SearchDatastore_Task, 'duration_secs': 0.015668} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 817.958945] env[61806]: DEBUG oslo_concurrency.lockutils [None req-4e219b6d-f08a-4088-99ff-2fcd194df36a tempest-ServerShowV257Test-1159790492 tempest-ServerShowV257Test-1159790492-project-member] Releasing lock "[datastore2] devstack-image-cache_base/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 817.959241] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-4e219b6d-f08a-4088-99ff-2fcd194df36a tempest-ServerShowV257Test-1159790492 tempest-ServerShowV257Test-1159790492-project-member] [instance: 616ec206-9804-469e-ab5c-41aea7f048aa] Processing image ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa {{(pid=61806) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 817.959506] env[61806]: DEBUG oslo_concurrency.lockutils [None req-4e219b6d-f08a-4088-99ff-2fcd194df36a tempest-ServerShowV257Test-1159790492 tempest-ServerShowV257Test-1159790492-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa.vmdk" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 817.959689] env[61806]: DEBUG oslo_concurrency.lockutils [None req-4e219b6d-f08a-4088-99ff-2fcd194df36a tempest-ServerShowV257Test-1159790492 tempest-ServerShowV257Test-1159790492-project-member] Acquired lock "[datastore2] devstack-image-cache_base/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa.vmdk" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 817.959896] env[61806]: DEBUG nova.virt.vmwareapi.ds_util [None req-4e219b6d-f08a-4088-99ff-2fcd194df36a tempest-ServerShowV257Test-1159790492 tempest-ServerShowV257Test-1159790492-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61806) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 817.960224] env[61806]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-fc8b6c36-b344-463e-809e-ded402189a78 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 817.971240] env[61806]: DEBUG nova.virt.vmwareapi.ds_util [None req-4e219b6d-f08a-4088-99ff-2fcd194df36a tempest-ServerShowV257Test-1159790492 tempest-ServerShowV257Test-1159790492-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61806) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 817.971534] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-4e219b6d-f08a-4088-99ff-2fcd194df36a tempest-ServerShowV257Test-1159790492 tempest-ServerShowV257Test-1159790492-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61806) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 817.975066] env[61806]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-fbd6f508-4533-4d93-9bd6-fe882f9c421b {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 817.976855] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8e1cca08-71a8-415e-94db-20fa18006268 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 817.984152] env[61806]: DEBUG oslo_vmware.api [None req-4e219b6d-f08a-4088-99ff-2fcd194df36a tempest-ServerShowV257Test-1159790492 tempest-ServerShowV257Test-1159790492-project-member] Waiting for the task: (returnval){ [ 817.984152] env[61806]: value = "session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]521bbda0-e65f-3763-fd13-0b7ae12d349a" [ 817.984152] env[61806]: _type = "Task" [ 817.984152] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 817.993448] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1ec3db50-d010-4066-88d9-522c8f8cb148 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 818.000272] env[61806]: DEBUG oslo_vmware.api [None req-4e219b6d-f08a-4088-99ff-2fcd194df36a tempest-ServerShowV257Test-1159790492 tempest-ServerShowV257Test-1159790492-project-member] Task: {'id': session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]521bbda0-e65f-3763-fd13-0b7ae12d349a, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 818.001337] env[61806]: DEBUG oslo_concurrency.lockutils [None req-c9a1ae7a-8313-4bdf-8de4-77929ce2d9bc tempest-ServerMetadataNegativeTestJSON-456453902 tempest-ServerMetadataNegativeTestJSON-456453902-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 818.035145] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a9329bde-e218-4553-b1e9-6f923407292b {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 818.044555] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-859b4afa-99ba-4dc9-8f34-ed1d534752ac {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 818.062882] env[61806]: DEBUG nova.compute.provider_tree [None req-7fa92f02-ad2b-4004-a9a7-cb08f83ec0f1 tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Inventory has not changed in ProviderTree for provider: a2858be1-fd22-4e08-979e-87ad25293407 {{(pid=61806) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 818.204904] env[61806]: DEBUG oslo_vmware.api [None req-7cf484b6-b263-4f99-a8c0-b55d49de0809 tempest-ServerShowV247Test-1218443059 tempest-ServerShowV247Test-1218443059-project-member] Task: {'id': task-1294609, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 818.260007] env[61806]: DEBUG oslo_vmware.api [None req-61b80e06-de2f-412b-be27-b9318eeb5c25 tempest-ServerAddressesTestJSON-661730579 tempest-ServerAddressesTestJSON-661730579-project-member] Task: {'id': session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]52fcdef8-7f17-3add-f2d4-9b38c12a25ae, 'name': SearchDatastore_Task, 'duration_secs': 0.041573} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 818.260895] env[61806]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-630bd673-ba08-4501-aaf2-4765db9d437a {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 818.267664] env[61806]: DEBUG oslo_vmware.api [None req-61b80e06-de2f-412b-be27-b9318eeb5c25 tempest-ServerAddressesTestJSON-661730579 tempest-ServerAddressesTestJSON-661730579-project-member] Waiting for the task: (returnval){ [ 818.267664] env[61806]: value = "session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]52c5a17e-95c2-ec91-bc29-da54311bdf3e" [ 818.267664] env[61806]: _type = "Task" [ 818.267664] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 818.276977] env[61806]: DEBUG oslo_vmware.api [None req-61b80e06-de2f-412b-be27-b9318eeb5c25 tempest-ServerAddressesTestJSON-661730579 tempest-ServerAddressesTestJSON-661730579-project-member] Task: {'id': session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]52c5a17e-95c2-ec91-bc29-da54311bdf3e, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 818.354205] env[61806]: DEBUG oslo_vmware.api [None req-5af76806-1cdf-4002-962d-34b8992ee9dc tempest-VolumesAdminNegativeTest-1387177911 tempest-VolumesAdminNegativeTest-1387177911-project-member] Task: {'id': task-1294608, 'name': ReconfigVM_Task} progress is 99%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 818.393385] env[61806]: DEBUG nova.compute.manager [req-181e5b0d-6b8f-42c3-82a0-d4b1df13cdd9 req-bb5a67e5-33b4-4d9a-b286-d43a54a42324 service nova] [instance: 43dcce0d-4fdb-4ee1-8309-92e92d2331a9] Received event network-vif-plugged-6361e800-fbf2-48e0-a11d-b00d9a9955b4 {{(pid=61806) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 818.393385] env[61806]: DEBUG oslo_concurrency.lockutils [req-181e5b0d-6b8f-42c3-82a0-d4b1df13cdd9 req-bb5a67e5-33b4-4d9a-b286-d43a54a42324 service nova] Acquiring lock "43dcce0d-4fdb-4ee1-8309-92e92d2331a9-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 818.393548] env[61806]: DEBUG oslo_concurrency.lockutils [req-181e5b0d-6b8f-42c3-82a0-d4b1df13cdd9 req-bb5a67e5-33b4-4d9a-b286-d43a54a42324 service nova] Lock "43dcce0d-4fdb-4ee1-8309-92e92d2331a9-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 818.393731] env[61806]: DEBUG oslo_concurrency.lockutils [req-181e5b0d-6b8f-42c3-82a0-d4b1df13cdd9 req-bb5a67e5-33b4-4d9a-b286-d43a54a42324 service nova] Lock "43dcce0d-4fdb-4ee1-8309-92e92d2331a9-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 818.393908] env[61806]: DEBUG nova.compute.manager [req-181e5b0d-6b8f-42c3-82a0-d4b1df13cdd9 req-bb5a67e5-33b4-4d9a-b286-d43a54a42324 service nova] [instance: 43dcce0d-4fdb-4ee1-8309-92e92d2331a9] No waiting events found dispatching network-vif-plugged-6361e800-fbf2-48e0-a11d-b00d9a9955b4 {{(pid=61806) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 818.394885] env[61806]: WARNING nova.compute.manager [req-181e5b0d-6b8f-42c3-82a0-d4b1df13cdd9 req-bb5a67e5-33b4-4d9a-b286-d43a54a42324 service nova] [instance: 43dcce0d-4fdb-4ee1-8309-92e92d2331a9] Received unexpected event network-vif-plugged-6361e800-fbf2-48e0-a11d-b00d9a9955b4 for instance with vm_state building and task_state spawning. [ 818.395221] env[61806]: DEBUG nova.compute.manager [req-181e5b0d-6b8f-42c3-82a0-d4b1df13cdd9 req-bb5a67e5-33b4-4d9a-b286-d43a54a42324 service nova] [instance: 43dcce0d-4fdb-4ee1-8309-92e92d2331a9] Received event network-changed-6361e800-fbf2-48e0-a11d-b00d9a9955b4 {{(pid=61806) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 818.395353] env[61806]: DEBUG nova.compute.manager [req-181e5b0d-6b8f-42c3-82a0-d4b1df13cdd9 req-bb5a67e5-33b4-4d9a-b286-d43a54a42324 service nova] [instance: 43dcce0d-4fdb-4ee1-8309-92e92d2331a9] Refreshing instance network info cache due to event network-changed-6361e800-fbf2-48e0-a11d-b00d9a9955b4. {{(pid=61806) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 818.395624] env[61806]: DEBUG oslo_concurrency.lockutils [req-181e5b0d-6b8f-42c3-82a0-d4b1df13cdd9 req-bb5a67e5-33b4-4d9a-b286-d43a54a42324 service nova] Acquiring lock "refresh_cache-43dcce0d-4fdb-4ee1-8309-92e92d2331a9" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 818.460894] env[61806]: DEBUG oslo_concurrency.lockutils [None req-8aa43bc8-f134-4682-8a54-ce9aa25e4ee5 tempest-AttachVolumeNegativeTest-1683033704 tempest-AttachVolumeNegativeTest-1683033704-project-member] Releasing lock "refresh_cache-43dcce0d-4fdb-4ee1-8309-92e92d2331a9" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 818.460894] env[61806]: DEBUG nova.compute.manager [None req-8aa43bc8-f134-4682-8a54-ce9aa25e4ee5 tempest-AttachVolumeNegativeTest-1683033704 tempest-AttachVolumeNegativeTest-1683033704-project-member] [instance: 43dcce0d-4fdb-4ee1-8309-92e92d2331a9] Instance network_info: |[{"id": "6361e800-fbf2-48e0-a11d-b00d9a9955b4", "address": "fa:16:3e:db:ef:a2", "network": {"id": "d636ca55-0737-48e4-8d9d-7fe0f797f49e", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-2090334349-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "22e47882c2c44ce492dff6b1c5d782e2", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6f493cd8-1cb4-42a1-8d56-bfa6ac7cf563", "external-id": "nsx-vlan-transportzone-931", "segmentation_id": 931, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap6361e800-fb", "ovs_interfaceid": "6361e800-fbf2-48e0-a11d-b00d9a9955b4", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61806) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 818.461143] env[61806]: DEBUG oslo_concurrency.lockutils [req-181e5b0d-6b8f-42c3-82a0-d4b1df13cdd9 req-bb5a67e5-33b4-4d9a-b286-d43a54a42324 service nova] Acquired lock "refresh_cache-43dcce0d-4fdb-4ee1-8309-92e92d2331a9" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 818.461143] env[61806]: DEBUG nova.network.neutron [req-181e5b0d-6b8f-42c3-82a0-d4b1df13cdd9 req-bb5a67e5-33b4-4d9a-b286-d43a54a42324 service nova] [instance: 43dcce0d-4fdb-4ee1-8309-92e92d2331a9] Refreshing network info cache for port 6361e800-fbf2-48e0-a11d-b00d9a9955b4 {{(pid=61806) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 818.462464] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-8aa43bc8-f134-4682-8a54-ce9aa25e4ee5 tempest-AttachVolumeNegativeTest-1683033704 tempest-AttachVolumeNegativeTest-1683033704-project-member] [instance: 43dcce0d-4fdb-4ee1-8309-92e92d2331a9] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:db:ef:a2', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '6f493cd8-1cb4-42a1-8d56-bfa6ac7cf563', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '6361e800-fbf2-48e0-a11d-b00d9a9955b4', 'vif_model': 'vmxnet3'}] {{(pid=61806) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 818.473837] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [None req-8aa43bc8-f134-4682-8a54-ce9aa25e4ee5 tempest-AttachVolumeNegativeTest-1683033704 tempest-AttachVolumeNegativeTest-1683033704-project-member] Creating folder: Project (22e47882c2c44ce492dff6b1c5d782e2). Parent ref: group-v277609. {{(pid=61806) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 818.474349] env[61806]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-b90e1085-e996-4656-9f41-ed0020d4cae6 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 818.491670] env[61806]: INFO nova.virt.vmwareapi.vm_util [None req-8aa43bc8-f134-4682-8a54-ce9aa25e4ee5 tempest-AttachVolumeNegativeTest-1683033704 tempest-AttachVolumeNegativeTest-1683033704-project-member] Created folder: Project (22e47882c2c44ce492dff6b1c5d782e2) in parent group-v277609. [ 818.491879] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [None req-8aa43bc8-f134-4682-8a54-ce9aa25e4ee5 tempest-AttachVolumeNegativeTest-1683033704 tempest-AttachVolumeNegativeTest-1683033704-project-member] Creating folder: Instances. Parent ref: group-v277658. {{(pid=61806) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 818.492535] env[61806]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-65deb051-c1ba-458c-a3d7-e33c6f98ed09 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 818.500162] env[61806]: DEBUG oslo_vmware.api [None req-4e219b6d-f08a-4088-99ff-2fcd194df36a tempest-ServerShowV257Test-1159790492 tempest-ServerShowV257Test-1159790492-project-member] Task: {'id': session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]521bbda0-e65f-3763-fd13-0b7ae12d349a, 'name': SearchDatastore_Task, 'duration_secs': 0.025723} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 818.500864] env[61806]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b5aa2826-37a1-45d5-86fc-2c758d2d6bfe {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 818.506712] env[61806]: DEBUG oslo_vmware.api [None req-4e219b6d-f08a-4088-99ff-2fcd194df36a tempest-ServerShowV257Test-1159790492 tempest-ServerShowV257Test-1159790492-project-member] Waiting for the task: (returnval){ [ 818.506712] env[61806]: value = "session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]521d3b9a-777b-6b59-5a4f-13768f4ee926" [ 818.506712] env[61806]: _type = "Task" [ 818.506712] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 818.511884] env[61806]: INFO nova.virt.vmwareapi.vm_util [None req-8aa43bc8-f134-4682-8a54-ce9aa25e4ee5 tempest-AttachVolumeNegativeTest-1683033704 tempest-AttachVolumeNegativeTest-1683033704-project-member] Created folder: Instances in parent group-v277658. [ 818.512150] env[61806]: DEBUG oslo.service.loopingcall [None req-8aa43bc8-f134-4682-8a54-ce9aa25e4ee5 tempest-AttachVolumeNegativeTest-1683033704 tempest-AttachVolumeNegativeTest-1683033704-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61806) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 818.512807] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 43dcce0d-4fdb-4ee1-8309-92e92d2331a9] Creating VM on the ESX host {{(pid=61806) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 818.513137] env[61806]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-00f8cdf8-ea94-41c1-93a9-57176c8dc470 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 818.531443] env[61806]: DEBUG oslo_vmware.api [None req-4e219b6d-f08a-4088-99ff-2fcd194df36a tempest-ServerShowV257Test-1159790492 tempest-ServerShowV257Test-1159790492-project-member] Task: {'id': session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]521d3b9a-777b-6b59-5a4f-13768f4ee926, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 818.537547] env[61806]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 818.537547] env[61806]: value = "task-1294612" [ 818.537547] env[61806]: _type = "Task" [ 818.537547] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 818.547248] env[61806]: DEBUG oslo_vmware.api [-] Task: {'id': task-1294612, 'name': CreateVM_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 818.566918] env[61806]: DEBUG nova.scheduler.client.report [None req-7fa92f02-ad2b-4004-a9a7-cb08f83ec0f1 tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Inventory has not changed for provider a2858be1-fd22-4e08-979e-87ad25293407 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 138, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61806) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 818.617110] env[61806]: DEBUG nova.compute.manager [None req-9c52c9e9-816e-4638-b13a-a63a9756140c tempest-SecurityGroupsTestJSON-1213485581 tempest-SecurityGroupsTestJSON-1213485581-project-member] [instance: b41531f2-f28c-4d82-9682-0b557bbaa491] Start spawning the instance on the hypervisor. {{(pid=61806) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 818.649773] env[61806]: DEBUG nova.virt.hardware [None req-9c52c9e9-816e-4638-b13a-a63a9756140c tempest-SecurityGroupsTestJSON-1213485581 tempest-SecurityGroupsTestJSON-1213485581-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-15T18:05:10Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-15T18:04:52Z,direct_url=,disk_format='vmdk',id=ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='84ba83fce2c349a6988173c5d6fc3b07',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-15T18:04:53Z,virtual_size=,visibility=), allow threads: False {{(pid=61806) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 818.650042] env[61806]: DEBUG nova.virt.hardware [None req-9c52c9e9-816e-4638-b13a-a63a9756140c tempest-SecurityGroupsTestJSON-1213485581 tempest-SecurityGroupsTestJSON-1213485581-project-member] Flavor limits 0:0:0 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 818.650213] env[61806]: DEBUG nova.virt.hardware [None req-9c52c9e9-816e-4638-b13a-a63a9756140c tempest-SecurityGroupsTestJSON-1213485581 tempest-SecurityGroupsTestJSON-1213485581-project-member] Image limits 0:0:0 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 818.650478] env[61806]: DEBUG nova.virt.hardware [None req-9c52c9e9-816e-4638-b13a-a63a9756140c tempest-SecurityGroupsTestJSON-1213485581 tempest-SecurityGroupsTestJSON-1213485581-project-member] Flavor pref 0:0:0 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 818.650721] env[61806]: DEBUG nova.virt.hardware [None req-9c52c9e9-816e-4638-b13a-a63a9756140c tempest-SecurityGroupsTestJSON-1213485581 tempest-SecurityGroupsTestJSON-1213485581-project-member] Image pref 0:0:0 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 818.650939] env[61806]: DEBUG nova.virt.hardware [None req-9c52c9e9-816e-4638-b13a-a63a9756140c tempest-SecurityGroupsTestJSON-1213485581 tempest-SecurityGroupsTestJSON-1213485581-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 818.651582] env[61806]: DEBUG nova.virt.hardware [None req-9c52c9e9-816e-4638-b13a-a63a9756140c tempest-SecurityGroupsTestJSON-1213485581 tempest-SecurityGroupsTestJSON-1213485581-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61806) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 818.651582] env[61806]: DEBUG nova.virt.hardware [None req-9c52c9e9-816e-4638-b13a-a63a9756140c tempest-SecurityGroupsTestJSON-1213485581 tempest-SecurityGroupsTestJSON-1213485581-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61806) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 818.651806] env[61806]: DEBUG nova.virt.hardware [None req-9c52c9e9-816e-4638-b13a-a63a9756140c tempest-SecurityGroupsTestJSON-1213485581 tempest-SecurityGroupsTestJSON-1213485581-project-member] Got 1 possible topologies {{(pid=61806) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 818.651896] env[61806]: DEBUG nova.virt.hardware [None req-9c52c9e9-816e-4638-b13a-a63a9756140c tempest-SecurityGroupsTestJSON-1213485581 tempest-SecurityGroupsTestJSON-1213485581-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61806) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 818.652043] env[61806]: DEBUG nova.virt.hardware [None req-9c52c9e9-816e-4638-b13a-a63a9756140c tempest-SecurityGroupsTestJSON-1213485581 tempest-SecurityGroupsTestJSON-1213485581-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61806) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 818.653525] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-af50d77b-f1c9-4825-a31a-4fac81fd10f9 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 818.663044] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1b6e90d6-b174-4cbf-9317-24feee3e80c7 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 818.704644] env[61806]: DEBUG oslo_vmware.api [None req-7cf484b6-b263-4f99-a8c0-b55d49de0809 tempest-ServerShowV247Test-1218443059 tempest-ServerShowV247Test-1218443059-project-member] Task: {'id': task-1294609, 'name': PowerOnVM_Task, 'duration_secs': 0.72879} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 818.704961] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [None req-7cf484b6-b263-4f99-a8c0-b55d49de0809 tempest-ServerShowV247Test-1218443059 tempest-ServerShowV247Test-1218443059-project-member] [instance: 4febd093-9f2b-494e-b175-e4693b0e3e0d] Powered on the VM {{(pid=61806) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 818.705577] env[61806]: DEBUG nova.compute.manager [None req-7cf484b6-b263-4f99-a8c0-b55d49de0809 tempest-ServerShowV247Test-1218443059 tempest-ServerShowV247Test-1218443059-project-member] [instance: 4febd093-9f2b-494e-b175-e4693b0e3e0d] Checking state {{(pid=61806) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 818.706662] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d6edef18-c19c-4bee-aadd-cde1d56adeef {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 818.779480] env[61806]: DEBUG oslo_vmware.api [None req-61b80e06-de2f-412b-be27-b9318eeb5c25 tempest-ServerAddressesTestJSON-661730579 tempest-ServerAddressesTestJSON-661730579-project-member] Task: {'id': session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]52c5a17e-95c2-ec91-bc29-da54311bdf3e, 'name': SearchDatastore_Task, 'duration_secs': 0.012938} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 818.779698] env[61806]: DEBUG oslo_concurrency.lockutils [None req-61b80e06-de2f-412b-be27-b9318eeb5c25 tempest-ServerAddressesTestJSON-661730579 tempest-ServerAddressesTestJSON-661730579-project-member] Releasing lock "[datastore1] devstack-image-cache_base/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa.vmdk" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 818.781166] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [None req-61b80e06-de2f-412b-be27-b9318eeb5c25 tempest-ServerAddressesTestJSON-661730579 tempest-ServerAddressesTestJSON-661730579-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa.vmdk to [datastore1] a842df44-d8a9-4376-b9fc-5ca19a68a4b7/a842df44-d8a9-4376-b9fc-5ca19a68a4b7.vmdk {{(pid=61806) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 818.781166] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-62d33f8b-0127-407a-980d-f23d7483dd54 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 818.789210] env[61806]: DEBUG oslo_vmware.api [None req-61b80e06-de2f-412b-be27-b9318eeb5c25 tempest-ServerAddressesTestJSON-661730579 tempest-ServerAddressesTestJSON-661730579-project-member] Waiting for the task: (returnval){ [ 818.789210] env[61806]: value = "task-1294613" [ 818.789210] env[61806]: _type = "Task" [ 818.789210] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 818.798642] env[61806]: DEBUG oslo_vmware.api [None req-61b80e06-de2f-412b-be27-b9318eeb5c25 tempest-ServerAddressesTestJSON-661730579 tempest-ServerAddressesTestJSON-661730579-project-member] Task: {'id': task-1294613, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 818.854582] env[61806]: DEBUG oslo_vmware.api [None req-5af76806-1cdf-4002-962d-34b8992ee9dc tempest-VolumesAdminNegativeTest-1387177911 tempest-VolumesAdminNegativeTest-1387177911-project-member] Task: {'id': task-1294608, 'name': ReconfigVM_Task} progress is 99%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 819.018926] env[61806]: DEBUG oslo_vmware.api [None req-4e219b6d-f08a-4088-99ff-2fcd194df36a tempest-ServerShowV257Test-1159790492 tempest-ServerShowV257Test-1159790492-project-member] Task: {'id': session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]521d3b9a-777b-6b59-5a4f-13768f4ee926, 'name': SearchDatastore_Task, 'duration_secs': 0.025305} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 819.019299] env[61806]: DEBUG oslo_concurrency.lockutils [None req-4e219b6d-f08a-4088-99ff-2fcd194df36a tempest-ServerShowV257Test-1159790492 tempest-ServerShowV257Test-1159790492-project-member] Releasing lock "[datastore2] devstack-image-cache_base/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa.vmdk" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 819.019544] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [None req-4e219b6d-f08a-4088-99ff-2fcd194df36a tempest-ServerShowV257Test-1159790492 tempest-ServerShowV257Test-1159790492-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa.vmdk to [datastore2] 616ec206-9804-469e-ab5c-41aea7f048aa/616ec206-9804-469e-ab5c-41aea7f048aa.vmdk {{(pid=61806) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 819.020159] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-ce04874d-03a8-4677-a8b8-cfdd1ddb4f51 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 819.028260] env[61806]: DEBUG oslo_vmware.api [None req-4e219b6d-f08a-4088-99ff-2fcd194df36a tempest-ServerShowV257Test-1159790492 tempest-ServerShowV257Test-1159790492-project-member] Waiting for the task: (returnval){ [ 819.028260] env[61806]: value = "task-1294614" [ 819.028260] env[61806]: _type = "Task" [ 819.028260] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 819.037515] env[61806]: DEBUG oslo_vmware.api [None req-4e219b6d-f08a-4088-99ff-2fcd194df36a tempest-ServerShowV257Test-1159790492 tempest-ServerShowV257Test-1159790492-project-member] Task: {'id': task-1294614, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 819.047615] env[61806]: DEBUG oslo_vmware.api [-] Task: {'id': task-1294612, 'name': CreateVM_Task, 'duration_secs': 0.472618} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 819.047789] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 43dcce0d-4fdb-4ee1-8309-92e92d2331a9] Created VM on the ESX host {{(pid=61806) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 819.048506] env[61806]: DEBUG oslo_concurrency.lockutils [None req-8aa43bc8-f134-4682-8a54-ce9aa25e4ee5 tempest-AttachVolumeNegativeTest-1683033704 tempest-AttachVolumeNegativeTest-1683033704-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 819.048688] env[61806]: DEBUG oslo_concurrency.lockutils [None req-8aa43bc8-f134-4682-8a54-ce9aa25e4ee5 tempest-AttachVolumeNegativeTest-1683033704 tempest-AttachVolumeNegativeTest-1683033704-project-member] Acquired lock "[datastore1] devstack-image-cache_base/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 819.049041] env[61806]: DEBUG oslo_concurrency.lockutils [None req-8aa43bc8-f134-4682-8a54-ce9aa25e4ee5 tempest-AttachVolumeNegativeTest-1683033704 tempest-AttachVolumeNegativeTest-1683033704-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 819.049322] env[61806]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-95d95e92-f1cb-4be8-a625-d83d05a393c6 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 819.054693] env[61806]: DEBUG oslo_vmware.api [None req-8aa43bc8-f134-4682-8a54-ce9aa25e4ee5 tempest-AttachVolumeNegativeTest-1683033704 tempest-AttachVolumeNegativeTest-1683033704-project-member] Waiting for the task: (returnval){ [ 819.054693] env[61806]: value = "session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]52a45b05-91b6-7da9-9c64-132a17feeb28" [ 819.054693] env[61806]: _type = "Task" [ 819.054693] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 819.066249] env[61806]: DEBUG oslo_vmware.api [None req-8aa43bc8-f134-4682-8a54-ce9aa25e4ee5 tempest-AttachVolumeNegativeTest-1683033704 tempest-AttachVolumeNegativeTest-1683033704-project-member] Task: {'id': session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]52a45b05-91b6-7da9-9c64-132a17feeb28, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 819.071177] env[61806]: DEBUG oslo_concurrency.lockutils [None req-7fa92f02-ad2b-4004-a9a7-cb08f83ec0f1 tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.480s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 819.071699] env[61806]: DEBUG nova.compute.manager [None req-7fa92f02-ad2b-4004-a9a7-cb08f83ec0f1 tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] [instance: 67d1bdd8-cf8f-4fb5-bd3c-42008e4d60e3] Start building networks asynchronously for instance. {{(pid=61806) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 819.077902] env[61806]: DEBUG oslo_concurrency.lockutils [None req-cbb90e6d-7dac-4c30-b860-1f0db7dd091e tempest-MultipleCreateTestJSON-1589015462 tempest-MultipleCreateTestJSON-1589015462-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 14.977s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 819.079622] env[61806]: INFO nova.compute.claims [None req-cbb90e6d-7dac-4c30-b860-1f0db7dd091e tempest-MultipleCreateTestJSON-1589015462 tempest-MultipleCreateTestJSON-1589015462-project-member] [instance: 25218e83-c1ab-4b97-ade1-7c32b6f99305] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 819.226224] env[61806]: DEBUG oslo_concurrency.lockutils [None req-7cf484b6-b263-4f99-a8c0-b55d49de0809 tempest-ServerShowV247Test-1218443059 tempest-ServerShowV247Test-1218443059-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 819.275881] env[61806]: DEBUG nova.network.neutron [req-181e5b0d-6b8f-42c3-82a0-d4b1df13cdd9 req-bb5a67e5-33b4-4d9a-b286-d43a54a42324 service nova] [instance: 43dcce0d-4fdb-4ee1-8309-92e92d2331a9] Updated VIF entry in instance network info cache for port 6361e800-fbf2-48e0-a11d-b00d9a9955b4. {{(pid=61806) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 819.276303] env[61806]: DEBUG nova.network.neutron [req-181e5b0d-6b8f-42c3-82a0-d4b1df13cdd9 req-bb5a67e5-33b4-4d9a-b286-d43a54a42324 service nova] [instance: 43dcce0d-4fdb-4ee1-8309-92e92d2331a9] Updating instance_info_cache with network_info: [{"id": "6361e800-fbf2-48e0-a11d-b00d9a9955b4", "address": "fa:16:3e:db:ef:a2", "network": {"id": "d636ca55-0737-48e4-8d9d-7fe0f797f49e", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-2090334349-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "22e47882c2c44ce492dff6b1c5d782e2", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6f493cd8-1cb4-42a1-8d56-bfa6ac7cf563", "external-id": "nsx-vlan-transportzone-931", "segmentation_id": 931, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap6361e800-fb", "ovs_interfaceid": "6361e800-fbf2-48e0-a11d-b00d9a9955b4", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61806) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 819.301175] env[61806]: DEBUG oslo_vmware.api [None req-61b80e06-de2f-412b-be27-b9318eeb5c25 tempest-ServerAddressesTestJSON-661730579 tempest-ServerAddressesTestJSON-661730579-project-member] Task: {'id': task-1294613, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 819.356328] env[61806]: DEBUG oslo_vmware.api [None req-5af76806-1cdf-4002-962d-34b8992ee9dc tempest-VolumesAdminNegativeTest-1387177911 tempest-VolumesAdminNegativeTest-1387177911-project-member] Task: {'id': task-1294608, 'name': ReconfigVM_Task, 'duration_secs': 1.708029} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 819.356635] env[61806]: DEBUG nova.virt.vmwareapi.volumeops [None req-5af76806-1cdf-4002-962d-34b8992ee9dc tempest-VolumesAdminNegativeTest-1387177911 tempest-VolumesAdminNegativeTest-1387177911-project-member] [instance: 4373b735-31cf-4b53-b655-38555cf212a5] Reconfigured VM instance instance-0000003d to attach disk [datastore1] 4373b735-31cf-4b53-b655-38555cf212a5/4373b735-31cf-4b53-b655-38555cf212a5.vmdk or device None with type sparse {{(pid=61806) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 819.357375] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-23bb9adf-7df7-438a-8c5f-166f72da41dc {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 819.365545] env[61806]: DEBUG oslo_vmware.api [None req-5af76806-1cdf-4002-962d-34b8992ee9dc tempest-VolumesAdminNegativeTest-1387177911 tempest-VolumesAdminNegativeTest-1387177911-project-member] Waiting for the task: (returnval){ [ 819.365545] env[61806]: value = "task-1294615" [ 819.365545] env[61806]: _type = "Task" [ 819.365545] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 819.377139] env[61806]: DEBUG oslo_vmware.api [None req-5af76806-1cdf-4002-962d-34b8992ee9dc tempest-VolumesAdminNegativeTest-1387177911 tempest-VolumesAdminNegativeTest-1387177911-project-member] Task: {'id': task-1294615, 'name': Rename_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 819.540191] env[61806]: DEBUG oslo_vmware.api [None req-4e219b6d-f08a-4088-99ff-2fcd194df36a tempest-ServerShowV257Test-1159790492 tempest-ServerShowV257Test-1159790492-project-member] Task: {'id': task-1294614, 'name': CopyVirtualDisk_Task} progress is 25%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 819.569931] env[61806]: DEBUG oslo_vmware.api [None req-8aa43bc8-f134-4682-8a54-ce9aa25e4ee5 tempest-AttachVolumeNegativeTest-1683033704 tempest-AttachVolumeNegativeTest-1683033704-project-member] Task: {'id': session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]52a45b05-91b6-7da9-9c64-132a17feeb28, 'name': SearchDatastore_Task, 'duration_secs': 0.06918} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 819.570338] env[61806]: DEBUG oslo_concurrency.lockutils [None req-8aa43bc8-f134-4682-8a54-ce9aa25e4ee5 tempest-AttachVolumeNegativeTest-1683033704 tempest-AttachVolumeNegativeTest-1683033704-project-member] Releasing lock "[datastore1] devstack-image-cache_base/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 819.570593] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-8aa43bc8-f134-4682-8a54-ce9aa25e4ee5 tempest-AttachVolumeNegativeTest-1683033704 tempest-AttachVolumeNegativeTest-1683033704-project-member] [instance: 43dcce0d-4fdb-4ee1-8309-92e92d2331a9] Processing image ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa {{(pid=61806) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 819.570998] env[61806]: DEBUG oslo_concurrency.lockutils [None req-8aa43bc8-f134-4682-8a54-ce9aa25e4ee5 tempest-AttachVolumeNegativeTest-1683033704 tempest-AttachVolumeNegativeTest-1683033704-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa.vmdk" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 819.571225] env[61806]: DEBUG oslo_concurrency.lockutils [None req-8aa43bc8-f134-4682-8a54-ce9aa25e4ee5 tempest-AttachVolumeNegativeTest-1683033704 tempest-AttachVolumeNegativeTest-1683033704-project-member] Acquired lock "[datastore1] devstack-image-cache_base/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa.vmdk" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 819.571436] env[61806]: DEBUG nova.virt.vmwareapi.ds_util [None req-8aa43bc8-f134-4682-8a54-ce9aa25e4ee5 tempest-AttachVolumeNegativeTest-1683033704 tempest-AttachVolumeNegativeTest-1683033704-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61806) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 819.571759] env[61806]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-97a22cb8-2264-4f8f-9a38-b8bce7a79b2a {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 819.577644] env[61806]: DEBUG nova.compute.utils [None req-7fa92f02-ad2b-4004-a9a7-cb08f83ec0f1 tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Using /dev/sd instead of None {{(pid=61806) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 819.579105] env[61806]: DEBUG nova.compute.manager [None req-7fa92f02-ad2b-4004-a9a7-cb08f83ec0f1 tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] [instance: 67d1bdd8-cf8f-4fb5-bd3c-42008e4d60e3] Allocating IP information in the background. {{(pid=61806) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 819.580074] env[61806]: DEBUG nova.network.neutron [None req-7fa92f02-ad2b-4004-a9a7-cb08f83ec0f1 tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] [instance: 67d1bdd8-cf8f-4fb5-bd3c-42008e4d60e3] allocate_for_instance() {{(pid=61806) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 819.588366] env[61806]: DEBUG nova.virt.vmwareapi.ds_util [None req-8aa43bc8-f134-4682-8a54-ce9aa25e4ee5 tempest-AttachVolumeNegativeTest-1683033704 tempest-AttachVolumeNegativeTest-1683033704-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61806) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 819.588556] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-8aa43bc8-f134-4682-8a54-ce9aa25e4ee5 tempest-AttachVolumeNegativeTest-1683033704 tempest-AttachVolumeNegativeTest-1683033704-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61806) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 819.589252] env[61806]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b27be685-edd0-484c-8022-6e4a2c4a4122 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 819.598691] env[61806]: DEBUG oslo_vmware.api [None req-8aa43bc8-f134-4682-8a54-ce9aa25e4ee5 tempest-AttachVolumeNegativeTest-1683033704 tempest-AttachVolumeNegativeTest-1683033704-project-member] Waiting for the task: (returnval){ [ 819.598691] env[61806]: value = "session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]525549a4-8d0f-207c-8c98-c5e236b74d2f" [ 819.598691] env[61806]: _type = "Task" [ 819.598691] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 819.608639] env[61806]: DEBUG oslo_vmware.api [None req-8aa43bc8-f134-4682-8a54-ce9aa25e4ee5 tempest-AttachVolumeNegativeTest-1683033704 tempest-AttachVolumeNegativeTest-1683033704-project-member] Task: {'id': session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]525549a4-8d0f-207c-8c98-c5e236b74d2f, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 819.674725] env[61806]: DEBUG nova.policy [None req-7fa92f02-ad2b-4004-a9a7-cb08f83ec0f1 tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '921aeb4e90e64f89888a9c65f873d66b', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'c1f2d161c50f4b70987b90e6364c3779', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61806) authorize /opt/stack/nova/nova/policy.py:201}} [ 819.780833] env[61806]: DEBUG oslo_concurrency.lockutils [req-181e5b0d-6b8f-42c3-82a0-d4b1df13cdd9 req-bb5a67e5-33b4-4d9a-b286-d43a54a42324 service nova] Releasing lock "refresh_cache-43dcce0d-4fdb-4ee1-8309-92e92d2331a9" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 819.802116] env[61806]: DEBUG oslo_vmware.api [None req-61b80e06-de2f-412b-be27-b9318eeb5c25 tempest-ServerAddressesTestJSON-661730579 tempest-ServerAddressesTestJSON-661730579-project-member] Task: {'id': task-1294613, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.596623} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 819.803403] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [None req-61b80e06-de2f-412b-be27-b9318eeb5c25 tempest-ServerAddressesTestJSON-661730579 tempest-ServerAddressesTestJSON-661730579-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa.vmdk to [datastore1] a842df44-d8a9-4376-b9fc-5ca19a68a4b7/a842df44-d8a9-4376-b9fc-5ca19a68a4b7.vmdk {{(pid=61806) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 819.803742] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-61b80e06-de2f-412b-be27-b9318eeb5c25 tempest-ServerAddressesTestJSON-661730579 tempest-ServerAddressesTestJSON-661730579-project-member] [instance: a842df44-d8a9-4376-b9fc-5ca19a68a4b7] Extending root virtual disk to 1048576 {{(pid=61806) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 819.804045] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-7878414d-31fc-47bb-91f6-c7de73b7cf5f {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 819.814309] env[61806]: DEBUG oslo_vmware.api [None req-61b80e06-de2f-412b-be27-b9318eeb5c25 tempest-ServerAddressesTestJSON-661730579 tempest-ServerAddressesTestJSON-661730579-project-member] Waiting for the task: (returnval){ [ 819.814309] env[61806]: value = "task-1294616" [ 819.814309] env[61806]: _type = "Task" [ 819.814309] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 819.823265] env[61806]: DEBUG oslo_vmware.api [None req-61b80e06-de2f-412b-be27-b9318eeb5c25 tempest-ServerAddressesTestJSON-661730579 tempest-ServerAddressesTestJSON-661730579-project-member] Task: {'id': task-1294616, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 819.876611] env[61806]: DEBUG oslo_vmware.api [None req-5af76806-1cdf-4002-962d-34b8992ee9dc tempest-VolumesAdminNegativeTest-1387177911 tempest-VolumesAdminNegativeTest-1387177911-project-member] Task: {'id': task-1294615, 'name': Rename_Task, 'duration_secs': 0.182699} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 819.876937] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [None req-5af76806-1cdf-4002-962d-34b8992ee9dc tempest-VolumesAdminNegativeTest-1387177911 tempest-VolumesAdminNegativeTest-1387177911-project-member] [instance: 4373b735-31cf-4b53-b655-38555cf212a5] Powering on the VM {{(pid=61806) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 819.878058] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-fb74a186-1d34-446b-b3b3-e396e2e67e03 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 819.885733] env[61806]: DEBUG oslo_vmware.api [None req-5af76806-1cdf-4002-962d-34b8992ee9dc tempest-VolumesAdminNegativeTest-1387177911 tempest-VolumesAdminNegativeTest-1387177911-project-member] Waiting for the task: (returnval){ [ 819.885733] env[61806]: value = "task-1294617" [ 819.885733] env[61806]: _type = "Task" [ 819.885733] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 819.894837] env[61806]: DEBUG oslo_vmware.api [None req-5af76806-1cdf-4002-962d-34b8992ee9dc tempest-VolumesAdminNegativeTest-1387177911 tempest-VolumesAdminNegativeTest-1387177911-project-member] Task: {'id': task-1294617, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 819.908111] env[61806]: DEBUG nova.compute.manager [req-cc3fbdb5-dd60-4bd7-b9be-c5c7d2655f83 req-06d36670-a26b-4e27-befe-b84161edd9dd service nova] [instance: b41531f2-f28c-4d82-9682-0b557bbaa491] Received event network-vif-plugged-513ca4b5-b62c-486a-9860-cf267a15c3e5 {{(pid=61806) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 819.908335] env[61806]: DEBUG oslo_concurrency.lockutils [req-cc3fbdb5-dd60-4bd7-b9be-c5c7d2655f83 req-06d36670-a26b-4e27-befe-b84161edd9dd service nova] Acquiring lock "b41531f2-f28c-4d82-9682-0b557bbaa491-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 819.908557] env[61806]: DEBUG oslo_concurrency.lockutils [req-cc3fbdb5-dd60-4bd7-b9be-c5c7d2655f83 req-06d36670-a26b-4e27-befe-b84161edd9dd service nova] Lock "b41531f2-f28c-4d82-9682-0b557bbaa491-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 819.908729] env[61806]: DEBUG oslo_concurrency.lockutils [req-cc3fbdb5-dd60-4bd7-b9be-c5c7d2655f83 req-06d36670-a26b-4e27-befe-b84161edd9dd service nova] Lock "b41531f2-f28c-4d82-9682-0b557bbaa491-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 819.908901] env[61806]: DEBUG nova.compute.manager [req-cc3fbdb5-dd60-4bd7-b9be-c5c7d2655f83 req-06d36670-a26b-4e27-befe-b84161edd9dd service nova] [instance: b41531f2-f28c-4d82-9682-0b557bbaa491] No waiting events found dispatching network-vif-plugged-513ca4b5-b62c-486a-9860-cf267a15c3e5 {{(pid=61806) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 819.909086] env[61806]: WARNING nova.compute.manager [req-cc3fbdb5-dd60-4bd7-b9be-c5c7d2655f83 req-06d36670-a26b-4e27-befe-b84161edd9dd service nova] [instance: b41531f2-f28c-4d82-9682-0b557bbaa491] Received unexpected event network-vif-plugged-513ca4b5-b62c-486a-9860-cf267a15c3e5 for instance with vm_state building and task_state spawning. [ 820.022737] env[61806]: DEBUG nova.network.neutron [None req-7fa92f02-ad2b-4004-a9a7-cb08f83ec0f1 tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] [instance: 67d1bdd8-cf8f-4fb5-bd3c-42008e4d60e3] Successfully created port: 859fb723-4a1c-4ea8-bd7c-0ed308f3a9e3 {{(pid=61806) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 820.039729] env[61806]: DEBUG oslo_vmware.api [None req-4e219b6d-f08a-4088-99ff-2fcd194df36a tempest-ServerShowV257Test-1159790492 tempest-ServerShowV257Test-1159790492-project-member] Task: {'id': task-1294614, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.730543} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 820.040017] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [None req-4e219b6d-f08a-4088-99ff-2fcd194df36a tempest-ServerShowV257Test-1159790492 tempest-ServerShowV257Test-1159790492-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa.vmdk to [datastore2] 616ec206-9804-469e-ab5c-41aea7f048aa/616ec206-9804-469e-ab5c-41aea7f048aa.vmdk {{(pid=61806) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 820.040246] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-4e219b6d-f08a-4088-99ff-2fcd194df36a tempest-ServerShowV257Test-1159790492 tempest-ServerShowV257Test-1159790492-project-member] [instance: 616ec206-9804-469e-ab5c-41aea7f048aa] Extending root virtual disk to 1048576 {{(pid=61806) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 820.040520] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-47c5c28e-fbe5-46cd-8e4f-9e98d4c26813 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 820.050045] env[61806]: DEBUG oslo_vmware.api [None req-4e219b6d-f08a-4088-99ff-2fcd194df36a tempest-ServerShowV257Test-1159790492 tempest-ServerShowV257Test-1159790492-project-member] Waiting for the task: (returnval){ [ 820.050045] env[61806]: value = "task-1294618" [ 820.050045] env[61806]: _type = "Task" [ 820.050045] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 820.059567] env[61806]: DEBUG oslo_vmware.api [None req-4e219b6d-f08a-4088-99ff-2fcd194df36a tempest-ServerShowV257Test-1159790492 tempest-ServerShowV257Test-1159790492-project-member] Task: {'id': task-1294618, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 820.082591] env[61806]: DEBUG nova.compute.manager [None req-7fa92f02-ad2b-4004-a9a7-cb08f83ec0f1 tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] [instance: 67d1bdd8-cf8f-4fb5-bd3c-42008e4d60e3] Start building block device mappings for instance. {{(pid=61806) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 820.111413] env[61806]: DEBUG oslo_vmware.api [None req-8aa43bc8-f134-4682-8a54-ce9aa25e4ee5 tempest-AttachVolumeNegativeTest-1683033704 tempest-AttachVolumeNegativeTest-1683033704-project-member] Task: {'id': session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]525549a4-8d0f-207c-8c98-c5e236b74d2f, 'name': SearchDatastore_Task, 'duration_secs': 0.020747} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 820.115025] env[61806]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-6f26804c-3690-48fd-b865-0d73ca37242a {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 820.119371] env[61806]: DEBUG oslo_vmware.api [None req-8aa43bc8-f134-4682-8a54-ce9aa25e4ee5 tempest-AttachVolumeNegativeTest-1683033704 tempest-AttachVolumeNegativeTest-1683033704-project-member] Waiting for the task: (returnval){ [ 820.119371] env[61806]: value = "session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]5295ccd1-b306-2880-64b2-37b0ccd0fdee" [ 820.119371] env[61806]: _type = "Task" [ 820.119371] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 820.138264] env[61806]: DEBUG oslo_vmware.api [None req-8aa43bc8-f134-4682-8a54-ce9aa25e4ee5 tempest-AttachVolumeNegativeTest-1683033704 tempest-AttachVolumeNegativeTest-1683033704-project-member] Task: {'id': session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]5295ccd1-b306-2880-64b2-37b0ccd0fdee, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 820.326785] env[61806]: DEBUG oslo_vmware.api [None req-61b80e06-de2f-412b-be27-b9318eeb5c25 tempest-ServerAddressesTestJSON-661730579 tempest-ServerAddressesTestJSON-661730579-project-member] Task: {'id': task-1294616, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.080035} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 820.327068] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-61b80e06-de2f-412b-be27-b9318eeb5c25 tempest-ServerAddressesTestJSON-661730579 tempest-ServerAddressesTestJSON-661730579-project-member] [instance: a842df44-d8a9-4376-b9fc-5ca19a68a4b7] Extended root virtual disk {{(pid=61806) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 820.329076] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-68681405-2227-4b2b-8ad9-565c5335b4a9 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 820.361089] env[61806]: DEBUG nova.virt.vmwareapi.volumeops [None req-61b80e06-de2f-412b-be27-b9318eeb5c25 tempest-ServerAddressesTestJSON-661730579 tempest-ServerAddressesTestJSON-661730579-project-member] [instance: a842df44-d8a9-4376-b9fc-5ca19a68a4b7] Reconfiguring VM instance instance-0000003e to attach disk [datastore1] a842df44-d8a9-4376-b9fc-5ca19a68a4b7/a842df44-d8a9-4376-b9fc-5ca19a68a4b7.vmdk or device None with type sparse {{(pid=61806) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 820.365158] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-9d99b0c9-47e0-4536-89b2-206507b9f4ec {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 820.381063] env[61806]: DEBUG nova.network.neutron [None req-9c52c9e9-816e-4638-b13a-a63a9756140c tempest-SecurityGroupsTestJSON-1213485581 tempest-SecurityGroupsTestJSON-1213485581-project-member] [instance: b41531f2-f28c-4d82-9682-0b557bbaa491] Successfully updated port: 513ca4b5-b62c-486a-9860-cf267a15c3e5 {{(pid=61806) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 820.391608] env[61806]: DEBUG oslo_vmware.api [None req-61b80e06-de2f-412b-be27-b9318eeb5c25 tempest-ServerAddressesTestJSON-661730579 tempest-ServerAddressesTestJSON-661730579-project-member] Waiting for the task: (returnval){ [ 820.391608] env[61806]: value = "task-1294619" [ 820.391608] env[61806]: _type = "Task" [ 820.391608] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 820.407373] env[61806]: DEBUG oslo_vmware.api [None req-5af76806-1cdf-4002-962d-34b8992ee9dc tempest-VolumesAdminNegativeTest-1387177911 tempest-VolumesAdminNegativeTest-1387177911-project-member] Task: {'id': task-1294617, 'name': PowerOnVM_Task, 'duration_secs': 0.492917} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 820.407930] env[61806]: DEBUG oslo_vmware.api [None req-61b80e06-de2f-412b-be27-b9318eeb5c25 tempest-ServerAddressesTestJSON-661730579 tempest-ServerAddressesTestJSON-661730579-project-member] Task: {'id': task-1294619, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 820.408200] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [None req-5af76806-1cdf-4002-962d-34b8992ee9dc tempest-VolumesAdminNegativeTest-1387177911 tempest-VolumesAdminNegativeTest-1387177911-project-member] [instance: 4373b735-31cf-4b53-b655-38555cf212a5] Powered on the VM {{(pid=61806) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 820.408405] env[61806]: INFO nova.compute.manager [None req-5af76806-1cdf-4002-962d-34b8992ee9dc tempest-VolumesAdminNegativeTest-1387177911 tempest-VolumesAdminNegativeTest-1387177911-project-member] [instance: 4373b735-31cf-4b53-b655-38555cf212a5] Took 9.10 seconds to spawn the instance on the hypervisor. [ 820.408592] env[61806]: DEBUG nova.compute.manager [None req-5af76806-1cdf-4002-962d-34b8992ee9dc tempest-VolumesAdminNegativeTest-1387177911 tempest-VolumesAdminNegativeTest-1387177911-project-member] [instance: 4373b735-31cf-4b53-b655-38555cf212a5] Checking state {{(pid=61806) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 820.409401] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0d83741c-d33d-469e-b12b-865e842dad11 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 820.442296] env[61806]: DEBUG nova.compute.manager [req-5353c2ee-5fa8-4404-b115-2e953c55ac8c req-7d4f70e6-4e1e-4dc9-81ff-8e1cc8f5bc30 service nova] [instance: 0758901a-7093-41d3-b0e2-5c519333abdd] Received event network-changed-5ed8d8cc-fa88-438d-80e3-22e0c009373d {{(pid=61806) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 820.442549] env[61806]: DEBUG nova.compute.manager [req-5353c2ee-5fa8-4404-b115-2e953c55ac8c req-7d4f70e6-4e1e-4dc9-81ff-8e1cc8f5bc30 service nova] [instance: 0758901a-7093-41d3-b0e2-5c519333abdd] Refreshing instance network info cache due to event network-changed-5ed8d8cc-fa88-438d-80e3-22e0c009373d. {{(pid=61806) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 820.442785] env[61806]: DEBUG oslo_concurrency.lockutils [req-5353c2ee-5fa8-4404-b115-2e953c55ac8c req-7d4f70e6-4e1e-4dc9-81ff-8e1cc8f5bc30 service nova] Acquiring lock "refresh_cache-0758901a-7093-41d3-b0e2-5c519333abdd" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 820.442957] env[61806]: DEBUG oslo_concurrency.lockutils [req-5353c2ee-5fa8-4404-b115-2e953c55ac8c req-7d4f70e6-4e1e-4dc9-81ff-8e1cc8f5bc30 service nova] Acquired lock "refresh_cache-0758901a-7093-41d3-b0e2-5c519333abdd" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 820.443148] env[61806]: DEBUG nova.network.neutron [req-5353c2ee-5fa8-4404-b115-2e953c55ac8c req-7d4f70e6-4e1e-4dc9-81ff-8e1cc8f5bc30 service nova] [instance: 0758901a-7093-41d3-b0e2-5c519333abdd] Refreshing network info cache for port 5ed8d8cc-fa88-438d-80e3-22e0c009373d {{(pid=61806) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 820.446439] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ebe6d185-1152-406a-9173-9a69304e57ff {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 820.455780] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3132b7fc-ae01-4714-9075-9cc7597282d5 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 820.490520] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e4fddc1e-0387-4e60-8dba-7ca9bf84159e {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 820.499268] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-73fc4be0-5a5c-42e1-97b5-ed9b9f0e14a2 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 820.514034] env[61806]: DEBUG nova.compute.provider_tree [None req-cbb90e6d-7dac-4c30-b860-1f0db7dd091e tempest-MultipleCreateTestJSON-1589015462 tempest-MultipleCreateTestJSON-1589015462-project-member] Inventory has not changed in ProviderTree for provider: a2858be1-fd22-4e08-979e-87ad25293407 {{(pid=61806) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 820.559122] env[61806]: DEBUG oslo_vmware.api [None req-4e219b6d-f08a-4088-99ff-2fcd194df36a tempest-ServerShowV257Test-1159790492 tempest-ServerShowV257Test-1159790492-project-member] Task: {'id': task-1294618, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.077852} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 820.559474] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-4e219b6d-f08a-4088-99ff-2fcd194df36a tempest-ServerShowV257Test-1159790492 tempest-ServerShowV257Test-1159790492-project-member] [instance: 616ec206-9804-469e-ab5c-41aea7f048aa] Extended root virtual disk {{(pid=61806) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 820.560271] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c2e2fc94-9f19-4dd6-be3c-8cd51eda0f60 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 820.580947] env[61806]: DEBUG nova.virt.vmwareapi.volumeops [None req-4e219b6d-f08a-4088-99ff-2fcd194df36a tempest-ServerShowV257Test-1159790492 tempest-ServerShowV257Test-1159790492-project-member] [instance: 616ec206-9804-469e-ab5c-41aea7f048aa] Reconfiguring VM instance instance-0000003b to attach disk [datastore2] 616ec206-9804-469e-ab5c-41aea7f048aa/616ec206-9804-469e-ab5c-41aea7f048aa.vmdk or device None with type sparse {{(pid=61806) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 820.581293] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-c64094a9-39f0-4480-9d79-902badfdd642 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 820.606587] env[61806]: DEBUG oslo_vmware.api [None req-4e219b6d-f08a-4088-99ff-2fcd194df36a tempest-ServerShowV257Test-1159790492 tempest-ServerShowV257Test-1159790492-project-member] Waiting for the task: (returnval){ [ 820.606587] env[61806]: value = "task-1294620" [ 820.606587] env[61806]: _type = "Task" [ 820.606587] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 820.615640] env[61806]: DEBUG oslo_vmware.api [None req-4e219b6d-f08a-4088-99ff-2fcd194df36a tempest-ServerShowV257Test-1159790492 tempest-ServerShowV257Test-1159790492-project-member] Task: {'id': task-1294620, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 820.633101] env[61806]: DEBUG oslo_vmware.api [None req-8aa43bc8-f134-4682-8a54-ce9aa25e4ee5 tempest-AttachVolumeNegativeTest-1683033704 tempest-AttachVolumeNegativeTest-1683033704-project-member] Task: {'id': session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]5295ccd1-b306-2880-64b2-37b0ccd0fdee, 'name': SearchDatastore_Task, 'duration_secs': 0.035804} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 820.633536] env[61806]: DEBUG oslo_concurrency.lockutils [None req-8aa43bc8-f134-4682-8a54-ce9aa25e4ee5 tempest-AttachVolumeNegativeTest-1683033704 tempest-AttachVolumeNegativeTest-1683033704-project-member] Releasing lock "[datastore1] devstack-image-cache_base/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa.vmdk" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 820.633994] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [None req-8aa43bc8-f134-4682-8a54-ce9aa25e4ee5 tempest-AttachVolumeNegativeTest-1683033704 tempest-AttachVolumeNegativeTest-1683033704-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa.vmdk to [datastore1] 43dcce0d-4fdb-4ee1-8309-92e92d2331a9/43dcce0d-4fdb-4ee1-8309-92e92d2331a9.vmdk {{(pid=61806) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 820.634358] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-cb858a8a-9869-41ef-b639-9db3eda650dd {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 820.642120] env[61806]: DEBUG oslo_vmware.api [None req-8aa43bc8-f134-4682-8a54-ce9aa25e4ee5 tempest-AttachVolumeNegativeTest-1683033704 tempest-AttachVolumeNegativeTest-1683033704-project-member] Waiting for the task: (returnval){ [ 820.642120] env[61806]: value = "task-1294621" [ 820.642120] env[61806]: _type = "Task" [ 820.642120] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 820.652331] env[61806]: DEBUG oslo_vmware.api [None req-8aa43bc8-f134-4682-8a54-ce9aa25e4ee5 tempest-AttachVolumeNegativeTest-1683033704 tempest-AttachVolumeNegativeTest-1683033704-project-member] Task: {'id': task-1294621, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 820.838763] env[61806]: DEBUG oslo_concurrency.lockutils [None req-b326e3ef-f103-433e-9385-2b5e433b613a tempest-ServerShowV247Test-1218443059 tempest-ServerShowV247Test-1218443059-project-member] Acquiring lock "4febd093-9f2b-494e-b175-e4693b0e3e0d" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 820.839111] env[61806]: DEBUG oslo_concurrency.lockutils [None req-b326e3ef-f103-433e-9385-2b5e433b613a tempest-ServerShowV247Test-1218443059 tempest-ServerShowV247Test-1218443059-project-member] Lock "4febd093-9f2b-494e-b175-e4693b0e3e0d" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 820.839395] env[61806]: DEBUG oslo_concurrency.lockutils [None req-b326e3ef-f103-433e-9385-2b5e433b613a tempest-ServerShowV247Test-1218443059 tempest-ServerShowV247Test-1218443059-project-member] Acquiring lock "4febd093-9f2b-494e-b175-e4693b0e3e0d-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 820.839620] env[61806]: DEBUG oslo_concurrency.lockutils [None req-b326e3ef-f103-433e-9385-2b5e433b613a tempest-ServerShowV247Test-1218443059 tempest-ServerShowV247Test-1218443059-project-member] Lock "4febd093-9f2b-494e-b175-e4693b0e3e0d-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 820.839846] env[61806]: DEBUG oslo_concurrency.lockutils [None req-b326e3ef-f103-433e-9385-2b5e433b613a tempest-ServerShowV247Test-1218443059 tempest-ServerShowV247Test-1218443059-project-member] Lock "4febd093-9f2b-494e-b175-e4693b0e3e0d-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 820.842316] env[61806]: INFO nova.compute.manager [None req-b326e3ef-f103-433e-9385-2b5e433b613a tempest-ServerShowV247Test-1218443059 tempest-ServerShowV247Test-1218443059-project-member] [instance: 4febd093-9f2b-494e-b175-e4693b0e3e0d] Terminating instance [ 820.845382] env[61806]: DEBUG oslo_concurrency.lockutils [None req-b326e3ef-f103-433e-9385-2b5e433b613a tempest-ServerShowV247Test-1218443059 tempest-ServerShowV247Test-1218443059-project-member] Acquiring lock "refresh_cache-4febd093-9f2b-494e-b175-e4693b0e3e0d" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 820.845586] env[61806]: DEBUG oslo_concurrency.lockutils [None req-b326e3ef-f103-433e-9385-2b5e433b613a tempest-ServerShowV247Test-1218443059 tempest-ServerShowV247Test-1218443059-project-member] Acquired lock "refresh_cache-4febd093-9f2b-494e-b175-e4693b0e3e0d" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 820.845783] env[61806]: DEBUG nova.network.neutron [None req-b326e3ef-f103-433e-9385-2b5e433b613a tempest-ServerShowV247Test-1218443059 tempest-ServerShowV247Test-1218443059-project-member] [instance: 4febd093-9f2b-494e-b175-e4693b0e3e0d] Building network info cache for instance {{(pid=61806) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 820.884224] env[61806]: DEBUG oslo_concurrency.lockutils [None req-9c52c9e9-816e-4638-b13a-a63a9756140c tempest-SecurityGroupsTestJSON-1213485581 tempest-SecurityGroupsTestJSON-1213485581-project-member] Acquiring lock "refresh_cache-b41531f2-f28c-4d82-9682-0b557bbaa491" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 820.884366] env[61806]: DEBUG oslo_concurrency.lockutils [None req-9c52c9e9-816e-4638-b13a-a63a9756140c tempest-SecurityGroupsTestJSON-1213485581 tempest-SecurityGroupsTestJSON-1213485581-project-member] Acquired lock "refresh_cache-b41531f2-f28c-4d82-9682-0b557bbaa491" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 820.884517] env[61806]: DEBUG nova.network.neutron [None req-9c52c9e9-816e-4638-b13a-a63a9756140c tempest-SecurityGroupsTestJSON-1213485581 tempest-SecurityGroupsTestJSON-1213485581-project-member] [instance: b41531f2-f28c-4d82-9682-0b557bbaa491] Building network info cache for instance {{(pid=61806) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 820.903858] env[61806]: DEBUG oslo_vmware.api [None req-61b80e06-de2f-412b-be27-b9318eeb5c25 tempest-ServerAddressesTestJSON-661730579 tempest-ServerAddressesTestJSON-661730579-project-member] Task: {'id': task-1294619, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 820.930727] env[61806]: INFO nova.compute.manager [None req-5af76806-1cdf-4002-962d-34b8992ee9dc tempest-VolumesAdminNegativeTest-1387177911 tempest-VolumesAdminNegativeTest-1387177911-project-member] [instance: 4373b735-31cf-4b53-b655-38555cf212a5] Took 27.22 seconds to build instance. [ 821.017526] env[61806]: DEBUG nova.scheduler.client.report [None req-cbb90e6d-7dac-4c30-b860-1f0db7dd091e tempest-MultipleCreateTestJSON-1589015462 tempest-MultipleCreateTestJSON-1589015462-project-member] Inventory has not changed for provider a2858be1-fd22-4e08-979e-87ad25293407 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 138, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61806) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 821.101332] env[61806]: DEBUG nova.compute.manager [None req-7fa92f02-ad2b-4004-a9a7-cb08f83ec0f1 tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] [instance: 67d1bdd8-cf8f-4fb5-bd3c-42008e4d60e3] Start spawning the instance on the hypervisor. {{(pid=61806) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 821.117815] env[61806]: DEBUG oslo_vmware.api [None req-4e219b6d-f08a-4088-99ff-2fcd194df36a tempest-ServerShowV257Test-1159790492 tempest-ServerShowV257Test-1159790492-project-member] Task: {'id': task-1294620, 'name': ReconfigVM_Task, 'duration_secs': 0.376571} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 821.118126] env[61806]: DEBUG nova.virt.vmwareapi.volumeops [None req-4e219b6d-f08a-4088-99ff-2fcd194df36a tempest-ServerShowV257Test-1159790492 tempest-ServerShowV257Test-1159790492-project-member] [instance: 616ec206-9804-469e-ab5c-41aea7f048aa] Reconfigured VM instance instance-0000003b to attach disk [datastore2] 616ec206-9804-469e-ab5c-41aea7f048aa/616ec206-9804-469e-ab5c-41aea7f048aa.vmdk or device None with type sparse {{(pid=61806) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 821.118757] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-57cc09d8-9e5f-40ea-bbb2-6060b57c9162 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 821.129033] env[61806]: DEBUG nova.virt.hardware [None req-7fa92f02-ad2b-4004-a9a7-cb08f83ec0f1 tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-15T18:05:10Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-15T18:04:52Z,direct_url=,disk_format='vmdk',id=ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='84ba83fce2c349a6988173c5d6fc3b07',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-15T18:04:53Z,virtual_size=,visibility=), allow threads: False {{(pid=61806) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 821.129292] env[61806]: DEBUG nova.virt.hardware [None req-7fa92f02-ad2b-4004-a9a7-cb08f83ec0f1 tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Flavor limits 0:0:0 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 821.129462] env[61806]: DEBUG nova.virt.hardware [None req-7fa92f02-ad2b-4004-a9a7-cb08f83ec0f1 tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Image limits 0:0:0 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 821.129676] env[61806]: DEBUG nova.virt.hardware [None req-7fa92f02-ad2b-4004-a9a7-cb08f83ec0f1 tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Flavor pref 0:0:0 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 821.129848] env[61806]: DEBUG nova.virt.hardware [None req-7fa92f02-ad2b-4004-a9a7-cb08f83ec0f1 tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Image pref 0:0:0 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 821.130026] env[61806]: DEBUG nova.virt.hardware [None req-7fa92f02-ad2b-4004-a9a7-cb08f83ec0f1 tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 821.130249] env[61806]: DEBUG nova.virt.hardware [None req-7fa92f02-ad2b-4004-a9a7-cb08f83ec0f1 tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61806) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 821.130454] env[61806]: DEBUG nova.virt.hardware [None req-7fa92f02-ad2b-4004-a9a7-cb08f83ec0f1 tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61806) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 821.130636] env[61806]: DEBUG nova.virt.hardware [None req-7fa92f02-ad2b-4004-a9a7-cb08f83ec0f1 tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Got 1 possible topologies {{(pid=61806) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 821.130838] env[61806]: DEBUG nova.virt.hardware [None req-7fa92f02-ad2b-4004-a9a7-cb08f83ec0f1 tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61806) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 821.131084] env[61806]: DEBUG nova.virt.hardware [None req-7fa92f02-ad2b-4004-a9a7-cb08f83ec0f1 tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61806) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 821.131925] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c1373aa3-c71b-4ed2-967a-d5c156d88bfe {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 821.136036] env[61806]: DEBUG oslo_vmware.api [None req-4e219b6d-f08a-4088-99ff-2fcd194df36a tempest-ServerShowV257Test-1159790492 tempest-ServerShowV257Test-1159790492-project-member] Waiting for the task: (returnval){ [ 821.136036] env[61806]: value = "task-1294622" [ 821.136036] env[61806]: _type = "Task" [ 821.136036] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 821.149297] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8b0b5403-4023-4104-abd3-483a558e51b3 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 821.159639] env[61806]: DEBUG oslo_vmware.api [None req-4e219b6d-f08a-4088-99ff-2fcd194df36a tempest-ServerShowV257Test-1159790492 tempest-ServerShowV257Test-1159790492-project-member] Task: {'id': task-1294622, 'name': Rename_Task} progress is 10%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 821.165631] env[61806]: DEBUG oslo_vmware.api [None req-8aa43bc8-f134-4682-8a54-ce9aa25e4ee5 tempest-AttachVolumeNegativeTest-1683033704 tempest-AttachVolumeNegativeTest-1683033704-project-member] Task: {'id': task-1294621, 'name': CopyVirtualDisk_Task} progress is 4%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 821.197142] env[61806]: DEBUG nova.network.neutron [req-5353c2ee-5fa8-4404-b115-2e953c55ac8c req-7d4f70e6-4e1e-4dc9-81ff-8e1cc8f5bc30 service nova] [instance: 0758901a-7093-41d3-b0e2-5c519333abdd] Updated VIF entry in instance network info cache for port 5ed8d8cc-fa88-438d-80e3-22e0c009373d. {{(pid=61806) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 821.197601] env[61806]: DEBUG nova.network.neutron [req-5353c2ee-5fa8-4404-b115-2e953c55ac8c req-7d4f70e6-4e1e-4dc9-81ff-8e1cc8f5bc30 service nova] [instance: 0758901a-7093-41d3-b0e2-5c519333abdd] Updating instance_info_cache with network_info: [{"id": "5ed8d8cc-fa88-438d-80e3-22e0c009373d", "address": "fa:16:3e:4b:68:4e", "network": {"id": "b087ee9c-cf27-42df-9c9a-a3652a50fbca", "bridge": "br-int", "label": "tempest-ServerActionsV293TestJSON-1492533501-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.177", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7e8bbf4274b240619d03f7903cc57ff5", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "92e4d027-e755-417b-8eea-9a8f24b85140", "external-id": "nsx-vlan-transportzone-756", "segmentation_id": 756, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap5ed8d8cc-fa", "ovs_interfaceid": "5ed8d8cc-fa88-438d-80e3-22e0c009373d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61806) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 821.364589] env[61806]: DEBUG nova.network.neutron [None req-b326e3ef-f103-433e-9385-2b5e433b613a tempest-ServerShowV247Test-1218443059 tempest-ServerShowV247Test-1218443059-project-member] [instance: 4febd093-9f2b-494e-b175-e4693b0e3e0d] Instance cache missing network info. {{(pid=61806) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 821.403780] env[61806]: DEBUG oslo_vmware.api [None req-61b80e06-de2f-412b-be27-b9318eeb5c25 tempest-ServerAddressesTestJSON-661730579 tempest-ServerAddressesTestJSON-661730579-project-member] Task: {'id': task-1294619, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 821.421887] env[61806]: DEBUG nova.network.neutron [None req-b326e3ef-f103-433e-9385-2b5e433b613a tempest-ServerShowV247Test-1218443059 tempest-ServerShowV247Test-1218443059-project-member] [instance: 4febd093-9f2b-494e-b175-e4693b0e3e0d] Updating instance_info_cache with network_info: [] {{(pid=61806) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 821.432282] env[61806]: DEBUG oslo_concurrency.lockutils [None req-5af76806-1cdf-4002-962d-34b8992ee9dc tempest-VolumesAdminNegativeTest-1387177911 tempest-VolumesAdminNegativeTest-1387177911-project-member] Lock "4373b735-31cf-4b53-b655-38555cf212a5" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 101.590s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 821.435151] env[61806]: DEBUG nova.network.neutron [None req-9c52c9e9-816e-4638-b13a-a63a9756140c tempest-SecurityGroupsTestJSON-1213485581 tempest-SecurityGroupsTestJSON-1213485581-project-member] [instance: b41531f2-f28c-4d82-9682-0b557bbaa491] Instance cache missing network info. {{(pid=61806) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 821.524036] env[61806]: DEBUG oslo_concurrency.lockutils [None req-cbb90e6d-7dac-4c30-b860-1f0db7dd091e tempest-MultipleCreateTestJSON-1589015462 tempest-MultipleCreateTestJSON-1589015462-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.446s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 821.524606] env[61806]: DEBUG nova.compute.manager [None req-cbb90e6d-7dac-4c30-b860-1f0db7dd091e tempest-MultipleCreateTestJSON-1589015462 tempest-MultipleCreateTestJSON-1589015462-project-member] [instance: 25218e83-c1ab-4b97-ade1-7c32b6f99305] Start building networks asynchronously for instance. {{(pid=61806) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 821.527545] env[61806]: DEBUG oslo_concurrency.lockutils [None req-fbc15506-7910-4d6f-bfbc-c3a1ce19fdcc tempest-ServerMetadataTestJSON-316746592 tempest-ServerMetadataTestJSON-316746592-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 13.075s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 821.527818] env[61806]: DEBUG nova.objects.instance [None req-fbc15506-7910-4d6f-bfbc-c3a1ce19fdcc tempest-ServerMetadataTestJSON-316746592 tempest-ServerMetadataTestJSON-316746592-project-member] Lazy-loading 'resources' on Instance uuid 4184e263-6f56-4bc1-99b2-a2c460531516 {{(pid=61806) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 821.537962] env[61806]: DEBUG nova.compute.manager [None req-8bfbcce0-8d5f-47b3-8909-c0045b6193b0 tempest-ServerDiagnosticsV248Test-1367482897 tempest-ServerDiagnosticsV248Test-1367482897-project-admin] [instance: a934d02d-26aa-4900-b473-a58489e5629e] Checking state {{(pid=61806) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 821.543615] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-61475b4a-832d-4c39-8ebf-d3e1904fcb9e {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 821.551539] env[61806]: INFO nova.compute.manager [None req-8bfbcce0-8d5f-47b3-8909-c0045b6193b0 tempest-ServerDiagnosticsV248Test-1367482897 tempest-ServerDiagnosticsV248Test-1367482897-project-admin] [instance: a934d02d-26aa-4900-b473-a58489e5629e] Retrieving diagnostics [ 821.552354] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cf013708-0c65-4436-a4ad-22b3b794aa67 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 821.651926] env[61806]: DEBUG oslo_vmware.api [None req-4e219b6d-f08a-4088-99ff-2fcd194df36a tempest-ServerShowV257Test-1159790492 tempest-ServerShowV257Test-1159790492-project-member] Task: {'id': task-1294622, 'name': Rename_Task, 'duration_secs': 0.149817} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 821.652840] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [None req-4e219b6d-f08a-4088-99ff-2fcd194df36a tempest-ServerShowV257Test-1159790492 tempest-ServerShowV257Test-1159790492-project-member] [instance: 616ec206-9804-469e-ab5c-41aea7f048aa] Powering on the VM {{(pid=61806) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 821.653147] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-f8299e79-fc68-4f06-867c-1c7dcab60d5e {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 821.658974] env[61806]: DEBUG oslo_vmware.api [None req-8aa43bc8-f134-4682-8a54-ce9aa25e4ee5 tempest-AttachVolumeNegativeTest-1683033704 tempest-AttachVolumeNegativeTest-1683033704-project-member] Task: {'id': task-1294621, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.867397} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 821.659280] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [None req-8aa43bc8-f134-4682-8a54-ce9aa25e4ee5 tempest-AttachVolumeNegativeTest-1683033704 tempest-AttachVolumeNegativeTest-1683033704-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa.vmdk to [datastore1] 43dcce0d-4fdb-4ee1-8309-92e92d2331a9/43dcce0d-4fdb-4ee1-8309-92e92d2331a9.vmdk {{(pid=61806) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 821.659505] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-8aa43bc8-f134-4682-8a54-ce9aa25e4ee5 tempest-AttachVolumeNegativeTest-1683033704 tempest-AttachVolumeNegativeTest-1683033704-project-member] [instance: 43dcce0d-4fdb-4ee1-8309-92e92d2331a9] Extending root virtual disk to 1048576 {{(pid=61806) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 821.659782] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-83c12339-6e49-4baf-b245-880038e1105e {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 821.664011] env[61806]: DEBUG oslo_vmware.api [None req-4e219b6d-f08a-4088-99ff-2fcd194df36a tempest-ServerShowV257Test-1159790492 tempest-ServerShowV257Test-1159790492-project-member] Waiting for the task: (returnval){ [ 821.664011] env[61806]: value = "task-1294623" [ 821.664011] env[61806]: _type = "Task" [ 821.664011] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 821.668907] env[61806]: DEBUG oslo_vmware.api [None req-8aa43bc8-f134-4682-8a54-ce9aa25e4ee5 tempest-AttachVolumeNegativeTest-1683033704 tempest-AttachVolumeNegativeTest-1683033704-project-member] Waiting for the task: (returnval){ [ 821.668907] env[61806]: value = "task-1294624" [ 821.668907] env[61806]: _type = "Task" [ 821.668907] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 821.676788] env[61806]: DEBUG oslo_vmware.api [None req-4e219b6d-f08a-4088-99ff-2fcd194df36a tempest-ServerShowV257Test-1159790492 tempest-ServerShowV257Test-1159790492-project-member] Task: {'id': task-1294623, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 821.683070] env[61806]: DEBUG oslo_vmware.api [None req-8aa43bc8-f134-4682-8a54-ce9aa25e4ee5 tempest-AttachVolumeNegativeTest-1683033704 tempest-AttachVolumeNegativeTest-1683033704-project-member] Task: {'id': task-1294624, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 821.688924] env[61806]: DEBUG nova.network.neutron [None req-9c52c9e9-816e-4638-b13a-a63a9756140c tempest-SecurityGroupsTestJSON-1213485581 tempest-SecurityGroupsTestJSON-1213485581-project-member] [instance: b41531f2-f28c-4d82-9682-0b557bbaa491] Updating instance_info_cache with network_info: [{"id": "513ca4b5-b62c-486a-9860-cf267a15c3e5", "address": "fa:16:3e:40:52:86", "network": {"id": "dfb52c93-e553-40d9-acb5-768cc6b67c89", "bridge": "br-int", "label": "tempest-SecurityGroupsTestJSON-1006891343-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "cfce93add38d40d79f294017f52e13d5", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6b6942d7-d4ab-4b2a-8d0f-76bf2a2478ad", "external-id": "nsx-vlan-transportzone-871", "segmentation_id": 871, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap513ca4b5-b6", "ovs_interfaceid": "513ca4b5-b62c-486a-9860-cf267a15c3e5", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61806) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 821.701351] env[61806]: DEBUG oslo_concurrency.lockutils [req-5353c2ee-5fa8-4404-b115-2e953c55ac8c req-7d4f70e6-4e1e-4dc9-81ff-8e1cc8f5bc30 service nova] Releasing lock "refresh_cache-0758901a-7093-41d3-b0e2-5c519333abdd" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 821.701712] env[61806]: DEBUG nova.compute.manager [req-5353c2ee-5fa8-4404-b115-2e953c55ac8c req-7d4f70e6-4e1e-4dc9-81ff-8e1cc8f5bc30 service nova] [instance: b41531f2-f28c-4d82-9682-0b557bbaa491] Received event network-changed-513ca4b5-b62c-486a-9860-cf267a15c3e5 {{(pid=61806) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 821.701951] env[61806]: DEBUG nova.compute.manager [req-5353c2ee-5fa8-4404-b115-2e953c55ac8c req-7d4f70e6-4e1e-4dc9-81ff-8e1cc8f5bc30 service nova] [instance: b41531f2-f28c-4d82-9682-0b557bbaa491] Refreshing instance network info cache due to event network-changed-513ca4b5-b62c-486a-9860-cf267a15c3e5. {{(pid=61806) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 821.702239] env[61806]: DEBUG oslo_concurrency.lockutils [req-5353c2ee-5fa8-4404-b115-2e953c55ac8c req-7d4f70e6-4e1e-4dc9-81ff-8e1cc8f5bc30 service nova] Acquiring lock "refresh_cache-b41531f2-f28c-4d82-9682-0b557bbaa491" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 821.904318] env[61806]: DEBUG oslo_vmware.api [None req-61b80e06-de2f-412b-be27-b9318eeb5c25 tempest-ServerAddressesTestJSON-661730579 tempest-ServerAddressesTestJSON-661730579-project-member] Task: {'id': task-1294619, 'name': ReconfigVM_Task, 'duration_secs': 1.15358} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 821.904950] env[61806]: DEBUG nova.virt.vmwareapi.volumeops [None req-61b80e06-de2f-412b-be27-b9318eeb5c25 tempest-ServerAddressesTestJSON-661730579 tempest-ServerAddressesTestJSON-661730579-project-member] [instance: a842df44-d8a9-4376-b9fc-5ca19a68a4b7] Reconfigured VM instance instance-0000003e to attach disk [datastore1] a842df44-d8a9-4376-b9fc-5ca19a68a4b7/a842df44-d8a9-4376-b9fc-5ca19a68a4b7.vmdk or device None with type sparse {{(pid=61806) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 821.905404] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-78407e54-1ba4-451d-af63-5209ecfcf80f {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 821.912666] env[61806]: DEBUG oslo_vmware.api [None req-61b80e06-de2f-412b-be27-b9318eeb5c25 tempest-ServerAddressesTestJSON-661730579 tempest-ServerAddressesTestJSON-661730579-project-member] Waiting for the task: (returnval){ [ 821.912666] env[61806]: value = "task-1294625" [ 821.912666] env[61806]: _type = "Task" [ 821.912666] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 821.921939] env[61806]: DEBUG oslo_vmware.api [None req-61b80e06-de2f-412b-be27-b9318eeb5c25 tempest-ServerAddressesTestJSON-661730579 tempest-ServerAddressesTestJSON-661730579-project-member] Task: {'id': task-1294625, 'name': Rename_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 821.924466] env[61806]: DEBUG oslo_concurrency.lockutils [None req-b326e3ef-f103-433e-9385-2b5e433b613a tempest-ServerShowV247Test-1218443059 tempest-ServerShowV247Test-1218443059-project-member] Releasing lock "refresh_cache-4febd093-9f2b-494e-b175-e4693b0e3e0d" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 821.924925] env[61806]: DEBUG nova.compute.manager [None req-b326e3ef-f103-433e-9385-2b5e433b613a tempest-ServerShowV247Test-1218443059 tempest-ServerShowV247Test-1218443059-project-member] [instance: 4febd093-9f2b-494e-b175-e4693b0e3e0d] Start destroying the instance on the hypervisor. {{(pid=61806) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 821.925064] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-b326e3ef-f103-433e-9385-2b5e433b613a tempest-ServerShowV247Test-1218443059 tempest-ServerShowV247Test-1218443059-project-member] [instance: 4febd093-9f2b-494e-b175-e4693b0e3e0d] Destroying instance {{(pid=61806) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 821.925814] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7e2ddbee-4e81-49e1-9287-26076263b84b {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 821.932868] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [None req-b326e3ef-f103-433e-9385-2b5e433b613a tempest-ServerShowV247Test-1218443059 tempest-ServerShowV247Test-1218443059-project-member] [instance: 4febd093-9f2b-494e-b175-e4693b0e3e0d] Powering off the VM {{(pid=61806) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 821.933182] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-8286934a-b91b-4df0-80c3-ee60073890a4 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 821.938452] env[61806]: DEBUG nova.compute.manager [None req-334c6338-dc9d-4c32-ad77-bd3fe007216c tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] [instance: be19fc44-ecbe-489c-9b6e-be8957bb1dd0] Starting instance... {{(pid=61806) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 821.942832] env[61806]: DEBUG oslo_vmware.api [None req-b326e3ef-f103-433e-9385-2b5e433b613a tempest-ServerShowV247Test-1218443059 tempest-ServerShowV247Test-1218443059-project-member] Waiting for the task: (returnval){ [ 821.942832] env[61806]: value = "task-1294626" [ 821.942832] env[61806]: _type = "Task" [ 821.942832] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 821.953904] env[61806]: DEBUG oslo_vmware.api [None req-b326e3ef-f103-433e-9385-2b5e433b613a tempest-ServerShowV247Test-1218443059 tempest-ServerShowV247Test-1218443059-project-member] Task: {'id': task-1294626, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 822.031605] env[61806]: DEBUG nova.compute.utils [None req-cbb90e6d-7dac-4c30-b860-1f0db7dd091e tempest-MultipleCreateTestJSON-1589015462 tempest-MultipleCreateTestJSON-1589015462-project-member] Using /dev/sd instead of None {{(pid=61806) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 822.033538] env[61806]: DEBUG nova.compute.manager [None req-cbb90e6d-7dac-4c30-b860-1f0db7dd091e tempest-MultipleCreateTestJSON-1589015462 tempest-MultipleCreateTestJSON-1589015462-project-member] [instance: 25218e83-c1ab-4b97-ade1-7c32b6f99305] Allocating IP information in the background. {{(pid=61806) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 822.033767] env[61806]: DEBUG nova.network.neutron [None req-cbb90e6d-7dac-4c30-b860-1f0db7dd091e tempest-MultipleCreateTestJSON-1589015462 tempest-MultipleCreateTestJSON-1589015462-project-member] [instance: 25218e83-c1ab-4b97-ade1-7c32b6f99305] allocate_for_instance() {{(pid=61806) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 822.061202] env[61806]: DEBUG nova.compute.manager [req-c2b98440-938e-4659-af40-99d29e26fa71 req-36ff828d-bf5c-4dcc-87fd-6d12a6565247 service nova] [instance: 67d1bdd8-cf8f-4fb5-bd3c-42008e4d60e3] Received event network-vif-plugged-859fb723-4a1c-4ea8-bd7c-0ed308f3a9e3 {{(pid=61806) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 822.061371] env[61806]: DEBUG oslo_concurrency.lockutils [req-c2b98440-938e-4659-af40-99d29e26fa71 req-36ff828d-bf5c-4dcc-87fd-6d12a6565247 service nova] Acquiring lock "67d1bdd8-cf8f-4fb5-bd3c-42008e4d60e3-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 822.061591] env[61806]: DEBUG oslo_concurrency.lockutils [req-c2b98440-938e-4659-af40-99d29e26fa71 req-36ff828d-bf5c-4dcc-87fd-6d12a6565247 service nova] Lock "67d1bdd8-cf8f-4fb5-bd3c-42008e4d60e3-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 822.061846] env[61806]: DEBUG oslo_concurrency.lockutils [req-c2b98440-938e-4659-af40-99d29e26fa71 req-36ff828d-bf5c-4dcc-87fd-6d12a6565247 service nova] Lock "67d1bdd8-cf8f-4fb5-bd3c-42008e4d60e3-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 822.061930] env[61806]: DEBUG nova.compute.manager [req-c2b98440-938e-4659-af40-99d29e26fa71 req-36ff828d-bf5c-4dcc-87fd-6d12a6565247 service nova] [instance: 67d1bdd8-cf8f-4fb5-bd3c-42008e4d60e3] No waiting events found dispatching network-vif-plugged-859fb723-4a1c-4ea8-bd7c-0ed308f3a9e3 {{(pid=61806) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 822.062137] env[61806]: WARNING nova.compute.manager [req-c2b98440-938e-4659-af40-99d29e26fa71 req-36ff828d-bf5c-4dcc-87fd-6d12a6565247 service nova] [instance: 67d1bdd8-cf8f-4fb5-bd3c-42008e4d60e3] Received unexpected event network-vif-plugged-859fb723-4a1c-4ea8-bd7c-0ed308f3a9e3 for instance with vm_state building and task_state spawning. [ 822.093708] env[61806]: DEBUG nova.policy [None req-cbb90e6d-7dac-4c30-b860-1f0db7dd091e tempest-MultipleCreateTestJSON-1589015462 tempest-MultipleCreateTestJSON-1589015462-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'b5a92ad3d119419085f56d7dea0975f5', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'dc40805c39f34d07be9d07861ad5841d', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61806) authorize /opt/stack/nova/nova/policy.py:201}} [ 822.184308] env[61806]: DEBUG oslo_vmware.api [None req-4e219b6d-f08a-4088-99ff-2fcd194df36a tempest-ServerShowV257Test-1159790492 tempest-ServerShowV257Test-1159790492-project-member] Task: {'id': task-1294623, 'name': PowerOnVM_Task, 'duration_secs': 0.491649} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 822.185189] env[61806]: DEBUG nova.network.neutron [None req-7fa92f02-ad2b-4004-a9a7-cb08f83ec0f1 tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] [instance: 67d1bdd8-cf8f-4fb5-bd3c-42008e4d60e3] Successfully updated port: 859fb723-4a1c-4ea8-bd7c-0ed308f3a9e3 {{(pid=61806) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 822.193607] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [None req-4e219b6d-f08a-4088-99ff-2fcd194df36a tempest-ServerShowV257Test-1159790492 tempest-ServerShowV257Test-1159790492-project-member] [instance: 616ec206-9804-469e-ab5c-41aea7f048aa] Powered on the VM {{(pid=61806) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 822.194145] env[61806]: DEBUG nova.compute.manager [None req-4e219b6d-f08a-4088-99ff-2fcd194df36a tempest-ServerShowV257Test-1159790492 tempest-ServerShowV257Test-1159790492-project-member] [instance: 616ec206-9804-469e-ab5c-41aea7f048aa] Checking state {{(pid=61806) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 822.197345] env[61806]: DEBUG oslo_concurrency.lockutils [None req-9c52c9e9-816e-4638-b13a-a63a9756140c tempest-SecurityGroupsTestJSON-1213485581 tempest-SecurityGroupsTestJSON-1213485581-project-member] Releasing lock "refresh_cache-b41531f2-f28c-4d82-9682-0b557bbaa491" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 822.197802] env[61806]: DEBUG nova.compute.manager [None req-9c52c9e9-816e-4638-b13a-a63a9756140c tempest-SecurityGroupsTestJSON-1213485581 tempest-SecurityGroupsTestJSON-1213485581-project-member] [instance: b41531f2-f28c-4d82-9682-0b557bbaa491] Instance network_info: |[{"id": "513ca4b5-b62c-486a-9860-cf267a15c3e5", "address": "fa:16:3e:40:52:86", "network": {"id": "dfb52c93-e553-40d9-acb5-768cc6b67c89", "bridge": "br-int", "label": "tempest-SecurityGroupsTestJSON-1006891343-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "cfce93add38d40d79f294017f52e13d5", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6b6942d7-d4ab-4b2a-8d0f-76bf2a2478ad", "external-id": "nsx-vlan-transportzone-871", "segmentation_id": 871, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap513ca4b5-b6", "ovs_interfaceid": "513ca4b5-b62c-486a-9860-cf267a15c3e5", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61806) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 822.203101] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0a4b25e6-6b23-46b2-b890-cc069c704757 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 822.205861] env[61806]: DEBUG oslo_concurrency.lockutils [req-5353c2ee-5fa8-4404-b115-2e953c55ac8c req-7d4f70e6-4e1e-4dc9-81ff-8e1cc8f5bc30 service nova] Acquired lock "refresh_cache-b41531f2-f28c-4d82-9682-0b557bbaa491" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 822.206065] env[61806]: DEBUG nova.network.neutron [req-5353c2ee-5fa8-4404-b115-2e953c55ac8c req-7d4f70e6-4e1e-4dc9-81ff-8e1cc8f5bc30 service nova] [instance: b41531f2-f28c-4d82-9682-0b557bbaa491] Refreshing network info cache for port 513ca4b5-b62c-486a-9860-cf267a15c3e5 {{(pid=61806) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 822.207790] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-9c52c9e9-816e-4638-b13a-a63a9756140c tempest-SecurityGroupsTestJSON-1213485581 tempest-SecurityGroupsTestJSON-1213485581-project-member] [instance: b41531f2-f28c-4d82-9682-0b557bbaa491] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:40:52:86', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '6b6942d7-d4ab-4b2a-8d0f-76bf2a2478ad', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '513ca4b5-b62c-486a-9860-cf267a15c3e5', 'vif_model': 'vmxnet3'}] {{(pid=61806) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 822.215187] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [None req-9c52c9e9-816e-4638-b13a-a63a9756140c tempest-SecurityGroupsTestJSON-1213485581 tempest-SecurityGroupsTestJSON-1213485581-project-member] Creating folder: Project (cfce93add38d40d79f294017f52e13d5). Parent ref: group-v277609. {{(pid=61806) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 822.216072] env[61806]: DEBUG oslo_vmware.api [None req-8aa43bc8-f134-4682-8a54-ce9aa25e4ee5 tempest-AttachVolumeNegativeTest-1683033704 tempest-AttachVolumeNegativeTest-1683033704-project-member] Task: {'id': task-1294624, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.076399} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 822.218977] env[61806]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-4acbd1cd-9892-429c-bc9a-dd104cde4e90 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 822.222189] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-8aa43bc8-f134-4682-8a54-ce9aa25e4ee5 tempest-AttachVolumeNegativeTest-1683033704 tempest-AttachVolumeNegativeTest-1683033704-project-member] [instance: 43dcce0d-4fdb-4ee1-8309-92e92d2331a9] Extended root virtual disk {{(pid=61806) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 822.223083] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a47065c9-400e-43cd-b271-10a79c462155 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 822.252627] env[61806]: DEBUG nova.virt.vmwareapi.volumeops [None req-8aa43bc8-f134-4682-8a54-ce9aa25e4ee5 tempest-AttachVolumeNegativeTest-1683033704 tempest-AttachVolumeNegativeTest-1683033704-project-member] [instance: 43dcce0d-4fdb-4ee1-8309-92e92d2331a9] Reconfiguring VM instance instance-0000003f to attach disk [datastore1] 43dcce0d-4fdb-4ee1-8309-92e92d2331a9/43dcce0d-4fdb-4ee1-8309-92e92d2331a9.vmdk or device None with type sparse {{(pid=61806) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 822.257310] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-ae6b3109-26fd-4a34-9603-a66c817a495b {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 822.272537] env[61806]: INFO nova.virt.vmwareapi.vm_util [None req-9c52c9e9-816e-4638-b13a-a63a9756140c tempest-SecurityGroupsTestJSON-1213485581 tempest-SecurityGroupsTestJSON-1213485581-project-member] Created folder: Project (cfce93add38d40d79f294017f52e13d5) in parent group-v277609. [ 822.273401] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [None req-9c52c9e9-816e-4638-b13a-a63a9756140c tempest-SecurityGroupsTestJSON-1213485581 tempest-SecurityGroupsTestJSON-1213485581-project-member] Creating folder: Instances. Parent ref: group-v277661. {{(pid=61806) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 822.274043] env[61806]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-5cf29ee0-efbc-4d6b-b7a5-abf78d53fda9 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 822.281478] env[61806]: DEBUG oslo_vmware.api [None req-8aa43bc8-f134-4682-8a54-ce9aa25e4ee5 tempest-AttachVolumeNegativeTest-1683033704 tempest-AttachVolumeNegativeTest-1683033704-project-member] Waiting for the task: (returnval){ [ 822.281478] env[61806]: value = "task-1294629" [ 822.281478] env[61806]: _type = "Task" [ 822.281478] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 822.286634] env[61806]: INFO nova.virt.vmwareapi.vm_util [None req-9c52c9e9-816e-4638-b13a-a63a9756140c tempest-SecurityGroupsTestJSON-1213485581 tempest-SecurityGroupsTestJSON-1213485581-project-member] Created folder: Instances in parent group-v277661. [ 822.286912] env[61806]: DEBUG oslo.service.loopingcall [None req-9c52c9e9-816e-4638-b13a-a63a9756140c tempest-SecurityGroupsTestJSON-1213485581 tempest-SecurityGroupsTestJSON-1213485581-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61806) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 822.287548] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: b41531f2-f28c-4d82-9682-0b557bbaa491] Creating VM on the ESX host {{(pid=61806) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 822.287807] env[61806]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-07385617-b50e-464e-b2e5-7cc29c6c7c5e {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 822.309337] env[61806]: DEBUG oslo_vmware.api [None req-8aa43bc8-f134-4682-8a54-ce9aa25e4ee5 tempest-AttachVolumeNegativeTest-1683033704 tempest-AttachVolumeNegativeTest-1683033704-project-member] Task: {'id': task-1294629, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 822.318574] env[61806]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 822.318574] env[61806]: value = "task-1294630" [ 822.318574] env[61806]: _type = "Task" [ 822.318574] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 822.329829] env[61806]: DEBUG oslo_vmware.api [-] Task: {'id': task-1294630, 'name': CreateVM_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 822.348023] env[61806]: DEBUG nova.compute.manager [req-e2a64c0c-3895-4c1e-b206-30922535a8df req-e021b049-0656-42f8-9a51-3eca4aeb4d7a service nova] [instance: 67d1bdd8-cf8f-4fb5-bd3c-42008e4d60e3] Received event network-changed-859fb723-4a1c-4ea8-bd7c-0ed308f3a9e3 {{(pid=61806) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 822.348023] env[61806]: DEBUG nova.compute.manager [req-e2a64c0c-3895-4c1e-b206-30922535a8df req-e021b049-0656-42f8-9a51-3eca4aeb4d7a service nova] [instance: 67d1bdd8-cf8f-4fb5-bd3c-42008e4d60e3] Refreshing instance network info cache due to event network-changed-859fb723-4a1c-4ea8-bd7c-0ed308f3a9e3. {{(pid=61806) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 822.348023] env[61806]: DEBUG oslo_concurrency.lockutils [req-e2a64c0c-3895-4c1e-b206-30922535a8df req-e021b049-0656-42f8-9a51-3eca4aeb4d7a service nova] Acquiring lock "refresh_cache-67d1bdd8-cf8f-4fb5-bd3c-42008e4d60e3" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 822.348023] env[61806]: DEBUG oslo_concurrency.lockutils [req-e2a64c0c-3895-4c1e-b206-30922535a8df req-e021b049-0656-42f8-9a51-3eca4aeb4d7a service nova] Acquired lock "refresh_cache-67d1bdd8-cf8f-4fb5-bd3c-42008e4d60e3" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 822.348023] env[61806]: DEBUG nova.network.neutron [req-e2a64c0c-3895-4c1e-b206-30922535a8df req-e021b049-0656-42f8-9a51-3eca4aeb4d7a service nova] [instance: 67d1bdd8-cf8f-4fb5-bd3c-42008e4d60e3] Refreshing network info cache for port 859fb723-4a1c-4ea8-bd7c-0ed308f3a9e3 {{(pid=61806) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 822.426973] env[61806]: DEBUG oslo_vmware.api [None req-61b80e06-de2f-412b-be27-b9318eeb5c25 tempest-ServerAddressesTestJSON-661730579 tempest-ServerAddressesTestJSON-661730579-project-member] Task: {'id': task-1294625, 'name': Rename_Task, 'duration_secs': 0.169867} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 822.427321] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [None req-61b80e06-de2f-412b-be27-b9318eeb5c25 tempest-ServerAddressesTestJSON-661730579 tempest-ServerAddressesTestJSON-661730579-project-member] [instance: a842df44-d8a9-4376-b9fc-5ca19a68a4b7] Powering on the VM {{(pid=61806) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 822.427584] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-97290c44-77ac-4631-8933-3342179635ed {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 822.430952] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a879b317-502c-4c9e-9637-be5e5b9f9124 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 822.435292] env[61806]: DEBUG oslo_vmware.api [None req-61b80e06-de2f-412b-be27-b9318eeb5c25 tempest-ServerAddressesTestJSON-661730579 tempest-ServerAddressesTestJSON-661730579-project-member] Waiting for the task: (returnval){ [ 822.435292] env[61806]: value = "task-1294631" [ 822.435292] env[61806]: _type = "Task" [ 822.435292] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 822.441693] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6e834dc0-b4f0-41b9-bcee-9a897857c6a9 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 822.449733] env[61806]: DEBUG oslo_vmware.api [None req-61b80e06-de2f-412b-be27-b9318eeb5c25 tempest-ServerAddressesTestJSON-661730579 tempest-ServerAddressesTestJSON-661730579-project-member] Task: {'id': task-1294631, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 822.467802] env[61806]: DEBUG oslo_vmware.api [None req-b326e3ef-f103-433e-9385-2b5e433b613a tempest-ServerShowV247Test-1218443059 tempest-ServerShowV247Test-1218443059-project-member] Task: {'id': task-1294626, 'name': PowerOffVM_Task, 'duration_secs': 0.129969} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 822.494688] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [None req-b326e3ef-f103-433e-9385-2b5e433b613a tempest-ServerShowV247Test-1218443059 tempest-ServerShowV247Test-1218443059-project-member] [instance: 4febd093-9f2b-494e-b175-e4693b0e3e0d] Powered off the VM {{(pid=61806) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 822.494865] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-b326e3ef-f103-433e-9385-2b5e433b613a tempest-ServerShowV247Test-1218443059 tempest-ServerShowV247Test-1218443059-project-member] [instance: 4febd093-9f2b-494e-b175-e4693b0e3e0d] Unregistering the VM {{(pid=61806) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 822.496706] env[61806]: DEBUG nova.network.neutron [None req-cbb90e6d-7dac-4c30-b860-1f0db7dd091e tempest-MultipleCreateTestJSON-1589015462 tempest-MultipleCreateTestJSON-1589015462-project-member] [instance: 25218e83-c1ab-4b97-ade1-7c32b6f99305] Successfully created port: c8809e29-ea3c-4ea0-bee1-0731daee293d {{(pid=61806) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 822.499255] env[61806]: DEBUG oslo_concurrency.lockutils [None req-334c6338-dc9d-4c32-ad77-bd3fe007216c tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 822.499517] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-cfc2f08d-3f2e-4645-89d8-221e255581b2 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 822.501651] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ce027b00-ff8c-4a82-9f32-449cad1a5bbe {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 822.511491] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-82281a8e-fdc8-4bf7-936d-7e42137c1c47 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 822.526668] env[61806]: DEBUG nova.compute.provider_tree [None req-fbc15506-7910-4d6f-bfbc-c3a1ce19fdcc tempest-ServerMetadataTestJSON-316746592 tempest-ServerMetadataTestJSON-316746592-project-member] Inventory has not changed in ProviderTree for provider: a2858be1-fd22-4e08-979e-87ad25293407 {{(pid=61806) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 822.533750] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-b326e3ef-f103-433e-9385-2b5e433b613a tempest-ServerShowV247Test-1218443059 tempest-ServerShowV247Test-1218443059-project-member] [instance: 4febd093-9f2b-494e-b175-e4693b0e3e0d] Unregistered the VM {{(pid=61806) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 822.534102] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-b326e3ef-f103-433e-9385-2b5e433b613a tempest-ServerShowV247Test-1218443059 tempest-ServerShowV247Test-1218443059-project-member] [instance: 4febd093-9f2b-494e-b175-e4693b0e3e0d] Deleting contents of the VM from datastore datastore1 {{(pid=61806) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 822.534397] env[61806]: DEBUG nova.virt.vmwareapi.ds_util [None req-b326e3ef-f103-433e-9385-2b5e433b613a tempest-ServerShowV247Test-1218443059 tempest-ServerShowV247Test-1218443059-project-member] Deleting the datastore file [datastore1] 4febd093-9f2b-494e-b175-e4693b0e3e0d {{(pid=61806) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 822.534732] env[61806]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-3368facf-6cc7-4ceb-b4ef-b63b9036c8a7 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 822.539674] env[61806]: DEBUG nova.compute.manager [None req-cbb90e6d-7dac-4c30-b860-1f0db7dd091e tempest-MultipleCreateTestJSON-1589015462 tempest-MultipleCreateTestJSON-1589015462-project-member] [instance: 25218e83-c1ab-4b97-ade1-7c32b6f99305] Start building block device mappings for instance. {{(pid=61806) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 822.546023] env[61806]: DEBUG oslo_vmware.api [None req-b326e3ef-f103-433e-9385-2b5e433b613a tempest-ServerShowV247Test-1218443059 tempest-ServerShowV247Test-1218443059-project-member] Waiting for the task: (returnval){ [ 822.546023] env[61806]: value = "task-1294633" [ 822.546023] env[61806]: _type = "Task" [ 822.546023] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 822.553799] env[61806]: DEBUG oslo_vmware.api [None req-b326e3ef-f103-433e-9385-2b5e433b613a tempest-ServerShowV247Test-1218443059 tempest-ServerShowV247Test-1218443059-project-member] Task: {'id': task-1294633, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 822.699014] env[61806]: DEBUG oslo_concurrency.lockutils [None req-7fa92f02-ad2b-4004-a9a7-cb08f83ec0f1 tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Acquiring lock "refresh_cache-67d1bdd8-cf8f-4fb5-bd3c-42008e4d60e3" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 822.699420] env[61806]: DEBUG oslo_concurrency.lockutils [None req-81ee2b3f-4b48-4fda-b966-c7fd3249b12e tempest-ServerDiagnosticsV248Test-463927155 tempest-ServerDiagnosticsV248Test-463927155-project-member] Acquiring lock "a934d02d-26aa-4900-b473-a58489e5629e" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 822.699659] env[61806]: DEBUG oslo_concurrency.lockutils [None req-81ee2b3f-4b48-4fda-b966-c7fd3249b12e tempest-ServerDiagnosticsV248Test-463927155 tempest-ServerDiagnosticsV248Test-463927155-project-member] Lock "a934d02d-26aa-4900-b473-a58489e5629e" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 822.699837] env[61806]: DEBUG oslo_concurrency.lockutils [None req-81ee2b3f-4b48-4fda-b966-c7fd3249b12e tempest-ServerDiagnosticsV248Test-463927155 tempest-ServerDiagnosticsV248Test-463927155-project-member] Acquiring lock "a934d02d-26aa-4900-b473-a58489e5629e-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 822.700040] env[61806]: DEBUG oslo_concurrency.lockutils [None req-81ee2b3f-4b48-4fda-b966-c7fd3249b12e tempest-ServerDiagnosticsV248Test-463927155 tempest-ServerDiagnosticsV248Test-463927155-project-member] Lock "a934d02d-26aa-4900-b473-a58489e5629e-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 822.700223] env[61806]: DEBUG oslo_concurrency.lockutils [None req-81ee2b3f-4b48-4fda-b966-c7fd3249b12e tempest-ServerDiagnosticsV248Test-463927155 tempest-ServerDiagnosticsV248Test-463927155-project-member] Lock "a934d02d-26aa-4900-b473-a58489e5629e-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 822.702414] env[61806]: INFO nova.compute.manager [None req-81ee2b3f-4b48-4fda-b966-c7fd3249b12e tempest-ServerDiagnosticsV248Test-463927155 tempest-ServerDiagnosticsV248Test-463927155-project-member] [instance: a934d02d-26aa-4900-b473-a58489e5629e] Terminating instance [ 822.704265] env[61806]: DEBUG oslo_concurrency.lockutils [None req-81ee2b3f-4b48-4fda-b966-c7fd3249b12e tempest-ServerDiagnosticsV248Test-463927155 tempest-ServerDiagnosticsV248Test-463927155-project-member] Acquiring lock "refresh_cache-a934d02d-26aa-4900-b473-a58489e5629e" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 822.704456] env[61806]: DEBUG oslo_concurrency.lockutils [None req-81ee2b3f-4b48-4fda-b966-c7fd3249b12e tempest-ServerDiagnosticsV248Test-463927155 tempest-ServerDiagnosticsV248Test-463927155-project-member] Acquired lock "refresh_cache-a934d02d-26aa-4900-b473-a58489e5629e" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 822.704703] env[61806]: DEBUG nova.network.neutron [None req-81ee2b3f-4b48-4fda-b966-c7fd3249b12e tempest-ServerDiagnosticsV248Test-463927155 tempest-ServerDiagnosticsV248Test-463927155-project-member] [instance: a934d02d-26aa-4900-b473-a58489e5629e] Building network info cache for instance {{(pid=61806) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 822.741723] env[61806]: DEBUG oslo_concurrency.lockutils [None req-4e219b6d-f08a-4088-99ff-2fcd194df36a tempest-ServerShowV257Test-1159790492 tempest-ServerShowV257Test-1159790492-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 822.794434] env[61806]: DEBUG oslo_vmware.api [None req-8aa43bc8-f134-4682-8a54-ce9aa25e4ee5 tempest-AttachVolumeNegativeTest-1683033704 tempest-AttachVolumeNegativeTest-1683033704-project-member] Task: {'id': task-1294629, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 822.843843] env[61806]: DEBUG oslo_vmware.api [-] Task: {'id': task-1294630, 'name': CreateVM_Task} progress is 25%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 822.911389] env[61806]: DEBUG nova.network.neutron [req-e2a64c0c-3895-4c1e-b206-30922535a8df req-e021b049-0656-42f8-9a51-3eca4aeb4d7a service nova] [instance: 67d1bdd8-cf8f-4fb5-bd3c-42008e4d60e3] Instance cache missing network info. {{(pid=61806) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 822.946785] env[61806]: DEBUG oslo_vmware.api [None req-61b80e06-de2f-412b-be27-b9318eeb5c25 tempest-ServerAddressesTestJSON-661730579 tempest-ServerAddressesTestJSON-661730579-project-member] Task: {'id': task-1294631, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 823.006626] env[61806]: DEBUG nova.network.neutron [req-e2a64c0c-3895-4c1e-b206-30922535a8df req-e021b049-0656-42f8-9a51-3eca4aeb4d7a service nova] [instance: 67d1bdd8-cf8f-4fb5-bd3c-42008e4d60e3] Updating instance_info_cache with network_info: [] {{(pid=61806) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 823.031716] env[61806]: DEBUG nova.scheduler.client.report [None req-fbc15506-7910-4d6f-bfbc-c3a1ce19fdcc tempest-ServerMetadataTestJSON-316746592 tempest-ServerMetadataTestJSON-316746592-project-member] Inventory has not changed for provider a2858be1-fd22-4e08-979e-87ad25293407 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 138, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61806) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 823.061662] env[61806]: DEBUG oslo_vmware.api [None req-b326e3ef-f103-433e-9385-2b5e433b613a tempest-ServerShowV247Test-1218443059 tempest-ServerShowV247Test-1218443059-project-member] Task: {'id': task-1294633, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.388813} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 823.063031] env[61806]: DEBUG nova.virt.vmwareapi.ds_util [None req-b326e3ef-f103-433e-9385-2b5e433b613a tempest-ServerShowV247Test-1218443059 tempest-ServerShowV247Test-1218443059-project-member] Deleted the datastore file {{(pid=61806) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 823.063031] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-b326e3ef-f103-433e-9385-2b5e433b613a tempest-ServerShowV247Test-1218443059 tempest-ServerShowV247Test-1218443059-project-member] [instance: 4febd093-9f2b-494e-b175-e4693b0e3e0d] Deleted contents of the VM from datastore datastore1 {{(pid=61806) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 823.063031] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-b326e3ef-f103-433e-9385-2b5e433b613a tempest-ServerShowV247Test-1218443059 tempest-ServerShowV247Test-1218443059-project-member] [instance: 4febd093-9f2b-494e-b175-e4693b0e3e0d] Instance destroyed {{(pid=61806) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 823.063031] env[61806]: INFO nova.compute.manager [None req-b326e3ef-f103-433e-9385-2b5e433b613a tempest-ServerShowV247Test-1218443059 tempest-ServerShowV247Test-1218443059-project-member] [instance: 4febd093-9f2b-494e-b175-e4693b0e3e0d] Took 1.14 seconds to destroy the instance on the hypervisor. [ 823.063261] env[61806]: DEBUG oslo.service.loopingcall [None req-b326e3ef-f103-433e-9385-2b5e433b613a tempest-ServerShowV247Test-1218443059 tempest-ServerShowV247Test-1218443059-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61806) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 823.064779] env[61806]: DEBUG nova.network.neutron [req-5353c2ee-5fa8-4404-b115-2e953c55ac8c req-7d4f70e6-4e1e-4dc9-81ff-8e1cc8f5bc30 service nova] [instance: b41531f2-f28c-4d82-9682-0b557bbaa491] Updated VIF entry in instance network info cache for port 513ca4b5-b62c-486a-9860-cf267a15c3e5. {{(pid=61806) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 823.065351] env[61806]: DEBUG nova.network.neutron [req-5353c2ee-5fa8-4404-b115-2e953c55ac8c req-7d4f70e6-4e1e-4dc9-81ff-8e1cc8f5bc30 service nova] [instance: b41531f2-f28c-4d82-9682-0b557bbaa491] Updating instance_info_cache with network_info: [{"id": "513ca4b5-b62c-486a-9860-cf267a15c3e5", "address": "fa:16:3e:40:52:86", "network": {"id": "dfb52c93-e553-40d9-acb5-768cc6b67c89", "bridge": "br-int", "label": "tempest-SecurityGroupsTestJSON-1006891343-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "cfce93add38d40d79f294017f52e13d5", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6b6942d7-d4ab-4b2a-8d0f-76bf2a2478ad", "external-id": "nsx-vlan-transportzone-871", "segmentation_id": 871, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap513ca4b5-b6", "ovs_interfaceid": "513ca4b5-b62c-486a-9860-cf267a15c3e5", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61806) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 823.067473] env[61806]: DEBUG nova.compute.manager [-] [instance: 4febd093-9f2b-494e-b175-e4693b0e3e0d] Deallocating network for instance {{(pid=61806) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 823.069031] env[61806]: DEBUG nova.network.neutron [-] [instance: 4febd093-9f2b-494e-b175-e4693b0e3e0d] deallocate_for_instance() {{(pid=61806) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 823.097446] env[61806]: DEBUG nova.network.neutron [-] [instance: 4febd093-9f2b-494e-b175-e4693b0e3e0d] Instance cache missing network info. {{(pid=61806) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 823.222644] env[61806]: DEBUG nova.network.neutron [None req-81ee2b3f-4b48-4fda-b966-c7fd3249b12e tempest-ServerDiagnosticsV248Test-463927155 tempest-ServerDiagnosticsV248Test-463927155-project-member] [instance: a934d02d-26aa-4900-b473-a58489e5629e] Instance cache missing network info. {{(pid=61806) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 823.290845] env[61806]: DEBUG nova.network.neutron [None req-81ee2b3f-4b48-4fda-b966-c7fd3249b12e tempest-ServerDiagnosticsV248Test-463927155 tempest-ServerDiagnosticsV248Test-463927155-project-member] [instance: a934d02d-26aa-4900-b473-a58489e5629e] Updating instance_info_cache with network_info: [] {{(pid=61806) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 823.297163] env[61806]: DEBUG oslo_vmware.api [None req-8aa43bc8-f134-4682-8a54-ce9aa25e4ee5 tempest-AttachVolumeNegativeTest-1683033704 tempest-AttachVolumeNegativeTest-1683033704-project-member] Task: {'id': task-1294629, 'name': ReconfigVM_Task, 'duration_secs': 0.83418} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 823.297594] env[61806]: DEBUG nova.virt.vmwareapi.volumeops [None req-8aa43bc8-f134-4682-8a54-ce9aa25e4ee5 tempest-AttachVolumeNegativeTest-1683033704 tempest-AttachVolumeNegativeTest-1683033704-project-member] [instance: 43dcce0d-4fdb-4ee1-8309-92e92d2331a9] Reconfigured VM instance instance-0000003f to attach disk [datastore1] 43dcce0d-4fdb-4ee1-8309-92e92d2331a9/43dcce0d-4fdb-4ee1-8309-92e92d2331a9.vmdk or device None with type sparse {{(pid=61806) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 823.298255] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-ac550da2-1603-4c69-ba56-9d9c3746fbb6 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 823.309570] env[61806]: DEBUG oslo_vmware.api [None req-8aa43bc8-f134-4682-8a54-ce9aa25e4ee5 tempest-AttachVolumeNegativeTest-1683033704 tempest-AttachVolumeNegativeTest-1683033704-project-member] Waiting for the task: (returnval){ [ 823.309570] env[61806]: value = "task-1294634" [ 823.309570] env[61806]: _type = "Task" [ 823.309570] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 823.322013] env[61806]: DEBUG oslo_vmware.api [None req-8aa43bc8-f134-4682-8a54-ce9aa25e4ee5 tempest-AttachVolumeNegativeTest-1683033704 tempest-AttachVolumeNegativeTest-1683033704-project-member] Task: {'id': task-1294634, 'name': Rename_Task} progress is 5%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 823.333544] env[61806]: DEBUG oslo_concurrency.lockutils [None req-f4c6ec76-6b50-4a2c-bb86-09b71e318ce8 tempest-ServerShowV257Test-1159790492 tempest-ServerShowV257Test-1159790492-project-member] Acquiring lock "616ec206-9804-469e-ab5c-41aea7f048aa" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 823.335213] env[61806]: DEBUG oslo_concurrency.lockutils [None req-f4c6ec76-6b50-4a2c-bb86-09b71e318ce8 tempest-ServerShowV257Test-1159790492 tempest-ServerShowV257Test-1159790492-project-member] Lock "616ec206-9804-469e-ab5c-41aea7f048aa" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 823.335213] env[61806]: DEBUG oslo_concurrency.lockutils [None req-f4c6ec76-6b50-4a2c-bb86-09b71e318ce8 tempest-ServerShowV257Test-1159790492 tempest-ServerShowV257Test-1159790492-project-member] Acquiring lock "616ec206-9804-469e-ab5c-41aea7f048aa-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 823.335213] env[61806]: DEBUG oslo_concurrency.lockutils [None req-f4c6ec76-6b50-4a2c-bb86-09b71e318ce8 tempest-ServerShowV257Test-1159790492 tempest-ServerShowV257Test-1159790492-project-member] Lock "616ec206-9804-469e-ab5c-41aea7f048aa-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 823.335213] env[61806]: DEBUG oslo_concurrency.lockutils [None req-f4c6ec76-6b50-4a2c-bb86-09b71e318ce8 tempest-ServerShowV257Test-1159790492 tempest-ServerShowV257Test-1159790492-project-member] Lock "616ec206-9804-469e-ab5c-41aea7f048aa-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 823.337215] env[61806]: DEBUG oslo_vmware.api [-] Task: {'id': task-1294630, 'name': CreateVM_Task, 'duration_secs': 0.71632} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 823.337481] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: b41531f2-f28c-4d82-9682-0b557bbaa491] Created VM on the ESX host {{(pid=61806) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 823.338279] env[61806]: DEBUG oslo_concurrency.lockutils [None req-9c52c9e9-816e-4638-b13a-a63a9756140c tempest-SecurityGroupsTestJSON-1213485581 tempest-SecurityGroupsTestJSON-1213485581-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 823.338494] env[61806]: DEBUG oslo_concurrency.lockutils [None req-9c52c9e9-816e-4638-b13a-a63a9756140c tempest-SecurityGroupsTestJSON-1213485581 tempest-SecurityGroupsTestJSON-1213485581-project-member] Acquired lock "[datastore2] devstack-image-cache_base/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 823.338862] env[61806]: DEBUG oslo_concurrency.lockutils [None req-9c52c9e9-816e-4638-b13a-a63a9756140c tempest-SecurityGroupsTestJSON-1213485581 tempest-SecurityGroupsTestJSON-1213485581-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 823.340729] env[61806]: INFO nova.compute.manager [None req-f4c6ec76-6b50-4a2c-bb86-09b71e318ce8 tempest-ServerShowV257Test-1159790492 tempest-ServerShowV257Test-1159790492-project-member] [instance: 616ec206-9804-469e-ab5c-41aea7f048aa] Terminating instance [ 823.342894] env[61806]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-3eb0201b-cf40-4922-aab9-f8a46e42260a {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 823.345330] env[61806]: DEBUG oslo_concurrency.lockutils [None req-f4c6ec76-6b50-4a2c-bb86-09b71e318ce8 tempest-ServerShowV257Test-1159790492 tempest-ServerShowV257Test-1159790492-project-member] Acquiring lock "refresh_cache-616ec206-9804-469e-ab5c-41aea7f048aa" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 823.345675] env[61806]: DEBUG oslo_concurrency.lockutils [None req-f4c6ec76-6b50-4a2c-bb86-09b71e318ce8 tempest-ServerShowV257Test-1159790492 tempest-ServerShowV257Test-1159790492-project-member] Acquired lock "refresh_cache-616ec206-9804-469e-ab5c-41aea7f048aa" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 823.346026] env[61806]: DEBUG nova.network.neutron [None req-f4c6ec76-6b50-4a2c-bb86-09b71e318ce8 tempest-ServerShowV257Test-1159790492 tempest-ServerShowV257Test-1159790492-project-member] [instance: 616ec206-9804-469e-ab5c-41aea7f048aa] Building network info cache for instance {{(pid=61806) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 823.351809] env[61806]: DEBUG oslo_vmware.api [None req-9c52c9e9-816e-4638-b13a-a63a9756140c tempest-SecurityGroupsTestJSON-1213485581 tempest-SecurityGroupsTestJSON-1213485581-project-member] Waiting for the task: (returnval){ [ 823.351809] env[61806]: value = "session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]52a8152f-81ea-b1f2-e787-1bb323997648" [ 823.351809] env[61806]: _type = "Task" [ 823.351809] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 823.361801] env[61806]: DEBUG oslo_vmware.api [None req-9c52c9e9-816e-4638-b13a-a63a9756140c tempest-SecurityGroupsTestJSON-1213485581 tempest-SecurityGroupsTestJSON-1213485581-project-member] Task: {'id': session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]52a8152f-81ea-b1f2-e787-1bb323997648, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 823.447246] env[61806]: DEBUG oslo_vmware.api [None req-61b80e06-de2f-412b-be27-b9318eeb5c25 tempest-ServerAddressesTestJSON-661730579 tempest-ServerAddressesTestJSON-661730579-project-member] Task: {'id': task-1294631, 'name': PowerOnVM_Task, 'duration_secs': 0.883885} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 823.447559] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [None req-61b80e06-de2f-412b-be27-b9318eeb5c25 tempest-ServerAddressesTestJSON-661730579 tempest-ServerAddressesTestJSON-661730579-project-member] [instance: a842df44-d8a9-4376-b9fc-5ca19a68a4b7] Powered on the VM {{(pid=61806) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 823.447777] env[61806]: INFO nova.compute.manager [None req-61b80e06-de2f-412b-be27-b9318eeb5c25 tempest-ServerAddressesTestJSON-661730579 tempest-ServerAddressesTestJSON-661730579-project-member] [instance: a842df44-d8a9-4376-b9fc-5ca19a68a4b7] Took 9.71 seconds to spawn the instance on the hypervisor. [ 823.447966] env[61806]: DEBUG nova.compute.manager [None req-61b80e06-de2f-412b-be27-b9318eeb5c25 tempest-ServerAddressesTestJSON-661730579 tempest-ServerAddressesTestJSON-661730579-project-member] [instance: a842df44-d8a9-4376-b9fc-5ca19a68a4b7] Checking state {{(pid=61806) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 823.448783] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a9a93478-a309-49cd-84b5-91b89b379397 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 823.509982] env[61806]: DEBUG oslo_concurrency.lockutils [req-e2a64c0c-3895-4c1e-b206-30922535a8df req-e021b049-0656-42f8-9a51-3eca4aeb4d7a service nova] Releasing lock "refresh_cache-67d1bdd8-cf8f-4fb5-bd3c-42008e4d60e3" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 823.510737] env[61806]: DEBUG oslo_concurrency.lockutils [None req-7fa92f02-ad2b-4004-a9a7-cb08f83ec0f1 tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Acquired lock "refresh_cache-67d1bdd8-cf8f-4fb5-bd3c-42008e4d60e3" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 823.510996] env[61806]: DEBUG nova.network.neutron [None req-7fa92f02-ad2b-4004-a9a7-cb08f83ec0f1 tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] [instance: 67d1bdd8-cf8f-4fb5-bd3c-42008e4d60e3] Building network info cache for instance {{(pid=61806) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 823.537741] env[61806]: DEBUG oslo_concurrency.lockutils [None req-fbc15506-7910-4d6f-bfbc-c3a1ce19fdcc tempest-ServerMetadataTestJSON-316746592 tempest-ServerMetadataTestJSON-316746592-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.010s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 823.540131] env[61806]: DEBUG oslo_concurrency.lockutils [None req-cbb90e6d-7dac-4c30-b860-1f0db7dd091e tempest-MultipleCreateTestJSON-1589015462 tempest-MultipleCreateTestJSON-1589015462-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 14.045s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 823.541681] env[61806]: INFO nova.compute.claims [None req-cbb90e6d-7dac-4c30-b860-1f0db7dd091e tempest-MultipleCreateTestJSON-1589015462 tempest-MultipleCreateTestJSON-1589015462-project-member] [instance: ef502a7d-7dc0-4319-8ce3-6a8701288237] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 823.549452] env[61806]: DEBUG nova.compute.manager [None req-cbb90e6d-7dac-4c30-b860-1f0db7dd091e tempest-MultipleCreateTestJSON-1589015462 tempest-MultipleCreateTestJSON-1589015462-project-member] [instance: 25218e83-c1ab-4b97-ade1-7c32b6f99305] Start spawning the instance on the hypervisor. {{(pid=61806) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 823.562357] env[61806]: INFO nova.scheduler.client.report [None req-fbc15506-7910-4d6f-bfbc-c3a1ce19fdcc tempest-ServerMetadataTestJSON-316746592 tempest-ServerMetadataTestJSON-316746592-project-member] Deleted allocations for instance 4184e263-6f56-4bc1-99b2-a2c460531516 [ 823.568639] env[61806]: DEBUG oslo_concurrency.lockutils [req-5353c2ee-5fa8-4404-b115-2e953c55ac8c req-7d4f70e6-4e1e-4dc9-81ff-8e1cc8f5bc30 service nova] Releasing lock "refresh_cache-b41531f2-f28c-4d82-9682-0b557bbaa491" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 823.578081] env[61806]: DEBUG nova.virt.hardware [None req-cbb90e6d-7dac-4c30-b860-1f0db7dd091e tempest-MultipleCreateTestJSON-1589015462 tempest-MultipleCreateTestJSON-1589015462-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-15T18:05:10Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-15T18:04:52Z,direct_url=,disk_format='vmdk',id=ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='84ba83fce2c349a6988173c5d6fc3b07',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-15T18:04:53Z,virtual_size=,visibility=), allow threads: False {{(pid=61806) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 823.578358] env[61806]: DEBUG nova.virt.hardware [None req-cbb90e6d-7dac-4c30-b860-1f0db7dd091e tempest-MultipleCreateTestJSON-1589015462 tempest-MultipleCreateTestJSON-1589015462-project-member] Flavor limits 0:0:0 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 823.578530] env[61806]: DEBUG nova.virt.hardware [None req-cbb90e6d-7dac-4c30-b860-1f0db7dd091e tempest-MultipleCreateTestJSON-1589015462 tempest-MultipleCreateTestJSON-1589015462-project-member] Image limits 0:0:0 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 823.578725] env[61806]: DEBUG nova.virt.hardware [None req-cbb90e6d-7dac-4c30-b860-1f0db7dd091e tempest-MultipleCreateTestJSON-1589015462 tempest-MultipleCreateTestJSON-1589015462-project-member] Flavor pref 0:0:0 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 823.578879] env[61806]: DEBUG nova.virt.hardware [None req-cbb90e6d-7dac-4c30-b860-1f0db7dd091e tempest-MultipleCreateTestJSON-1589015462 tempest-MultipleCreateTestJSON-1589015462-project-member] Image pref 0:0:0 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 823.579043] env[61806]: DEBUG nova.virt.hardware [None req-cbb90e6d-7dac-4c30-b860-1f0db7dd091e tempest-MultipleCreateTestJSON-1589015462 tempest-MultipleCreateTestJSON-1589015462-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 823.579267] env[61806]: DEBUG nova.virt.hardware [None req-cbb90e6d-7dac-4c30-b860-1f0db7dd091e tempest-MultipleCreateTestJSON-1589015462 tempest-MultipleCreateTestJSON-1589015462-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61806) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 823.579433] env[61806]: DEBUG nova.virt.hardware [None req-cbb90e6d-7dac-4c30-b860-1f0db7dd091e tempest-MultipleCreateTestJSON-1589015462 tempest-MultipleCreateTestJSON-1589015462-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61806) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 823.579611] env[61806]: DEBUG nova.virt.hardware [None req-cbb90e6d-7dac-4c30-b860-1f0db7dd091e tempest-MultipleCreateTestJSON-1589015462 tempest-MultipleCreateTestJSON-1589015462-project-member] Got 1 possible topologies {{(pid=61806) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 823.579784] env[61806]: DEBUG nova.virt.hardware [None req-cbb90e6d-7dac-4c30-b860-1f0db7dd091e tempest-MultipleCreateTestJSON-1589015462 tempest-MultipleCreateTestJSON-1589015462-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61806) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 823.579964] env[61806]: DEBUG nova.virt.hardware [None req-cbb90e6d-7dac-4c30-b860-1f0db7dd091e tempest-MultipleCreateTestJSON-1589015462 tempest-MultipleCreateTestJSON-1589015462-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61806) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 823.580867] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7e7e6cd0-eb74-4c71-a7c6-2e4a0d3e1cd4 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 823.590237] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-435b0344-4404-4b6c-93d2-261c2661cfad {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 823.598715] env[61806]: DEBUG nova.network.neutron [-] [instance: 4febd093-9f2b-494e-b175-e4693b0e3e0d] Updating instance_info_cache with network_info: [] {{(pid=61806) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 823.798516] env[61806]: DEBUG oslo_concurrency.lockutils [None req-81ee2b3f-4b48-4fda-b966-c7fd3249b12e tempest-ServerDiagnosticsV248Test-463927155 tempest-ServerDiagnosticsV248Test-463927155-project-member] Releasing lock "refresh_cache-a934d02d-26aa-4900-b473-a58489e5629e" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 823.798948] env[61806]: DEBUG nova.compute.manager [None req-81ee2b3f-4b48-4fda-b966-c7fd3249b12e tempest-ServerDiagnosticsV248Test-463927155 tempest-ServerDiagnosticsV248Test-463927155-project-member] [instance: a934d02d-26aa-4900-b473-a58489e5629e] Start destroying the instance on the hypervisor. {{(pid=61806) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 823.799220] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-81ee2b3f-4b48-4fda-b966-c7fd3249b12e tempest-ServerDiagnosticsV248Test-463927155 tempest-ServerDiagnosticsV248Test-463927155-project-member] [instance: a934d02d-26aa-4900-b473-a58489e5629e] Destroying instance {{(pid=61806) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 823.800521] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-328b0696-96e3-493a-aadd-9d3d541e5b29 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 823.808909] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [None req-81ee2b3f-4b48-4fda-b966-c7fd3249b12e tempest-ServerDiagnosticsV248Test-463927155 tempest-ServerDiagnosticsV248Test-463927155-project-member] [instance: a934d02d-26aa-4900-b473-a58489e5629e] Powering off the VM {{(pid=61806) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 823.809537] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-2790a1ef-81e5-4ab0-936f-143df484146b {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 823.819747] env[61806]: DEBUG oslo_vmware.api [None req-8aa43bc8-f134-4682-8a54-ce9aa25e4ee5 tempest-AttachVolumeNegativeTest-1683033704 tempest-AttachVolumeNegativeTest-1683033704-project-member] Task: {'id': task-1294634, 'name': Rename_Task} progress is 99%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 823.821438] env[61806]: DEBUG oslo_vmware.api [None req-81ee2b3f-4b48-4fda-b966-c7fd3249b12e tempest-ServerDiagnosticsV248Test-463927155 tempest-ServerDiagnosticsV248Test-463927155-project-member] Waiting for the task: (returnval){ [ 823.821438] env[61806]: value = "task-1294635" [ 823.821438] env[61806]: _type = "Task" [ 823.821438] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 823.832901] env[61806]: DEBUG oslo_vmware.api [None req-81ee2b3f-4b48-4fda-b966-c7fd3249b12e tempest-ServerDiagnosticsV248Test-463927155 tempest-ServerDiagnosticsV248Test-463927155-project-member] Task: {'id': task-1294635, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 823.862979] env[61806]: DEBUG oslo_vmware.api [None req-9c52c9e9-816e-4638-b13a-a63a9756140c tempest-SecurityGroupsTestJSON-1213485581 tempest-SecurityGroupsTestJSON-1213485581-project-member] Task: {'id': session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]52a8152f-81ea-b1f2-e787-1bb323997648, 'name': SearchDatastore_Task, 'duration_secs': 0.014123} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 823.863319] env[61806]: DEBUG oslo_concurrency.lockutils [None req-9c52c9e9-816e-4638-b13a-a63a9756140c tempest-SecurityGroupsTestJSON-1213485581 tempest-SecurityGroupsTestJSON-1213485581-project-member] Releasing lock "[datastore2] devstack-image-cache_base/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 823.863630] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-9c52c9e9-816e-4638-b13a-a63a9756140c tempest-SecurityGroupsTestJSON-1213485581 tempest-SecurityGroupsTestJSON-1213485581-project-member] [instance: b41531f2-f28c-4d82-9682-0b557bbaa491] Processing image ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa {{(pid=61806) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 823.863900] env[61806]: DEBUG oslo_concurrency.lockutils [None req-9c52c9e9-816e-4638-b13a-a63a9756140c tempest-SecurityGroupsTestJSON-1213485581 tempest-SecurityGroupsTestJSON-1213485581-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa.vmdk" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 823.864079] env[61806]: DEBUG oslo_concurrency.lockutils [None req-9c52c9e9-816e-4638-b13a-a63a9756140c tempest-SecurityGroupsTestJSON-1213485581 tempest-SecurityGroupsTestJSON-1213485581-project-member] Acquired lock "[datastore2] devstack-image-cache_base/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa.vmdk" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 823.864274] env[61806]: DEBUG nova.virt.vmwareapi.ds_util [None req-9c52c9e9-816e-4638-b13a-a63a9756140c tempest-SecurityGroupsTestJSON-1213485581 tempest-SecurityGroupsTestJSON-1213485581-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61806) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 823.864541] env[61806]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-a20e7500-d7ba-43d6-9cac-61757dc23ae5 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 823.868849] env[61806]: DEBUG nova.network.neutron [None req-f4c6ec76-6b50-4a2c-bb86-09b71e318ce8 tempest-ServerShowV257Test-1159790492 tempest-ServerShowV257Test-1159790492-project-member] [instance: 616ec206-9804-469e-ab5c-41aea7f048aa] Instance cache missing network info. {{(pid=61806) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 823.874648] env[61806]: DEBUG nova.virt.vmwareapi.ds_util [None req-9c52c9e9-816e-4638-b13a-a63a9756140c tempest-SecurityGroupsTestJSON-1213485581 tempest-SecurityGroupsTestJSON-1213485581-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61806) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 823.874915] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-9c52c9e9-816e-4638-b13a-a63a9756140c tempest-SecurityGroupsTestJSON-1213485581 tempest-SecurityGroupsTestJSON-1213485581-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61806) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 823.875720] env[61806]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-1a0f9eed-a439-4488-8553-7457485dddc2 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 823.881814] env[61806]: DEBUG oslo_vmware.api [None req-9c52c9e9-816e-4638-b13a-a63a9756140c tempest-SecurityGroupsTestJSON-1213485581 tempest-SecurityGroupsTestJSON-1213485581-project-member] Waiting for the task: (returnval){ [ 823.881814] env[61806]: value = "session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]521eaf46-f015-e26d-7279-383a83b0d818" [ 823.881814] env[61806]: _type = "Task" [ 823.881814] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 823.890690] env[61806]: DEBUG oslo_vmware.api [None req-9c52c9e9-816e-4638-b13a-a63a9756140c tempest-SecurityGroupsTestJSON-1213485581 tempest-SecurityGroupsTestJSON-1213485581-project-member] Task: {'id': session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]521eaf46-f015-e26d-7279-383a83b0d818, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 823.974270] env[61806]: INFO nova.compute.manager [None req-61b80e06-de2f-412b-be27-b9318eeb5c25 tempest-ServerAddressesTestJSON-661730579 tempest-ServerAddressesTestJSON-661730579-project-member] [instance: a842df44-d8a9-4376-b9fc-5ca19a68a4b7] Took 29.34 seconds to build instance. [ 823.974270] env[61806]: DEBUG nova.network.neutron [None req-f4c6ec76-6b50-4a2c-bb86-09b71e318ce8 tempest-ServerShowV257Test-1159790492 tempest-ServerShowV257Test-1159790492-project-member] [instance: 616ec206-9804-469e-ab5c-41aea7f048aa] Updating instance_info_cache with network_info: [] {{(pid=61806) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 824.069936] env[61806]: DEBUG oslo_concurrency.lockutils [None req-fbc15506-7910-4d6f-bfbc-c3a1ce19fdcc tempest-ServerMetadataTestJSON-316746592 tempest-ServerMetadataTestJSON-316746592-project-member] Lock "4184e263-6f56-4bc1-99b2-a2c460531516" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 19.068s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 824.074973] env[61806]: DEBUG nova.network.neutron [None req-7fa92f02-ad2b-4004-a9a7-cb08f83ec0f1 tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] [instance: 67d1bdd8-cf8f-4fb5-bd3c-42008e4d60e3] Instance cache missing network info. {{(pid=61806) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 824.108365] env[61806]: INFO nova.compute.manager [-] [instance: 4febd093-9f2b-494e-b175-e4693b0e3e0d] Took 1.04 seconds to deallocate network for instance. [ 824.321328] env[61806]: DEBUG oslo_vmware.api [None req-8aa43bc8-f134-4682-8a54-ce9aa25e4ee5 tempest-AttachVolumeNegativeTest-1683033704 tempest-AttachVolumeNegativeTest-1683033704-project-member] Task: {'id': task-1294634, 'name': Rename_Task} progress is 99%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 824.322310] env[61806]: DEBUG nova.network.neutron [None req-7fa92f02-ad2b-4004-a9a7-cb08f83ec0f1 tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] [instance: 67d1bdd8-cf8f-4fb5-bd3c-42008e4d60e3] Updating instance_info_cache with network_info: [{"id": "859fb723-4a1c-4ea8-bd7c-0ed308f3a9e3", "address": "fa:16:3e:3c:40:d6", "network": {"id": "746a03b6-c1a6-4d27-98b9-0bee6376634f", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-1971382587-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c1f2d161c50f4b70987b90e6364c3779", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6a31b224-77d7-48a4-af87-312758611b8e", "external-id": "nsx-vlan-transportzone-761", "segmentation_id": 761, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap859fb723-4a", "ovs_interfaceid": "859fb723-4a1c-4ea8-bd7c-0ed308f3a9e3", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61806) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 824.330791] env[61806]: DEBUG nova.network.neutron [None req-cbb90e6d-7dac-4c30-b860-1f0db7dd091e tempest-MultipleCreateTestJSON-1589015462 tempest-MultipleCreateTestJSON-1589015462-project-member] [instance: 25218e83-c1ab-4b97-ade1-7c32b6f99305] Successfully updated port: c8809e29-ea3c-4ea0-bee1-0731daee293d {{(pid=61806) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 824.335447] env[61806]: DEBUG oslo_vmware.api [None req-81ee2b3f-4b48-4fda-b966-c7fd3249b12e tempest-ServerDiagnosticsV248Test-463927155 tempest-ServerDiagnosticsV248Test-463927155-project-member] Task: {'id': task-1294635, 'name': PowerOffVM_Task, 'duration_secs': 0.193696} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 824.335753] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [None req-81ee2b3f-4b48-4fda-b966-c7fd3249b12e tempest-ServerDiagnosticsV248Test-463927155 tempest-ServerDiagnosticsV248Test-463927155-project-member] [instance: a934d02d-26aa-4900-b473-a58489e5629e] Powered off the VM {{(pid=61806) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 824.335930] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-81ee2b3f-4b48-4fda-b966-c7fd3249b12e tempest-ServerDiagnosticsV248Test-463927155 tempest-ServerDiagnosticsV248Test-463927155-project-member] [instance: a934d02d-26aa-4900-b473-a58489e5629e] Unregistering the VM {{(pid=61806) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 824.339199] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-0ed7a6b6-a402-4e18-b7b6-53a6c28faeca {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 824.368369] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-81ee2b3f-4b48-4fda-b966-c7fd3249b12e tempest-ServerDiagnosticsV248Test-463927155 tempest-ServerDiagnosticsV248Test-463927155-project-member] [instance: a934d02d-26aa-4900-b473-a58489e5629e] Unregistered the VM {{(pid=61806) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 824.368718] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-81ee2b3f-4b48-4fda-b966-c7fd3249b12e tempest-ServerDiagnosticsV248Test-463927155 tempest-ServerDiagnosticsV248Test-463927155-project-member] [instance: a934d02d-26aa-4900-b473-a58489e5629e] Deleting contents of the VM from datastore datastore2 {{(pid=61806) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 824.368925] env[61806]: DEBUG nova.virt.vmwareapi.ds_util [None req-81ee2b3f-4b48-4fda-b966-c7fd3249b12e tempest-ServerDiagnosticsV248Test-463927155 tempest-ServerDiagnosticsV248Test-463927155-project-member] Deleting the datastore file [datastore2] a934d02d-26aa-4900-b473-a58489e5629e {{(pid=61806) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 824.369380] env[61806]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-9552e1c6-5f8d-4350-a606-dc341377b3b8 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 824.377918] env[61806]: DEBUG oslo_vmware.api [None req-81ee2b3f-4b48-4fda-b966-c7fd3249b12e tempest-ServerDiagnosticsV248Test-463927155 tempest-ServerDiagnosticsV248Test-463927155-project-member] Waiting for the task: (returnval){ [ 824.377918] env[61806]: value = "task-1294637" [ 824.377918] env[61806]: _type = "Task" [ 824.377918] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 824.394916] env[61806]: DEBUG oslo_vmware.api [None req-81ee2b3f-4b48-4fda-b966-c7fd3249b12e tempest-ServerDiagnosticsV248Test-463927155 tempest-ServerDiagnosticsV248Test-463927155-project-member] Task: {'id': task-1294637, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 824.395076] env[61806]: DEBUG oslo_vmware.api [None req-9c52c9e9-816e-4638-b13a-a63a9756140c tempest-SecurityGroupsTestJSON-1213485581 tempest-SecurityGroupsTestJSON-1213485581-project-member] Task: {'id': session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]521eaf46-f015-e26d-7279-383a83b0d818, 'name': SearchDatastore_Task, 'duration_secs': 0.013353} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 824.395853] env[61806]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d401b67c-0237-4be1-9131-fa4e98990dff {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 824.403046] env[61806]: DEBUG oslo_vmware.api [None req-9c52c9e9-816e-4638-b13a-a63a9756140c tempest-SecurityGroupsTestJSON-1213485581 tempest-SecurityGroupsTestJSON-1213485581-project-member] Waiting for the task: (returnval){ [ 824.403046] env[61806]: value = "session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]5216735e-762f-907c-cd93-68cb23a8ac1c" [ 824.403046] env[61806]: _type = "Task" [ 824.403046] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 824.412990] env[61806]: DEBUG oslo_vmware.api [None req-9c52c9e9-816e-4638-b13a-a63a9756140c tempest-SecurityGroupsTestJSON-1213485581 tempest-SecurityGroupsTestJSON-1213485581-project-member] Task: {'id': session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]5216735e-762f-907c-cd93-68cb23a8ac1c, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 824.475921] env[61806]: DEBUG oslo_concurrency.lockutils [None req-61b80e06-de2f-412b-be27-b9318eeb5c25 tempest-ServerAddressesTestJSON-661730579 tempest-ServerAddressesTestJSON-661730579-project-member] Lock "a842df44-d8a9-4376-b9fc-5ca19a68a4b7" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 104.494s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 824.477711] env[61806]: DEBUG oslo_concurrency.lockutils [None req-f4c6ec76-6b50-4a2c-bb86-09b71e318ce8 tempest-ServerShowV257Test-1159790492 tempest-ServerShowV257Test-1159790492-project-member] Releasing lock "refresh_cache-616ec206-9804-469e-ab5c-41aea7f048aa" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 824.477879] env[61806]: DEBUG nova.compute.manager [None req-f4c6ec76-6b50-4a2c-bb86-09b71e318ce8 tempest-ServerShowV257Test-1159790492 tempest-ServerShowV257Test-1159790492-project-member] [instance: 616ec206-9804-469e-ab5c-41aea7f048aa] Start destroying the instance on the hypervisor. {{(pid=61806) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 824.477997] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-f4c6ec76-6b50-4a2c-bb86-09b71e318ce8 tempest-ServerShowV257Test-1159790492 tempest-ServerShowV257Test-1159790492-project-member] [instance: 616ec206-9804-469e-ab5c-41aea7f048aa] Destroying instance {{(pid=61806) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 824.479409] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e4e96fd6-4039-492d-b08b-949e867ed903 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 824.506530] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [None req-f4c6ec76-6b50-4a2c-bb86-09b71e318ce8 tempest-ServerShowV257Test-1159790492 tempest-ServerShowV257Test-1159790492-project-member] [instance: 616ec206-9804-469e-ab5c-41aea7f048aa] Powering off the VM {{(pid=61806) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 824.506920] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-713917d4-1754-491a-aae1-2dae93bc6d8f {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 824.512031] env[61806]: DEBUG nova.compute.manager [req-fd4c6b15-6d49-4512-a701-36ff2780835c req-a1df8c6d-978a-40eb-93c3-1ed7a6109bfc service nova] [instance: 25218e83-c1ab-4b97-ade1-7c32b6f99305] Received event network-vif-plugged-c8809e29-ea3c-4ea0-bee1-0731daee293d {{(pid=61806) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 824.512598] env[61806]: DEBUG oslo_concurrency.lockutils [req-fd4c6b15-6d49-4512-a701-36ff2780835c req-a1df8c6d-978a-40eb-93c3-1ed7a6109bfc service nova] Acquiring lock "25218e83-c1ab-4b97-ade1-7c32b6f99305-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 824.512598] env[61806]: DEBUG oslo_concurrency.lockutils [req-fd4c6b15-6d49-4512-a701-36ff2780835c req-a1df8c6d-978a-40eb-93c3-1ed7a6109bfc service nova] Lock "25218e83-c1ab-4b97-ade1-7c32b6f99305-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 824.512848] env[61806]: DEBUG oslo_concurrency.lockutils [req-fd4c6b15-6d49-4512-a701-36ff2780835c req-a1df8c6d-978a-40eb-93c3-1ed7a6109bfc service nova] Lock "25218e83-c1ab-4b97-ade1-7c32b6f99305-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 824.513409] env[61806]: DEBUG nova.compute.manager [req-fd4c6b15-6d49-4512-a701-36ff2780835c req-a1df8c6d-978a-40eb-93c3-1ed7a6109bfc service nova] [instance: 25218e83-c1ab-4b97-ade1-7c32b6f99305] No waiting events found dispatching network-vif-plugged-c8809e29-ea3c-4ea0-bee1-0731daee293d {{(pid=61806) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 824.513409] env[61806]: WARNING nova.compute.manager [req-fd4c6b15-6d49-4512-a701-36ff2780835c req-a1df8c6d-978a-40eb-93c3-1ed7a6109bfc service nova] [instance: 25218e83-c1ab-4b97-ade1-7c32b6f99305] Received unexpected event network-vif-plugged-c8809e29-ea3c-4ea0-bee1-0731daee293d for instance with vm_state building and task_state spawning. [ 824.513597] env[61806]: DEBUG nova.compute.manager [req-fd4c6b15-6d49-4512-a701-36ff2780835c req-a1df8c6d-978a-40eb-93c3-1ed7a6109bfc service nova] [instance: 25218e83-c1ab-4b97-ade1-7c32b6f99305] Received event network-changed-c8809e29-ea3c-4ea0-bee1-0731daee293d {{(pid=61806) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 824.513823] env[61806]: DEBUG nova.compute.manager [req-fd4c6b15-6d49-4512-a701-36ff2780835c req-a1df8c6d-978a-40eb-93c3-1ed7a6109bfc service nova] [instance: 25218e83-c1ab-4b97-ade1-7c32b6f99305] Refreshing instance network info cache due to event network-changed-c8809e29-ea3c-4ea0-bee1-0731daee293d. {{(pid=61806) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 824.514086] env[61806]: DEBUG oslo_concurrency.lockutils [req-fd4c6b15-6d49-4512-a701-36ff2780835c req-a1df8c6d-978a-40eb-93c3-1ed7a6109bfc service nova] Acquiring lock "refresh_cache-25218e83-c1ab-4b97-ade1-7c32b6f99305" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 824.514256] env[61806]: DEBUG oslo_concurrency.lockutils [req-fd4c6b15-6d49-4512-a701-36ff2780835c req-a1df8c6d-978a-40eb-93c3-1ed7a6109bfc service nova] Acquired lock "refresh_cache-25218e83-c1ab-4b97-ade1-7c32b6f99305" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 824.514426] env[61806]: DEBUG nova.network.neutron [req-fd4c6b15-6d49-4512-a701-36ff2780835c req-a1df8c6d-978a-40eb-93c3-1ed7a6109bfc service nova] [instance: 25218e83-c1ab-4b97-ade1-7c32b6f99305] Refreshing network info cache for port c8809e29-ea3c-4ea0-bee1-0731daee293d {{(pid=61806) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 824.517959] env[61806]: DEBUG oslo_vmware.api [None req-f4c6ec76-6b50-4a2c-bb86-09b71e318ce8 tempest-ServerShowV257Test-1159790492 tempest-ServerShowV257Test-1159790492-project-member] Waiting for the task: (returnval){ [ 824.517959] env[61806]: value = "task-1294638" [ 824.517959] env[61806]: _type = "Task" [ 824.517959] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 824.532188] env[61806]: DEBUG oslo_vmware.api [None req-f4c6ec76-6b50-4a2c-bb86-09b71e318ce8 tempest-ServerShowV257Test-1159790492 tempest-ServerShowV257Test-1159790492-project-member] Task: {'id': task-1294638, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 824.616964] env[61806]: DEBUG oslo_concurrency.lockutils [None req-b326e3ef-f103-433e-9385-2b5e433b613a tempest-ServerShowV247Test-1218443059 tempest-ServerShowV247Test-1218443059-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 824.729393] env[61806]: DEBUG oslo_concurrency.lockutils [None req-0b05357b-24cd-44f2-8a41-86cdceeb41ee tempest-VolumesAdminNegativeTest-1387177911 tempest-VolumesAdminNegativeTest-1387177911-project-member] Acquiring lock "4373b735-31cf-4b53-b655-38555cf212a5" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 824.729676] env[61806]: DEBUG oslo_concurrency.lockutils [None req-0b05357b-24cd-44f2-8a41-86cdceeb41ee tempest-VolumesAdminNegativeTest-1387177911 tempest-VolumesAdminNegativeTest-1387177911-project-member] Lock "4373b735-31cf-4b53-b655-38555cf212a5" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 824.729898] env[61806]: DEBUG oslo_concurrency.lockutils [None req-0b05357b-24cd-44f2-8a41-86cdceeb41ee tempest-VolumesAdminNegativeTest-1387177911 tempest-VolumesAdminNegativeTest-1387177911-project-member] Acquiring lock "4373b735-31cf-4b53-b655-38555cf212a5-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 824.730096] env[61806]: DEBUG oslo_concurrency.lockutils [None req-0b05357b-24cd-44f2-8a41-86cdceeb41ee tempest-VolumesAdminNegativeTest-1387177911 tempest-VolumesAdminNegativeTest-1387177911-project-member] Lock "4373b735-31cf-4b53-b655-38555cf212a5-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 824.730415] env[61806]: DEBUG oslo_concurrency.lockutils [None req-0b05357b-24cd-44f2-8a41-86cdceeb41ee tempest-VolumesAdminNegativeTest-1387177911 tempest-VolumesAdminNegativeTest-1387177911-project-member] Lock "4373b735-31cf-4b53-b655-38555cf212a5-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 824.732409] env[61806]: INFO nova.compute.manager [None req-0b05357b-24cd-44f2-8a41-86cdceeb41ee tempest-VolumesAdminNegativeTest-1387177911 tempest-VolumesAdminNegativeTest-1387177911-project-member] [instance: 4373b735-31cf-4b53-b655-38555cf212a5] Terminating instance [ 824.737054] env[61806]: DEBUG nova.compute.manager [None req-0b05357b-24cd-44f2-8a41-86cdceeb41ee tempest-VolumesAdminNegativeTest-1387177911 tempest-VolumesAdminNegativeTest-1387177911-project-member] [instance: 4373b735-31cf-4b53-b655-38555cf212a5] Start destroying the instance on the hypervisor. {{(pid=61806) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 824.737351] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-0b05357b-24cd-44f2-8a41-86cdceeb41ee tempest-VolumesAdminNegativeTest-1387177911 tempest-VolumesAdminNegativeTest-1387177911-project-member] [instance: 4373b735-31cf-4b53-b655-38555cf212a5] Destroying instance {{(pid=61806) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 824.738245] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-273a9bb1-34c0-4708-afab-bb16e06dce25 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 824.748629] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [None req-0b05357b-24cd-44f2-8a41-86cdceeb41ee tempest-VolumesAdminNegativeTest-1387177911 tempest-VolumesAdminNegativeTest-1387177911-project-member] [instance: 4373b735-31cf-4b53-b655-38555cf212a5] Powering off the VM {{(pid=61806) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 824.751587] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-06338b7b-9be6-4307-8249-cbf7136fc1ef {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 824.760914] env[61806]: DEBUG oslo_vmware.api [None req-0b05357b-24cd-44f2-8a41-86cdceeb41ee tempest-VolumesAdminNegativeTest-1387177911 tempest-VolumesAdminNegativeTest-1387177911-project-member] Waiting for the task: (returnval){ [ 824.760914] env[61806]: value = "task-1294639" [ 824.760914] env[61806]: _type = "Task" [ 824.760914] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 824.774409] env[61806]: DEBUG oslo_vmware.api [None req-0b05357b-24cd-44f2-8a41-86cdceeb41ee tempest-VolumesAdminNegativeTest-1387177911 tempest-VolumesAdminNegativeTest-1387177911-project-member] Task: {'id': task-1294639, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 824.823108] env[61806]: DEBUG oslo_vmware.api [None req-8aa43bc8-f134-4682-8a54-ce9aa25e4ee5 tempest-AttachVolumeNegativeTest-1683033704 tempest-AttachVolumeNegativeTest-1683033704-project-member] Task: {'id': task-1294634, 'name': Rename_Task, 'duration_secs': 1.225197} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 824.823416] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [None req-8aa43bc8-f134-4682-8a54-ce9aa25e4ee5 tempest-AttachVolumeNegativeTest-1683033704 tempest-AttachVolumeNegativeTest-1683033704-project-member] [instance: 43dcce0d-4fdb-4ee1-8309-92e92d2331a9] Powering on the VM {{(pid=61806) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 824.823669] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-7af9d928-5564-43d2-a3c9-6375c9986ad9 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 824.827913] env[61806]: DEBUG oslo_concurrency.lockutils [None req-7fa92f02-ad2b-4004-a9a7-cb08f83ec0f1 tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Releasing lock "refresh_cache-67d1bdd8-cf8f-4fb5-bd3c-42008e4d60e3" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 824.828284] env[61806]: DEBUG nova.compute.manager [None req-7fa92f02-ad2b-4004-a9a7-cb08f83ec0f1 tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] [instance: 67d1bdd8-cf8f-4fb5-bd3c-42008e4d60e3] Instance network_info: |[{"id": "859fb723-4a1c-4ea8-bd7c-0ed308f3a9e3", "address": "fa:16:3e:3c:40:d6", "network": {"id": "746a03b6-c1a6-4d27-98b9-0bee6376634f", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-1971382587-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c1f2d161c50f4b70987b90e6364c3779", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6a31b224-77d7-48a4-af87-312758611b8e", "external-id": "nsx-vlan-transportzone-761", "segmentation_id": 761, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap859fb723-4a", "ovs_interfaceid": "859fb723-4a1c-4ea8-bd7c-0ed308f3a9e3", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61806) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 824.828675] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-7fa92f02-ad2b-4004-a9a7-cb08f83ec0f1 tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] [instance: 67d1bdd8-cf8f-4fb5-bd3c-42008e4d60e3] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:3c:40:d6', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '6a31b224-77d7-48a4-af87-312758611b8e', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '859fb723-4a1c-4ea8-bd7c-0ed308f3a9e3', 'vif_model': 'vmxnet3'}] {{(pid=61806) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 824.836867] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [None req-7fa92f02-ad2b-4004-a9a7-cb08f83ec0f1 tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Creating folder: Project (c1f2d161c50f4b70987b90e6364c3779). Parent ref: group-v277609. {{(pid=61806) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 824.841767] env[61806]: DEBUG oslo_concurrency.lockutils [None req-cbb90e6d-7dac-4c30-b860-1f0db7dd091e tempest-MultipleCreateTestJSON-1589015462 tempest-MultipleCreateTestJSON-1589015462-project-member] Acquiring lock "refresh_cache-25218e83-c1ab-4b97-ade1-7c32b6f99305" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 824.841767] env[61806]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-e3105cb0-cbcb-4fce-85f1-3aa7e6844d40 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 824.842288] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ff070cba-975f-4a9f-84d5-ed98f8206ba3 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 824.849190] env[61806]: DEBUG oslo_vmware.api [None req-8aa43bc8-f134-4682-8a54-ce9aa25e4ee5 tempest-AttachVolumeNegativeTest-1683033704 tempest-AttachVolumeNegativeTest-1683033704-project-member] Waiting for the task: (returnval){ [ 824.849190] env[61806]: value = "task-1294640" [ 824.849190] env[61806]: _type = "Task" [ 824.849190] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 824.858997] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f7566224-26a6-41ad-a6c7-58b7258d1ac6 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 824.867046] env[61806]: INFO nova.virt.vmwareapi.vm_util [None req-7fa92f02-ad2b-4004-a9a7-cb08f83ec0f1 tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Created folder: Project (c1f2d161c50f4b70987b90e6364c3779) in parent group-v277609. [ 824.867046] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [None req-7fa92f02-ad2b-4004-a9a7-cb08f83ec0f1 tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Creating folder: Instances. Parent ref: group-v277664. {{(pid=61806) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 824.872535] env[61806]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-74f9ee4a-7ee8-4153-824d-32718a9e5fd3 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 824.872535] env[61806]: DEBUG oslo_vmware.api [None req-8aa43bc8-f134-4682-8a54-ce9aa25e4ee5 tempest-AttachVolumeNegativeTest-1683033704 tempest-AttachVolumeNegativeTest-1683033704-project-member] Task: {'id': task-1294640, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 824.906654] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-73f5898e-a9f7-4c88-8b9b-1d51d628151b {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 824.909881] env[61806]: INFO nova.virt.vmwareapi.vm_util [None req-7fa92f02-ad2b-4004-a9a7-cb08f83ec0f1 tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Created folder: Instances in parent group-v277664. [ 824.909881] env[61806]: DEBUG oslo.service.loopingcall [None req-7fa92f02-ad2b-4004-a9a7-cb08f83ec0f1 tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61806) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 824.910453] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 67d1bdd8-cf8f-4fb5-bd3c-42008e4d60e3] Creating VM on the ESX host {{(pid=61806) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 824.914098] env[61806]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-308b5381-111d-4547-8f25-89b6f4a02ad4 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 824.937147] env[61806]: DEBUG oslo_vmware.api [None req-81ee2b3f-4b48-4fda-b966-c7fd3249b12e tempest-ServerDiagnosticsV248Test-463927155 tempest-ServerDiagnosticsV248Test-463927155-project-member] Task: {'id': task-1294637, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.226085} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 824.938989] env[61806]: DEBUG nova.virt.vmwareapi.ds_util [None req-81ee2b3f-4b48-4fda-b966-c7fd3249b12e tempest-ServerDiagnosticsV248Test-463927155 tempest-ServerDiagnosticsV248Test-463927155-project-member] Deleted the datastore file {{(pid=61806) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 824.939186] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-81ee2b3f-4b48-4fda-b966-c7fd3249b12e tempest-ServerDiagnosticsV248Test-463927155 tempest-ServerDiagnosticsV248Test-463927155-project-member] [instance: a934d02d-26aa-4900-b473-a58489e5629e] Deleted contents of the VM from datastore datastore2 {{(pid=61806) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 824.939373] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-81ee2b3f-4b48-4fda-b966-c7fd3249b12e tempest-ServerDiagnosticsV248Test-463927155 tempest-ServerDiagnosticsV248Test-463927155-project-member] [instance: a934d02d-26aa-4900-b473-a58489e5629e] Instance destroyed {{(pid=61806) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 824.939554] env[61806]: INFO nova.compute.manager [None req-81ee2b3f-4b48-4fda-b966-c7fd3249b12e tempest-ServerDiagnosticsV248Test-463927155 tempest-ServerDiagnosticsV248Test-463927155-project-member] [instance: a934d02d-26aa-4900-b473-a58489e5629e] Took 1.14 seconds to destroy the instance on the hypervisor. [ 824.939961] env[61806]: DEBUG oslo.service.loopingcall [None req-81ee2b3f-4b48-4fda-b966-c7fd3249b12e tempest-ServerDiagnosticsV248Test-463927155 tempest-ServerDiagnosticsV248Test-463927155-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61806) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 824.941191] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d6f58dd0-ba7e-427d-8f7e-948666ceee0a {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 824.950429] env[61806]: DEBUG nova.compute.manager [-] [instance: a934d02d-26aa-4900-b473-a58489e5629e] Deallocating network for instance {{(pid=61806) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 824.950587] env[61806]: DEBUG nova.network.neutron [-] [instance: a934d02d-26aa-4900-b473-a58489e5629e] deallocate_for_instance() {{(pid=61806) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 824.952304] env[61806]: DEBUG oslo_vmware.api [None req-9c52c9e9-816e-4638-b13a-a63a9756140c tempest-SecurityGroupsTestJSON-1213485581 tempest-SecurityGroupsTestJSON-1213485581-project-member] Task: {'id': session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]5216735e-762f-907c-cd93-68cb23a8ac1c, 'name': SearchDatastore_Task, 'duration_secs': 0.011373} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 824.953486] env[61806]: DEBUG oslo_concurrency.lockutils [None req-9c52c9e9-816e-4638-b13a-a63a9756140c tempest-SecurityGroupsTestJSON-1213485581 tempest-SecurityGroupsTestJSON-1213485581-project-member] Releasing lock "[datastore2] devstack-image-cache_base/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa.vmdk" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 824.953830] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [None req-9c52c9e9-816e-4638-b13a-a63a9756140c tempest-SecurityGroupsTestJSON-1213485581 tempest-SecurityGroupsTestJSON-1213485581-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa.vmdk to [datastore2] b41531f2-f28c-4d82-9682-0b557bbaa491/b41531f2-f28c-4d82-9682-0b557bbaa491.vmdk {{(pid=61806) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 824.955136] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-1fb9d5d9-0faa-4e12-ac03-8d5fcc2994ab {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 824.965821] env[61806]: DEBUG nova.compute.provider_tree [None req-cbb90e6d-7dac-4c30-b860-1f0db7dd091e tempest-MultipleCreateTestJSON-1589015462 tempest-MultipleCreateTestJSON-1589015462-project-member] Inventory has not changed in ProviderTree for provider: a2858be1-fd22-4e08-979e-87ad25293407 {{(pid=61806) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 824.968837] env[61806]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 824.968837] env[61806]: value = "task-1294643" [ 824.968837] env[61806]: _type = "Task" [ 824.968837] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 824.972952] env[61806]: DEBUG nova.network.neutron [-] [instance: a934d02d-26aa-4900-b473-a58489e5629e] Instance cache missing network info. {{(pid=61806) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 824.977040] env[61806]: DEBUG oslo_vmware.api [None req-9c52c9e9-816e-4638-b13a-a63a9756140c tempest-SecurityGroupsTestJSON-1213485581 tempest-SecurityGroupsTestJSON-1213485581-project-member] Waiting for the task: (returnval){ [ 824.977040] env[61806]: value = "task-1294644" [ 824.977040] env[61806]: _type = "Task" [ 824.977040] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 824.985169] env[61806]: DEBUG nova.compute.manager [None req-4b919e4f-1f0b-40e0-b482-0547e22d55ff tempest-ImagesTestJSON-618852556 tempest-ImagesTestJSON-618852556-project-member] [instance: 28e0baab-8516-42e3-acc2-9b8eb5192f57] Starting instance... {{(pid=61806) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 824.988014] env[61806]: DEBUG oslo_vmware.api [-] Task: {'id': task-1294643, 'name': CreateVM_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 824.998548] env[61806]: DEBUG oslo_vmware.api [None req-9c52c9e9-816e-4638-b13a-a63a9756140c tempest-SecurityGroupsTestJSON-1213485581 tempest-SecurityGroupsTestJSON-1213485581-project-member] Task: {'id': task-1294644, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 825.033852] env[61806]: DEBUG oslo_vmware.api [None req-f4c6ec76-6b50-4a2c-bb86-09b71e318ce8 tempest-ServerShowV257Test-1159790492 tempest-ServerShowV257Test-1159790492-project-member] Task: {'id': task-1294638, 'name': PowerOffVM_Task, 'duration_secs': 0.291555} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 825.034339] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [None req-f4c6ec76-6b50-4a2c-bb86-09b71e318ce8 tempest-ServerShowV257Test-1159790492 tempest-ServerShowV257Test-1159790492-project-member] [instance: 616ec206-9804-469e-ab5c-41aea7f048aa] Powered off the VM {{(pid=61806) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 825.034628] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-f4c6ec76-6b50-4a2c-bb86-09b71e318ce8 tempest-ServerShowV257Test-1159790492 tempest-ServerShowV257Test-1159790492-project-member] [instance: 616ec206-9804-469e-ab5c-41aea7f048aa] Unregistering the VM {{(pid=61806) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 825.034949] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-e9d8dc81-aba1-491c-aaac-a556e2142547 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 825.066915] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-f4c6ec76-6b50-4a2c-bb86-09b71e318ce8 tempest-ServerShowV257Test-1159790492 tempest-ServerShowV257Test-1159790492-project-member] [instance: 616ec206-9804-469e-ab5c-41aea7f048aa] Unregistered the VM {{(pid=61806) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 825.067148] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-f4c6ec76-6b50-4a2c-bb86-09b71e318ce8 tempest-ServerShowV257Test-1159790492 tempest-ServerShowV257Test-1159790492-project-member] [instance: 616ec206-9804-469e-ab5c-41aea7f048aa] Deleting contents of the VM from datastore datastore2 {{(pid=61806) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 825.067363] env[61806]: DEBUG nova.virt.vmwareapi.ds_util [None req-f4c6ec76-6b50-4a2c-bb86-09b71e318ce8 tempest-ServerShowV257Test-1159790492 tempest-ServerShowV257Test-1159790492-project-member] Deleting the datastore file [datastore2] 616ec206-9804-469e-ab5c-41aea7f048aa {{(pid=61806) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 825.067660] env[61806]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-36d82afd-a39e-41e8-8717-ba0e094a177e {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 825.073614] env[61806]: DEBUG nova.network.neutron [req-fd4c6b15-6d49-4512-a701-36ff2780835c req-a1df8c6d-978a-40eb-93c3-1ed7a6109bfc service nova] [instance: 25218e83-c1ab-4b97-ade1-7c32b6f99305] Instance cache missing network info. {{(pid=61806) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 825.079244] env[61806]: DEBUG oslo_vmware.api [None req-f4c6ec76-6b50-4a2c-bb86-09b71e318ce8 tempest-ServerShowV257Test-1159790492 tempest-ServerShowV257Test-1159790492-project-member] Waiting for the task: (returnval){ [ 825.079244] env[61806]: value = "task-1294646" [ 825.079244] env[61806]: _type = "Task" [ 825.079244] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 825.085649] env[61806]: DEBUG oslo_vmware.api [None req-f4c6ec76-6b50-4a2c-bb86-09b71e318ce8 tempest-ServerShowV257Test-1159790492 tempest-ServerShowV257Test-1159790492-project-member] Task: {'id': task-1294646, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 825.166089] env[61806]: DEBUG nova.network.neutron [req-fd4c6b15-6d49-4512-a701-36ff2780835c req-a1df8c6d-978a-40eb-93c3-1ed7a6109bfc service nova] [instance: 25218e83-c1ab-4b97-ade1-7c32b6f99305] Updating instance_info_cache with network_info: [] {{(pid=61806) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 825.225615] env[61806]: DEBUG oslo_concurrency.lockutils [None req-90367f73-05d7-4a77-94a7-20cc6ea5d878 tempest-ServerAddressesTestJSON-661730579 tempest-ServerAddressesTestJSON-661730579-project-member] Acquiring lock "a842df44-d8a9-4376-b9fc-5ca19a68a4b7" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 825.226057] env[61806]: DEBUG oslo_concurrency.lockutils [None req-90367f73-05d7-4a77-94a7-20cc6ea5d878 tempest-ServerAddressesTestJSON-661730579 tempest-ServerAddressesTestJSON-661730579-project-member] Lock "a842df44-d8a9-4376-b9fc-5ca19a68a4b7" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 825.226311] env[61806]: DEBUG oslo_concurrency.lockutils [None req-90367f73-05d7-4a77-94a7-20cc6ea5d878 tempest-ServerAddressesTestJSON-661730579 tempest-ServerAddressesTestJSON-661730579-project-member] Acquiring lock "a842df44-d8a9-4376-b9fc-5ca19a68a4b7-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 825.226619] env[61806]: DEBUG oslo_concurrency.lockutils [None req-90367f73-05d7-4a77-94a7-20cc6ea5d878 tempest-ServerAddressesTestJSON-661730579 tempest-ServerAddressesTestJSON-661730579-project-member] Lock "a842df44-d8a9-4376-b9fc-5ca19a68a4b7-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 825.226909] env[61806]: DEBUG oslo_concurrency.lockutils [None req-90367f73-05d7-4a77-94a7-20cc6ea5d878 tempest-ServerAddressesTestJSON-661730579 tempest-ServerAddressesTestJSON-661730579-project-member] Lock "a842df44-d8a9-4376-b9fc-5ca19a68a4b7-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 825.229698] env[61806]: INFO nova.compute.manager [None req-90367f73-05d7-4a77-94a7-20cc6ea5d878 tempest-ServerAddressesTestJSON-661730579 tempest-ServerAddressesTestJSON-661730579-project-member] [instance: a842df44-d8a9-4376-b9fc-5ca19a68a4b7] Terminating instance [ 825.232132] env[61806]: DEBUG nova.compute.manager [None req-90367f73-05d7-4a77-94a7-20cc6ea5d878 tempest-ServerAddressesTestJSON-661730579 tempest-ServerAddressesTestJSON-661730579-project-member] [instance: a842df44-d8a9-4376-b9fc-5ca19a68a4b7] Start destroying the instance on the hypervisor. {{(pid=61806) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 825.232901] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-90367f73-05d7-4a77-94a7-20cc6ea5d878 tempest-ServerAddressesTestJSON-661730579 tempest-ServerAddressesTestJSON-661730579-project-member] [instance: a842df44-d8a9-4376-b9fc-5ca19a68a4b7] Destroying instance {{(pid=61806) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 825.233928] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7ab22b3f-789b-4922-baba-49aa9b3b7753 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 825.243986] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [None req-90367f73-05d7-4a77-94a7-20cc6ea5d878 tempest-ServerAddressesTestJSON-661730579 tempest-ServerAddressesTestJSON-661730579-project-member] [instance: a842df44-d8a9-4376-b9fc-5ca19a68a4b7] Powering off the VM {{(pid=61806) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 825.244383] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-487397f0-841a-4e31-8414-604f28b6bcfb {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 825.252789] env[61806]: DEBUG oslo_vmware.api [None req-90367f73-05d7-4a77-94a7-20cc6ea5d878 tempest-ServerAddressesTestJSON-661730579 tempest-ServerAddressesTestJSON-661730579-project-member] Waiting for the task: (returnval){ [ 825.252789] env[61806]: value = "task-1294647" [ 825.252789] env[61806]: _type = "Task" [ 825.252789] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 825.262659] env[61806]: DEBUG oslo_vmware.api [None req-90367f73-05d7-4a77-94a7-20cc6ea5d878 tempest-ServerAddressesTestJSON-661730579 tempest-ServerAddressesTestJSON-661730579-project-member] Task: {'id': task-1294647, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 825.272945] env[61806]: DEBUG oslo_vmware.api [None req-0b05357b-24cd-44f2-8a41-86cdceeb41ee tempest-VolumesAdminNegativeTest-1387177911 tempest-VolumesAdminNegativeTest-1387177911-project-member] Task: {'id': task-1294639, 'name': PowerOffVM_Task, 'duration_secs': 0.237824} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 825.273069] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [None req-0b05357b-24cd-44f2-8a41-86cdceeb41ee tempest-VolumesAdminNegativeTest-1387177911 tempest-VolumesAdminNegativeTest-1387177911-project-member] [instance: 4373b735-31cf-4b53-b655-38555cf212a5] Powered off the VM {{(pid=61806) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 825.273911] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-0b05357b-24cd-44f2-8a41-86cdceeb41ee tempest-VolumesAdminNegativeTest-1387177911 tempest-VolumesAdminNegativeTest-1387177911-project-member] [instance: 4373b735-31cf-4b53-b655-38555cf212a5] Unregistering the VM {{(pid=61806) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 825.274434] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-8119cbe1-98f7-436b-a4f6-c29a93058753 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 825.364023] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-0b05357b-24cd-44f2-8a41-86cdceeb41ee tempest-VolumesAdminNegativeTest-1387177911 tempest-VolumesAdminNegativeTest-1387177911-project-member] [instance: 4373b735-31cf-4b53-b655-38555cf212a5] Unregistered the VM {{(pid=61806) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 825.364023] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-0b05357b-24cd-44f2-8a41-86cdceeb41ee tempest-VolumesAdminNegativeTest-1387177911 tempest-VolumesAdminNegativeTest-1387177911-project-member] [instance: 4373b735-31cf-4b53-b655-38555cf212a5] Deleting contents of the VM from datastore datastore1 {{(pid=61806) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 825.364023] env[61806]: DEBUG nova.virt.vmwareapi.ds_util [None req-0b05357b-24cd-44f2-8a41-86cdceeb41ee tempest-VolumesAdminNegativeTest-1387177911 tempest-VolumesAdminNegativeTest-1387177911-project-member] Deleting the datastore file [datastore1] 4373b735-31cf-4b53-b655-38555cf212a5 {{(pid=61806) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 825.364023] env[61806]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-3249396e-14de-421f-b301-a679e3926db9 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 825.374348] env[61806]: DEBUG oslo_vmware.api [None req-8aa43bc8-f134-4682-8a54-ce9aa25e4ee5 tempest-AttachVolumeNegativeTest-1683033704 tempest-AttachVolumeNegativeTest-1683033704-project-member] Task: {'id': task-1294640, 'name': PowerOnVM_Task} progress is 89%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 825.382221] env[61806]: DEBUG oslo_vmware.api [None req-0b05357b-24cd-44f2-8a41-86cdceeb41ee tempest-VolumesAdminNegativeTest-1387177911 tempest-VolumesAdminNegativeTest-1387177911-project-member] Waiting for the task: (returnval){ [ 825.382221] env[61806]: value = "task-1294649" [ 825.382221] env[61806]: _type = "Task" [ 825.382221] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 825.396179] env[61806]: DEBUG oslo_vmware.api [None req-0b05357b-24cd-44f2-8a41-86cdceeb41ee tempest-VolumesAdminNegativeTest-1387177911 tempest-VolumesAdminNegativeTest-1387177911-project-member] Task: {'id': task-1294649, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 825.470963] env[61806]: DEBUG nova.scheduler.client.report [None req-cbb90e6d-7dac-4c30-b860-1f0db7dd091e tempest-MultipleCreateTestJSON-1589015462 tempest-MultipleCreateTestJSON-1589015462-project-member] Inventory has not changed for provider a2858be1-fd22-4e08-979e-87ad25293407 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 138, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61806) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 825.480250] env[61806]: DEBUG nova.network.neutron [-] [instance: a934d02d-26aa-4900-b473-a58489e5629e] Updating instance_info_cache with network_info: [] {{(pid=61806) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 825.513464] env[61806]: DEBUG oslo_vmware.api [-] Task: {'id': task-1294643, 'name': CreateVM_Task} progress is 99%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 825.514901] env[61806]: DEBUG oslo_vmware.api [None req-9c52c9e9-816e-4638-b13a-a63a9756140c tempest-SecurityGroupsTestJSON-1213485581 tempest-SecurityGroupsTestJSON-1213485581-project-member] Task: {'id': task-1294644, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 825.516323] env[61806]: DEBUG oslo_concurrency.lockutils [None req-4b919e4f-1f0b-40e0-b482-0547e22d55ff tempest-ImagesTestJSON-618852556 tempest-ImagesTestJSON-618852556-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 825.589204] env[61806]: DEBUG oslo_vmware.api [None req-f4c6ec76-6b50-4a2c-bb86-09b71e318ce8 tempest-ServerShowV257Test-1159790492 tempest-ServerShowV257Test-1159790492-project-member] Task: {'id': task-1294646, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 825.669948] env[61806]: DEBUG oslo_concurrency.lockutils [req-fd4c6b15-6d49-4512-a701-36ff2780835c req-a1df8c6d-978a-40eb-93c3-1ed7a6109bfc service nova] Releasing lock "refresh_cache-25218e83-c1ab-4b97-ade1-7c32b6f99305" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 825.669948] env[61806]: DEBUG oslo_concurrency.lockutils [None req-cbb90e6d-7dac-4c30-b860-1f0db7dd091e tempest-MultipleCreateTestJSON-1589015462 tempest-MultipleCreateTestJSON-1589015462-project-member] Acquired lock "refresh_cache-25218e83-c1ab-4b97-ade1-7c32b6f99305" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 825.670148] env[61806]: DEBUG nova.network.neutron [None req-cbb90e6d-7dac-4c30-b860-1f0db7dd091e tempest-MultipleCreateTestJSON-1589015462 tempest-MultipleCreateTestJSON-1589015462-project-member] [instance: 25218e83-c1ab-4b97-ade1-7c32b6f99305] Building network info cache for instance {{(pid=61806) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 825.765394] env[61806]: DEBUG oslo_vmware.api [None req-90367f73-05d7-4a77-94a7-20cc6ea5d878 tempest-ServerAddressesTestJSON-661730579 tempest-ServerAddressesTestJSON-661730579-project-member] Task: {'id': task-1294647, 'name': PowerOffVM_Task, 'duration_secs': 0.385055} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 825.767015] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [None req-90367f73-05d7-4a77-94a7-20cc6ea5d878 tempest-ServerAddressesTestJSON-661730579 tempest-ServerAddressesTestJSON-661730579-project-member] [instance: a842df44-d8a9-4376-b9fc-5ca19a68a4b7] Powered off the VM {{(pid=61806) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 825.767015] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-90367f73-05d7-4a77-94a7-20cc6ea5d878 tempest-ServerAddressesTestJSON-661730579 tempest-ServerAddressesTestJSON-661730579-project-member] [instance: a842df44-d8a9-4376-b9fc-5ca19a68a4b7] Unregistering the VM {{(pid=61806) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 825.767015] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-eabe4bb1-4cdc-4673-b45c-bbff0c38c19d {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 825.841097] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-90367f73-05d7-4a77-94a7-20cc6ea5d878 tempest-ServerAddressesTestJSON-661730579 tempest-ServerAddressesTestJSON-661730579-project-member] [instance: a842df44-d8a9-4376-b9fc-5ca19a68a4b7] Unregistered the VM {{(pid=61806) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 825.841895] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-90367f73-05d7-4a77-94a7-20cc6ea5d878 tempest-ServerAddressesTestJSON-661730579 tempest-ServerAddressesTestJSON-661730579-project-member] [instance: a842df44-d8a9-4376-b9fc-5ca19a68a4b7] Deleting contents of the VM from datastore datastore1 {{(pid=61806) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 825.842269] env[61806]: DEBUG nova.virt.vmwareapi.ds_util [None req-90367f73-05d7-4a77-94a7-20cc6ea5d878 tempest-ServerAddressesTestJSON-661730579 tempest-ServerAddressesTestJSON-661730579-project-member] Deleting the datastore file [datastore1] a842df44-d8a9-4376-b9fc-5ca19a68a4b7 {{(pid=61806) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 825.842635] env[61806]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-153a4dd1-d02e-462a-8c96-a85ea9b7ddc9 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 825.850788] env[61806]: DEBUG oslo_vmware.api [None req-90367f73-05d7-4a77-94a7-20cc6ea5d878 tempest-ServerAddressesTestJSON-661730579 tempest-ServerAddressesTestJSON-661730579-project-member] Waiting for the task: (returnval){ [ 825.850788] env[61806]: value = "task-1294651" [ 825.850788] env[61806]: _type = "Task" [ 825.850788] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 825.862838] env[61806]: DEBUG oslo_vmware.api [None req-8aa43bc8-f134-4682-8a54-ce9aa25e4ee5 tempest-AttachVolumeNegativeTest-1683033704 tempest-AttachVolumeNegativeTest-1683033704-project-member] Task: {'id': task-1294640, 'name': PowerOnVM_Task, 'duration_secs': 0.753485} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 825.865818] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [None req-8aa43bc8-f134-4682-8a54-ce9aa25e4ee5 tempest-AttachVolumeNegativeTest-1683033704 tempest-AttachVolumeNegativeTest-1683033704-project-member] [instance: 43dcce0d-4fdb-4ee1-8309-92e92d2331a9] Powered on the VM {{(pid=61806) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 825.866071] env[61806]: INFO nova.compute.manager [None req-8aa43bc8-f134-4682-8a54-ce9aa25e4ee5 tempest-AttachVolumeNegativeTest-1683033704 tempest-AttachVolumeNegativeTest-1683033704-project-member] [instance: 43dcce0d-4fdb-4ee1-8309-92e92d2331a9] Took 9.66 seconds to spawn the instance on the hypervisor. [ 825.866320] env[61806]: DEBUG nova.compute.manager [None req-8aa43bc8-f134-4682-8a54-ce9aa25e4ee5 tempest-AttachVolumeNegativeTest-1683033704 tempest-AttachVolumeNegativeTest-1683033704-project-member] [instance: 43dcce0d-4fdb-4ee1-8309-92e92d2331a9] Checking state {{(pid=61806) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 825.866682] env[61806]: DEBUG oslo_vmware.api [None req-90367f73-05d7-4a77-94a7-20cc6ea5d878 tempest-ServerAddressesTestJSON-661730579 tempest-ServerAddressesTestJSON-661730579-project-member] Task: {'id': task-1294651, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 825.867337] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5f12f7aa-a201-40b6-803c-f92969bbe07c {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 825.894117] env[61806]: DEBUG oslo_vmware.api [None req-0b05357b-24cd-44f2-8a41-86cdceeb41ee tempest-VolumesAdminNegativeTest-1387177911 tempest-VolumesAdminNegativeTest-1387177911-project-member] Task: {'id': task-1294649, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.45593} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 825.894385] env[61806]: DEBUG nova.virt.vmwareapi.ds_util [None req-0b05357b-24cd-44f2-8a41-86cdceeb41ee tempest-VolumesAdminNegativeTest-1387177911 tempest-VolumesAdminNegativeTest-1387177911-project-member] Deleted the datastore file {{(pid=61806) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 825.895103] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-0b05357b-24cd-44f2-8a41-86cdceeb41ee tempest-VolumesAdminNegativeTest-1387177911 tempest-VolumesAdminNegativeTest-1387177911-project-member] [instance: 4373b735-31cf-4b53-b655-38555cf212a5] Deleted contents of the VM from datastore datastore1 {{(pid=61806) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 825.895103] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-0b05357b-24cd-44f2-8a41-86cdceeb41ee tempest-VolumesAdminNegativeTest-1387177911 tempest-VolumesAdminNegativeTest-1387177911-project-member] [instance: 4373b735-31cf-4b53-b655-38555cf212a5] Instance destroyed {{(pid=61806) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 825.895103] env[61806]: INFO nova.compute.manager [None req-0b05357b-24cd-44f2-8a41-86cdceeb41ee tempest-VolumesAdminNegativeTest-1387177911 tempest-VolumesAdminNegativeTest-1387177911-project-member] [instance: 4373b735-31cf-4b53-b655-38555cf212a5] Took 1.16 seconds to destroy the instance on the hypervisor. [ 825.895221] env[61806]: DEBUG oslo.service.loopingcall [None req-0b05357b-24cd-44f2-8a41-86cdceeb41ee tempest-VolumesAdminNegativeTest-1387177911 tempest-VolumesAdminNegativeTest-1387177911-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61806) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 825.895386] env[61806]: DEBUG nova.compute.manager [-] [instance: 4373b735-31cf-4b53-b655-38555cf212a5] Deallocating network for instance {{(pid=61806) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 825.895481] env[61806]: DEBUG nova.network.neutron [-] [instance: 4373b735-31cf-4b53-b655-38555cf212a5] deallocate_for_instance() {{(pid=61806) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 825.981739] env[61806]: DEBUG oslo_concurrency.lockutils [None req-cbb90e6d-7dac-4c30-b860-1f0db7dd091e tempest-MultipleCreateTestJSON-1589015462 tempest-MultipleCreateTestJSON-1589015462-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.441s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 825.981862] env[61806]: DEBUG nova.compute.manager [None req-cbb90e6d-7dac-4c30-b860-1f0db7dd091e tempest-MultipleCreateTestJSON-1589015462 tempest-MultipleCreateTestJSON-1589015462-project-member] [instance: ef502a7d-7dc0-4319-8ce3-6a8701288237] Start building networks asynchronously for instance. {{(pid=61806) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 825.984972] env[61806]: DEBUG oslo_concurrency.lockutils [None req-0ea3a38f-3871-4412-a6bc-d21c2086d262 tempest-AttachInterfacesV270Test-1974652806 tempest-AttachInterfacesV270Test-1974652806-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 13.525s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 825.986441] env[61806]: INFO nova.compute.claims [None req-0ea3a38f-3871-4412-a6bc-d21c2086d262 tempest-AttachInterfacesV270Test-1974652806 tempest-AttachInterfacesV270Test-1974652806-project-member] [instance: 117139cd-4ed4-4bea-8ad3-f4ac32c692e4] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 825.990168] env[61806]: INFO nova.compute.manager [-] [instance: a934d02d-26aa-4900-b473-a58489e5629e] Took 1.04 seconds to deallocate network for instance. [ 825.998618] env[61806]: DEBUG oslo_vmware.api [-] Task: {'id': task-1294643, 'name': CreateVM_Task, 'duration_secs': 0.705353} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 826.002415] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 67d1bdd8-cf8f-4fb5-bd3c-42008e4d60e3] Created VM on the ESX host {{(pid=61806) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 826.003973] env[61806]: DEBUG oslo_concurrency.lockutils [None req-7fa92f02-ad2b-4004-a9a7-cb08f83ec0f1 tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 826.003973] env[61806]: DEBUG oslo_concurrency.lockutils [None req-7fa92f02-ad2b-4004-a9a7-cb08f83ec0f1 tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Acquired lock "[datastore1] devstack-image-cache_base/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 826.003973] env[61806]: DEBUG oslo_concurrency.lockutils [None req-7fa92f02-ad2b-4004-a9a7-cb08f83ec0f1 tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 826.004283] env[61806]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f6cab7ce-d6fb-414b-b16c-cbe7a40dd3ef {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 826.009219] env[61806]: DEBUG oslo_vmware.api [None req-9c52c9e9-816e-4638-b13a-a63a9756140c tempest-SecurityGroupsTestJSON-1213485581 tempest-SecurityGroupsTestJSON-1213485581-project-member] Task: {'id': task-1294644, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.628469} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 826.009843] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [None req-9c52c9e9-816e-4638-b13a-a63a9756140c tempest-SecurityGroupsTestJSON-1213485581 tempest-SecurityGroupsTestJSON-1213485581-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa.vmdk to [datastore2] b41531f2-f28c-4d82-9682-0b557bbaa491/b41531f2-f28c-4d82-9682-0b557bbaa491.vmdk {{(pid=61806) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 826.010247] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-9c52c9e9-816e-4638-b13a-a63a9756140c tempest-SecurityGroupsTestJSON-1213485581 tempest-SecurityGroupsTestJSON-1213485581-project-member] [instance: b41531f2-f28c-4d82-9682-0b557bbaa491] Extending root virtual disk to 1048576 {{(pid=61806) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 826.010447] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-11c517f8-3915-490e-82db-b98d153e1f96 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 826.013639] env[61806]: DEBUG oslo_vmware.api [None req-7fa92f02-ad2b-4004-a9a7-cb08f83ec0f1 tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Waiting for the task: (returnval){ [ 826.013639] env[61806]: value = "session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]525c2a95-4d4e-4b78-583b-80a927fd1ad5" [ 826.013639] env[61806]: _type = "Task" [ 826.013639] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 826.020956] env[61806]: DEBUG oslo_vmware.api [None req-9c52c9e9-816e-4638-b13a-a63a9756140c tempest-SecurityGroupsTestJSON-1213485581 tempest-SecurityGroupsTestJSON-1213485581-project-member] Waiting for the task: (returnval){ [ 826.020956] env[61806]: value = "task-1294652" [ 826.020956] env[61806]: _type = "Task" [ 826.020956] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 826.025144] env[61806]: DEBUG oslo_vmware.api [None req-7fa92f02-ad2b-4004-a9a7-cb08f83ec0f1 tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Task: {'id': session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]525c2a95-4d4e-4b78-583b-80a927fd1ad5, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 826.033891] env[61806]: DEBUG oslo_vmware.api [None req-9c52c9e9-816e-4638-b13a-a63a9756140c tempest-SecurityGroupsTestJSON-1213485581 tempest-SecurityGroupsTestJSON-1213485581-project-member] Task: {'id': task-1294652, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 826.087969] env[61806]: DEBUG oslo_vmware.api [None req-f4c6ec76-6b50-4a2c-bb86-09b71e318ce8 tempest-ServerShowV257Test-1159790492 tempest-ServerShowV257Test-1159790492-project-member] Task: {'id': task-1294646, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.525456} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 826.087969] env[61806]: DEBUG nova.virt.vmwareapi.ds_util [None req-f4c6ec76-6b50-4a2c-bb86-09b71e318ce8 tempest-ServerShowV257Test-1159790492 tempest-ServerShowV257Test-1159790492-project-member] Deleted the datastore file {{(pid=61806) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 826.087969] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-f4c6ec76-6b50-4a2c-bb86-09b71e318ce8 tempest-ServerShowV257Test-1159790492 tempest-ServerShowV257Test-1159790492-project-member] [instance: 616ec206-9804-469e-ab5c-41aea7f048aa] Deleted contents of the VM from datastore datastore2 {{(pid=61806) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 826.088142] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-f4c6ec76-6b50-4a2c-bb86-09b71e318ce8 tempest-ServerShowV257Test-1159790492 tempest-ServerShowV257Test-1159790492-project-member] [instance: 616ec206-9804-469e-ab5c-41aea7f048aa] Instance destroyed {{(pid=61806) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 826.088331] env[61806]: INFO nova.compute.manager [None req-f4c6ec76-6b50-4a2c-bb86-09b71e318ce8 tempest-ServerShowV257Test-1159790492 tempest-ServerShowV257Test-1159790492-project-member] [instance: 616ec206-9804-469e-ab5c-41aea7f048aa] Took 1.61 seconds to destroy the instance on the hypervisor. [ 826.088481] env[61806]: DEBUG oslo.service.loopingcall [None req-f4c6ec76-6b50-4a2c-bb86-09b71e318ce8 tempest-ServerShowV257Test-1159790492 tempest-ServerShowV257Test-1159790492-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61806) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 826.088676] env[61806]: DEBUG nova.compute.manager [-] [instance: 616ec206-9804-469e-ab5c-41aea7f048aa] Deallocating network for instance {{(pid=61806) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 826.088768] env[61806]: DEBUG nova.network.neutron [-] [instance: 616ec206-9804-469e-ab5c-41aea7f048aa] deallocate_for_instance() {{(pid=61806) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 826.112709] env[61806]: DEBUG nova.network.neutron [-] [instance: 616ec206-9804-469e-ab5c-41aea7f048aa] Instance cache missing network info. {{(pid=61806) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 826.259302] env[61806]: DEBUG nova.network.neutron [None req-cbb90e6d-7dac-4c30-b860-1f0db7dd091e tempest-MultipleCreateTestJSON-1589015462 tempest-MultipleCreateTestJSON-1589015462-project-member] [instance: 25218e83-c1ab-4b97-ade1-7c32b6f99305] Instance cache missing network info. {{(pid=61806) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 826.365251] env[61806]: DEBUG oslo_vmware.api [None req-90367f73-05d7-4a77-94a7-20cc6ea5d878 tempest-ServerAddressesTestJSON-661730579 tempest-ServerAddressesTestJSON-661730579-project-member] Task: {'id': task-1294651, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 826.400385] env[61806]: INFO nova.compute.manager [None req-8aa43bc8-f134-4682-8a54-ce9aa25e4ee5 tempest-AttachVolumeNegativeTest-1683033704 tempest-AttachVolumeNegativeTest-1683033704-project-member] [instance: 43dcce0d-4fdb-4ee1-8309-92e92d2331a9] Took 27.07 seconds to build instance. [ 826.494297] env[61806]: DEBUG nova.compute.utils [None req-cbb90e6d-7dac-4c30-b860-1f0db7dd091e tempest-MultipleCreateTestJSON-1589015462 tempest-MultipleCreateTestJSON-1589015462-project-member] Using /dev/sd instead of None {{(pid=61806) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 826.503979] env[61806]: DEBUG nova.compute.manager [None req-cbb90e6d-7dac-4c30-b860-1f0db7dd091e tempest-MultipleCreateTestJSON-1589015462 tempest-MultipleCreateTestJSON-1589015462-project-member] [instance: ef502a7d-7dc0-4319-8ce3-6a8701288237] Allocating IP information in the background. {{(pid=61806) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 826.503979] env[61806]: DEBUG nova.network.neutron [None req-cbb90e6d-7dac-4c30-b860-1f0db7dd091e tempest-MultipleCreateTestJSON-1589015462 tempest-MultipleCreateTestJSON-1589015462-project-member] [instance: ef502a7d-7dc0-4319-8ce3-6a8701288237] allocate_for_instance() {{(pid=61806) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 826.505658] env[61806]: DEBUG oslo_concurrency.lockutils [None req-81ee2b3f-4b48-4fda-b966-c7fd3249b12e tempest-ServerDiagnosticsV248Test-463927155 tempest-ServerDiagnosticsV248Test-463927155-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 826.530419] env[61806]: DEBUG oslo_vmware.api [None req-7fa92f02-ad2b-4004-a9a7-cb08f83ec0f1 tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Task: {'id': session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]525c2a95-4d4e-4b78-583b-80a927fd1ad5, 'name': SearchDatastore_Task, 'duration_secs': 0.013809} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 826.536356] env[61806]: DEBUG oslo_concurrency.lockutils [None req-7fa92f02-ad2b-4004-a9a7-cb08f83ec0f1 tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Releasing lock "[datastore1] devstack-image-cache_base/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 826.536744] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-7fa92f02-ad2b-4004-a9a7-cb08f83ec0f1 tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] [instance: 67d1bdd8-cf8f-4fb5-bd3c-42008e4d60e3] Processing image ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa {{(pid=61806) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 826.537364] env[61806]: DEBUG oslo_concurrency.lockutils [None req-7fa92f02-ad2b-4004-a9a7-cb08f83ec0f1 tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa.vmdk" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 826.537364] env[61806]: DEBUG oslo_concurrency.lockutils [None req-7fa92f02-ad2b-4004-a9a7-cb08f83ec0f1 tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Acquired lock "[datastore1] devstack-image-cache_base/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa.vmdk" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 826.537569] env[61806]: DEBUG nova.virt.vmwareapi.ds_util [None req-7fa92f02-ad2b-4004-a9a7-cb08f83ec0f1 tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61806) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 826.538058] env[61806]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-bd9b3276-f482-401d-9b27-14d8bca5dff6 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 826.543810] env[61806]: DEBUG nova.compute.manager [req-45bb828c-3c67-42f2-9d45-b0ce633fbb71 req-5b620900-c107-4df3-a352-b4c96533441f service nova] [instance: 4373b735-31cf-4b53-b655-38555cf212a5] Received event network-vif-deleted-2df33c67-a8f5-4af4-ae74-29dd266d5054 {{(pid=61806) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 826.544087] env[61806]: INFO nova.compute.manager [req-45bb828c-3c67-42f2-9d45-b0ce633fbb71 req-5b620900-c107-4df3-a352-b4c96533441f service nova] [instance: 4373b735-31cf-4b53-b655-38555cf212a5] Neutron deleted interface 2df33c67-a8f5-4af4-ae74-29dd266d5054; detaching it from the instance and deleting it from the info cache [ 826.544363] env[61806]: DEBUG nova.network.neutron [req-45bb828c-3c67-42f2-9d45-b0ce633fbb71 req-5b620900-c107-4df3-a352-b4c96533441f service nova] [instance: 4373b735-31cf-4b53-b655-38555cf212a5] Updating instance_info_cache with network_info: [] {{(pid=61806) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 826.551527] env[61806]: DEBUG oslo_vmware.api [None req-9c52c9e9-816e-4638-b13a-a63a9756140c tempest-SecurityGroupsTestJSON-1213485581 tempest-SecurityGroupsTestJSON-1213485581-project-member] Task: {'id': task-1294652, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.069629} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 826.552833] env[61806]: DEBUG nova.network.neutron [None req-cbb90e6d-7dac-4c30-b860-1f0db7dd091e tempest-MultipleCreateTestJSON-1589015462 tempest-MultipleCreateTestJSON-1589015462-project-member] [instance: 25218e83-c1ab-4b97-ade1-7c32b6f99305] Updating instance_info_cache with network_info: [{"id": "c8809e29-ea3c-4ea0-bee1-0731daee293d", "address": "fa:16:3e:6d:0c:29", "network": {"id": "56fd57d0-aec6-497c-806f-868ec544c85f", "bridge": "br-int", "label": "tempest-MultipleCreateTestJSON-480554183-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "dc40805c39f34d07be9d07861ad5841d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "359850cc-b061-4c9c-a61c-eb42e0f7c359", "external-id": "nsx-vlan-transportzone-113", "segmentation_id": 113, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc8809e29-ea", "ovs_interfaceid": "c8809e29-ea3c-4ea0-bee1-0731daee293d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61806) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 826.553944] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-9c52c9e9-816e-4638-b13a-a63a9756140c tempest-SecurityGroupsTestJSON-1213485581 tempest-SecurityGroupsTestJSON-1213485581-project-member] [instance: b41531f2-f28c-4d82-9682-0b557bbaa491] Extended root virtual disk {{(pid=61806) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 826.556093] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-962b1904-eb81-411f-855f-d292ba6ffdee {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 826.562015] env[61806]: DEBUG nova.virt.vmwareapi.ds_util [None req-7fa92f02-ad2b-4004-a9a7-cb08f83ec0f1 tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61806) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 826.562912] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-7fa92f02-ad2b-4004-a9a7-cb08f83ec0f1 tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61806) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 826.563963] env[61806]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-eb948612-8932-4240-b646-0937e8842ad3 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 826.586582] env[61806]: DEBUG nova.policy [None req-cbb90e6d-7dac-4c30-b860-1f0db7dd091e tempest-MultipleCreateTestJSON-1589015462 tempest-MultipleCreateTestJSON-1589015462-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'b5a92ad3d119419085f56d7dea0975f5', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'dc40805c39f34d07be9d07861ad5841d', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61806) authorize /opt/stack/nova/nova/policy.py:201}} [ 826.600313] env[61806]: DEBUG nova.virt.vmwareapi.volumeops [None req-9c52c9e9-816e-4638-b13a-a63a9756140c tempest-SecurityGroupsTestJSON-1213485581 tempest-SecurityGroupsTestJSON-1213485581-project-member] [instance: b41531f2-f28c-4d82-9682-0b557bbaa491] Reconfiguring VM instance instance-00000040 to attach disk [datastore2] b41531f2-f28c-4d82-9682-0b557bbaa491/b41531f2-f28c-4d82-9682-0b557bbaa491.vmdk or device None with type sparse {{(pid=61806) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 826.601354] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-bee23eea-81eb-402b-af53-9a02957b326e {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 826.618035] env[61806]: DEBUG nova.network.neutron [-] [instance: 616ec206-9804-469e-ab5c-41aea7f048aa] Updating instance_info_cache with network_info: [] {{(pid=61806) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 826.619382] env[61806]: DEBUG oslo_vmware.api [None req-7fa92f02-ad2b-4004-a9a7-cb08f83ec0f1 tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Waiting for the task: (returnval){ [ 826.619382] env[61806]: value = "session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]52e92c46-39ba-7c22-b566-705c94e40478" [ 826.619382] env[61806]: _type = "Task" [ 826.619382] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 826.625227] env[61806]: DEBUG oslo_vmware.api [None req-9c52c9e9-816e-4638-b13a-a63a9756140c tempest-SecurityGroupsTestJSON-1213485581 tempest-SecurityGroupsTestJSON-1213485581-project-member] Waiting for the task: (returnval){ [ 826.625227] env[61806]: value = "task-1294653" [ 826.625227] env[61806]: _type = "Task" [ 826.625227] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 826.628629] env[61806]: DEBUG oslo_vmware.api [None req-7fa92f02-ad2b-4004-a9a7-cb08f83ec0f1 tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Task: {'id': session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]52e92c46-39ba-7c22-b566-705c94e40478, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 826.637633] env[61806]: DEBUG oslo_vmware.api [None req-9c52c9e9-816e-4638-b13a-a63a9756140c tempest-SecurityGroupsTestJSON-1213485581 tempest-SecurityGroupsTestJSON-1213485581-project-member] Task: {'id': task-1294653, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 826.865670] env[61806]: DEBUG oslo_vmware.api [None req-90367f73-05d7-4a77-94a7-20cc6ea5d878 tempest-ServerAddressesTestJSON-661730579 tempest-ServerAddressesTestJSON-661730579-project-member] Task: {'id': task-1294651, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.581558} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 826.865956] env[61806]: DEBUG nova.virt.vmwareapi.ds_util [None req-90367f73-05d7-4a77-94a7-20cc6ea5d878 tempest-ServerAddressesTestJSON-661730579 tempest-ServerAddressesTestJSON-661730579-project-member] Deleted the datastore file {{(pid=61806) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 826.866199] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-90367f73-05d7-4a77-94a7-20cc6ea5d878 tempest-ServerAddressesTestJSON-661730579 tempest-ServerAddressesTestJSON-661730579-project-member] [instance: a842df44-d8a9-4376-b9fc-5ca19a68a4b7] Deleted contents of the VM from datastore datastore1 {{(pid=61806) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 826.866388] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-90367f73-05d7-4a77-94a7-20cc6ea5d878 tempest-ServerAddressesTestJSON-661730579 tempest-ServerAddressesTestJSON-661730579-project-member] [instance: a842df44-d8a9-4376-b9fc-5ca19a68a4b7] Instance destroyed {{(pid=61806) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 826.866563] env[61806]: INFO nova.compute.manager [None req-90367f73-05d7-4a77-94a7-20cc6ea5d878 tempest-ServerAddressesTestJSON-661730579 tempest-ServerAddressesTestJSON-661730579-project-member] [instance: a842df44-d8a9-4376-b9fc-5ca19a68a4b7] Took 1.63 seconds to destroy the instance on the hypervisor. [ 826.866915] env[61806]: DEBUG oslo.service.loopingcall [None req-90367f73-05d7-4a77-94a7-20cc6ea5d878 tempest-ServerAddressesTestJSON-661730579 tempest-ServerAddressesTestJSON-661730579-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61806) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 826.867153] env[61806]: DEBUG nova.compute.manager [-] [instance: a842df44-d8a9-4376-b9fc-5ca19a68a4b7] Deallocating network for instance {{(pid=61806) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 826.867244] env[61806]: DEBUG nova.network.neutron [-] [instance: a842df44-d8a9-4376-b9fc-5ca19a68a4b7] deallocate_for_instance() {{(pid=61806) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 826.903279] env[61806]: DEBUG oslo_concurrency.lockutils [None req-8aa43bc8-f134-4682-8a54-ce9aa25e4ee5 tempest-AttachVolumeNegativeTest-1683033704 tempest-AttachVolumeNegativeTest-1683033704-project-member] Lock "43dcce0d-4fdb-4ee1-8309-92e92d2331a9" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 98.634s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 826.933854] env[61806]: DEBUG nova.network.neutron [-] [instance: 4373b735-31cf-4b53-b655-38555cf212a5] Updating instance_info_cache with network_info: [] {{(pid=61806) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 827.002064] env[61806]: DEBUG nova.compute.manager [None req-cbb90e6d-7dac-4c30-b860-1f0db7dd091e tempest-MultipleCreateTestJSON-1589015462 tempest-MultipleCreateTestJSON-1589015462-project-member] [instance: ef502a7d-7dc0-4319-8ce3-6a8701288237] Start building block device mappings for instance. {{(pid=61806) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 827.052841] env[61806]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-66cc5ecf-750c-46b7-8b45-f31f3d3a4028 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 827.057533] env[61806]: DEBUG oslo_concurrency.lockutils [None req-cbb90e6d-7dac-4c30-b860-1f0db7dd091e tempest-MultipleCreateTestJSON-1589015462 tempest-MultipleCreateTestJSON-1589015462-project-member] Releasing lock "refresh_cache-25218e83-c1ab-4b97-ade1-7c32b6f99305" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 827.057533] env[61806]: DEBUG nova.compute.manager [None req-cbb90e6d-7dac-4c30-b860-1f0db7dd091e tempest-MultipleCreateTestJSON-1589015462 tempest-MultipleCreateTestJSON-1589015462-project-member] [instance: 25218e83-c1ab-4b97-ade1-7c32b6f99305] Instance network_info: |[{"id": "c8809e29-ea3c-4ea0-bee1-0731daee293d", "address": "fa:16:3e:6d:0c:29", "network": {"id": "56fd57d0-aec6-497c-806f-868ec544c85f", "bridge": "br-int", "label": "tempest-MultipleCreateTestJSON-480554183-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "dc40805c39f34d07be9d07861ad5841d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "359850cc-b061-4c9c-a61c-eb42e0f7c359", "external-id": "nsx-vlan-transportzone-113", "segmentation_id": 113, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc8809e29-ea", "ovs_interfaceid": "c8809e29-ea3c-4ea0-bee1-0731daee293d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61806) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 827.057676] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-cbb90e6d-7dac-4c30-b860-1f0db7dd091e tempest-MultipleCreateTestJSON-1589015462 tempest-MultipleCreateTestJSON-1589015462-project-member] [instance: 25218e83-c1ab-4b97-ade1-7c32b6f99305] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:6d:0c:29', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '359850cc-b061-4c9c-a61c-eb42e0f7c359', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'c8809e29-ea3c-4ea0-bee1-0731daee293d', 'vif_model': 'vmxnet3'}] {{(pid=61806) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 827.067087] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [None req-cbb90e6d-7dac-4c30-b860-1f0db7dd091e tempest-MultipleCreateTestJSON-1589015462 tempest-MultipleCreateTestJSON-1589015462-project-member] Creating folder: Project (dc40805c39f34d07be9d07861ad5841d). Parent ref: group-v277609. {{(pid=61806) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 827.069984] env[61806]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-448cfd7a-3fd5-4e07-abfa-a6a54dacd33b {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 827.076339] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f8c42d0e-c705-4db1-9eb4-bba3550408c1 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 827.092177] env[61806]: INFO nova.virt.vmwareapi.vm_util [None req-cbb90e6d-7dac-4c30-b860-1f0db7dd091e tempest-MultipleCreateTestJSON-1589015462 tempest-MultipleCreateTestJSON-1589015462-project-member] Created folder: Project (dc40805c39f34d07be9d07861ad5841d) in parent group-v277609. [ 827.092487] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [None req-cbb90e6d-7dac-4c30-b860-1f0db7dd091e tempest-MultipleCreateTestJSON-1589015462 tempest-MultipleCreateTestJSON-1589015462-project-member] Creating folder: Instances. Parent ref: group-v277667. {{(pid=61806) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 827.093344] env[61806]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-67bb3b76-1d6f-47d7-b537-7090ee3252ed {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 827.113445] env[61806]: INFO nova.virt.vmwareapi.vm_util [None req-cbb90e6d-7dac-4c30-b860-1f0db7dd091e tempest-MultipleCreateTestJSON-1589015462 tempest-MultipleCreateTestJSON-1589015462-project-member] Created folder: Instances in parent group-v277667. [ 827.113820] env[61806]: DEBUG oslo.service.loopingcall [None req-cbb90e6d-7dac-4c30-b860-1f0db7dd091e tempest-MultipleCreateTestJSON-1589015462 tempest-MultipleCreateTestJSON-1589015462-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61806) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 827.114111] env[61806]: DEBUG nova.compute.manager [req-45bb828c-3c67-42f2-9d45-b0ce633fbb71 req-5b620900-c107-4df3-a352-b4c96533441f service nova] [instance: 4373b735-31cf-4b53-b655-38555cf212a5] Detach interface failed, port_id=2df33c67-a8f5-4af4-ae74-29dd266d5054, reason: Instance 4373b735-31cf-4b53-b655-38555cf212a5 could not be found. {{(pid=61806) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10963}} [ 827.114561] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 25218e83-c1ab-4b97-ade1-7c32b6f99305] Creating VM on the ESX host {{(pid=61806) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 827.114828] env[61806]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-a6181516-8f98-44ba-ae12-f4dd2c1ce953 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 827.132486] env[61806]: INFO nova.compute.manager [-] [instance: 616ec206-9804-469e-ab5c-41aea7f048aa] Took 1.04 seconds to deallocate network for instance. [ 827.151126] env[61806]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 827.151126] env[61806]: value = "task-1294656" [ 827.151126] env[61806]: _type = "Task" [ 827.151126] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 827.159939] env[61806]: DEBUG oslo_vmware.api [None req-7fa92f02-ad2b-4004-a9a7-cb08f83ec0f1 tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Task: {'id': session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]52e92c46-39ba-7c22-b566-705c94e40478, 'name': SearchDatastore_Task, 'duration_secs': 0.047556} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 827.159939] env[61806]: DEBUG oslo_vmware.api [None req-9c52c9e9-816e-4638-b13a-a63a9756140c tempest-SecurityGroupsTestJSON-1213485581 tempest-SecurityGroupsTestJSON-1213485581-project-member] Task: {'id': task-1294653, 'name': ReconfigVM_Task, 'duration_secs': 0.306961} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 827.161926] env[61806]: DEBUG nova.virt.vmwareapi.volumeops [None req-9c52c9e9-816e-4638-b13a-a63a9756140c tempest-SecurityGroupsTestJSON-1213485581 tempest-SecurityGroupsTestJSON-1213485581-project-member] [instance: b41531f2-f28c-4d82-9682-0b557bbaa491] Reconfigured VM instance instance-00000040 to attach disk [datastore2] b41531f2-f28c-4d82-9682-0b557bbaa491/b41531f2-f28c-4d82-9682-0b557bbaa491.vmdk or device None with type sparse {{(pid=61806) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 827.162686] env[61806]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f16aee29-0220-40da-af93-aff95a69962d {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 827.165544] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-dce0a0ab-6a06-418f-ac02-1aa422d19c05 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 827.176461] env[61806]: DEBUG oslo_vmware.api [-] Task: {'id': task-1294656, 'name': CreateVM_Task} progress is 6%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 827.178789] env[61806]: DEBUG oslo_vmware.api [None req-7fa92f02-ad2b-4004-a9a7-cb08f83ec0f1 tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Waiting for the task: (returnval){ [ 827.178789] env[61806]: value = "session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]52bd2579-938d-9a76-7127-36b359cad520" [ 827.178789] env[61806]: _type = "Task" [ 827.178789] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 827.179043] env[61806]: DEBUG oslo_vmware.api [None req-9c52c9e9-816e-4638-b13a-a63a9756140c tempest-SecurityGroupsTestJSON-1213485581 tempest-SecurityGroupsTestJSON-1213485581-project-member] Waiting for the task: (returnval){ [ 827.179043] env[61806]: value = "task-1294657" [ 827.179043] env[61806]: _type = "Task" [ 827.179043] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 827.195857] env[61806]: DEBUG oslo_vmware.api [None req-7fa92f02-ad2b-4004-a9a7-cb08f83ec0f1 tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Task: {'id': session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]52bd2579-938d-9a76-7127-36b359cad520, 'name': SearchDatastore_Task, 'duration_secs': 0.011799} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 827.199211] env[61806]: DEBUG oslo_concurrency.lockutils [None req-7fa92f02-ad2b-4004-a9a7-cb08f83ec0f1 tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Releasing lock "[datastore1] devstack-image-cache_base/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa.vmdk" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 827.199496] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [None req-7fa92f02-ad2b-4004-a9a7-cb08f83ec0f1 tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa.vmdk to [datastore1] 67d1bdd8-cf8f-4fb5-bd3c-42008e4d60e3/67d1bdd8-cf8f-4fb5-bd3c-42008e4d60e3.vmdk {{(pid=61806) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 827.199791] env[61806]: DEBUG oslo_vmware.api [None req-9c52c9e9-816e-4638-b13a-a63a9756140c tempest-SecurityGroupsTestJSON-1213485581 tempest-SecurityGroupsTestJSON-1213485581-project-member] Task: {'id': task-1294657, 'name': Rename_Task} progress is 5%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 827.202755] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-87564fc6-5297-4d67-aa47-37ae72f36829 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 827.213194] env[61806]: DEBUG oslo_vmware.api [None req-7fa92f02-ad2b-4004-a9a7-cb08f83ec0f1 tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Waiting for the task: (returnval){ [ 827.213194] env[61806]: value = "task-1294658" [ 827.213194] env[61806]: _type = "Task" [ 827.213194] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 827.224113] env[61806]: DEBUG oslo_vmware.api [None req-7fa92f02-ad2b-4004-a9a7-cb08f83ec0f1 tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Task: {'id': task-1294658, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 827.368910] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-64729b59-d8ad-4d3f-8682-265b8b0bbe1a {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 827.380027] env[61806]: DEBUG nova.network.neutron [None req-cbb90e6d-7dac-4c30-b860-1f0db7dd091e tempest-MultipleCreateTestJSON-1589015462 tempest-MultipleCreateTestJSON-1589015462-project-member] [instance: ef502a7d-7dc0-4319-8ce3-6a8701288237] Successfully created port: 9e500f22-fafc-4f40-ab22-09b4330b6d4f {{(pid=61806) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 827.388065] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e05793ee-beff-481a-b061-f3fc7050f0e7 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 827.435463] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b9533cc3-3cb2-4a81-bb7b-b565398a394d {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 827.439839] env[61806]: INFO nova.compute.manager [-] [instance: 4373b735-31cf-4b53-b655-38555cf212a5] Took 1.54 seconds to deallocate network for instance. [ 827.447264] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b7f5d1b8-e4f1-4257-aaca-ff58e5634c81 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 827.463704] env[61806]: DEBUG nova.compute.provider_tree [None req-0ea3a38f-3871-4412-a6bc-d21c2086d262 tempest-AttachInterfacesV270Test-1974652806 tempest-AttachInterfacesV270Test-1974652806-project-member] Inventory has not changed in ProviderTree for provider: a2858be1-fd22-4e08-979e-87ad25293407 {{(pid=61806) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 827.654149] env[61806]: DEBUG oslo_concurrency.lockutils [None req-f4c6ec76-6b50-4a2c-bb86-09b71e318ce8 tempest-ServerShowV257Test-1159790492 tempest-ServerShowV257Test-1159790492-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 827.670719] env[61806]: DEBUG oslo_vmware.api [-] Task: {'id': task-1294656, 'name': CreateVM_Task} progress is 99%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 827.690935] env[61806]: DEBUG oslo_vmware.api [None req-9c52c9e9-816e-4638-b13a-a63a9756140c tempest-SecurityGroupsTestJSON-1213485581 tempest-SecurityGroupsTestJSON-1213485581-project-member] Task: {'id': task-1294657, 'name': Rename_Task, 'duration_secs': 0.206005} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 827.691280] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [None req-9c52c9e9-816e-4638-b13a-a63a9756140c tempest-SecurityGroupsTestJSON-1213485581 tempest-SecurityGroupsTestJSON-1213485581-project-member] [instance: b41531f2-f28c-4d82-9682-0b557bbaa491] Powering on the VM {{(pid=61806) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 827.691573] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-0b1eb9e3-2877-45db-bf2d-265611d2ba63 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 827.700343] env[61806]: DEBUG oslo_vmware.api [None req-9c52c9e9-816e-4638-b13a-a63a9756140c tempest-SecurityGroupsTestJSON-1213485581 tempest-SecurityGroupsTestJSON-1213485581-project-member] Waiting for the task: (returnval){ [ 827.700343] env[61806]: value = "task-1294659" [ 827.700343] env[61806]: _type = "Task" [ 827.700343] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 827.711529] env[61806]: DEBUG oslo_vmware.api [None req-9c52c9e9-816e-4638-b13a-a63a9756140c tempest-SecurityGroupsTestJSON-1213485581 tempest-SecurityGroupsTestJSON-1213485581-project-member] Task: {'id': task-1294659, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 827.723957] env[61806]: DEBUG oslo_vmware.api [None req-7fa92f02-ad2b-4004-a9a7-cb08f83ec0f1 tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Task: {'id': task-1294658, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 827.897795] env[61806]: DEBUG nova.compute.manager [req-7f971af8-3759-4bbf-95ed-53b249044cf5 req-41ed0fe2-1580-4eef-b65d-fff0c9848f38 service nova] [instance: 43dcce0d-4fdb-4ee1-8309-92e92d2331a9] Received event network-changed-6361e800-fbf2-48e0-a11d-b00d9a9955b4 {{(pid=61806) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 827.898054] env[61806]: DEBUG nova.compute.manager [req-7f971af8-3759-4bbf-95ed-53b249044cf5 req-41ed0fe2-1580-4eef-b65d-fff0c9848f38 service nova] [instance: 43dcce0d-4fdb-4ee1-8309-92e92d2331a9] Refreshing instance network info cache due to event network-changed-6361e800-fbf2-48e0-a11d-b00d9a9955b4. {{(pid=61806) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 827.898343] env[61806]: DEBUG oslo_concurrency.lockutils [req-7f971af8-3759-4bbf-95ed-53b249044cf5 req-41ed0fe2-1580-4eef-b65d-fff0c9848f38 service nova] Acquiring lock "refresh_cache-43dcce0d-4fdb-4ee1-8309-92e92d2331a9" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 827.898443] env[61806]: DEBUG oslo_concurrency.lockutils [req-7f971af8-3759-4bbf-95ed-53b249044cf5 req-41ed0fe2-1580-4eef-b65d-fff0c9848f38 service nova] Acquired lock "refresh_cache-43dcce0d-4fdb-4ee1-8309-92e92d2331a9" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 827.898684] env[61806]: DEBUG nova.network.neutron [req-7f971af8-3759-4bbf-95ed-53b249044cf5 req-41ed0fe2-1580-4eef-b65d-fff0c9848f38 service nova] [instance: 43dcce0d-4fdb-4ee1-8309-92e92d2331a9] Refreshing network info cache for port 6361e800-fbf2-48e0-a11d-b00d9a9955b4 {{(pid=61806) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 827.947610] env[61806]: DEBUG oslo_concurrency.lockutils [None req-0b05357b-24cd-44f2-8a41-86cdceeb41ee tempest-VolumesAdminNegativeTest-1387177911 tempest-VolumesAdminNegativeTest-1387177911-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 827.968130] env[61806]: DEBUG nova.scheduler.client.report [None req-0ea3a38f-3871-4412-a6bc-d21c2086d262 tempest-AttachInterfacesV270Test-1974652806 tempest-AttachInterfacesV270Test-1974652806-project-member] Inventory has not changed for provider a2858be1-fd22-4e08-979e-87ad25293407 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 138, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61806) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 828.020468] env[61806]: DEBUG nova.compute.manager [None req-cbb90e6d-7dac-4c30-b860-1f0db7dd091e tempest-MultipleCreateTestJSON-1589015462 tempest-MultipleCreateTestJSON-1589015462-project-member] [instance: ef502a7d-7dc0-4319-8ce3-6a8701288237] Start spawning the instance on the hypervisor. {{(pid=61806) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 828.045342] env[61806]: DEBUG nova.network.neutron [-] [instance: a842df44-d8a9-4376-b9fc-5ca19a68a4b7] Updating instance_info_cache with network_info: [] {{(pid=61806) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 828.061630] env[61806]: DEBUG nova.virt.hardware [None req-cbb90e6d-7dac-4c30-b860-1f0db7dd091e tempest-MultipleCreateTestJSON-1589015462 tempest-MultipleCreateTestJSON-1589015462-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-15T18:05:10Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-15T18:04:52Z,direct_url=,disk_format='vmdk',id=ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='84ba83fce2c349a6988173c5d6fc3b07',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-15T18:04:53Z,virtual_size=,visibility=), allow threads: False {{(pid=61806) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 828.061630] env[61806]: DEBUG nova.virt.hardware [None req-cbb90e6d-7dac-4c30-b860-1f0db7dd091e tempest-MultipleCreateTestJSON-1589015462 tempest-MultipleCreateTestJSON-1589015462-project-member] Flavor limits 0:0:0 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 828.061630] env[61806]: DEBUG nova.virt.hardware [None req-cbb90e6d-7dac-4c30-b860-1f0db7dd091e tempest-MultipleCreateTestJSON-1589015462 tempest-MultipleCreateTestJSON-1589015462-project-member] Image limits 0:0:0 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 828.061782] env[61806]: DEBUG nova.virt.hardware [None req-cbb90e6d-7dac-4c30-b860-1f0db7dd091e tempest-MultipleCreateTestJSON-1589015462 tempest-MultipleCreateTestJSON-1589015462-project-member] Flavor pref 0:0:0 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 828.061782] env[61806]: DEBUG nova.virt.hardware [None req-cbb90e6d-7dac-4c30-b860-1f0db7dd091e tempest-MultipleCreateTestJSON-1589015462 tempest-MultipleCreateTestJSON-1589015462-project-member] Image pref 0:0:0 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 828.061782] env[61806]: DEBUG nova.virt.hardware [None req-cbb90e6d-7dac-4c30-b860-1f0db7dd091e tempest-MultipleCreateTestJSON-1589015462 tempest-MultipleCreateTestJSON-1589015462-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 828.061782] env[61806]: DEBUG nova.virt.hardware [None req-cbb90e6d-7dac-4c30-b860-1f0db7dd091e tempest-MultipleCreateTestJSON-1589015462 tempest-MultipleCreateTestJSON-1589015462-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61806) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 828.061782] env[61806]: DEBUG nova.virt.hardware [None req-cbb90e6d-7dac-4c30-b860-1f0db7dd091e tempest-MultipleCreateTestJSON-1589015462 tempest-MultipleCreateTestJSON-1589015462-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61806) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 828.061928] env[61806]: DEBUG nova.virt.hardware [None req-cbb90e6d-7dac-4c30-b860-1f0db7dd091e tempest-MultipleCreateTestJSON-1589015462 tempest-MultipleCreateTestJSON-1589015462-project-member] Got 1 possible topologies {{(pid=61806) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 828.062130] env[61806]: DEBUG nova.virt.hardware [None req-cbb90e6d-7dac-4c30-b860-1f0db7dd091e tempest-MultipleCreateTestJSON-1589015462 tempest-MultipleCreateTestJSON-1589015462-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61806) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 828.062369] env[61806]: DEBUG nova.virt.hardware [None req-cbb90e6d-7dac-4c30-b860-1f0db7dd091e tempest-MultipleCreateTestJSON-1589015462 tempest-MultipleCreateTestJSON-1589015462-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61806) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 828.063295] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-667fa354-e30c-4269-88a8-856320367f20 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 828.073137] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1ab9fad6-252c-4659-b346-08ebddcae520 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 828.169671] env[61806]: DEBUG oslo_vmware.api [-] Task: {'id': task-1294656, 'name': CreateVM_Task, 'duration_secs': 0.800903} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 828.169845] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 25218e83-c1ab-4b97-ade1-7c32b6f99305] Created VM on the ESX host {{(pid=61806) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 828.170606] env[61806]: DEBUG oslo_concurrency.lockutils [None req-cbb90e6d-7dac-4c30-b860-1f0db7dd091e tempest-MultipleCreateTestJSON-1589015462 tempest-MultipleCreateTestJSON-1589015462-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 828.170786] env[61806]: DEBUG oslo_concurrency.lockutils [None req-cbb90e6d-7dac-4c30-b860-1f0db7dd091e tempest-MultipleCreateTestJSON-1589015462 tempest-MultipleCreateTestJSON-1589015462-project-member] Acquired lock "[datastore2] devstack-image-cache_base/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 828.171119] env[61806]: DEBUG oslo_concurrency.lockutils [None req-cbb90e6d-7dac-4c30-b860-1f0db7dd091e tempest-MultipleCreateTestJSON-1589015462 tempest-MultipleCreateTestJSON-1589015462-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 828.171388] env[61806]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-39454413-522b-46bc-93e4-0cca0f037d05 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 828.176833] env[61806]: DEBUG oslo_vmware.api [None req-cbb90e6d-7dac-4c30-b860-1f0db7dd091e tempest-MultipleCreateTestJSON-1589015462 tempest-MultipleCreateTestJSON-1589015462-project-member] Waiting for the task: (returnval){ [ 828.176833] env[61806]: value = "session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]52826bba-3fb5-c302-1dc7-a3669dc1a1a3" [ 828.176833] env[61806]: _type = "Task" [ 828.176833] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 828.186161] env[61806]: DEBUG oslo_vmware.api [None req-cbb90e6d-7dac-4c30-b860-1f0db7dd091e tempest-MultipleCreateTestJSON-1589015462 tempest-MultipleCreateTestJSON-1589015462-project-member] Task: {'id': session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]52826bba-3fb5-c302-1dc7-a3669dc1a1a3, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 828.211602] env[61806]: DEBUG oslo_vmware.api [None req-9c52c9e9-816e-4638-b13a-a63a9756140c tempest-SecurityGroupsTestJSON-1213485581 tempest-SecurityGroupsTestJSON-1213485581-project-member] Task: {'id': task-1294659, 'name': PowerOnVM_Task} progress is 89%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 828.228031] env[61806]: DEBUG oslo_vmware.api [None req-7fa92f02-ad2b-4004-a9a7-cb08f83ec0f1 tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Task: {'id': task-1294658, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.622009} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 828.228031] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [None req-7fa92f02-ad2b-4004-a9a7-cb08f83ec0f1 tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa.vmdk to [datastore1] 67d1bdd8-cf8f-4fb5-bd3c-42008e4d60e3/67d1bdd8-cf8f-4fb5-bd3c-42008e4d60e3.vmdk {{(pid=61806) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 828.228031] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-7fa92f02-ad2b-4004-a9a7-cb08f83ec0f1 tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] [instance: 67d1bdd8-cf8f-4fb5-bd3c-42008e4d60e3] Extending root virtual disk to 1048576 {{(pid=61806) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 828.228031] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-b1fb8ec4-f62f-4746-b140-cdf766f70c5e {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 828.234204] env[61806]: DEBUG oslo_vmware.api [None req-7fa92f02-ad2b-4004-a9a7-cb08f83ec0f1 tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Waiting for the task: (returnval){ [ 828.234204] env[61806]: value = "task-1294660" [ 828.234204] env[61806]: _type = "Task" [ 828.234204] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 828.245375] env[61806]: DEBUG oslo_vmware.api [None req-7fa92f02-ad2b-4004-a9a7-cb08f83ec0f1 tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Task: {'id': task-1294660, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 828.482127] env[61806]: DEBUG oslo_concurrency.lockutils [None req-0ea3a38f-3871-4412-a6bc-d21c2086d262 tempest-AttachInterfacesV270Test-1974652806 tempest-AttachInterfacesV270Test-1974652806-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.497s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 828.482463] env[61806]: DEBUG nova.compute.manager [None req-0ea3a38f-3871-4412-a6bc-d21c2086d262 tempest-AttachInterfacesV270Test-1974652806 tempest-AttachInterfacesV270Test-1974652806-project-member] [instance: 117139cd-4ed4-4bea-8ad3-f4ac32c692e4] Start building networks asynchronously for instance. {{(pid=61806) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 828.485854] env[61806]: DEBUG oslo_concurrency.lockutils [None req-4e0c8443-0fd4-459b-a4c5-7cf373aa1089 tempest-ServerAddressesNegativeTestJSON-1526246390 tempest-ServerAddressesNegativeTestJSON-1526246390-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 14.117s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 828.487264] env[61806]: INFO nova.compute.claims [None req-4e0c8443-0fd4-459b-a4c5-7cf373aa1089 tempest-ServerAddressesNegativeTestJSON-1526246390 tempest-ServerAddressesNegativeTestJSON-1526246390-project-member] [instance: c8033551-6591-4e37-ae78-4efe7145b10b] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 828.552583] env[61806]: INFO nova.compute.manager [-] [instance: a842df44-d8a9-4376-b9fc-5ca19a68a4b7] Took 1.69 seconds to deallocate network for instance. [ 828.576511] env[61806]: DEBUG nova.compute.manager [req-9af22353-dc8f-40d5-b55d-ef3600be4011 req-e3f0b7e5-a094-4f86-9d54-06c4693a3b87 service nova] [instance: a842df44-d8a9-4376-b9fc-5ca19a68a4b7] Received event network-vif-deleted-cb3cf8c0-2bc5-4308-899d-d6d2734f7922 {{(pid=61806) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 828.660252] env[61806]: DEBUG nova.network.neutron [req-7f971af8-3759-4bbf-95ed-53b249044cf5 req-41ed0fe2-1580-4eef-b65d-fff0c9848f38 service nova] [instance: 43dcce0d-4fdb-4ee1-8309-92e92d2331a9] Updated VIF entry in instance network info cache for port 6361e800-fbf2-48e0-a11d-b00d9a9955b4. {{(pid=61806) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 828.660670] env[61806]: DEBUG nova.network.neutron [req-7f971af8-3759-4bbf-95ed-53b249044cf5 req-41ed0fe2-1580-4eef-b65d-fff0c9848f38 service nova] [instance: 43dcce0d-4fdb-4ee1-8309-92e92d2331a9] Updating instance_info_cache with network_info: [{"id": "6361e800-fbf2-48e0-a11d-b00d9a9955b4", "address": "fa:16:3e:db:ef:a2", "network": {"id": "d636ca55-0737-48e4-8d9d-7fe0f797f49e", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-2090334349-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.164", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "22e47882c2c44ce492dff6b1c5d782e2", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6f493cd8-1cb4-42a1-8d56-bfa6ac7cf563", "external-id": "nsx-vlan-transportzone-931", "segmentation_id": 931, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap6361e800-fb", "ovs_interfaceid": "6361e800-fbf2-48e0-a11d-b00d9a9955b4", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61806) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 828.687543] env[61806]: DEBUG oslo_vmware.api [None req-cbb90e6d-7dac-4c30-b860-1f0db7dd091e tempest-MultipleCreateTestJSON-1589015462 tempest-MultipleCreateTestJSON-1589015462-project-member] Task: {'id': session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]52826bba-3fb5-c302-1dc7-a3669dc1a1a3, 'name': SearchDatastore_Task, 'duration_secs': 0.012489} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 828.687905] env[61806]: DEBUG oslo_concurrency.lockutils [None req-cbb90e6d-7dac-4c30-b860-1f0db7dd091e tempest-MultipleCreateTestJSON-1589015462 tempest-MultipleCreateTestJSON-1589015462-project-member] Releasing lock "[datastore2] devstack-image-cache_base/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 828.688173] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-cbb90e6d-7dac-4c30-b860-1f0db7dd091e tempest-MultipleCreateTestJSON-1589015462 tempest-MultipleCreateTestJSON-1589015462-project-member] [instance: 25218e83-c1ab-4b97-ade1-7c32b6f99305] Processing image ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa {{(pid=61806) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 828.688427] env[61806]: DEBUG oslo_concurrency.lockutils [None req-cbb90e6d-7dac-4c30-b860-1f0db7dd091e tempest-MultipleCreateTestJSON-1589015462 tempest-MultipleCreateTestJSON-1589015462-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa.vmdk" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 828.688581] env[61806]: DEBUG oslo_concurrency.lockutils [None req-cbb90e6d-7dac-4c30-b860-1f0db7dd091e tempest-MultipleCreateTestJSON-1589015462 tempest-MultipleCreateTestJSON-1589015462-project-member] Acquired lock "[datastore2] devstack-image-cache_base/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa.vmdk" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 828.688770] env[61806]: DEBUG nova.virt.vmwareapi.ds_util [None req-cbb90e6d-7dac-4c30-b860-1f0db7dd091e tempest-MultipleCreateTestJSON-1589015462 tempest-MultipleCreateTestJSON-1589015462-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61806) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 828.689064] env[61806]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-beed4ff4-c043-4dae-80c1-60135cb3b684 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 828.706182] env[61806]: DEBUG nova.virt.vmwareapi.ds_util [None req-cbb90e6d-7dac-4c30-b860-1f0db7dd091e tempest-MultipleCreateTestJSON-1589015462 tempest-MultipleCreateTestJSON-1589015462-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61806) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 828.706394] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-cbb90e6d-7dac-4c30-b860-1f0db7dd091e tempest-MultipleCreateTestJSON-1589015462 tempest-MultipleCreateTestJSON-1589015462-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61806) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 828.707664] env[61806]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-a73fb996-50af-43ba-9409-4dfdf5546e36 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 828.713076] env[61806]: DEBUG oslo_vmware.api [None req-9c52c9e9-816e-4638-b13a-a63a9756140c tempest-SecurityGroupsTestJSON-1213485581 tempest-SecurityGroupsTestJSON-1213485581-project-member] Task: {'id': task-1294659, 'name': PowerOnVM_Task, 'duration_secs': 0.584025} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 828.713653] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [None req-9c52c9e9-816e-4638-b13a-a63a9756140c tempest-SecurityGroupsTestJSON-1213485581 tempest-SecurityGroupsTestJSON-1213485581-project-member] [instance: b41531f2-f28c-4d82-9682-0b557bbaa491] Powered on the VM {{(pid=61806) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 828.713874] env[61806]: INFO nova.compute.manager [None req-9c52c9e9-816e-4638-b13a-a63a9756140c tempest-SecurityGroupsTestJSON-1213485581 tempest-SecurityGroupsTestJSON-1213485581-project-member] [instance: b41531f2-f28c-4d82-9682-0b557bbaa491] Took 10.10 seconds to spawn the instance on the hypervisor. [ 828.714079] env[61806]: DEBUG nova.compute.manager [None req-9c52c9e9-816e-4638-b13a-a63a9756140c tempest-SecurityGroupsTestJSON-1213485581 tempest-SecurityGroupsTestJSON-1213485581-project-member] [instance: b41531f2-f28c-4d82-9682-0b557bbaa491] Checking state {{(pid=61806) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 828.714848] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5d0746d6-6121-46ae-aa34-24bfe16240de {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 828.718106] env[61806]: DEBUG oslo_vmware.api [None req-cbb90e6d-7dac-4c30-b860-1f0db7dd091e tempest-MultipleCreateTestJSON-1589015462 tempest-MultipleCreateTestJSON-1589015462-project-member] Waiting for the task: (returnval){ [ 828.718106] env[61806]: value = "session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]521a8eb0-3594-fa67-4e48-a57d4d6c5f14" [ 828.718106] env[61806]: _type = "Task" [ 828.718106] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 828.729379] env[61806]: DEBUG oslo_vmware.api [None req-cbb90e6d-7dac-4c30-b860-1f0db7dd091e tempest-MultipleCreateTestJSON-1589015462 tempest-MultipleCreateTestJSON-1589015462-project-member] Task: {'id': session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]521a8eb0-3594-fa67-4e48-a57d4d6c5f14, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 828.743249] env[61806]: DEBUG oslo_vmware.api [None req-7fa92f02-ad2b-4004-a9a7-cb08f83ec0f1 tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Task: {'id': task-1294660, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.121069} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 828.743496] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-7fa92f02-ad2b-4004-a9a7-cb08f83ec0f1 tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] [instance: 67d1bdd8-cf8f-4fb5-bd3c-42008e4d60e3] Extended root virtual disk {{(pid=61806) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 828.744241] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-90636549-867d-4f66-b4c3-0efe4963ddfc {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 828.767722] env[61806]: DEBUG nova.virt.vmwareapi.volumeops [None req-7fa92f02-ad2b-4004-a9a7-cb08f83ec0f1 tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] [instance: 67d1bdd8-cf8f-4fb5-bd3c-42008e4d60e3] Reconfiguring VM instance instance-00000041 to attach disk [datastore1] 67d1bdd8-cf8f-4fb5-bd3c-42008e4d60e3/67d1bdd8-cf8f-4fb5-bd3c-42008e4d60e3.vmdk or device None with type sparse {{(pid=61806) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 828.768621] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-f7741c08-fd2b-4c7a-9802-fbe9f320e967 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 828.788617] env[61806]: DEBUG oslo_vmware.api [None req-7fa92f02-ad2b-4004-a9a7-cb08f83ec0f1 tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Waiting for the task: (returnval){ [ 828.788617] env[61806]: value = "task-1294661" [ 828.788617] env[61806]: _type = "Task" [ 828.788617] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 828.797498] env[61806]: DEBUG oslo_vmware.api [None req-7fa92f02-ad2b-4004-a9a7-cb08f83ec0f1 tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Task: {'id': task-1294661, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 828.996171] env[61806]: DEBUG nova.compute.utils [None req-0ea3a38f-3871-4412-a6bc-d21c2086d262 tempest-AttachInterfacesV270Test-1974652806 tempest-AttachInterfacesV270Test-1974652806-project-member] Using /dev/sd instead of None {{(pid=61806) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 829.001020] env[61806]: DEBUG nova.compute.manager [None req-0ea3a38f-3871-4412-a6bc-d21c2086d262 tempest-AttachInterfacesV270Test-1974652806 tempest-AttachInterfacesV270Test-1974652806-project-member] [instance: 117139cd-4ed4-4bea-8ad3-f4ac32c692e4] Allocating IP information in the background. {{(pid=61806) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 829.001020] env[61806]: DEBUG nova.network.neutron [None req-0ea3a38f-3871-4412-a6bc-d21c2086d262 tempest-AttachInterfacesV270Test-1974652806 tempest-AttachInterfacesV270Test-1974652806-project-member] [instance: 117139cd-4ed4-4bea-8ad3-f4ac32c692e4] allocate_for_instance() {{(pid=61806) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 829.060993] env[61806]: DEBUG oslo_concurrency.lockutils [None req-90367f73-05d7-4a77-94a7-20cc6ea5d878 tempest-ServerAddressesTestJSON-661730579 tempest-ServerAddressesTestJSON-661730579-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 829.062665] env[61806]: DEBUG nova.policy [None req-0ea3a38f-3871-4412-a6bc-d21c2086d262 tempest-AttachInterfacesV270Test-1974652806 tempest-AttachInterfacesV270Test-1974652806-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'd4b2fa37c4a7432588bea9813ea67596', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'b0499ad8c8bb4b6d80b523fa85388e53', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61806) authorize /opt/stack/nova/nova/policy.py:201}} [ 829.162959] env[61806]: DEBUG oslo_concurrency.lockutils [req-7f971af8-3759-4bbf-95ed-53b249044cf5 req-41ed0fe2-1580-4eef-b65d-fff0c9848f38 service nova] Releasing lock "refresh_cache-43dcce0d-4fdb-4ee1-8309-92e92d2331a9" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 829.207933] env[61806]: DEBUG nova.network.neutron [None req-cbb90e6d-7dac-4c30-b860-1f0db7dd091e tempest-MultipleCreateTestJSON-1589015462 tempest-MultipleCreateTestJSON-1589015462-project-member] [instance: ef502a7d-7dc0-4319-8ce3-6a8701288237] Successfully updated port: 9e500f22-fafc-4f40-ab22-09b4330b6d4f {{(pid=61806) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 829.237955] env[61806]: DEBUG oslo_vmware.api [None req-cbb90e6d-7dac-4c30-b860-1f0db7dd091e tempest-MultipleCreateTestJSON-1589015462 tempest-MultipleCreateTestJSON-1589015462-project-member] Task: {'id': session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]521a8eb0-3594-fa67-4e48-a57d4d6c5f14, 'name': SearchDatastore_Task, 'duration_secs': 0.055871} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 829.245590] env[61806]: INFO nova.compute.manager [None req-9c52c9e9-816e-4638-b13a-a63a9756140c tempest-SecurityGroupsTestJSON-1213485581 tempest-SecurityGroupsTestJSON-1213485581-project-member] [instance: b41531f2-f28c-4d82-9682-0b557bbaa491] Took 28.25 seconds to build instance. [ 829.250141] env[61806]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-fa2318ee-91dd-437c-869e-41ced67d0605 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 829.256874] env[61806]: DEBUG oslo_vmware.api [None req-cbb90e6d-7dac-4c30-b860-1f0db7dd091e tempest-MultipleCreateTestJSON-1589015462 tempest-MultipleCreateTestJSON-1589015462-project-member] Waiting for the task: (returnval){ [ 829.256874] env[61806]: value = "session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]5251118c-3312-4a28-31c8-31d8c923b9d2" [ 829.256874] env[61806]: _type = "Task" [ 829.256874] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 829.266669] env[61806]: DEBUG oslo_vmware.api [None req-cbb90e6d-7dac-4c30-b860-1f0db7dd091e tempest-MultipleCreateTestJSON-1589015462 tempest-MultipleCreateTestJSON-1589015462-project-member] Task: {'id': session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]5251118c-3312-4a28-31c8-31d8c923b9d2, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 829.302765] env[61806]: DEBUG oslo_vmware.api [None req-7fa92f02-ad2b-4004-a9a7-cb08f83ec0f1 tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Task: {'id': task-1294661, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 829.458839] env[61806]: DEBUG nova.network.neutron [None req-0ea3a38f-3871-4412-a6bc-d21c2086d262 tempest-AttachInterfacesV270Test-1974652806 tempest-AttachInterfacesV270Test-1974652806-project-member] [instance: 117139cd-4ed4-4bea-8ad3-f4ac32c692e4] Successfully created port: f984af48-8d6f-48b1-a3b3-1f988c36b872 {{(pid=61806) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 829.504020] env[61806]: DEBUG nova.compute.manager [None req-0ea3a38f-3871-4412-a6bc-d21c2086d262 tempest-AttachInterfacesV270Test-1974652806 tempest-AttachInterfacesV270Test-1974652806-project-member] [instance: 117139cd-4ed4-4bea-8ad3-f4ac32c692e4] Start building block device mappings for instance. {{(pid=61806) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 829.710930] env[61806]: DEBUG oslo_concurrency.lockutils [None req-cbb90e6d-7dac-4c30-b860-1f0db7dd091e tempest-MultipleCreateTestJSON-1589015462 tempest-MultipleCreateTestJSON-1589015462-project-member] Acquiring lock "refresh_cache-ef502a7d-7dc0-4319-8ce3-6a8701288237" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 829.711019] env[61806]: DEBUG oslo_concurrency.lockutils [None req-cbb90e6d-7dac-4c30-b860-1f0db7dd091e tempest-MultipleCreateTestJSON-1589015462 tempest-MultipleCreateTestJSON-1589015462-project-member] Acquired lock "refresh_cache-ef502a7d-7dc0-4319-8ce3-6a8701288237" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 829.711221] env[61806]: DEBUG nova.network.neutron [None req-cbb90e6d-7dac-4c30-b860-1f0db7dd091e tempest-MultipleCreateTestJSON-1589015462 tempest-MultipleCreateTestJSON-1589015462-project-member] [instance: ef502a7d-7dc0-4319-8ce3-6a8701288237] Building network info cache for instance {{(pid=61806) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 829.747684] env[61806]: DEBUG oslo_concurrency.lockutils [None req-9c52c9e9-816e-4638-b13a-a63a9756140c tempest-SecurityGroupsTestJSON-1213485581 tempest-SecurityGroupsTestJSON-1213485581-project-member] Lock "b41531f2-f28c-4d82-9682-0b557bbaa491" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 94.954s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 829.764902] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-337c2938-e423-4eb8-a96c-957a0b1a3c4f {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 829.779243] env[61806]: DEBUG oslo_vmware.api [None req-cbb90e6d-7dac-4c30-b860-1f0db7dd091e tempest-MultipleCreateTestJSON-1589015462 tempest-MultipleCreateTestJSON-1589015462-project-member] Task: {'id': session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]5251118c-3312-4a28-31c8-31d8c923b9d2, 'name': SearchDatastore_Task, 'duration_secs': 0.035591} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 829.780781] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-97efe851-dcab-499b-95c9-d2ad8f9b1254 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 829.785593] env[61806]: DEBUG oslo_concurrency.lockutils [None req-cbb90e6d-7dac-4c30-b860-1f0db7dd091e tempest-MultipleCreateTestJSON-1589015462 tempest-MultipleCreateTestJSON-1589015462-project-member] Releasing lock "[datastore2] devstack-image-cache_base/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa.vmdk" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 829.785999] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [None req-cbb90e6d-7dac-4c30-b860-1f0db7dd091e tempest-MultipleCreateTestJSON-1589015462 tempest-MultipleCreateTestJSON-1589015462-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa.vmdk to [datastore2] 25218e83-c1ab-4b97-ade1-7c32b6f99305/25218e83-c1ab-4b97-ade1-7c32b6f99305.vmdk {{(pid=61806) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 829.786373] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-aaa2b628-fd1f-4803-ab56-060508d55dc1 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 829.827698] env[61806]: DEBUG oslo_vmware.api [None req-cbb90e6d-7dac-4c30-b860-1f0db7dd091e tempest-MultipleCreateTestJSON-1589015462 tempest-MultipleCreateTestJSON-1589015462-project-member] Waiting for the task: (returnval){ [ 829.827698] env[61806]: value = "task-1294662" [ 829.827698] env[61806]: _type = "Task" [ 829.827698] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 829.829470] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f629dba9-8c9c-4a19-819c-9b079bdbb2fd {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 829.838381] env[61806]: DEBUG oslo_vmware.api [None req-7fa92f02-ad2b-4004-a9a7-cb08f83ec0f1 tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Task: {'id': task-1294661, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 829.846966] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-511953b6-fe4e-4f46-8044-4d7ebf7cc63b {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 829.850633] env[61806]: DEBUG oslo_vmware.api [None req-cbb90e6d-7dac-4c30-b860-1f0db7dd091e tempest-MultipleCreateTestJSON-1589015462 tempest-MultipleCreateTestJSON-1589015462-project-member] Task: {'id': task-1294662, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 829.860811] env[61806]: DEBUG nova.compute.provider_tree [None req-4e0c8443-0fd4-459b-a4c5-7cf373aa1089 tempest-ServerAddressesNegativeTestJSON-1526246390 tempest-ServerAddressesNegativeTestJSON-1526246390-project-member] Inventory has not changed in ProviderTree for provider: a2858be1-fd22-4e08-979e-87ad25293407 {{(pid=61806) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 830.042147] env[61806]: DEBUG nova.compute.manager [req-f94c69e0-c504-417f-b00a-34860b38c484 req-dd3c842d-e8d1-42ca-8bbf-0a7c76912bab service nova] [instance: b41531f2-f28c-4d82-9682-0b557bbaa491] Received event network-changed-513ca4b5-b62c-486a-9860-cf267a15c3e5 {{(pid=61806) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 830.042147] env[61806]: DEBUG nova.compute.manager [req-f94c69e0-c504-417f-b00a-34860b38c484 req-dd3c842d-e8d1-42ca-8bbf-0a7c76912bab service nova] [instance: b41531f2-f28c-4d82-9682-0b557bbaa491] Refreshing instance network info cache due to event network-changed-513ca4b5-b62c-486a-9860-cf267a15c3e5. {{(pid=61806) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 830.042147] env[61806]: DEBUG oslo_concurrency.lockutils [req-f94c69e0-c504-417f-b00a-34860b38c484 req-dd3c842d-e8d1-42ca-8bbf-0a7c76912bab service nova] Acquiring lock "refresh_cache-b41531f2-f28c-4d82-9682-0b557bbaa491" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 830.042700] env[61806]: DEBUG oslo_concurrency.lockutils [req-f94c69e0-c504-417f-b00a-34860b38c484 req-dd3c842d-e8d1-42ca-8bbf-0a7c76912bab service nova] Acquired lock "refresh_cache-b41531f2-f28c-4d82-9682-0b557bbaa491" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 830.042820] env[61806]: DEBUG nova.network.neutron [req-f94c69e0-c504-417f-b00a-34860b38c484 req-dd3c842d-e8d1-42ca-8bbf-0a7c76912bab service nova] [instance: b41531f2-f28c-4d82-9682-0b557bbaa491] Refreshing network info cache for port 513ca4b5-b62c-486a-9860-cf267a15c3e5 {{(pid=61806) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 830.063123] env[61806]: DEBUG oslo_concurrency.lockutils [None req-6d96025a-fbfa-435a-871d-fe0c696f1d71 tempest-SecurityGroupsTestJSON-1213485581 tempest-SecurityGroupsTestJSON-1213485581-project-member] Acquiring lock "b41531f2-f28c-4d82-9682-0b557bbaa491" by "nova.compute.manager.ComputeManager.reboot_instance..do_reboot_instance" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 830.063401] env[61806]: DEBUG oslo_concurrency.lockutils [None req-6d96025a-fbfa-435a-871d-fe0c696f1d71 tempest-SecurityGroupsTestJSON-1213485581 tempest-SecurityGroupsTestJSON-1213485581-project-member] Lock "b41531f2-f28c-4d82-9682-0b557bbaa491" acquired by "nova.compute.manager.ComputeManager.reboot_instance..do_reboot_instance" :: waited 0.000s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 830.063657] env[61806]: INFO nova.compute.manager [None req-6d96025a-fbfa-435a-871d-fe0c696f1d71 tempest-SecurityGroupsTestJSON-1213485581 tempest-SecurityGroupsTestJSON-1213485581-project-member] [instance: b41531f2-f28c-4d82-9682-0b557bbaa491] Rebooting instance [ 830.259989] env[61806]: DEBUG nova.network.neutron [None req-cbb90e6d-7dac-4c30-b860-1f0db7dd091e tempest-MultipleCreateTestJSON-1589015462 tempest-MultipleCreateTestJSON-1589015462-project-member] [instance: ef502a7d-7dc0-4319-8ce3-6a8701288237] Instance cache missing network info. {{(pid=61806) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 830.302499] env[61806]: DEBUG oslo_vmware.api [None req-7fa92f02-ad2b-4004-a9a7-cb08f83ec0f1 tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Task: {'id': task-1294661, 'name': ReconfigVM_Task, 'duration_secs': 1.221996} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 830.305308] env[61806]: DEBUG nova.virt.vmwareapi.volumeops [None req-7fa92f02-ad2b-4004-a9a7-cb08f83ec0f1 tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] [instance: 67d1bdd8-cf8f-4fb5-bd3c-42008e4d60e3] Reconfigured VM instance instance-00000041 to attach disk [datastore1] 67d1bdd8-cf8f-4fb5-bd3c-42008e4d60e3/67d1bdd8-cf8f-4fb5-bd3c-42008e4d60e3.vmdk or device None with type sparse {{(pid=61806) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 830.306390] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-b6f61411-3f81-4226-b733-49cfbb4c793d {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 830.317635] env[61806]: DEBUG oslo_vmware.api [None req-7fa92f02-ad2b-4004-a9a7-cb08f83ec0f1 tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Waiting for the task: (returnval){ [ 830.317635] env[61806]: value = "task-1294663" [ 830.317635] env[61806]: _type = "Task" [ 830.317635] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 830.331180] env[61806]: DEBUG oslo_vmware.api [None req-7fa92f02-ad2b-4004-a9a7-cb08f83ec0f1 tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Task: {'id': task-1294663, 'name': Rename_Task} progress is 5%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 830.343492] env[61806]: DEBUG oslo_vmware.api [None req-cbb90e6d-7dac-4c30-b860-1f0db7dd091e tempest-MultipleCreateTestJSON-1589015462 tempest-MultipleCreateTestJSON-1589015462-project-member] Task: {'id': task-1294662, 'name': CopyVirtualDisk_Task} progress is 25%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 830.364295] env[61806]: DEBUG nova.scheduler.client.report [None req-4e0c8443-0fd4-459b-a4c5-7cf373aa1089 tempest-ServerAddressesNegativeTestJSON-1526246390 tempest-ServerAddressesNegativeTestJSON-1526246390-project-member] Inventory has not changed for provider a2858be1-fd22-4e08-979e-87ad25293407 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 138, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61806) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 830.441561] env[61806]: DEBUG nova.network.neutron [None req-cbb90e6d-7dac-4c30-b860-1f0db7dd091e tempest-MultipleCreateTestJSON-1589015462 tempest-MultipleCreateTestJSON-1589015462-project-member] [instance: ef502a7d-7dc0-4319-8ce3-6a8701288237] Updating instance_info_cache with network_info: [{"id": "9e500f22-fafc-4f40-ab22-09b4330b6d4f", "address": "fa:16:3e:27:ee:7f", "network": {"id": "56fd57d0-aec6-497c-806f-868ec544c85f", "bridge": "br-int", "label": "tempest-MultipleCreateTestJSON-480554183-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "dc40805c39f34d07be9d07861ad5841d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "359850cc-b061-4c9c-a61c-eb42e0f7c359", "external-id": "nsx-vlan-transportzone-113", "segmentation_id": 113, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9e500f22-fa", "ovs_interfaceid": "9e500f22-fafc-4f40-ab22-09b4330b6d4f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61806) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 830.514246] env[61806]: DEBUG nova.compute.manager [None req-0ea3a38f-3871-4412-a6bc-d21c2086d262 tempest-AttachInterfacesV270Test-1974652806 tempest-AttachInterfacesV270Test-1974652806-project-member] [instance: 117139cd-4ed4-4bea-8ad3-f4ac32c692e4] Start spawning the instance on the hypervisor. {{(pid=61806) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 830.541958] env[61806]: DEBUG nova.virt.hardware [None req-0ea3a38f-3871-4412-a6bc-d21c2086d262 tempest-AttachInterfacesV270Test-1974652806 tempest-AttachInterfacesV270Test-1974652806-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-15T18:05:10Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-15T18:04:52Z,direct_url=,disk_format='vmdk',id=ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='84ba83fce2c349a6988173c5d6fc3b07',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-15T18:04:53Z,virtual_size=,visibility=), allow threads: False {{(pid=61806) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 830.542269] env[61806]: DEBUG nova.virt.hardware [None req-0ea3a38f-3871-4412-a6bc-d21c2086d262 tempest-AttachInterfacesV270Test-1974652806 tempest-AttachInterfacesV270Test-1974652806-project-member] Flavor limits 0:0:0 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 830.542465] env[61806]: DEBUG nova.virt.hardware [None req-0ea3a38f-3871-4412-a6bc-d21c2086d262 tempest-AttachInterfacesV270Test-1974652806 tempest-AttachInterfacesV270Test-1974652806-project-member] Image limits 0:0:0 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 830.542730] env[61806]: DEBUG nova.virt.hardware [None req-0ea3a38f-3871-4412-a6bc-d21c2086d262 tempest-AttachInterfacesV270Test-1974652806 tempest-AttachInterfacesV270Test-1974652806-project-member] Flavor pref 0:0:0 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 830.542895] env[61806]: DEBUG nova.virt.hardware [None req-0ea3a38f-3871-4412-a6bc-d21c2086d262 tempest-AttachInterfacesV270Test-1974652806 tempest-AttachInterfacesV270Test-1974652806-project-member] Image pref 0:0:0 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 830.543065] env[61806]: DEBUG nova.virt.hardware [None req-0ea3a38f-3871-4412-a6bc-d21c2086d262 tempest-AttachInterfacesV270Test-1974652806 tempest-AttachInterfacesV270Test-1974652806-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 830.543289] env[61806]: DEBUG nova.virt.hardware [None req-0ea3a38f-3871-4412-a6bc-d21c2086d262 tempest-AttachInterfacesV270Test-1974652806 tempest-AttachInterfacesV270Test-1974652806-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61806) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 830.543457] env[61806]: DEBUG nova.virt.hardware [None req-0ea3a38f-3871-4412-a6bc-d21c2086d262 tempest-AttachInterfacesV270Test-1974652806 tempest-AttachInterfacesV270Test-1974652806-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61806) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 830.543660] env[61806]: DEBUG nova.virt.hardware [None req-0ea3a38f-3871-4412-a6bc-d21c2086d262 tempest-AttachInterfacesV270Test-1974652806 tempest-AttachInterfacesV270Test-1974652806-project-member] Got 1 possible topologies {{(pid=61806) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 830.543842] env[61806]: DEBUG nova.virt.hardware [None req-0ea3a38f-3871-4412-a6bc-d21c2086d262 tempest-AttachInterfacesV270Test-1974652806 tempest-AttachInterfacesV270Test-1974652806-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61806) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 830.544037] env[61806]: DEBUG nova.virt.hardware [None req-0ea3a38f-3871-4412-a6bc-d21c2086d262 tempest-AttachInterfacesV270Test-1974652806 tempest-AttachInterfacesV270Test-1974652806-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61806) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 830.544912] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-91baa7b1-ac10-44a3-932b-2f0e9e74a1c3 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 830.555026] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dbce45bc-f693-46a6-9d22-ad578993ca9b {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 830.596718] env[61806]: DEBUG oslo_concurrency.lockutils [None req-6d96025a-fbfa-435a-871d-fe0c696f1d71 tempest-SecurityGroupsTestJSON-1213485581 tempest-SecurityGroupsTestJSON-1213485581-project-member] Acquiring lock "refresh_cache-b41531f2-f28c-4d82-9682-0b557bbaa491" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 830.609797] env[61806]: DEBUG nova.compute.manager [req-954128ad-9c63-4aa4-aaac-c35aa507d91d req-da93a2c5-7057-4025-aa14-e2d2d675b0ac service nova] [instance: ef502a7d-7dc0-4319-8ce3-6a8701288237] Received event network-vif-plugged-9e500f22-fafc-4f40-ab22-09b4330b6d4f {{(pid=61806) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 830.610085] env[61806]: DEBUG oslo_concurrency.lockutils [req-954128ad-9c63-4aa4-aaac-c35aa507d91d req-da93a2c5-7057-4025-aa14-e2d2d675b0ac service nova] Acquiring lock "ef502a7d-7dc0-4319-8ce3-6a8701288237-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 830.610238] env[61806]: DEBUG oslo_concurrency.lockutils [req-954128ad-9c63-4aa4-aaac-c35aa507d91d req-da93a2c5-7057-4025-aa14-e2d2d675b0ac service nova] Lock "ef502a7d-7dc0-4319-8ce3-6a8701288237-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 830.610472] env[61806]: DEBUG oslo_concurrency.lockutils [req-954128ad-9c63-4aa4-aaac-c35aa507d91d req-da93a2c5-7057-4025-aa14-e2d2d675b0ac service nova] Lock "ef502a7d-7dc0-4319-8ce3-6a8701288237-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 830.610668] env[61806]: DEBUG nova.compute.manager [req-954128ad-9c63-4aa4-aaac-c35aa507d91d req-da93a2c5-7057-4025-aa14-e2d2d675b0ac service nova] [instance: ef502a7d-7dc0-4319-8ce3-6a8701288237] No waiting events found dispatching network-vif-plugged-9e500f22-fafc-4f40-ab22-09b4330b6d4f {{(pid=61806) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 830.610861] env[61806]: WARNING nova.compute.manager [req-954128ad-9c63-4aa4-aaac-c35aa507d91d req-da93a2c5-7057-4025-aa14-e2d2d675b0ac service nova] [instance: ef502a7d-7dc0-4319-8ce3-6a8701288237] Received unexpected event network-vif-plugged-9e500f22-fafc-4f40-ab22-09b4330b6d4f for instance with vm_state building and task_state spawning. [ 830.611034] env[61806]: DEBUG nova.compute.manager [req-954128ad-9c63-4aa4-aaac-c35aa507d91d req-da93a2c5-7057-4025-aa14-e2d2d675b0ac service nova] [instance: ef502a7d-7dc0-4319-8ce3-6a8701288237] Received event network-changed-9e500f22-fafc-4f40-ab22-09b4330b6d4f {{(pid=61806) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 830.611201] env[61806]: DEBUG nova.compute.manager [req-954128ad-9c63-4aa4-aaac-c35aa507d91d req-da93a2c5-7057-4025-aa14-e2d2d675b0ac service nova] [instance: ef502a7d-7dc0-4319-8ce3-6a8701288237] Refreshing instance network info cache due to event network-changed-9e500f22-fafc-4f40-ab22-09b4330b6d4f. {{(pid=61806) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 830.611433] env[61806]: DEBUG oslo_concurrency.lockutils [req-954128ad-9c63-4aa4-aaac-c35aa507d91d req-da93a2c5-7057-4025-aa14-e2d2d675b0ac service nova] Acquiring lock "refresh_cache-ef502a7d-7dc0-4319-8ce3-6a8701288237" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 830.832092] env[61806]: DEBUG oslo_vmware.api [None req-7fa92f02-ad2b-4004-a9a7-cb08f83ec0f1 tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Task: {'id': task-1294663, 'name': Rename_Task, 'duration_secs': 0.256052} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 830.833376] env[61806]: DEBUG nova.network.neutron [req-f94c69e0-c504-417f-b00a-34860b38c484 req-dd3c842d-e8d1-42ca-8bbf-0a7c76912bab service nova] [instance: b41531f2-f28c-4d82-9682-0b557bbaa491] Updated VIF entry in instance network info cache for port 513ca4b5-b62c-486a-9860-cf267a15c3e5. {{(pid=61806) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 830.833850] env[61806]: DEBUG nova.network.neutron [req-f94c69e0-c504-417f-b00a-34860b38c484 req-dd3c842d-e8d1-42ca-8bbf-0a7c76912bab service nova] [instance: b41531f2-f28c-4d82-9682-0b557bbaa491] Updating instance_info_cache with network_info: [{"id": "513ca4b5-b62c-486a-9860-cf267a15c3e5", "address": "fa:16:3e:40:52:86", "network": {"id": "dfb52c93-e553-40d9-acb5-768cc6b67c89", "bridge": "br-int", "label": "tempest-SecurityGroupsTestJSON-1006891343-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "cfce93add38d40d79f294017f52e13d5", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6b6942d7-d4ab-4b2a-8d0f-76bf2a2478ad", "external-id": "nsx-vlan-transportzone-871", "segmentation_id": 871, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap513ca4b5-b6", "ovs_interfaceid": "513ca4b5-b62c-486a-9860-cf267a15c3e5", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61806) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 830.839533] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [None req-7fa92f02-ad2b-4004-a9a7-cb08f83ec0f1 tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] [instance: 67d1bdd8-cf8f-4fb5-bd3c-42008e4d60e3] Powering on the VM {{(pid=61806) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 830.840141] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-d77f00e0-e3c0-4244-979c-ba9c16f6415b {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 830.850136] env[61806]: DEBUG oslo_vmware.api [None req-cbb90e6d-7dac-4c30-b860-1f0db7dd091e tempest-MultipleCreateTestJSON-1589015462 tempest-MultipleCreateTestJSON-1589015462-project-member] Task: {'id': task-1294662, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.727635} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 830.853490] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [None req-cbb90e6d-7dac-4c30-b860-1f0db7dd091e tempest-MultipleCreateTestJSON-1589015462 tempest-MultipleCreateTestJSON-1589015462-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa.vmdk to [datastore2] 25218e83-c1ab-4b97-ade1-7c32b6f99305/25218e83-c1ab-4b97-ade1-7c32b6f99305.vmdk {{(pid=61806) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 830.853490] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-cbb90e6d-7dac-4c30-b860-1f0db7dd091e tempest-MultipleCreateTestJSON-1589015462 tempest-MultipleCreateTestJSON-1589015462-project-member] [instance: 25218e83-c1ab-4b97-ade1-7c32b6f99305] Extending root virtual disk to 1048576 {{(pid=61806) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 830.853490] env[61806]: DEBUG oslo_vmware.api [None req-7fa92f02-ad2b-4004-a9a7-cb08f83ec0f1 tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Waiting for the task: (returnval){ [ 830.853490] env[61806]: value = "task-1294664" [ 830.853490] env[61806]: _type = "Task" [ 830.853490] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 830.853490] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-39a3ef08-4be3-44b7-aaf5-b6bbf30e25cb {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 830.868424] env[61806]: DEBUG oslo_vmware.api [None req-7fa92f02-ad2b-4004-a9a7-cb08f83ec0f1 tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Task: {'id': task-1294664, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 830.870736] env[61806]: DEBUG oslo_concurrency.lockutils [None req-4e0c8443-0fd4-459b-a4c5-7cf373aa1089 tempest-ServerAddressesNegativeTestJSON-1526246390 tempest-ServerAddressesNegativeTestJSON-1526246390-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.385s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 830.871464] env[61806]: DEBUG nova.compute.manager [None req-4e0c8443-0fd4-459b-a4c5-7cf373aa1089 tempest-ServerAddressesNegativeTestJSON-1526246390 tempest-ServerAddressesNegativeTestJSON-1526246390-project-member] [instance: c8033551-6591-4e37-ae78-4efe7145b10b] Start building networks asynchronously for instance. {{(pid=61806) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 830.875872] env[61806]: DEBUG oslo_vmware.api [None req-cbb90e6d-7dac-4c30-b860-1f0db7dd091e tempest-MultipleCreateTestJSON-1589015462 tempest-MultipleCreateTestJSON-1589015462-project-member] Waiting for the task: (returnval){ [ 830.875872] env[61806]: value = "task-1294665" [ 830.875872] env[61806]: _type = "Task" [ 830.875872] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 830.876375] env[61806]: DEBUG oslo_concurrency.lockutils [None req-c9a1ae7a-8313-4bdf-8de4-77929ce2d9bc tempest-ServerMetadataNegativeTestJSON-456453902 tempest-ServerMetadataNegativeTestJSON-456453902-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 12.875s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 830.878507] env[61806]: INFO nova.compute.claims [None req-c9a1ae7a-8313-4bdf-8de4-77929ce2d9bc tempest-ServerMetadataNegativeTestJSON-456453902 tempest-ServerMetadataNegativeTestJSON-456453902-project-member] [instance: 7f4907d0-178d-452a-8149-030becde8779] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 830.902567] env[61806]: DEBUG oslo_vmware.api [None req-cbb90e6d-7dac-4c30-b860-1f0db7dd091e tempest-MultipleCreateTestJSON-1589015462 tempest-MultipleCreateTestJSON-1589015462-project-member] Task: {'id': task-1294665, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 830.945053] env[61806]: DEBUG oslo_concurrency.lockutils [None req-cbb90e6d-7dac-4c30-b860-1f0db7dd091e tempest-MultipleCreateTestJSON-1589015462 tempest-MultipleCreateTestJSON-1589015462-project-member] Releasing lock "refresh_cache-ef502a7d-7dc0-4319-8ce3-6a8701288237" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 830.945508] env[61806]: DEBUG nova.compute.manager [None req-cbb90e6d-7dac-4c30-b860-1f0db7dd091e tempest-MultipleCreateTestJSON-1589015462 tempest-MultipleCreateTestJSON-1589015462-project-member] [instance: ef502a7d-7dc0-4319-8ce3-6a8701288237] Instance network_info: |[{"id": "9e500f22-fafc-4f40-ab22-09b4330b6d4f", "address": "fa:16:3e:27:ee:7f", "network": {"id": "56fd57d0-aec6-497c-806f-868ec544c85f", "bridge": "br-int", "label": "tempest-MultipleCreateTestJSON-480554183-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "dc40805c39f34d07be9d07861ad5841d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "359850cc-b061-4c9c-a61c-eb42e0f7c359", "external-id": "nsx-vlan-transportzone-113", "segmentation_id": 113, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9e500f22-fa", "ovs_interfaceid": "9e500f22-fafc-4f40-ab22-09b4330b6d4f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61806) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 830.946380] env[61806]: DEBUG oslo_concurrency.lockutils [req-954128ad-9c63-4aa4-aaac-c35aa507d91d req-da93a2c5-7057-4025-aa14-e2d2d675b0ac service nova] Acquired lock "refresh_cache-ef502a7d-7dc0-4319-8ce3-6a8701288237" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 830.946551] env[61806]: DEBUG nova.network.neutron [req-954128ad-9c63-4aa4-aaac-c35aa507d91d req-da93a2c5-7057-4025-aa14-e2d2d675b0ac service nova] [instance: ef502a7d-7dc0-4319-8ce3-6a8701288237] Refreshing network info cache for port 9e500f22-fafc-4f40-ab22-09b4330b6d4f {{(pid=61806) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 830.948017] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-cbb90e6d-7dac-4c30-b860-1f0db7dd091e tempest-MultipleCreateTestJSON-1589015462 tempest-MultipleCreateTestJSON-1589015462-project-member] [instance: ef502a7d-7dc0-4319-8ce3-6a8701288237] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:27:ee:7f', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '359850cc-b061-4c9c-a61c-eb42e0f7c359', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '9e500f22-fafc-4f40-ab22-09b4330b6d4f', 'vif_model': 'vmxnet3'}] {{(pid=61806) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 830.958711] env[61806]: DEBUG oslo.service.loopingcall [None req-cbb90e6d-7dac-4c30-b860-1f0db7dd091e tempest-MultipleCreateTestJSON-1589015462 tempest-MultipleCreateTestJSON-1589015462-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61806) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 830.962106] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: ef502a7d-7dc0-4319-8ce3-6a8701288237] Creating VM on the ESX host {{(pid=61806) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 830.963111] env[61806]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-ff461a3c-d8b9-4692-b9df-8e884fd75fa7 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 830.987016] env[61806]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 830.987016] env[61806]: value = "task-1294666" [ 830.987016] env[61806]: _type = "Task" [ 830.987016] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 830.997974] env[61806]: DEBUG oslo_vmware.api [-] Task: {'id': task-1294666, 'name': CreateVM_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 831.095433] env[61806]: DEBUG nova.network.neutron [None req-0ea3a38f-3871-4412-a6bc-d21c2086d262 tempest-AttachInterfacesV270Test-1974652806 tempest-AttachInterfacesV270Test-1974652806-project-member] [instance: 117139cd-4ed4-4bea-8ad3-f4ac32c692e4] Successfully updated port: f984af48-8d6f-48b1-a3b3-1f988c36b872 {{(pid=61806) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 831.271190] env[61806]: DEBUG nova.network.neutron [req-954128ad-9c63-4aa4-aaac-c35aa507d91d req-da93a2c5-7057-4025-aa14-e2d2d675b0ac service nova] [instance: ef502a7d-7dc0-4319-8ce3-6a8701288237] Updated VIF entry in instance network info cache for port 9e500f22-fafc-4f40-ab22-09b4330b6d4f. {{(pid=61806) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 831.271573] env[61806]: DEBUG nova.network.neutron [req-954128ad-9c63-4aa4-aaac-c35aa507d91d req-da93a2c5-7057-4025-aa14-e2d2d675b0ac service nova] [instance: ef502a7d-7dc0-4319-8ce3-6a8701288237] Updating instance_info_cache with network_info: [{"id": "9e500f22-fafc-4f40-ab22-09b4330b6d4f", "address": "fa:16:3e:27:ee:7f", "network": {"id": "56fd57d0-aec6-497c-806f-868ec544c85f", "bridge": "br-int", "label": "tempest-MultipleCreateTestJSON-480554183-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "dc40805c39f34d07be9d07861ad5841d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "359850cc-b061-4c9c-a61c-eb42e0f7c359", "external-id": "nsx-vlan-transportzone-113", "segmentation_id": 113, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9e500f22-fa", "ovs_interfaceid": "9e500f22-fafc-4f40-ab22-09b4330b6d4f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61806) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 831.341472] env[61806]: DEBUG oslo_concurrency.lockutils [req-f94c69e0-c504-417f-b00a-34860b38c484 req-dd3c842d-e8d1-42ca-8bbf-0a7c76912bab service nova] Releasing lock "refresh_cache-b41531f2-f28c-4d82-9682-0b557bbaa491" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 831.342461] env[61806]: DEBUG oslo_concurrency.lockutils [None req-6d96025a-fbfa-435a-871d-fe0c696f1d71 tempest-SecurityGroupsTestJSON-1213485581 tempest-SecurityGroupsTestJSON-1213485581-project-member] Acquired lock "refresh_cache-b41531f2-f28c-4d82-9682-0b557bbaa491" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 831.342461] env[61806]: DEBUG nova.network.neutron [None req-6d96025a-fbfa-435a-871d-fe0c696f1d71 tempest-SecurityGroupsTestJSON-1213485581 tempest-SecurityGroupsTestJSON-1213485581-project-member] [instance: b41531f2-f28c-4d82-9682-0b557bbaa491] Building network info cache for instance {{(pid=61806) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 831.366341] env[61806]: DEBUG oslo_vmware.api [None req-7fa92f02-ad2b-4004-a9a7-cb08f83ec0f1 tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Task: {'id': task-1294664, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 831.378915] env[61806]: DEBUG nova.compute.utils [None req-4e0c8443-0fd4-459b-a4c5-7cf373aa1089 tempest-ServerAddressesNegativeTestJSON-1526246390 tempest-ServerAddressesNegativeTestJSON-1526246390-project-member] Using /dev/sd instead of None {{(pid=61806) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 831.380638] env[61806]: DEBUG nova.compute.manager [None req-4e0c8443-0fd4-459b-a4c5-7cf373aa1089 tempest-ServerAddressesNegativeTestJSON-1526246390 tempest-ServerAddressesNegativeTestJSON-1526246390-project-member] [instance: c8033551-6591-4e37-ae78-4efe7145b10b] Allocating IP information in the background. {{(pid=61806) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 831.380956] env[61806]: DEBUG nova.network.neutron [None req-4e0c8443-0fd4-459b-a4c5-7cf373aa1089 tempest-ServerAddressesNegativeTestJSON-1526246390 tempest-ServerAddressesNegativeTestJSON-1526246390-project-member] [instance: c8033551-6591-4e37-ae78-4efe7145b10b] allocate_for_instance() {{(pid=61806) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 831.401257] env[61806]: DEBUG oslo_vmware.api [None req-cbb90e6d-7dac-4c30-b860-1f0db7dd091e tempest-MultipleCreateTestJSON-1589015462 tempest-MultipleCreateTestJSON-1589015462-project-member] Task: {'id': task-1294665, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.085137} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 831.401749] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-cbb90e6d-7dac-4c30-b860-1f0db7dd091e tempest-MultipleCreateTestJSON-1589015462 tempest-MultipleCreateTestJSON-1589015462-project-member] [instance: 25218e83-c1ab-4b97-ade1-7c32b6f99305] Extended root virtual disk {{(pid=61806) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 831.402673] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-db0149e3-9918-433e-ac94-c26906b28e14 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 831.428039] env[61806]: DEBUG nova.virt.vmwareapi.volumeops [None req-cbb90e6d-7dac-4c30-b860-1f0db7dd091e tempest-MultipleCreateTestJSON-1589015462 tempest-MultipleCreateTestJSON-1589015462-project-member] [instance: 25218e83-c1ab-4b97-ade1-7c32b6f99305] Reconfiguring VM instance instance-00000042 to attach disk [datastore2] 25218e83-c1ab-4b97-ade1-7c32b6f99305/25218e83-c1ab-4b97-ade1-7c32b6f99305.vmdk or device None with type sparse {{(pid=61806) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 831.429048] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-3c962128-189b-4838-b902-b6b71cfbe92c {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 831.445118] env[61806]: DEBUG nova.policy [None req-4e0c8443-0fd4-459b-a4c5-7cf373aa1089 tempest-ServerAddressesNegativeTestJSON-1526246390 tempest-ServerAddressesNegativeTestJSON-1526246390-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '9d9826926cc74e0c8aacabe4dd8375d6', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'dd7c12f41e8b4ba489480ffa3959b7ce', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61806) authorize /opt/stack/nova/nova/policy.py:201}} [ 831.453794] env[61806]: DEBUG oslo_vmware.api [None req-cbb90e6d-7dac-4c30-b860-1f0db7dd091e tempest-MultipleCreateTestJSON-1589015462 tempest-MultipleCreateTestJSON-1589015462-project-member] Waiting for the task: (returnval){ [ 831.453794] env[61806]: value = "task-1294667" [ 831.453794] env[61806]: _type = "Task" [ 831.453794] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 831.464207] env[61806]: DEBUG oslo_vmware.api [None req-cbb90e6d-7dac-4c30-b860-1f0db7dd091e tempest-MultipleCreateTestJSON-1589015462 tempest-MultipleCreateTestJSON-1589015462-project-member] Task: {'id': task-1294667, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 831.500300] env[61806]: DEBUG oslo_vmware.api [-] Task: {'id': task-1294666, 'name': CreateVM_Task} progress is 99%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 831.602250] env[61806]: DEBUG oslo_concurrency.lockutils [None req-0ea3a38f-3871-4412-a6bc-d21c2086d262 tempest-AttachInterfacesV270Test-1974652806 tempest-AttachInterfacesV270Test-1974652806-project-member] Acquiring lock "refresh_cache-117139cd-4ed4-4bea-8ad3-f4ac32c692e4" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 831.602948] env[61806]: DEBUG oslo_concurrency.lockutils [None req-0ea3a38f-3871-4412-a6bc-d21c2086d262 tempest-AttachInterfacesV270Test-1974652806 tempest-AttachInterfacesV270Test-1974652806-project-member] Acquired lock "refresh_cache-117139cd-4ed4-4bea-8ad3-f4ac32c692e4" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 831.602948] env[61806]: DEBUG nova.network.neutron [None req-0ea3a38f-3871-4412-a6bc-d21c2086d262 tempest-AttachInterfacesV270Test-1974652806 tempest-AttachInterfacesV270Test-1974652806-project-member] [instance: 117139cd-4ed4-4bea-8ad3-f4ac32c692e4] Building network info cache for instance {{(pid=61806) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 831.774490] env[61806]: DEBUG oslo_concurrency.lockutils [req-954128ad-9c63-4aa4-aaac-c35aa507d91d req-da93a2c5-7057-4025-aa14-e2d2d675b0ac service nova] Releasing lock "refresh_cache-ef502a7d-7dc0-4319-8ce3-6a8701288237" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 831.836299] env[61806]: DEBUG nova.network.neutron [None req-4e0c8443-0fd4-459b-a4c5-7cf373aa1089 tempest-ServerAddressesNegativeTestJSON-1526246390 tempest-ServerAddressesNegativeTestJSON-1526246390-project-member] [instance: c8033551-6591-4e37-ae78-4efe7145b10b] Successfully created port: e8bfbf94-495b-4d13-8d2c-7a02e7a90387 {{(pid=61806) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 831.866329] env[61806]: DEBUG oslo_vmware.api [None req-7fa92f02-ad2b-4004-a9a7-cb08f83ec0f1 tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Task: {'id': task-1294664, 'name': PowerOnVM_Task, 'duration_secs': 0.761792} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 831.866611] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [None req-7fa92f02-ad2b-4004-a9a7-cb08f83ec0f1 tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] [instance: 67d1bdd8-cf8f-4fb5-bd3c-42008e4d60e3] Powered on the VM {{(pid=61806) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 831.866828] env[61806]: INFO nova.compute.manager [None req-7fa92f02-ad2b-4004-a9a7-cb08f83ec0f1 tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] [instance: 67d1bdd8-cf8f-4fb5-bd3c-42008e4d60e3] Took 10.77 seconds to spawn the instance on the hypervisor. [ 831.867017] env[61806]: DEBUG nova.compute.manager [None req-7fa92f02-ad2b-4004-a9a7-cb08f83ec0f1 tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] [instance: 67d1bdd8-cf8f-4fb5-bd3c-42008e4d60e3] Checking state {{(pid=61806) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 831.867798] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ba9bb84b-411b-48c1-a695-69524df5e421 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 831.889882] env[61806]: DEBUG nova.compute.manager [None req-4e0c8443-0fd4-459b-a4c5-7cf373aa1089 tempest-ServerAddressesNegativeTestJSON-1526246390 tempest-ServerAddressesNegativeTestJSON-1526246390-project-member] [instance: c8033551-6591-4e37-ae78-4efe7145b10b] Start building block device mappings for instance. {{(pid=61806) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 831.965885] env[61806]: DEBUG oslo_vmware.api [None req-cbb90e6d-7dac-4c30-b860-1f0db7dd091e tempest-MultipleCreateTestJSON-1589015462 tempest-MultipleCreateTestJSON-1589015462-project-member] Task: {'id': task-1294667, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 832.004021] env[61806]: DEBUG oslo_vmware.api [-] Task: {'id': task-1294666, 'name': CreateVM_Task, 'duration_secs': 0.576874} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 832.004021] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: ef502a7d-7dc0-4319-8ce3-6a8701288237] Created VM on the ESX host {{(pid=61806) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 832.004021] env[61806]: DEBUG oslo_concurrency.lockutils [None req-cbb90e6d-7dac-4c30-b860-1f0db7dd091e tempest-MultipleCreateTestJSON-1589015462 tempest-MultipleCreateTestJSON-1589015462-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 832.004021] env[61806]: DEBUG oslo_concurrency.lockutils [None req-cbb90e6d-7dac-4c30-b860-1f0db7dd091e tempest-MultipleCreateTestJSON-1589015462 tempest-MultipleCreateTestJSON-1589015462-project-member] Acquired lock "[datastore2] devstack-image-cache_base/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 832.004021] env[61806]: DEBUG oslo_concurrency.lockutils [None req-cbb90e6d-7dac-4c30-b860-1f0db7dd091e tempest-MultipleCreateTestJSON-1589015462 tempest-MultipleCreateTestJSON-1589015462-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 832.004021] env[61806]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-1b1d3c5f-811e-4db6-8212-881c70abce74 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 832.010326] env[61806]: DEBUG oslo_vmware.api [None req-cbb90e6d-7dac-4c30-b860-1f0db7dd091e tempest-MultipleCreateTestJSON-1589015462 tempest-MultipleCreateTestJSON-1589015462-project-member] Waiting for the task: (returnval){ [ 832.010326] env[61806]: value = "session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]521f5ca8-f196-e7af-1f04-f560e5c8fef0" [ 832.010326] env[61806]: _type = "Task" [ 832.010326] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 832.018461] env[61806]: DEBUG oslo_vmware.api [None req-cbb90e6d-7dac-4c30-b860-1f0db7dd091e tempest-MultipleCreateTestJSON-1589015462 tempest-MultipleCreateTestJSON-1589015462-project-member] Task: {'id': session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]521f5ca8-f196-e7af-1f04-f560e5c8fef0, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 832.141861] env[61806]: DEBUG nova.network.neutron [None req-0ea3a38f-3871-4412-a6bc-d21c2086d262 tempest-AttachInterfacesV270Test-1974652806 tempest-AttachInterfacesV270Test-1974652806-project-member] [instance: 117139cd-4ed4-4bea-8ad3-f4ac32c692e4] Instance cache missing network info. {{(pid=61806) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 832.169805] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-420eeb06-d99e-4db0-b85e-bd31dfbe5b0a {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 832.173241] env[61806]: DEBUG nova.network.neutron [None req-6d96025a-fbfa-435a-871d-fe0c696f1d71 tempest-SecurityGroupsTestJSON-1213485581 tempest-SecurityGroupsTestJSON-1213485581-project-member] [instance: b41531f2-f28c-4d82-9682-0b557bbaa491] Updating instance_info_cache with network_info: [{"id": "513ca4b5-b62c-486a-9860-cf267a15c3e5", "address": "fa:16:3e:40:52:86", "network": {"id": "dfb52c93-e553-40d9-acb5-768cc6b67c89", "bridge": "br-int", "label": "tempest-SecurityGroupsTestJSON-1006891343-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "cfce93add38d40d79f294017f52e13d5", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6b6942d7-d4ab-4b2a-8d0f-76bf2a2478ad", "external-id": "nsx-vlan-transportzone-871", "segmentation_id": 871, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap513ca4b5-b6", "ovs_interfaceid": "513ca4b5-b62c-486a-9860-cf267a15c3e5", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61806) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 832.180455] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-21aade20-562b-4962-afe3-ad6964e7ba92 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 832.217040] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5420c9e4-30c7-4b30-b70a-b71eea2c5f86 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 832.225868] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-74bada02-363b-47bb-ba21-944c38684f80 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 832.241873] env[61806]: DEBUG nova.compute.provider_tree [None req-c9a1ae7a-8313-4bdf-8de4-77929ce2d9bc tempest-ServerMetadataNegativeTestJSON-456453902 tempest-ServerMetadataNegativeTestJSON-456453902-project-member] Inventory has not changed in ProviderTree for provider: a2858be1-fd22-4e08-979e-87ad25293407 {{(pid=61806) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 832.375352] env[61806]: DEBUG nova.network.neutron [None req-0ea3a38f-3871-4412-a6bc-d21c2086d262 tempest-AttachInterfacesV270Test-1974652806 tempest-AttachInterfacesV270Test-1974652806-project-member] [instance: 117139cd-4ed4-4bea-8ad3-f4ac32c692e4] Updating instance_info_cache with network_info: [{"id": "f984af48-8d6f-48b1-a3b3-1f988c36b872", "address": "fa:16:3e:66:04:70", "network": {"id": "b3bed2c3-cf75-47f9-a112-c9f6417fc40d", "bridge": "br-int", "label": "tempest-AttachInterfacesV270Test-1376920766-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b0499ad8c8bb4b6d80b523fa85388e53", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "2c06e3c2-8edb-4cf0-be6b-45dfe059c00b", "external-id": "nsx-vlan-transportzone-264", "segmentation_id": 264, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf984af48-8d", "ovs_interfaceid": "f984af48-8d6f-48b1-a3b3-1f988c36b872", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61806) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 832.386075] env[61806]: INFO nova.compute.manager [None req-7fa92f02-ad2b-4004-a9a7-cb08f83ec0f1 tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] [instance: 67d1bdd8-cf8f-4fb5-bd3c-42008e4d60e3] Took 29.81 seconds to build instance. [ 832.464711] env[61806]: DEBUG oslo_vmware.api [None req-cbb90e6d-7dac-4c30-b860-1f0db7dd091e tempest-MultipleCreateTestJSON-1589015462 tempest-MultipleCreateTestJSON-1589015462-project-member] Task: {'id': task-1294667, 'name': ReconfigVM_Task, 'duration_secs': 0.797881} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 832.465337] env[61806]: DEBUG nova.virt.vmwareapi.volumeops [None req-cbb90e6d-7dac-4c30-b860-1f0db7dd091e tempest-MultipleCreateTestJSON-1589015462 tempest-MultipleCreateTestJSON-1589015462-project-member] [instance: 25218e83-c1ab-4b97-ade1-7c32b6f99305] Reconfigured VM instance instance-00000042 to attach disk [datastore2] 25218e83-c1ab-4b97-ade1-7c32b6f99305/25218e83-c1ab-4b97-ade1-7c32b6f99305.vmdk or device None with type sparse {{(pid=61806) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 832.465715] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-b851c59b-e239-4cf1-b65d-a91baf167614 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 832.472664] env[61806]: DEBUG oslo_vmware.api [None req-cbb90e6d-7dac-4c30-b860-1f0db7dd091e tempest-MultipleCreateTestJSON-1589015462 tempest-MultipleCreateTestJSON-1589015462-project-member] Waiting for the task: (returnval){ [ 832.472664] env[61806]: value = "task-1294668" [ 832.472664] env[61806]: _type = "Task" [ 832.472664] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 832.480746] env[61806]: DEBUG oslo_vmware.api [None req-cbb90e6d-7dac-4c30-b860-1f0db7dd091e tempest-MultipleCreateTestJSON-1589015462 tempest-MultipleCreateTestJSON-1589015462-project-member] Task: {'id': task-1294668, 'name': Rename_Task} progress is 5%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 832.520019] env[61806]: DEBUG oslo_vmware.api [None req-cbb90e6d-7dac-4c30-b860-1f0db7dd091e tempest-MultipleCreateTestJSON-1589015462 tempest-MultipleCreateTestJSON-1589015462-project-member] Task: {'id': session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]521f5ca8-f196-e7af-1f04-f560e5c8fef0, 'name': SearchDatastore_Task, 'duration_secs': 0.073838} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 832.520426] env[61806]: DEBUG oslo_concurrency.lockutils [None req-cbb90e6d-7dac-4c30-b860-1f0db7dd091e tempest-MultipleCreateTestJSON-1589015462 tempest-MultipleCreateTestJSON-1589015462-project-member] Releasing lock "[datastore2] devstack-image-cache_base/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 832.520896] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-cbb90e6d-7dac-4c30-b860-1f0db7dd091e tempest-MultipleCreateTestJSON-1589015462 tempest-MultipleCreateTestJSON-1589015462-project-member] [instance: ef502a7d-7dc0-4319-8ce3-6a8701288237] Processing image ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa {{(pid=61806) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 832.520987] env[61806]: DEBUG oslo_concurrency.lockutils [None req-cbb90e6d-7dac-4c30-b860-1f0db7dd091e tempest-MultipleCreateTestJSON-1589015462 tempest-MultipleCreateTestJSON-1589015462-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa.vmdk" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 832.521118] env[61806]: DEBUG oslo_concurrency.lockutils [None req-cbb90e6d-7dac-4c30-b860-1f0db7dd091e tempest-MultipleCreateTestJSON-1589015462 tempest-MultipleCreateTestJSON-1589015462-project-member] Acquired lock "[datastore2] devstack-image-cache_base/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa.vmdk" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 832.521358] env[61806]: DEBUG nova.virt.vmwareapi.ds_util [None req-cbb90e6d-7dac-4c30-b860-1f0db7dd091e tempest-MultipleCreateTestJSON-1589015462 tempest-MultipleCreateTestJSON-1589015462-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61806) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 832.521770] env[61806]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-a7a4289d-5b0e-4443-8110-ab093e90ecb9 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 832.531549] env[61806]: DEBUG nova.virt.vmwareapi.ds_util [None req-cbb90e6d-7dac-4c30-b860-1f0db7dd091e tempest-MultipleCreateTestJSON-1589015462 tempest-MultipleCreateTestJSON-1589015462-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61806) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 832.531756] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-cbb90e6d-7dac-4c30-b860-1f0db7dd091e tempest-MultipleCreateTestJSON-1589015462 tempest-MultipleCreateTestJSON-1589015462-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61806) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 832.532543] env[61806]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-208c4b3d-9a58-44d6-9333-3eafa1ca489b {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 832.537463] env[61806]: DEBUG oslo_vmware.api [None req-cbb90e6d-7dac-4c30-b860-1f0db7dd091e tempest-MultipleCreateTestJSON-1589015462 tempest-MultipleCreateTestJSON-1589015462-project-member] Waiting for the task: (returnval){ [ 832.537463] env[61806]: value = "session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]52ba4f41-f8c1-ff86-42fe-6a48c533f9be" [ 832.537463] env[61806]: _type = "Task" [ 832.537463] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 832.544898] env[61806]: DEBUG oslo_vmware.api [None req-cbb90e6d-7dac-4c30-b860-1f0db7dd091e tempest-MultipleCreateTestJSON-1589015462 tempest-MultipleCreateTestJSON-1589015462-project-member] Task: {'id': session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]52ba4f41-f8c1-ff86-42fe-6a48c533f9be, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 832.635222] env[61806]: DEBUG nova.compute.manager [req-33cd6231-823f-4493-b038-5a7830e007b9 req-2414cc26-5318-4c9a-b26b-617a7de33cab service nova] [instance: 117139cd-4ed4-4bea-8ad3-f4ac32c692e4] Received event network-vif-plugged-f984af48-8d6f-48b1-a3b3-1f988c36b872 {{(pid=61806) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 832.635762] env[61806]: DEBUG oslo_concurrency.lockutils [req-33cd6231-823f-4493-b038-5a7830e007b9 req-2414cc26-5318-4c9a-b26b-617a7de33cab service nova] Acquiring lock "117139cd-4ed4-4bea-8ad3-f4ac32c692e4-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 832.635762] env[61806]: DEBUG oslo_concurrency.lockutils [req-33cd6231-823f-4493-b038-5a7830e007b9 req-2414cc26-5318-4c9a-b26b-617a7de33cab service nova] Lock "117139cd-4ed4-4bea-8ad3-f4ac32c692e4-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 832.635866] env[61806]: DEBUG oslo_concurrency.lockutils [req-33cd6231-823f-4493-b038-5a7830e007b9 req-2414cc26-5318-4c9a-b26b-617a7de33cab service nova] Lock "117139cd-4ed4-4bea-8ad3-f4ac32c692e4-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 832.636050] env[61806]: DEBUG nova.compute.manager [req-33cd6231-823f-4493-b038-5a7830e007b9 req-2414cc26-5318-4c9a-b26b-617a7de33cab service nova] [instance: 117139cd-4ed4-4bea-8ad3-f4ac32c692e4] No waiting events found dispatching network-vif-plugged-f984af48-8d6f-48b1-a3b3-1f988c36b872 {{(pid=61806) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 832.636229] env[61806]: WARNING nova.compute.manager [req-33cd6231-823f-4493-b038-5a7830e007b9 req-2414cc26-5318-4c9a-b26b-617a7de33cab service nova] [instance: 117139cd-4ed4-4bea-8ad3-f4ac32c692e4] Received unexpected event network-vif-plugged-f984af48-8d6f-48b1-a3b3-1f988c36b872 for instance with vm_state building and task_state spawning. [ 832.636398] env[61806]: DEBUG nova.compute.manager [req-33cd6231-823f-4493-b038-5a7830e007b9 req-2414cc26-5318-4c9a-b26b-617a7de33cab service nova] [instance: 117139cd-4ed4-4bea-8ad3-f4ac32c692e4] Received event network-changed-f984af48-8d6f-48b1-a3b3-1f988c36b872 {{(pid=61806) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 832.636558] env[61806]: DEBUG nova.compute.manager [req-33cd6231-823f-4493-b038-5a7830e007b9 req-2414cc26-5318-4c9a-b26b-617a7de33cab service nova] [instance: 117139cd-4ed4-4bea-8ad3-f4ac32c692e4] Refreshing instance network info cache due to event network-changed-f984af48-8d6f-48b1-a3b3-1f988c36b872. {{(pid=61806) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 832.636733] env[61806]: DEBUG oslo_concurrency.lockutils [req-33cd6231-823f-4493-b038-5a7830e007b9 req-2414cc26-5318-4c9a-b26b-617a7de33cab service nova] Acquiring lock "refresh_cache-117139cd-4ed4-4bea-8ad3-f4ac32c692e4" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 832.675590] env[61806]: DEBUG oslo_concurrency.lockutils [None req-6d96025a-fbfa-435a-871d-fe0c696f1d71 tempest-SecurityGroupsTestJSON-1213485581 tempest-SecurityGroupsTestJSON-1213485581-project-member] Releasing lock "refresh_cache-b41531f2-f28c-4d82-9682-0b557bbaa491" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 832.677683] env[61806]: DEBUG nova.compute.manager [None req-6d96025a-fbfa-435a-871d-fe0c696f1d71 tempest-SecurityGroupsTestJSON-1213485581 tempest-SecurityGroupsTestJSON-1213485581-project-member] [instance: b41531f2-f28c-4d82-9682-0b557bbaa491] Checking state {{(pid=61806) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 832.678575] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7320c863-eadc-412f-8be8-295501ee1a83 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 832.744234] env[61806]: DEBUG nova.scheduler.client.report [None req-c9a1ae7a-8313-4bdf-8de4-77929ce2d9bc tempest-ServerMetadataNegativeTestJSON-456453902 tempest-ServerMetadataNegativeTestJSON-456453902-project-member] Inventory has not changed for provider a2858be1-fd22-4e08-979e-87ad25293407 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 138, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61806) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 832.883126] env[61806]: DEBUG oslo_concurrency.lockutils [None req-0ea3a38f-3871-4412-a6bc-d21c2086d262 tempest-AttachInterfacesV270Test-1974652806 tempest-AttachInterfacesV270Test-1974652806-project-member] Releasing lock "refresh_cache-117139cd-4ed4-4bea-8ad3-f4ac32c692e4" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 832.883468] env[61806]: DEBUG nova.compute.manager [None req-0ea3a38f-3871-4412-a6bc-d21c2086d262 tempest-AttachInterfacesV270Test-1974652806 tempest-AttachInterfacesV270Test-1974652806-project-member] [instance: 117139cd-4ed4-4bea-8ad3-f4ac32c692e4] Instance network_info: |[{"id": "f984af48-8d6f-48b1-a3b3-1f988c36b872", "address": "fa:16:3e:66:04:70", "network": {"id": "b3bed2c3-cf75-47f9-a112-c9f6417fc40d", "bridge": "br-int", "label": "tempest-AttachInterfacesV270Test-1376920766-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b0499ad8c8bb4b6d80b523fa85388e53", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "2c06e3c2-8edb-4cf0-be6b-45dfe059c00b", "external-id": "nsx-vlan-transportzone-264", "segmentation_id": 264, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf984af48-8d", "ovs_interfaceid": "f984af48-8d6f-48b1-a3b3-1f988c36b872", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61806) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 832.883863] env[61806]: DEBUG oslo_concurrency.lockutils [req-33cd6231-823f-4493-b038-5a7830e007b9 req-2414cc26-5318-4c9a-b26b-617a7de33cab service nova] Acquired lock "refresh_cache-117139cd-4ed4-4bea-8ad3-f4ac32c692e4" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 832.884116] env[61806]: DEBUG nova.network.neutron [req-33cd6231-823f-4493-b038-5a7830e007b9 req-2414cc26-5318-4c9a-b26b-617a7de33cab service nova] [instance: 117139cd-4ed4-4bea-8ad3-f4ac32c692e4] Refreshing network info cache for port f984af48-8d6f-48b1-a3b3-1f988c36b872 {{(pid=61806) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 832.885450] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-0ea3a38f-3871-4412-a6bc-d21c2086d262 tempest-AttachInterfacesV270Test-1974652806 tempest-AttachInterfacesV270Test-1974652806-project-member] [instance: 117139cd-4ed4-4bea-8ad3-f4ac32c692e4] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:66:04:70', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '2c06e3c2-8edb-4cf0-be6b-45dfe059c00b', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'f984af48-8d6f-48b1-a3b3-1f988c36b872', 'vif_model': 'vmxnet3'}] {{(pid=61806) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 832.893014] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [None req-0ea3a38f-3871-4412-a6bc-d21c2086d262 tempest-AttachInterfacesV270Test-1974652806 tempest-AttachInterfacesV270Test-1974652806-project-member] Creating folder: Project (b0499ad8c8bb4b6d80b523fa85388e53). Parent ref: group-v277609. {{(pid=61806) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 832.896180] env[61806]: DEBUG oslo_concurrency.lockutils [None req-7fa92f02-ad2b-4004-a9a7-cb08f83ec0f1 tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Lock "67d1bdd8-cf8f-4fb5-bd3c-42008e4d60e3" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 94.728s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 832.896434] env[61806]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-16bbdc09-1809-45bc-a924-0abf9ee84c2d {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 832.899707] env[61806]: DEBUG nova.compute.manager [None req-4e0c8443-0fd4-459b-a4c5-7cf373aa1089 tempest-ServerAddressesNegativeTestJSON-1526246390 tempest-ServerAddressesNegativeTestJSON-1526246390-project-member] [instance: c8033551-6591-4e37-ae78-4efe7145b10b] Start spawning the instance on the hypervisor. {{(pid=61806) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 832.909964] env[61806]: INFO nova.virt.vmwareapi.vm_util [None req-0ea3a38f-3871-4412-a6bc-d21c2086d262 tempest-AttachInterfacesV270Test-1974652806 tempest-AttachInterfacesV270Test-1974652806-project-member] Created folder: Project (b0499ad8c8bb4b6d80b523fa85388e53) in parent group-v277609. [ 832.910264] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [None req-0ea3a38f-3871-4412-a6bc-d21c2086d262 tempest-AttachInterfacesV270Test-1974652806 tempest-AttachInterfacesV270Test-1974652806-project-member] Creating folder: Instances. Parent ref: group-v277671. {{(pid=61806) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 832.910527] env[61806]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-b3ee03d9-42ec-468e-b39a-e183654ea893 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 832.918869] env[61806]: INFO nova.virt.vmwareapi.vm_util [None req-0ea3a38f-3871-4412-a6bc-d21c2086d262 tempest-AttachInterfacesV270Test-1974652806 tempest-AttachInterfacesV270Test-1974652806-project-member] Created folder: Instances in parent group-v277671. [ 832.919084] env[61806]: DEBUG oslo.service.loopingcall [None req-0ea3a38f-3871-4412-a6bc-d21c2086d262 tempest-AttachInterfacesV270Test-1974652806 tempest-AttachInterfacesV270Test-1974652806-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61806) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 832.920961] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 117139cd-4ed4-4bea-8ad3-f4ac32c692e4] Creating VM on the ESX host {{(pid=61806) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 832.921368] env[61806]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-95a5599c-c17d-48c5-a1ed-5a8bbcd87a65 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 832.938085] env[61806]: DEBUG nova.virt.hardware [None req-4e0c8443-0fd4-459b-a4c5-7cf373aa1089 tempest-ServerAddressesNegativeTestJSON-1526246390 tempest-ServerAddressesNegativeTestJSON-1526246390-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-15T18:05:10Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-15T18:04:52Z,direct_url=,disk_format='vmdk',id=ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='84ba83fce2c349a6988173c5d6fc3b07',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-15T18:04:53Z,virtual_size=,visibility=), allow threads: False {{(pid=61806) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 832.938343] env[61806]: DEBUG nova.virt.hardware [None req-4e0c8443-0fd4-459b-a4c5-7cf373aa1089 tempest-ServerAddressesNegativeTestJSON-1526246390 tempest-ServerAddressesNegativeTestJSON-1526246390-project-member] Flavor limits 0:0:0 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 832.938505] env[61806]: DEBUG nova.virt.hardware [None req-4e0c8443-0fd4-459b-a4c5-7cf373aa1089 tempest-ServerAddressesNegativeTestJSON-1526246390 tempest-ServerAddressesNegativeTestJSON-1526246390-project-member] Image limits 0:0:0 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 832.938692] env[61806]: DEBUG nova.virt.hardware [None req-4e0c8443-0fd4-459b-a4c5-7cf373aa1089 tempest-ServerAddressesNegativeTestJSON-1526246390 tempest-ServerAddressesNegativeTestJSON-1526246390-project-member] Flavor pref 0:0:0 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 832.938848] env[61806]: DEBUG nova.virt.hardware [None req-4e0c8443-0fd4-459b-a4c5-7cf373aa1089 tempest-ServerAddressesNegativeTestJSON-1526246390 tempest-ServerAddressesNegativeTestJSON-1526246390-project-member] Image pref 0:0:0 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 832.938997] env[61806]: DEBUG nova.virt.hardware [None req-4e0c8443-0fd4-459b-a4c5-7cf373aa1089 tempest-ServerAddressesNegativeTestJSON-1526246390 tempest-ServerAddressesNegativeTestJSON-1526246390-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 832.939227] env[61806]: DEBUG nova.virt.hardware [None req-4e0c8443-0fd4-459b-a4c5-7cf373aa1089 tempest-ServerAddressesNegativeTestJSON-1526246390 tempest-ServerAddressesNegativeTestJSON-1526246390-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61806) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 832.939396] env[61806]: DEBUG nova.virt.hardware [None req-4e0c8443-0fd4-459b-a4c5-7cf373aa1089 tempest-ServerAddressesNegativeTestJSON-1526246390 tempest-ServerAddressesNegativeTestJSON-1526246390-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61806) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 832.939568] env[61806]: DEBUG nova.virt.hardware [None req-4e0c8443-0fd4-459b-a4c5-7cf373aa1089 tempest-ServerAddressesNegativeTestJSON-1526246390 tempest-ServerAddressesNegativeTestJSON-1526246390-project-member] Got 1 possible topologies {{(pid=61806) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 832.939740] env[61806]: DEBUG nova.virt.hardware [None req-4e0c8443-0fd4-459b-a4c5-7cf373aa1089 tempest-ServerAddressesNegativeTestJSON-1526246390 tempest-ServerAddressesNegativeTestJSON-1526246390-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61806) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 832.939921] env[61806]: DEBUG nova.virt.hardware [None req-4e0c8443-0fd4-459b-a4c5-7cf373aa1089 tempest-ServerAddressesNegativeTestJSON-1526246390 tempest-ServerAddressesNegativeTestJSON-1526246390-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61806) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 832.943072] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bb187531-8303-418e-ab3c-63824295d02e {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 832.952341] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-36566edb-2c3c-48a8-9773-183f3fd21124 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 832.956064] env[61806]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 832.956064] env[61806]: value = "task-1294671" [ 832.956064] env[61806]: _type = "Task" [ 832.956064] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 832.973542] env[61806]: DEBUG oslo_vmware.api [-] Task: {'id': task-1294671, 'name': CreateVM_Task} progress is 10%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 832.983341] env[61806]: DEBUG oslo_vmware.api [None req-cbb90e6d-7dac-4c30-b860-1f0db7dd091e tempest-MultipleCreateTestJSON-1589015462 tempest-MultipleCreateTestJSON-1589015462-project-member] Task: {'id': task-1294668, 'name': Rename_Task, 'duration_secs': 0.392686} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 832.983638] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [None req-cbb90e6d-7dac-4c30-b860-1f0db7dd091e tempest-MultipleCreateTestJSON-1589015462 tempest-MultipleCreateTestJSON-1589015462-project-member] [instance: 25218e83-c1ab-4b97-ade1-7c32b6f99305] Powering on the VM {{(pid=61806) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 832.983924] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-aadb8450-a77e-48ef-8791-8831a26458bf {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 832.989815] env[61806]: DEBUG oslo_vmware.api [None req-cbb90e6d-7dac-4c30-b860-1f0db7dd091e tempest-MultipleCreateTestJSON-1589015462 tempest-MultipleCreateTestJSON-1589015462-project-member] Waiting for the task: (returnval){ [ 832.989815] env[61806]: value = "task-1294672" [ 832.989815] env[61806]: _type = "Task" [ 832.989815] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 832.997377] env[61806]: DEBUG oslo_vmware.api [None req-cbb90e6d-7dac-4c30-b860-1f0db7dd091e tempest-MultipleCreateTestJSON-1589015462 tempest-MultipleCreateTestJSON-1589015462-project-member] Task: {'id': task-1294672, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 833.050804] env[61806]: DEBUG oslo_vmware.api [None req-cbb90e6d-7dac-4c30-b860-1f0db7dd091e tempest-MultipleCreateTestJSON-1589015462 tempest-MultipleCreateTestJSON-1589015462-project-member] Task: {'id': session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]52ba4f41-f8c1-ff86-42fe-6a48c533f9be, 'name': SearchDatastore_Task, 'duration_secs': 0.022884} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 833.051624] env[61806]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-703d34b7-3e6f-48cf-aba9-9c7debfc54bc {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 833.057172] env[61806]: DEBUG oslo_vmware.api [None req-cbb90e6d-7dac-4c30-b860-1f0db7dd091e tempest-MultipleCreateTestJSON-1589015462 tempest-MultipleCreateTestJSON-1589015462-project-member] Waiting for the task: (returnval){ [ 833.057172] env[61806]: value = "session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]5211ba9f-69dc-37eb-7693-86f8e6380416" [ 833.057172] env[61806]: _type = "Task" [ 833.057172] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 833.065118] env[61806]: DEBUG oslo_vmware.api [None req-cbb90e6d-7dac-4c30-b860-1f0db7dd091e tempest-MultipleCreateTestJSON-1589015462 tempest-MultipleCreateTestJSON-1589015462-project-member] Task: {'id': session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]5211ba9f-69dc-37eb-7693-86f8e6380416, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 833.148915] env[61806]: DEBUG nova.network.neutron [req-33cd6231-823f-4493-b038-5a7830e007b9 req-2414cc26-5318-4c9a-b26b-617a7de33cab service nova] [instance: 117139cd-4ed4-4bea-8ad3-f4ac32c692e4] Updated VIF entry in instance network info cache for port f984af48-8d6f-48b1-a3b3-1f988c36b872. {{(pid=61806) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 833.149515] env[61806]: DEBUG nova.network.neutron [req-33cd6231-823f-4493-b038-5a7830e007b9 req-2414cc26-5318-4c9a-b26b-617a7de33cab service nova] [instance: 117139cd-4ed4-4bea-8ad3-f4ac32c692e4] Updating instance_info_cache with network_info: [{"id": "f984af48-8d6f-48b1-a3b3-1f988c36b872", "address": "fa:16:3e:66:04:70", "network": {"id": "b3bed2c3-cf75-47f9-a112-c9f6417fc40d", "bridge": "br-int", "label": "tempest-AttachInterfacesV270Test-1376920766-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b0499ad8c8bb4b6d80b523fa85388e53", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "2c06e3c2-8edb-4cf0-be6b-45dfe059c00b", "external-id": "nsx-vlan-transportzone-264", "segmentation_id": 264, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf984af48-8d", "ovs_interfaceid": "f984af48-8d6f-48b1-a3b3-1f988c36b872", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61806) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 833.249748] env[61806]: DEBUG oslo_concurrency.lockutils [None req-c9a1ae7a-8313-4bdf-8de4-77929ce2d9bc tempest-ServerMetadataNegativeTestJSON-456453902 tempest-ServerMetadataNegativeTestJSON-456453902-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.373s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 833.250399] env[61806]: DEBUG nova.compute.manager [None req-c9a1ae7a-8313-4bdf-8de4-77929ce2d9bc tempest-ServerMetadataNegativeTestJSON-456453902 tempest-ServerMetadataNegativeTestJSON-456453902-project-member] [instance: 7f4907d0-178d-452a-8149-030becde8779] Start building networks asynchronously for instance. {{(pid=61806) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 833.253428] env[61806]: DEBUG oslo_concurrency.lockutils [None req-7cf484b6-b263-4f99-a8c0-b55d49de0809 tempest-ServerShowV247Test-1218443059 tempest-ServerShowV247Test-1218443059-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: waited 14.027s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 833.253705] env[61806]: DEBUG nova.objects.instance [None req-7cf484b6-b263-4f99-a8c0-b55d49de0809 tempest-ServerShowV247Test-1218443059 tempest-ServerShowV247Test-1218443059-project-member] [instance: 4febd093-9f2b-494e-b175-e4693b0e3e0d] Trying to apply a migration context that does not seem to be set for this instance {{(pid=61806) apply_migration_context /opt/stack/nova/nova/objects/instance.py:1067}} [ 833.466009] env[61806]: DEBUG oslo_vmware.api [-] Task: {'id': task-1294671, 'name': CreateVM_Task} progress is 25%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 833.501098] env[61806]: DEBUG oslo_vmware.api [None req-cbb90e6d-7dac-4c30-b860-1f0db7dd091e tempest-MultipleCreateTestJSON-1589015462 tempest-MultipleCreateTestJSON-1589015462-project-member] Task: {'id': task-1294672, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 833.566827] env[61806]: DEBUG oslo_vmware.api [None req-cbb90e6d-7dac-4c30-b860-1f0db7dd091e tempest-MultipleCreateTestJSON-1589015462 tempest-MultipleCreateTestJSON-1589015462-project-member] Task: {'id': session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]5211ba9f-69dc-37eb-7693-86f8e6380416, 'name': SearchDatastore_Task, 'duration_secs': 0.016457} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 833.567120] env[61806]: DEBUG oslo_concurrency.lockutils [None req-cbb90e6d-7dac-4c30-b860-1f0db7dd091e tempest-MultipleCreateTestJSON-1589015462 tempest-MultipleCreateTestJSON-1589015462-project-member] Releasing lock "[datastore2] devstack-image-cache_base/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa.vmdk" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 833.567379] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [None req-cbb90e6d-7dac-4c30-b860-1f0db7dd091e tempest-MultipleCreateTestJSON-1589015462 tempest-MultipleCreateTestJSON-1589015462-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa.vmdk to [datastore2] ef502a7d-7dc0-4319-8ce3-6a8701288237/ef502a7d-7dc0-4319-8ce3-6a8701288237.vmdk {{(pid=61806) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 833.567660] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-9768b8d4-bff3-403a-b9ef-ea2a8343d6fb {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 833.574215] env[61806]: DEBUG oslo_vmware.api [None req-cbb90e6d-7dac-4c30-b860-1f0db7dd091e tempest-MultipleCreateTestJSON-1589015462 tempest-MultipleCreateTestJSON-1589015462-project-member] Waiting for the task: (returnval){ [ 833.574215] env[61806]: value = "task-1294673" [ 833.574215] env[61806]: _type = "Task" [ 833.574215] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 833.581882] env[61806]: DEBUG oslo_vmware.api [None req-cbb90e6d-7dac-4c30-b860-1f0db7dd091e tempest-MultipleCreateTestJSON-1589015462 tempest-MultipleCreateTestJSON-1589015462-project-member] Task: {'id': task-1294673, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 833.626871] env[61806]: DEBUG nova.network.neutron [None req-4e0c8443-0fd4-459b-a4c5-7cf373aa1089 tempest-ServerAddressesNegativeTestJSON-1526246390 tempest-ServerAddressesNegativeTestJSON-1526246390-project-member] [instance: c8033551-6591-4e37-ae78-4efe7145b10b] Successfully updated port: e8bfbf94-495b-4d13-8d2c-7a02e7a90387 {{(pid=61806) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 833.652126] env[61806]: DEBUG oslo_concurrency.lockutils [req-33cd6231-823f-4493-b038-5a7830e007b9 req-2414cc26-5318-4c9a-b26b-617a7de33cab service nova] Releasing lock "refresh_cache-117139cd-4ed4-4bea-8ad3-f4ac32c692e4" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 833.695533] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c7bff500-3004-4cd3-8bdf-b98f007c70b2 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 833.703300] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-6d96025a-fbfa-435a-871d-fe0c696f1d71 tempest-SecurityGroupsTestJSON-1213485581 tempest-SecurityGroupsTestJSON-1213485581-project-member] [instance: b41531f2-f28c-4d82-9682-0b557bbaa491] Doing hard reboot of VM {{(pid=61806) reboot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1064}} [ 833.703558] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ResetVM_Task with opID=oslo.vmware-d34fb956-6b3b-464f-a20d-1623a4aa1fa5 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 833.709536] env[61806]: DEBUG oslo_vmware.api [None req-6d96025a-fbfa-435a-871d-fe0c696f1d71 tempest-SecurityGroupsTestJSON-1213485581 tempest-SecurityGroupsTestJSON-1213485581-project-member] Waiting for the task: (returnval){ [ 833.709536] env[61806]: value = "task-1294674" [ 833.709536] env[61806]: _type = "Task" [ 833.709536] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 833.717117] env[61806]: DEBUG oslo_vmware.api [None req-6d96025a-fbfa-435a-871d-fe0c696f1d71 tempest-SecurityGroupsTestJSON-1213485581 tempest-SecurityGroupsTestJSON-1213485581-project-member] Task: {'id': task-1294674, 'name': ResetVM_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 833.758776] env[61806]: DEBUG nova.compute.utils [None req-c9a1ae7a-8313-4bdf-8de4-77929ce2d9bc tempest-ServerMetadataNegativeTestJSON-456453902 tempest-ServerMetadataNegativeTestJSON-456453902-project-member] Using /dev/sd instead of None {{(pid=61806) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 833.763163] env[61806]: DEBUG nova.compute.manager [None req-c9a1ae7a-8313-4bdf-8de4-77929ce2d9bc tempest-ServerMetadataNegativeTestJSON-456453902 tempest-ServerMetadataNegativeTestJSON-456453902-project-member] [instance: 7f4907d0-178d-452a-8149-030becde8779] Allocating IP information in the background. {{(pid=61806) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 833.763385] env[61806]: DEBUG nova.network.neutron [None req-c9a1ae7a-8313-4bdf-8de4-77929ce2d9bc tempest-ServerMetadataNegativeTestJSON-456453902 tempest-ServerMetadataNegativeTestJSON-456453902-project-member] [instance: 7f4907d0-178d-452a-8149-030becde8779] allocate_for_instance() {{(pid=61806) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 833.801566] env[61806]: DEBUG nova.policy [None req-c9a1ae7a-8313-4bdf-8de4-77929ce2d9bc tempest-ServerMetadataNegativeTestJSON-456453902 tempest-ServerMetadataNegativeTestJSON-456453902-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'd2291801d3304256b303355dd94a60d4', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '42d6862cb4734bd9a4ffcc64070c92a3', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61806) authorize /opt/stack/nova/nova/policy.py:201}} [ 833.968660] env[61806]: DEBUG oslo_vmware.api [-] Task: {'id': task-1294671, 'name': CreateVM_Task} progress is 25%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 834.001458] env[61806]: DEBUG oslo_vmware.api [None req-cbb90e6d-7dac-4c30-b860-1f0db7dd091e tempest-MultipleCreateTestJSON-1589015462 tempest-MultipleCreateTestJSON-1589015462-project-member] Task: {'id': task-1294672, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 834.079731] env[61806]: DEBUG nova.network.neutron [None req-c9a1ae7a-8313-4bdf-8de4-77929ce2d9bc tempest-ServerMetadataNegativeTestJSON-456453902 tempest-ServerMetadataNegativeTestJSON-456453902-project-member] [instance: 7f4907d0-178d-452a-8149-030becde8779] Successfully created port: 718d0c6d-5962-4b66-a13f-963aa96b2985 {{(pid=61806) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 834.088959] env[61806]: DEBUG oslo_vmware.api [None req-cbb90e6d-7dac-4c30-b860-1f0db7dd091e tempest-MultipleCreateTestJSON-1589015462 tempest-MultipleCreateTestJSON-1589015462-project-member] Task: {'id': task-1294673, 'name': CopyVirtualDisk_Task} progress is 25%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 834.129911] env[61806]: DEBUG oslo_concurrency.lockutils [None req-4e0c8443-0fd4-459b-a4c5-7cf373aa1089 tempest-ServerAddressesNegativeTestJSON-1526246390 tempest-ServerAddressesNegativeTestJSON-1526246390-project-member] Acquiring lock "refresh_cache-c8033551-6591-4e37-ae78-4efe7145b10b" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 834.130479] env[61806]: DEBUG oslo_concurrency.lockutils [None req-4e0c8443-0fd4-459b-a4c5-7cf373aa1089 tempest-ServerAddressesNegativeTestJSON-1526246390 tempest-ServerAddressesNegativeTestJSON-1526246390-project-member] Acquired lock "refresh_cache-c8033551-6591-4e37-ae78-4efe7145b10b" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 834.130479] env[61806]: DEBUG nova.network.neutron [None req-4e0c8443-0fd4-459b-a4c5-7cf373aa1089 tempest-ServerAddressesNegativeTestJSON-1526246390 tempest-ServerAddressesNegativeTestJSON-1526246390-project-member] [instance: c8033551-6591-4e37-ae78-4efe7145b10b] Building network info cache for instance {{(pid=61806) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 834.222257] env[61806]: DEBUG oslo_vmware.api [None req-6d96025a-fbfa-435a-871d-fe0c696f1d71 tempest-SecurityGroupsTestJSON-1213485581 tempest-SecurityGroupsTestJSON-1213485581-project-member] Task: {'id': task-1294674, 'name': ResetVM_Task, 'duration_secs': 0.109789} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 834.222687] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-6d96025a-fbfa-435a-871d-fe0c696f1d71 tempest-SecurityGroupsTestJSON-1213485581 tempest-SecurityGroupsTestJSON-1213485581-project-member] [instance: b41531f2-f28c-4d82-9682-0b557bbaa491] Did hard reboot of VM {{(pid=61806) reboot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1068}} [ 834.222793] env[61806]: DEBUG nova.compute.manager [None req-6d96025a-fbfa-435a-871d-fe0c696f1d71 tempest-SecurityGroupsTestJSON-1213485581 tempest-SecurityGroupsTestJSON-1213485581-project-member] [instance: b41531f2-f28c-4d82-9682-0b557bbaa491] Checking state {{(pid=61806) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 834.223648] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b21b65d0-2209-4ad7-8854-983d52990a1f {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 834.267256] env[61806]: DEBUG oslo_concurrency.lockutils [None req-7cf484b6-b263-4f99-a8c0-b55d49de0809 tempest-ServerShowV247Test-1218443059 tempest-ServerShowV247Test-1218443059-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: held 1.014s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 834.268700] env[61806]: DEBUG nova.compute.manager [None req-c9a1ae7a-8313-4bdf-8de4-77929ce2d9bc tempest-ServerMetadataNegativeTestJSON-456453902 tempest-ServerMetadataNegativeTestJSON-456453902-project-member] [instance: 7f4907d0-178d-452a-8149-030becde8779] Start building block device mappings for instance. {{(pid=61806) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 834.271987] env[61806]: DEBUG oslo_concurrency.lockutils [None req-334c6338-dc9d-4c32-ad77-bd3fe007216c tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 11.773s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 834.274716] env[61806]: INFO nova.compute.claims [None req-334c6338-dc9d-4c32-ad77-bd3fe007216c tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] [instance: be19fc44-ecbe-489c-9b6e-be8957bb1dd0] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 834.469737] env[61806]: DEBUG oslo_vmware.api [-] Task: {'id': task-1294671, 'name': CreateVM_Task} progress is 25%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 834.502937] env[61806]: DEBUG oslo_vmware.api [None req-cbb90e6d-7dac-4c30-b860-1f0db7dd091e tempest-MultipleCreateTestJSON-1589015462 tempest-MultipleCreateTestJSON-1589015462-project-member] Task: {'id': task-1294672, 'name': PowerOnVM_Task} progress is 88%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 834.585770] env[61806]: DEBUG oslo_vmware.api [None req-cbb90e6d-7dac-4c30-b860-1f0db7dd091e tempest-MultipleCreateTestJSON-1589015462 tempest-MultipleCreateTestJSON-1589015462-project-member] Task: {'id': task-1294673, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.704637} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 834.586166] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [None req-cbb90e6d-7dac-4c30-b860-1f0db7dd091e tempest-MultipleCreateTestJSON-1589015462 tempest-MultipleCreateTestJSON-1589015462-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa.vmdk to [datastore2] ef502a7d-7dc0-4319-8ce3-6a8701288237/ef502a7d-7dc0-4319-8ce3-6a8701288237.vmdk {{(pid=61806) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 834.586564] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-cbb90e6d-7dac-4c30-b860-1f0db7dd091e tempest-MultipleCreateTestJSON-1589015462 tempest-MultipleCreateTestJSON-1589015462-project-member] [instance: ef502a7d-7dc0-4319-8ce3-6a8701288237] Extending root virtual disk to 1048576 {{(pid=61806) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 834.586804] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-4d1086ef-5869-4a90-b111-823d3fd8f007 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 834.595026] env[61806]: DEBUG oslo_vmware.api [None req-cbb90e6d-7dac-4c30-b860-1f0db7dd091e tempest-MultipleCreateTestJSON-1589015462 tempest-MultipleCreateTestJSON-1589015462-project-member] Waiting for the task: (returnval){ [ 834.595026] env[61806]: value = "task-1294675" [ 834.595026] env[61806]: _type = "Task" [ 834.595026] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 834.603996] env[61806]: DEBUG oslo_vmware.api [None req-cbb90e6d-7dac-4c30-b860-1f0db7dd091e tempest-MultipleCreateTestJSON-1589015462 tempest-MultipleCreateTestJSON-1589015462-project-member] Task: {'id': task-1294675, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 834.685388] env[61806]: DEBUG nova.network.neutron [None req-4e0c8443-0fd4-459b-a4c5-7cf373aa1089 tempest-ServerAddressesNegativeTestJSON-1526246390 tempest-ServerAddressesNegativeTestJSON-1526246390-project-member] [instance: c8033551-6591-4e37-ae78-4efe7145b10b] Instance cache missing network info. {{(pid=61806) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 834.740138] env[61806]: DEBUG oslo_concurrency.lockutils [None req-6d96025a-fbfa-435a-871d-fe0c696f1d71 tempest-SecurityGroupsTestJSON-1213485581 tempest-SecurityGroupsTestJSON-1213485581-project-member] Lock "b41531f2-f28c-4d82-9682-0b557bbaa491" "released" by "nova.compute.manager.ComputeManager.reboot_instance..do_reboot_instance" :: held 4.676s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 834.751419] env[61806]: DEBUG nova.compute.manager [req-dccfed4b-c313-4c44-bcf9-f1f8979c828c req-6a7a8dc0-1edb-43d7-bbd6-a0ecc58ac377 service nova] [instance: 67d1bdd8-cf8f-4fb5-bd3c-42008e4d60e3] Received event network-changed-859fb723-4a1c-4ea8-bd7c-0ed308f3a9e3 {{(pid=61806) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 834.751679] env[61806]: DEBUG nova.compute.manager [req-dccfed4b-c313-4c44-bcf9-f1f8979c828c req-6a7a8dc0-1edb-43d7-bbd6-a0ecc58ac377 service nova] [instance: 67d1bdd8-cf8f-4fb5-bd3c-42008e4d60e3] Refreshing instance network info cache due to event network-changed-859fb723-4a1c-4ea8-bd7c-0ed308f3a9e3. {{(pid=61806) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 834.751903] env[61806]: DEBUG oslo_concurrency.lockutils [req-dccfed4b-c313-4c44-bcf9-f1f8979c828c req-6a7a8dc0-1edb-43d7-bbd6-a0ecc58ac377 service nova] Acquiring lock "refresh_cache-67d1bdd8-cf8f-4fb5-bd3c-42008e4d60e3" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 834.752073] env[61806]: DEBUG oslo_concurrency.lockutils [req-dccfed4b-c313-4c44-bcf9-f1f8979c828c req-6a7a8dc0-1edb-43d7-bbd6-a0ecc58ac377 service nova] Acquired lock "refresh_cache-67d1bdd8-cf8f-4fb5-bd3c-42008e4d60e3" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 834.752261] env[61806]: DEBUG nova.network.neutron [req-dccfed4b-c313-4c44-bcf9-f1f8979c828c req-6a7a8dc0-1edb-43d7-bbd6-a0ecc58ac377 service nova] [instance: 67d1bdd8-cf8f-4fb5-bd3c-42008e4d60e3] Refreshing network info cache for port 859fb723-4a1c-4ea8-bd7c-0ed308f3a9e3 {{(pid=61806) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 834.865015] env[61806]: DEBUG nova.network.neutron [None req-4e0c8443-0fd4-459b-a4c5-7cf373aa1089 tempest-ServerAddressesNegativeTestJSON-1526246390 tempest-ServerAddressesNegativeTestJSON-1526246390-project-member] [instance: c8033551-6591-4e37-ae78-4efe7145b10b] Updating instance_info_cache with network_info: [{"id": "e8bfbf94-495b-4d13-8d2c-7a02e7a90387", "address": "fa:16:3e:25:8e:4f", "network": {"id": "c8fe2900-c167-4b7d-8df5-914d5fe8cab2", "bridge": "br-int", "label": "tempest-ServerAddressesNegativeTestJSON-1486853118-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "dd7c12f41e8b4ba489480ffa3959b7ce", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "1323cb03-8367-485a-962e-131af8eba474", "external-id": "nsx-vlan-transportzone-41", "segmentation_id": 41, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape8bfbf94-49", "ovs_interfaceid": "e8bfbf94-495b-4d13-8d2c-7a02e7a90387", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61806) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 834.967827] env[61806]: DEBUG oslo_vmware.api [-] Task: {'id': task-1294671, 'name': CreateVM_Task, 'duration_secs': 1.666059} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 834.967980] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 117139cd-4ed4-4bea-8ad3-f4ac32c692e4] Created VM on the ESX host {{(pid=61806) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 834.968712] env[61806]: DEBUG oslo_concurrency.lockutils [None req-0ea3a38f-3871-4412-a6bc-d21c2086d262 tempest-AttachInterfacesV270Test-1974652806 tempest-AttachInterfacesV270Test-1974652806-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 834.968892] env[61806]: DEBUG oslo_concurrency.lockutils [None req-0ea3a38f-3871-4412-a6bc-d21c2086d262 tempest-AttachInterfacesV270Test-1974652806 tempest-AttachInterfacesV270Test-1974652806-project-member] Acquired lock "[datastore2] devstack-image-cache_base/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 834.969254] env[61806]: DEBUG oslo_concurrency.lockutils [None req-0ea3a38f-3871-4412-a6bc-d21c2086d262 tempest-AttachInterfacesV270Test-1974652806 tempest-AttachInterfacesV270Test-1974652806-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 834.969534] env[61806]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-0b822d48-6965-43c0-a3f3-8502538cb5d5 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 834.974373] env[61806]: DEBUG oslo_vmware.api [None req-0ea3a38f-3871-4412-a6bc-d21c2086d262 tempest-AttachInterfacesV270Test-1974652806 tempest-AttachInterfacesV270Test-1974652806-project-member] Waiting for the task: (returnval){ [ 834.974373] env[61806]: value = "session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]52362009-54d5-c22b-f262-f78f89814cc2" [ 834.974373] env[61806]: _type = "Task" [ 834.974373] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 834.983666] env[61806]: DEBUG oslo_vmware.api [None req-0ea3a38f-3871-4412-a6bc-d21c2086d262 tempest-AttachInterfacesV270Test-1974652806 tempest-AttachInterfacesV270Test-1974652806-project-member] Task: {'id': session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]52362009-54d5-c22b-f262-f78f89814cc2, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 835.001916] env[61806]: DEBUG oslo_vmware.api [None req-cbb90e6d-7dac-4c30-b860-1f0db7dd091e tempest-MultipleCreateTestJSON-1589015462 tempest-MultipleCreateTestJSON-1589015462-project-member] Task: {'id': task-1294672, 'name': PowerOnVM_Task, 'duration_secs': 1.756365} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 835.002198] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [None req-cbb90e6d-7dac-4c30-b860-1f0db7dd091e tempest-MultipleCreateTestJSON-1589015462 tempest-MultipleCreateTestJSON-1589015462-project-member] [instance: 25218e83-c1ab-4b97-ade1-7c32b6f99305] Powered on the VM {{(pid=61806) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 835.002412] env[61806]: INFO nova.compute.manager [None req-cbb90e6d-7dac-4c30-b860-1f0db7dd091e tempest-MultipleCreateTestJSON-1589015462 tempest-MultipleCreateTestJSON-1589015462-project-member] [instance: 25218e83-c1ab-4b97-ade1-7c32b6f99305] Took 11.45 seconds to spawn the instance on the hypervisor. [ 835.002631] env[61806]: DEBUG nova.compute.manager [None req-cbb90e6d-7dac-4c30-b860-1f0db7dd091e tempest-MultipleCreateTestJSON-1589015462 tempest-MultipleCreateTestJSON-1589015462-project-member] [instance: 25218e83-c1ab-4b97-ade1-7c32b6f99305] Checking state {{(pid=61806) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 835.003403] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-434e0586-2dcc-4e96-9d50-94213579c552 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 835.104282] env[61806]: DEBUG oslo_vmware.api [None req-cbb90e6d-7dac-4c30-b860-1f0db7dd091e tempest-MultipleCreateTestJSON-1589015462 tempest-MultipleCreateTestJSON-1589015462-project-member] Task: {'id': task-1294675, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.080097} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 835.104561] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-cbb90e6d-7dac-4c30-b860-1f0db7dd091e tempest-MultipleCreateTestJSON-1589015462 tempest-MultipleCreateTestJSON-1589015462-project-member] [instance: ef502a7d-7dc0-4319-8ce3-6a8701288237] Extended root virtual disk {{(pid=61806) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 835.105329] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b1b0bc86-d4ff-46ea-ad00-abe513350827 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 835.127217] env[61806]: DEBUG nova.virt.vmwareapi.volumeops [None req-cbb90e6d-7dac-4c30-b860-1f0db7dd091e tempest-MultipleCreateTestJSON-1589015462 tempest-MultipleCreateTestJSON-1589015462-project-member] [instance: ef502a7d-7dc0-4319-8ce3-6a8701288237] Reconfiguring VM instance instance-00000043 to attach disk [datastore2] ef502a7d-7dc0-4319-8ce3-6a8701288237/ef502a7d-7dc0-4319-8ce3-6a8701288237.vmdk or device None with type sparse {{(pid=61806) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 835.127472] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-78f9802a-9a4d-4538-9219-e50d62279bad {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 835.153305] env[61806]: DEBUG oslo_vmware.api [None req-cbb90e6d-7dac-4c30-b860-1f0db7dd091e tempest-MultipleCreateTestJSON-1589015462 tempest-MultipleCreateTestJSON-1589015462-project-member] Waiting for the task: (returnval){ [ 835.153305] env[61806]: value = "task-1294676" [ 835.153305] env[61806]: _type = "Task" [ 835.153305] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 835.165475] env[61806]: DEBUG oslo_vmware.api [None req-cbb90e6d-7dac-4c30-b860-1f0db7dd091e tempest-MultipleCreateTestJSON-1589015462 tempest-MultipleCreateTestJSON-1589015462-project-member] Task: {'id': task-1294676, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 835.284896] env[61806]: DEBUG nova.compute.manager [None req-c9a1ae7a-8313-4bdf-8de4-77929ce2d9bc tempest-ServerMetadataNegativeTestJSON-456453902 tempest-ServerMetadataNegativeTestJSON-456453902-project-member] [instance: 7f4907d0-178d-452a-8149-030becde8779] Start spawning the instance on the hypervisor. {{(pid=61806) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 835.311510] env[61806]: DEBUG nova.virt.hardware [None req-c9a1ae7a-8313-4bdf-8de4-77929ce2d9bc tempest-ServerMetadataNegativeTestJSON-456453902 tempest-ServerMetadataNegativeTestJSON-456453902-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-15T18:05:10Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-15T18:04:52Z,direct_url=,disk_format='vmdk',id=ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='84ba83fce2c349a6988173c5d6fc3b07',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-15T18:04:53Z,virtual_size=,visibility=), allow threads: False {{(pid=61806) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 835.311881] env[61806]: DEBUG nova.virt.hardware [None req-c9a1ae7a-8313-4bdf-8de4-77929ce2d9bc tempest-ServerMetadataNegativeTestJSON-456453902 tempest-ServerMetadataNegativeTestJSON-456453902-project-member] Flavor limits 0:0:0 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 835.312140] env[61806]: DEBUG nova.virt.hardware [None req-c9a1ae7a-8313-4bdf-8de4-77929ce2d9bc tempest-ServerMetadataNegativeTestJSON-456453902 tempest-ServerMetadataNegativeTestJSON-456453902-project-member] Image limits 0:0:0 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 835.312403] env[61806]: DEBUG nova.virt.hardware [None req-c9a1ae7a-8313-4bdf-8de4-77929ce2d9bc tempest-ServerMetadataNegativeTestJSON-456453902 tempest-ServerMetadataNegativeTestJSON-456453902-project-member] Flavor pref 0:0:0 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 835.312665] env[61806]: DEBUG nova.virt.hardware [None req-c9a1ae7a-8313-4bdf-8de4-77929ce2d9bc tempest-ServerMetadataNegativeTestJSON-456453902 tempest-ServerMetadataNegativeTestJSON-456453902-project-member] Image pref 0:0:0 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 835.312900] env[61806]: DEBUG nova.virt.hardware [None req-c9a1ae7a-8313-4bdf-8de4-77929ce2d9bc tempest-ServerMetadataNegativeTestJSON-456453902 tempest-ServerMetadataNegativeTestJSON-456453902-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 835.313256] env[61806]: DEBUG nova.virt.hardware [None req-c9a1ae7a-8313-4bdf-8de4-77929ce2d9bc tempest-ServerMetadataNegativeTestJSON-456453902 tempest-ServerMetadataNegativeTestJSON-456453902-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61806) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 835.313442] env[61806]: DEBUG nova.virt.hardware [None req-c9a1ae7a-8313-4bdf-8de4-77929ce2d9bc tempest-ServerMetadataNegativeTestJSON-456453902 tempest-ServerMetadataNegativeTestJSON-456453902-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61806) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 835.313686] env[61806]: DEBUG nova.virt.hardware [None req-c9a1ae7a-8313-4bdf-8de4-77929ce2d9bc tempest-ServerMetadataNegativeTestJSON-456453902 tempest-ServerMetadataNegativeTestJSON-456453902-project-member] Got 1 possible topologies {{(pid=61806) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 835.313911] env[61806]: DEBUG nova.virt.hardware [None req-c9a1ae7a-8313-4bdf-8de4-77929ce2d9bc tempest-ServerMetadataNegativeTestJSON-456453902 tempest-ServerMetadataNegativeTestJSON-456453902-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61806) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 835.314252] env[61806]: DEBUG nova.virt.hardware [None req-c9a1ae7a-8313-4bdf-8de4-77929ce2d9bc tempest-ServerMetadataNegativeTestJSON-456453902 tempest-ServerMetadataNegativeTestJSON-456453902-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61806) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 835.315170] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-91163eeb-e524-4ef0-b750-5345177575d1 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 835.328462] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7fb3cd6d-fa05-4d10-95f9-644c4d8fa560 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 835.367570] env[61806]: DEBUG oslo_concurrency.lockutils [None req-4e0c8443-0fd4-459b-a4c5-7cf373aa1089 tempest-ServerAddressesNegativeTestJSON-1526246390 tempest-ServerAddressesNegativeTestJSON-1526246390-project-member] Releasing lock "refresh_cache-c8033551-6591-4e37-ae78-4efe7145b10b" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 835.367888] env[61806]: DEBUG nova.compute.manager [None req-4e0c8443-0fd4-459b-a4c5-7cf373aa1089 tempest-ServerAddressesNegativeTestJSON-1526246390 tempest-ServerAddressesNegativeTestJSON-1526246390-project-member] [instance: c8033551-6591-4e37-ae78-4efe7145b10b] Instance network_info: |[{"id": "e8bfbf94-495b-4d13-8d2c-7a02e7a90387", "address": "fa:16:3e:25:8e:4f", "network": {"id": "c8fe2900-c167-4b7d-8df5-914d5fe8cab2", "bridge": "br-int", "label": "tempest-ServerAddressesNegativeTestJSON-1486853118-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "dd7c12f41e8b4ba489480ffa3959b7ce", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "1323cb03-8367-485a-962e-131af8eba474", "external-id": "nsx-vlan-transportzone-41", "segmentation_id": 41, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape8bfbf94-49", "ovs_interfaceid": "e8bfbf94-495b-4d13-8d2c-7a02e7a90387", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61806) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 835.368384] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-4e0c8443-0fd4-459b-a4c5-7cf373aa1089 tempest-ServerAddressesNegativeTestJSON-1526246390 tempest-ServerAddressesNegativeTestJSON-1526246390-project-member] [instance: c8033551-6591-4e37-ae78-4efe7145b10b] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:25:8e:4f', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '1323cb03-8367-485a-962e-131af8eba474', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'e8bfbf94-495b-4d13-8d2c-7a02e7a90387', 'vif_model': 'vmxnet3'}] {{(pid=61806) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 835.376153] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [None req-4e0c8443-0fd4-459b-a4c5-7cf373aa1089 tempest-ServerAddressesNegativeTestJSON-1526246390 tempest-ServerAddressesNegativeTestJSON-1526246390-project-member] Creating folder: Project (dd7c12f41e8b4ba489480ffa3959b7ce). Parent ref: group-v277609. {{(pid=61806) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 835.376461] env[61806]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-73122544-3ecd-440a-b0f7-f625bd3f2959 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 835.391838] env[61806]: INFO nova.virt.vmwareapi.vm_util [None req-4e0c8443-0fd4-459b-a4c5-7cf373aa1089 tempest-ServerAddressesNegativeTestJSON-1526246390 tempest-ServerAddressesNegativeTestJSON-1526246390-project-member] Created folder: Project (dd7c12f41e8b4ba489480ffa3959b7ce) in parent group-v277609. [ 835.392076] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [None req-4e0c8443-0fd4-459b-a4c5-7cf373aa1089 tempest-ServerAddressesNegativeTestJSON-1526246390 tempest-ServerAddressesNegativeTestJSON-1526246390-project-member] Creating folder: Instances. Parent ref: group-v277674. {{(pid=61806) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 835.392339] env[61806]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-7835d842-78b7-4188-aa15-cbe616b726ea {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 835.400937] env[61806]: INFO nova.virt.vmwareapi.vm_util [None req-4e0c8443-0fd4-459b-a4c5-7cf373aa1089 tempest-ServerAddressesNegativeTestJSON-1526246390 tempest-ServerAddressesNegativeTestJSON-1526246390-project-member] Created folder: Instances in parent group-v277674. [ 835.401229] env[61806]: DEBUG oslo.service.loopingcall [None req-4e0c8443-0fd4-459b-a4c5-7cf373aa1089 tempest-ServerAddressesNegativeTestJSON-1526246390 tempest-ServerAddressesNegativeTestJSON-1526246390-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61806) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 835.401434] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: c8033551-6591-4e37-ae78-4efe7145b10b] Creating VM on the ESX host {{(pid=61806) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 835.401649] env[61806]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-3125fffe-e87c-4338-bd3a-57f73fedefdb {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 835.425416] env[61806]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 835.425416] env[61806]: value = "task-1294679" [ 835.425416] env[61806]: _type = "Task" [ 835.425416] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 835.433886] env[61806]: DEBUG oslo_vmware.api [-] Task: {'id': task-1294679, 'name': CreateVM_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 835.485333] env[61806]: DEBUG oslo_vmware.api [None req-0ea3a38f-3871-4412-a6bc-d21c2086d262 tempest-AttachInterfacesV270Test-1974652806 tempest-AttachInterfacesV270Test-1974652806-project-member] Task: {'id': session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]52362009-54d5-c22b-f262-f78f89814cc2, 'name': SearchDatastore_Task, 'duration_secs': 0.042541} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 835.488608] env[61806]: DEBUG oslo_concurrency.lockutils [None req-0ea3a38f-3871-4412-a6bc-d21c2086d262 tempest-AttachInterfacesV270Test-1974652806 tempest-AttachInterfacesV270Test-1974652806-project-member] Releasing lock "[datastore2] devstack-image-cache_base/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 835.488871] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-0ea3a38f-3871-4412-a6bc-d21c2086d262 tempest-AttachInterfacesV270Test-1974652806 tempest-AttachInterfacesV270Test-1974652806-project-member] [instance: 117139cd-4ed4-4bea-8ad3-f4ac32c692e4] Processing image ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa {{(pid=61806) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 835.489132] env[61806]: DEBUG oslo_concurrency.lockutils [None req-0ea3a38f-3871-4412-a6bc-d21c2086d262 tempest-AttachInterfacesV270Test-1974652806 tempest-AttachInterfacesV270Test-1974652806-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa.vmdk" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 835.489307] env[61806]: DEBUG oslo_concurrency.lockutils [None req-0ea3a38f-3871-4412-a6bc-d21c2086d262 tempest-AttachInterfacesV270Test-1974652806 tempest-AttachInterfacesV270Test-1974652806-project-member] Acquired lock "[datastore2] devstack-image-cache_base/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa.vmdk" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 835.489501] env[61806]: DEBUG nova.virt.vmwareapi.ds_util [None req-0ea3a38f-3871-4412-a6bc-d21c2086d262 tempest-AttachInterfacesV270Test-1974652806 tempest-AttachInterfacesV270Test-1974652806-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61806) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 835.490354] env[61806]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-eed48664-6a4a-4bd3-be26-0e6d71a626d7 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 835.501972] env[61806]: DEBUG nova.virt.vmwareapi.ds_util [None req-0ea3a38f-3871-4412-a6bc-d21c2086d262 tempest-AttachInterfacesV270Test-1974652806 tempest-AttachInterfacesV270Test-1974652806-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61806) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 835.502228] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-0ea3a38f-3871-4412-a6bc-d21c2086d262 tempest-AttachInterfacesV270Test-1974652806 tempest-AttachInterfacesV270Test-1974652806-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61806) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 835.505363] env[61806]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-6a091dc5-68ca-4a7b-88e0-7ae9dfc1f99b {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 835.511622] env[61806]: DEBUG oslo_vmware.api [None req-0ea3a38f-3871-4412-a6bc-d21c2086d262 tempest-AttachInterfacesV270Test-1974652806 tempest-AttachInterfacesV270Test-1974652806-project-member] Waiting for the task: (returnval){ [ 835.511622] env[61806]: value = "session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]52f23547-3747-c9e4-a96c-2c911d3619f7" [ 835.511622] env[61806]: _type = "Task" [ 835.511622] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 835.529532] env[61806]: DEBUG oslo_vmware.api [None req-0ea3a38f-3871-4412-a6bc-d21c2086d262 tempest-AttachInterfacesV270Test-1974652806 tempest-AttachInterfacesV270Test-1974652806-project-member] Task: {'id': session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]52f23547-3747-c9e4-a96c-2c911d3619f7, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 835.533132] env[61806]: INFO nova.compute.manager [None req-cbb90e6d-7dac-4c30-b860-1f0db7dd091e tempest-MultipleCreateTestJSON-1589015462 tempest-MultipleCreateTestJSON-1589015462-project-member] [instance: 25218e83-c1ab-4b97-ade1-7c32b6f99305] Took 31.46 seconds to build instance. [ 835.573673] env[61806]: DEBUG nova.network.neutron [req-dccfed4b-c313-4c44-bcf9-f1f8979c828c req-6a7a8dc0-1edb-43d7-bbd6-a0ecc58ac377 service nova] [instance: 67d1bdd8-cf8f-4fb5-bd3c-42008e4d60e3] Updated VIF entry in instance network info cache for port 859fb723-4a1c-4ea8-bd7c-0ed308f3a9e3. {{(pid=61806) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 835.574105] env[61806]: DEBUG nova.network.neutron [req-dccfed4b-c313-4c44-bcf9-f1f8979c828c req-6a7a8dc0-1edb-43d7-bbd6-a0ecc58ac377 service nova] [instance: 67d1bdd8-cf8f-4fb5-bd3c-42008e4d60e3] Updating instance_info_cache with network_info: [{"id": "859fb723-4a1c-4ea8-bd7c-0ed308f3a9e3", "address": "fa:16:3e:3c:40:d6", "network": {"id": "746a03b6-c1a6-4d27-98b9-0bee6376634f", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-1971382587-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.245", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c1f2d161c50f4b70987b90e6364c3779", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6a31b224-77d7-48a4-af87-312758611b8e", "external-id": "nsx-vlan-transportzone-761", "segmentation_id": 761, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap859fb723-4a", "ovs_interfaceid": "859fb723-4a1c-4ea8-bd7c-0ed308f3a9e3", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61806) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 835.625719] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-46fe3657-84df-4697-8954-98bb10c834ee {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 835.633836] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-286ded2e-cec3-423a-904d-b4386e3800f1 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 835.672063] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-71d4b63a-ba71-418b-8570-6590454d90b4 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 835.680543] env[61806]: DEBUG oslo_vmware.api [None req-cbb90e6d-7dac-4c30-b860-1f0db7dd091e tempest-MultipleCreateTestJSON-1589015462 tempest-MultipleCreateTestJSON-1589015462-project-member] Task: {'id': task-1294676, 'name': ReconfigVM_Task, 'duration_secs': 0.526287} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 835.682817] env[61806]: DEBUG nova.virt.vmwareapi.volumeops [None req-cbb90e6d-7dac-4c30-b860-1f0db7dd091e tempest-MultipleCreateTestJSON-1589015462 tempest-MultipleCreateTestJSON-1589015462-project-member] [instance: ef502a7d-7dc0-4319-8ce3-6a8701288237] Reconfigured VM instance instance-00000043 to attach disk [datastore2] ef502a7d-7dc0-4319-8ce3-6a8701288237/ef502a7d-7dc0-4319-8ce3-6a8701288237.vmdk or device None with type sparse {{(pid=61806) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 835.683519] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-05dedf57-11c1-439e-817a-7a4f4475d399 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 835.686065] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3b1834dd-d793-4f54-94a5-3215402f9d28 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 835.700919] env[61806]: DEBUG nova.compute.provider_tree [None req-334c6338-dc9d-4c32-ad77-bd3fe007216c tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Inventory has not changed in ProviderTree for provider: a2858be1-fd22-4e08-979e-87ad25293407 {{(pid=61806) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 835.705291] env[61806]: DEBUG oslo_vmware.api [None req-cbb90e6d-7dac-4c30-b860-1f0db7dd091e tempest-MultipleCreateTestJSON-1589015462 tempest-MultipleCreateTestJSON-1589015462-project-member] Waiting for the task: (returnval){ [ 835.705291] env[61806]: value = "task-1294680" [ 835.705291] env[61806]: _type = "Task" [ 835.705291] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 835.716828] env[61806]: DEBUG oslo_vmware.api [None req-cbb90e6d-7dac-4c30-b860-1f0db7dd091e tempest-MultipleCreateTestJSON-1589015462 tempest-MultipleCreateTestJSON-1589015462-project-member] Task: {'id': task-1294680, 'name': Rename_Task} progress is 14%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 835.881100] env[61806]: DEBUG nova.network.neutron [None req-c9a1ae7a-8313-4bdf-8de4-77929ce2d9bc tempest-ServerMetadataNegativeTestJSON-456453902 tempest-ServerMetadataNegativeTestJSON-456453902-project-member] [instance: 7f4907d0-178d-452a-8149-030becde8779] Successfully updated port: 718d0c6d-5962-4b66-a13f-963aa96b2985 {{(pid=61806) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 835.936441] env[61806]: DEBUG oslo_vmware.api [-] Task: {'id': task-1294679, 'name': CreateVM_Task, 'duration_secs': 0.443238} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 835.936530] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: c8033551-6591-4e37-ae78-4efe7145b10b] Created VM on the ESX host {{(pid=61806) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 835.937229] env[61806]: DEBUG oslo_concurrency.lockutils [None req-4e0c8443-0fd4-459b-a4c5-7cf373aa1089 tempest-ServerAddressesNegativeTestJSON-1526246390 tempest-ServerAddressesNegativeTestJSON-1526246390-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 835.937410] env[61806]: DEBUG oslo_concurrency.lockutils [None req-4e0c8443-0fd4-459b-a4c5-7cf373aa1089 tempest-ServerAddressesNegativeTestJSON-1526246390 tempest-ServerAddressesNegativeTestJSON-1526246390-project-member] Acquired lock "[datastore2] devstack-image-cache_base/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 835.938354] env[61806]: DEBUG oslo_concurrency.lockutils [None req-4e0c8443-0fd4-459b-a4c5-7cf373aa1089 tempest-ServerAddressesNegativeTestJSON-1526246390 tempest-ServerAddressesNegativeTestJSON-1526246390-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 835.938354] env[61806]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-a7e1a1e9-ed38-4ef5-aa9b-190ae1af0fa6 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 835.943363] env[61806]: DEBUG oslo_vmware.api [None req-4e0c8443-0fd4-459b-a4c5-7cf373aa1089 tempest-ServerAddressesNegativeTestJSON-1526246390 tempest-ServerAddressesNegativeTestJSON-1526246390-project-member] Waiting for the task: (returnval){ [ 835.943363] env[61806]: value = "session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]525083e0-7896-4c8e-8300-13a96508fe18" [ 835.943363] env[61806]: _type = "Task" [ 835.943363] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 835.951723] env[61806]: DEBUG oslo_vmware.api [None req-4e0c8443-0fd4-459b-a4c5-7cf373aa1089 tempest-ServerAddressesNegativeTestJSON-1526246390 tempest-ServerAddressesNegativeTestJSON-1526246390-project-member] Task: {'id': session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]525083e0-7896-4c8e-8300-13a96508fe18, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 835.973957] env[61806]: DEBUG nova.compute.manager [req-b856ac6a-7846-4611-a003-a75556e76da7 req-475d8cec-205e-4d72-8c36-2570184072e4 service nova] [instance: b41531f2-f28c-4d82-9682-0b557bbaa491] Received event network-changed-513ca4b5-b62c-486a-9860-cf267a15c3e5 {{(pid=61806) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 835.974440] env[61806]: DEBUG nova.compute.manager [req-b856ac6a-7846-4611-a003-a75556e76da7 req-475d8cec-205e-4d72-8c36-2570184072e4 service nova] [instance: b41531f2-f28c-4d82-9682-0b557bbaa491] Refreshing instance network info cache due to event network-changed-513ca4b5-b62c-486a-9860-cf267a15c3e5. {{(pid=61806) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 835.974440] env[61806]: DEBUG oslo_concurrency.lockutils [req-b856ac6a-7846-4611-a003-a75556e76da7 req-475d8cec-205e-4d72-8c36-2570184072e4 service nova] Acquiring lock "refresh_cache-b41531f2-f28c-4d82-9682-0b557bbaa491" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 835.974626] env[61806]: DEBUG oslo_concurrency.lockutils [req-b856ac6a-7846-4611-a003-a75556e76da7 req-475d8cec-205e-4d72-8c36-2570184072e4 service nova] Acquired lock "refresh_cache-b41531f2-f28c-4d82-9682-0b557bbaa491" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 835.974626] env[61806]: DEBUG nova.network.neutron [req-b856ac6a-7846-4611-a003-a75556e76da7 req-475d8cec-205e-4d72-8c36-2570184072e4 service nova] [instance: b41531f2-f28c-4d82-9682-0b557bbaa491] Refreshing network info cache for port 513ca4b5-b62c-486a-9860-cf267a15c3e5 {{(pid=61806) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 836.021059] env[61806]: DEBUG oslo_vmware.api [None req-0ea3a38f-3871-4412-a6bc-d21c2086d262 tempest-AttachInterfacesV270Test-1974652806 tempest-AttachInterfacesV270Test-1974652806-project-member] Task: {'id': session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]52f23547-3747-c9e4-a96c-2c911d3619f7, 'name': SearchDatastore_Task, 'duration_secs': 0.013763} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 836.021874] env[61806]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-6b7c51d8-f9d9-4201-acf2-8b6ac0bda636 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 836.027285] env[61806]: DEBUG oslo_vmware.api [None req-0ea3a38f-3871-4412-a6bc-d21c2086d262 tempest-AttachInterfacesV270Test-1974652806 tempest-AttachInterfacesV270Test-1974652806-project-member] Waiting for the task: (returnval){ [ 836.027285] env[61806]: value = "session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]52c49ba9-2727-bf9e-498e-21e229c7b57c" [ 836.027285] env[61806]: _type = "Task" [ 836.027285] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 836.035850] env[61806]: DEBUG oslo_concurrency.lockutils [None req-cbb90e6d-7dac-4c30-b860-1f0db7dd091e tempest-MultipleCreateTestJSON-1589015462 tempest-MultipleCreateTestJSON-1589015462-project-member] Lock "25218e83-c1ab-4b97-ade1-7c32b6f99305" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 78.757s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 836.036103] env[61806]: DEBUG oslo_vmware.api [None req-0ea3a38f-3871-4412-a6bc-d21c2086d262 tempest-AttachInterfacesV270Test-1974652806 tempest-AttachInterfacesV270Test-1974652806-project-member] Task: {'id': session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]52c49ba9-2727-bf9e-498e-21e229c7b57c, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 836.076681] env[61806]: DEBUG oslo_concurrency.lockutils [req-dccfed4b-c313-4c44-bcf9-f1f8979c828c req-6a7a8dc0-1edb-43d7-bbd6-a0ecc58ac377 service nova] Releasing lock "refresh_cache-67d1bdd8-cf8f-4fb5-bd3c-42008e4d60e3" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 836.076935] env[61806]: DEBUG nova.compute.manager [req-dccfed4b-c313-4c44-bcf9-f1f8979c828c req-6a7a8dc0-1edb-43d7-bbd6-a0ecc58ac377 service nova] [instance: c8033551-6591-4e37-ae78-4efe7145b10b] Received event network-vif-plugged-e8bfbf94-495b-4d13-8d2c-7a02e7a90387 {{(pid=61806) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 836.077158] env[61806]: DEBUG oslo_concurrency.lockutils [req-dccfed4b-c313-4c44-bcf9-f1f8979c828c req-6a7a8dc0-1edb-43d7-bbd6-a0ecc58ac377 service nova] Acquiring lock "c8033551-6591-4e37-ae78-4efe7145b10b-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 836.077370] env[61806]: DEBUG oslo_concurrency.lockutils [req-dccfed4b-c313-4c44-bcf9-f1f8979c828c req-6a7a8dc0-1edb-43d7-bbd6-a0ecc58ac377 service nova] Lock "c8033551-6591-4e37-ae78-4efe7145b10b-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 836.077537] env[61806]: DEBUG oslo_concurrency.lockutils [req-dccfed4b-c313-4c44-bcf9-f1f8979c828c req-6a7a8dc0-1edb-43d7-bbd6-a0ecc58ac377 service nova] Lock "c8033551-6591-4e37-ae78-4efe7145b10b-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 836.077711] env[61806]: DEBUG nova.compute.manager [req-dccfed4b-c313-4c44-bcf9-f1f8979c828c req-6a7a8dc0-1edb-43d7-bbd6-a0ecc58ac377 service nova] [instance: c8033551-6591-4e37-ae78-4efe7145b10b] No waiting events found dispatching network-vif-plugged-e8bfbf94-495b-4d13-8d2c-7a02e7a90387 {{(pid=61806) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 836.078240] env[61806]: WARNING nova.compute.manager [req-dccfed4b-c313-4c44-bcf9-f1f8979c828c req-6a7a8dc0-1edb-43d7-bbd6-a0ecc58ac377 service nova] [instance: c8033551-6591-4e37-ae78-4efe7145b10b] Received unexpected event network-vif-plugged-e8bfbf94-495b-4d13-8d2c-7a02e7a90387 for instance with vm_state building and task_state spawning. [ 836.078487] env[61806]: DEBUG nova.compute.manager [req-dccfed4b-c313-4c44-bcf9-f1f8979c828c req-6a7a8dc0-1edb-43d7-bbd6-a0ecc58ac377 service nova] [instance: c8033551-6591-4e37-ae78-4efe7145b10b] Received event network-changed-e8bfbf94-495b-4d13-8d2c-7a02e7a90387 {{(pid=61806) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 836.078668] env[61806]: DEBUG nova.compute.manager [req-dccfed4b-c313-4c44-bcf9-f1f8979c828c req-6a7a8dc0-1edb-43d7-bbd6-a0ecc58ac377 service nova] [instance: c8033551-6591-4e37-ae78-4efe7145b10b] Refreshing instance network info cache due to event network-changed-e8bfbf94-495b-4d13-8d2c-7a02e7a90387. {{(pid=61806) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 836.078867] env[61806]: DEBUG oslo_concurrency.lockutils [req-dccfed4b-c313-4c44-bcf9-f1f8979c828c req-6a7a8dc0-1edb-43d7-bbd6-a0ecc58ac377 service nova] Acquiring lock "refresh_cache-c8033551-6591-4e37-ae78-4efe7145b10b" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 836.079027] env[61806]: DEBUG oslo_concurrency.lockutils [req-dccfed4b-c313-4c44-bcf9-f1f8979c828c req-6a7a8dc0-1edb-43d7-bbd6-a0ecc58ac377 service nova] Acquired lock "refresh_cache-c8033551-6591-4e37-ae78-4efe7145b10b" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 836.079200] env[61806]: DEBUG nova.network.neutron [req-dccfed4b-c313-4c44-bcf9-f1f8979c828c req-6a7a8dc0-1edb-43d7-bbd6-a0ecc58ac377 service nova] [instance: c8033551-6591-4e37-ae78-4efe7145b10b] Refreshing network info cache for port e8bfbf94-495b-4d13-8d2c-7a02e7a90387 {{(pid=61806) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 836.205222] env[61806]: DEBUG nova.scheduler.client.report [None req-334c6338-dc9d-4c32-ad77-bd3fe007216c tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Inventory has not changed for provider a2858be1-fd22-4e08-979e-87ad25293407 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 138, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61806) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 836.218569] env[61806]: DEBUG oslo_vmware.api [None req-cbb90e6d-7dac-4c30-b860-1f0db7dd091e tempest-MultipleCreateTestJSON-1589015462 tempest-MultipleCreateTestJSON-1589015462-project-member] Task: {'id': task-1294680, 'name': Rename_Task, 'duration_secs': 0.179325} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 836.219244] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [None req-cbb90e6d-7dac-4c30-b860-1f0db7dd091e tempest-MultipleCreateTestJSON-1589015462 tempest-MultipleCreateTestJSON-1589015462-project-member] [instance: ef502a7d-7dc0-4319-8ce3-6a8701288237] Powering on the VM {{(pid=61806) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 836.222142] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-96fb7c0a-0575-41c3-87fd-40188223a41c {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 836.228954] env[61806]: DEBUG oslo_vmware.api [None req-cbb90e6d-7dac-4c30-b860-1f0db7dd091e tempest-MultipleCreateTestJSON-1589015462 tempest-MultipleCreateTestJSON-1589015462-project-member] Waiting for the task: (returnval){ [ 836.228954] env[61806]: value = "task-1294681" [ 836.228954] env[61806]: _type = "Task" [ 836.228954] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 836.239112] env[61806]: DEBUG oslo_vmware.api [None req-cbb90e6d-7dac-4c30-b860-1f0db7dd091e tempest-MultipleCreateTestJSON-1589015462 tempest-MultipleCreateTestJSON-1589015462-project-member] Task: {'id': task-1294681, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 836.386117] env[61806]: DEBUG oslo_concurrency.lockutils [None req-c9a1ae7a-8313-4bdf-8de4-77929ce2d9bc tempest-ServerMetadataNegativeTestJSON-456453902 tempest-ServerMetadataNegativeTestJSON-456453902-project-member] Acquiring lock "refresh_cache-7f4907d0-178d-452a-8149-030becde8779" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 836.386231] env[61806]: DEBUG oslo_concurrency.lockutils [None req-c9a1ae7a-8313-4bdf-8de4-77929ce2d9bc tempest-ServerMetadataNegativeTestJSON-456453902 tempest-ServerMetadataNegativeTestJSON-456453902-project-member] Acquired lock "refresh_cache-7f4907d0-178d-452a-8149-030becde8779" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 836.386431] env[61806]: DEBUG nova.network.neutron [None req-c9a1ae7a-8313-4bdf-8de4-77929ce2d9bc tempest-ServerMetadataNegativeTestJSON-456453902 tempest-ServerMetadataNegativeTestJSON-456453902-project-member] [instance: 7f4907d0-178d-452a-8149-030becde8779] Building network info cache for instance {{(pid=61806) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 836.453592] env[61806]: DEBUG oslo_vmware.api [None req-4e0c8443-0fd4-459b-a4c5-7cf373aa1089 tempest-ServerAddressesNegativeTestJSON-1526246390 tempest-ServerAddressesNegativeTestJSON-1526246390-project-member] Task: {'id': session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]525083e0-7896-4c8e-8300-13a96508fe18, 'name': SearchDatastore_Task, 'duration_secs': 0.08298} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 836.453851] env[61806]: DEBUG oslo_concurrency.lockutils [None req-4e0c8443-0fd4-459b-a4c5-7cf373aa1089 tempest-ServerAddressesNegativeTestJSON-1526246390 tempest-ServerAddressesNegativeTestJSON-1526246390-project-member] Releasing lock "[datastore2] devstack-image-cache_base/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 836.454104] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-4e0c8443-0fd4-459b-a4c5-7cf373aa1089 tempest-ServerAddressesNegativeTestJSON-1526246390 tempest-ServerAddressesNegativeTestJSON-1526246390-project-member] [instance: c8033551-6591-4e37-ae78-4efe7145b10b] Processing image ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa {{(pid=61806) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 836.454328] env[61806]: DEBUG oslo_concurrency.lockutils [None req-4e0c8443-0fd4-459b-a4c5-7cf373aa1089 tempest-ServerAddressesNegativeTestJSON-1526246390 tempest-ServerAddressesNegativeTestJSON-1526246390-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa.vmdk" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 836.537979] env[61806]: DEBUG oslo_vmware.api [None req-0ea3a38f-3871-4412-a6bc-d21c2086d262 tempest-AttachInterfacesV270Test-1974652806 tempest-AttachInterfacesV270Test-1974652806-project-member] Task: {'id': session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]52c49ba9-2727-bf9e-498e-21e229c7b57c, 'name': SearchDatastore_Task, 'duration_secs': 0.06694} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 836.538257] env[61806]: DEBUG oslo_concurrency.lockutils [None req-0ea3a38f-3871-4412-a6bc-d21c2086d262 tempest-AttachInterfacesV270Test-1974652806 tempest-AttachInterfacesV270Test-1974652806-project-member] Releasing lock "[datastore2] devstack-image-cache_base/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa.vmdk" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 836.538516] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [None req-0ea3a38f-3871-4412-a6bc-d21c2086d262 tempest-AttachInterfacesV270Test-1974652806 tempest-AttachInterfacesV270Test-1974652806-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa.vmdk to [datastore2] 117139cd-4ed4-4bea-8ad3-f4ac32c692e4/117139cd-4ed4-4bea-8ad3-f4ac32c692e4.vmdk {{(pid=61806) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 836.538795] env[61806]: DEBUG oslo_concurrency.lockutils [None req-4e0c8443-0fd4-459b-a4c5-7cf373aa1089 tempest-ServerAddressesNegativeTestJSON-1526246390 tempest-ServerAddressesNegativeTestJSON-1526246390-project-member] Acquired lock "[datastore2] devstack-image-cache_base/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa.vmdk" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 836.538987] env[61806]: DEBUG nova.virt.vmwareapi.ds_util [None req-4e0c8443-0fd4-459b-a4c5-7cf373aa1089 tempest-ServerAddressesNegativeTestJSON-1526246390 tempest-ServerAddressesNegativeTestJSON-1526246390-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61806) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 836.539215] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-f4603ddd-1b3d-4e85-81eb-1ffe5629d29c {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 836.541066] env[61806]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-2b7170d8-3721-47bc-b188-c618f5121265 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 836.546888] env[61806]: DEBUG oslo_vmware.api [None req-0ea3a38f-3871-4412-a6bc-d21c2086d262 tempest-AttachInterfacesV270Test-1974652806 tempest-AttachInterfacesV270Test-1974652806-project-member] Waiting for the task: (returnval){ [ 836.546888] env[61806]: value = "task-1294682" [ 836.546888] env[61806]: _type = "Task" [ 836.546888] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 836.550561] env[61806]: DEBUG nova.virt.vmwareapi.ds_util [None req-4e0c8443-0fd4-459b-a4c5-7cf373aa1089 tempest-ServerAddressesNegativeTestJSON-1526246390 tempest-ServerAddressesNegativeTestJSON-1526246390-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61806) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 836.550739] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-4e0c8443-0fd4-459b-a4c5-7cf373aa1089 tempest-ServerAddressesNegativeTestJSON-1526246390 tempest-ServerAddressesNegativeTestJSON-1526246390-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61806) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 836.551714] env[61806]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-2c2332c1-b41e-4cd8-b428-edcb13a36470 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 836.556855] env[61806]: DEBUG oslo_vmware.api [None req-0ea3a38f-3871-4412-a6bc-d21c2086d262 tempest-AttachInterfacesV270Test-1974652806 tempest-AttachInterfacesV270Test-1974652806-project-member] Task: {'id': task-1294682, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 836.559677] env[61806]: DEBUG oslo_vmware.api [None req-4e0c8443-0fd4-459b-a4c5-7cf373aa1089 tempest-ServerAddressesNegativeTestJSON-1526246390 tempest-ServerAddressesNegativeTestJSON-1526246390-project-member] Waiting for the task: (returnval){ [ 836.559677] env[61806]: value = "session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]52476bc5-cdc4-32c7-3fa5-7a8153e34fa5" [ 836.559677] env[61806]: _type = "Task" [ 836.559677] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 836.568921] env[61806]: DEBUG oslo_vmware.api [None req-4e0c8443-0fd4-459b-a4c5-7cf373aa1089 tempest-ServerAddressesNegativeTestJSON-1526246390 tempest-ServerAddressesNegativeTestJSON-1526246390-project-member] Task: {'id': session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]52476bc5-cdc4-32c7-3fa5-7a8153e34fa5, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 836.699750] env[61806]: DEBUG nova.network.neutron [req-b856ac6a-7846-4611-a003-a75556e76da7 req-475d8cec-205e-4d72-8c36-2570184072e4 service nova] [instance: b41531f2-f28c-4d82-9682-0b557bbaa491] Updated VIF entry in instance network info cache for port 513ca4b5-b62c-486a-9860-cf267a15c3e5. {{(pid=61806) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 836.700169] env[61806]: DEBUG nova.network.neutron [req-b856ac6a-7846-4611-a003-a75556e76da7 req-475d8cec-205e-4d72-8c36-2570184072e4 service nova] [instance: b41531f2-f28c-4d82-9682-0b557bbaa491] Updating instance_info_cache with network_info: [{"id": "513ca4b5-b62c-486a-9860-cf267a15c3e5", "address": "fa:16:3e:40:52:86", "network": {"id": "dfb52c93-e553-40d9-acb5-768cc6b67c89", "bridge": "br-int", "label": "tempest-SecurityGroupsTestJSON-1006891343-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "cfce93add38d40d79f294017f52e13d5", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6b6942d7-d4ab-4b2a-8d0f-76bf2a2478ad", "external-id": "nsx-vlan-transportzone-871", "segmentation_id": 871, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap513ca4b5-b6", "ovs_interfaceid": "513ca4b5-b62c-486a-9860-cf267a15c3e5", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61806) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 836.713696] env[61806]: DEBUG oslo_concurrency.lockutils [None req-334c6338-dc9d-4c32-ad77-bd3fe007216c tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.441s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 836.714115] env[61806]: DEBUG nova.compute.manager [None req-334c6338-dc9d-4c32-ad77-bd3fe007216c tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] [instance: be19fc44-ecbe-489c-9b6e-be8957bb1dd0] Start building networks asynchronously for instance. {{(pid=61806) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 836.717064] env[61806]: DEBUG oslo_concurrency.lockutils [None req-4e219b6d-f08a-4088-99ff-2fcd194df36a tempest-ServerShowV257Test-1159790492 tempest-ServerShowV257Test-1159790492-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: waited 13.975s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 836.717196] env[61806]: DEBUG nova.objects.instance [None req-4e219b6d-f08a-4088-99ff-2fcd194df36a tempest-ServerShowV257Test-1159790492 tempest-ServerShowV257Test-1159790492-project-member] [instance: 616ec206-9804-469e-ab5c-41aea7f048aa] Trying to apply a migration context that does not seem to be set for this instance {{(pid=61806) apply_migration_context /opt/stack/nova/nova/objects/instance.py:1067}} [ 836.740588] env[61806]: DEBUG oslo_vmware.api [None req-cbb90e6d-7dac-4c30-b860-1f0db7dd091e tempest-MultipleCreateTestJSON-1589015462 tempest-MultipleCreateTestJSON-1589015462-project-member] Task: {'id': task-1294681, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 836.749249] env[61806]: DEBUG oslo_concurrency.lockutils [None req-fd7f690e-61c7-4f8c-a053-2b904c459e5b tempest-SecurityGroupsTestJSON-1213485581 tempest-SecurityGroupsTestJSON-1213485581-project-member] Acquiring lock "b41531f2-f28c-4d82-9682-0b557bbaa491" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 836.749556] env[61806]: DEBUG oslo_concurrency.lockutils [None req-fd7f690e-61c7-4f8c-a053-2b904c459e5b tempest-SecurityGroupsTestJSON-1213485581 tempest-SecurityGroupsTestJSON-1213485581-project-member] Lock "b41531f2-f28c-4d82-9682-0b557bbaa491" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 836.749786] env[61806]: DEBUG oslo_concurrency.lockutils [None req-fd7f690e-61c7-4f8c-a053-2b904c459e5b tempest-SecurityGroupsTestJSON-1213485581 tempest-SecurityGroupsTestJSON-1213485581-project-member] Acquiring lock "b41531f2-f28c-4d82-9682-0b557bbaa491-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 836.750062] env[61806]: DEBUG oslo_concurrency.lockutils [None req-fd7f690e-61c7-4f8c-a053-2b904c459e5b tempest-SecurityGroupsTestJSON-1213485581 tempest-SecurityGroupsTestJSON-1213485581-project-member] Lock "b41531f2-f28c-4d82-9682-0b557bbaa491-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 836.750220] env[61806]: DEBUG oslo_concurrency.lockutils [None req-fd7f690e-61c7-4f8c-a053-2b904c459e5b tempest-SecurityGroupsTestJSON-1213485581 tempest-SecurityGroupsTestJSON-1213485581-project-member] Lock "b41531f2-f28c-4d82-9682-0b557bbaa491-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 836.754814] env[61806]: INFO nova.compute.manager [None req-fd7f690e-61c7-4f8c-a053-2b904c459e5b tempest-SecurityGroupsTestJSON-1213485581 tempest-SecurityGroupsTestJSON-1213485581-project-member] [instance: b41531f2-f28c-4d82-9682-0b557bbaa491] Terminating instance [ 836.756804] env[61806]: DEBUG nova.compute.manager [None req-fd7f690e-61c7-4f8c-a053-2b904c459e5b tempest-SecurityGroupsTestJSON-1213485581 tempest-SecurityGroupsTestJSON-1213485581-project-member] [instance: b41531f2-f28c-4d82-9682-0b557bbaa491] Start destroying the instance on the hypervisor. {{(pid=61806) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 836.757166] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-fd7f690e-61c7-4f8c-a053-2b904c459e5b tempest-SecurityGroupsTestJSON-1213485581 tempest-SecurityGroupsTestJSON-1213485581-project-member] [instance: b41531f2-f28c-4d82-9682-0b557bbaa491] Destroying instance {{(pid=61806) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 836.758435] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0d503cf3-344c-4019-9270-8c02088d0741 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 836.766971] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [None req-fd7f690e-61c7-4f8c-a053-2b904c459e5b tempest-SecurityGroupsTestJSON-1213485581 tempest-SecurityGroupsTestJSON-1213485581-project-member] [instance: b41531f2-f28c-4d82-9682-0b557bbaa491] Powering off the VM {{(pid=61806) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 836.767258] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-6516e95f-30f0-44e9-ba06-ded7df28b8af {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 836.773404] env[61806]: DEBUG oslo_vmware.api [None req-fd7f690e-61c7-4f8c-a053-2b904c459e5b tempest-SecurityGroupsTestJSON-1213485581 tempest-SecurityGroupsTestJSON-1213485581-project-member] Waiting for the task: (returnval){ [ 836.773404] env[61806]: value = "task-1294683" [ 836.773404] env[61806]: _type = "Task" [ 836.773404] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 836.783108] env[61806]: DEBUG oslo_vmware.api [None req-fd7f690e-61c7-4f8c-a053-2b904c459e5b tempest-SecurityGroupsTestJSON-1213485581 tempest-SecurityGroupsTestJSON-1213485581-project-member] Task: {'id': task-1294683, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 836.786605] env[61806]: DEBUG nova.network.neutron [req-dccfed4b-c313-4c44-bcf9-f1f8979c828c req-6a7a8dc0-1edb-43d7-bbd6-a0ecc58ac377 service nova] [instance: c8033551-6591-4e37-ae78-4efe7145b10b] Updated VIF entry in instance network info cache for port e8bfbf94-495b-4d13-8d2c-7a02e7a90387. {{(pid=61806) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 836.786947] env[61806]: DEBUG nova.network.neutron [req-dccfed4b-c313-4c44-bcf9-f1f8979c828c req-6a7a8dc0-1edb-43d7-bbd6-a0ecc58ac377 service nova] [instance: c8033551-6591-4e37-ae78-4efe7145b10b] Updating instance_info_cache with network_info: [{"id": "e8bfbf94-495b-4d13-8d2c-7a02e7a90387", "address": "fa:16:3e:25:8e:4f", "network": {"id": "c8fe2900-c167-4b7d-8df5-914d5fe8cab2", "bridge": "br-int", "label": "tempest-ServerAddressesNegativeTestJSON-1486853118-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "dd7c12f41e8b4ba489480ffa3959b7ce", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "1323cb03-8367-485a-962e-131af8eba474", "external-id": "nsx-vlan-transportzone-41", "segmentation_id": 41, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape8bfbf94-49", "ovs_interfaceid": "e8bfbf94-495b-4d13-8d2c-7a02e7a90387", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61806) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 836.822291] env[61806]: DEBUG nova.compute.manager [req-ed667bab-0d2f-4f1b-8d09-0a10dd9f9ed6 req-900021cf-01b4-4cd2-9fec-d2be18885ed0 service nova] [instance: 7f4907d0-178d-452a-8149-030becde8779] Received event network-vif-plugged-718d0c6d-5962-4b66-a13f-963aa96b2985 {{(pid=61806) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 836.822407] env[61806]: DEBUG oslo_concurrency.lockutils [req-ed667bab-0d2f-4f1b-8d09-0a10dd9f9ed6 req-900021cf-01b4-4cd2-9fec-d2be18885ed0 service nova] Acquiring lock "7f4907d0-178d-452a-8149-030becde8779-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 836.822691] env[61806]: DEBUG oslo_concurrency.lockutils [req-ed667bab-0d2f-4f1b-8d09-0a10dd9f9ed6 req-900021cf-01b4-4cd2-9fec-d2be18885ed0 service nova] Lock "7f4907d0-178d-452a-8149-030becde8779-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 836.822915] env[61806]: DEBUG oslo_concurrency.lockutils [req-ed667bab-0d2f-4f1b-8d09-0a10dd9f9ed6 req-900021cf-01b4-4cd2-9fec-d2be18885ed0 service nova] Lock "7f4907d0-178d-452a-8149-030becde8779-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 836.823062] env[61806]: DEBUG nova.compute.manager [req-ed667bab-0d2f-4f1b-8d09-0a10dd9f9ed6 req-900021cf-01b4-4cd2-9fec-d2be18885ed0 service nova] [instance: 7f4907d0-178d-452a-8149-030becde8779] No waiting events found dispatching network-vif-plugged-718d0c6d-5962-4b66-a13f-963aa96b2985 {{(pid=61806) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 836.823236] env[61806]: WARNING nova.compute.manager [req-ed667bab-0d2f-4f1b-8d09-0a10dd9f9ed6 req-900021cf-01b4-4cd2-9fec-d2be18885ed0 service nova] [instance: 7f4907d0-178d-452a-8149-030becde8779] Received unexpected event network-vif-plugged-718d0c6d-5962-4b66-a13f-963aa96b2985 for instance with vm_state building and task_state spawning. [ 836.823404] env[61806]: DEBUG nova.compute.manager [req-ed667bab-0d2f-4f1b-8d09-0a10dd9f9ed6 req-900021cf-01b4-4cd2-9fec-d2be18885ed0 service nova] [instance: 7f4907d0-178d-452a-8149-030becde8779] Received event network-changed-718d0c6d-5962-4b66-a13f-963aa96b2985 {{(pid=61806) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 836.823575] env[61806]: DEBUG nova.compute.manager [req-ed667bab-0d2f-4f1b-8d09-0a10dd9f9ed6 req-900021cf-01b4-4cd2-9fec-d2be18885ed0 service nova] [instance: 7f4907d0-178d-452a-8149-030becde8779] Refreshing instance network info cache due to event network-changed-718d0c6d-5962-4b66-a13f-963aa96b2985. {{(pid=61806) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 836.823771] env[61806]: DEBUG oslo_concurrency.lockutils [req-ed667bab-0d2f-4f1b-8d09-0a10dd9f9ed6 req-900021cf-01b4-4cd2-9fec-d2be18885ed0 service nova] Acquiring lock "refresh_cache-7f4907d0-178d-452a-8149-030becde8779" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 836.921393] env[61806]: DEBUG nova.network.neutron [None req-c9a1ae7a-8313-4bdf-8de4-77929ce2d9bc tempest-ServerMetadataNegativeTestJSON-456453902 tempest-ServerMetadataNegativeTestJSON-456453902-project-member] [instance: 7f4907d0-178d-452a-8149-030becde8779] Instance cache missing network info. {{(pid=61806) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 836.927212] env[61806]: DEBUG oslo_concurrency.lockutils [None req-247c941c-7e95-4297-9247-f35b645a15da tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] Acquiring lock "interface-2a3f81e3-b84f-4370-a598-2f6a607c3d47-None" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 836.927506] env[61806]: DEBUG oslo_concurrency.lockutils [None req-247c941c-7e95-4297-9247-f35b645a15da tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] Lock "interface-2a3f81e3-b84f-4370-a598-2f6a607c3d47-None" acquired by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: waited 0.000s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 836.927888] env[61806]: DEBUG nova.objects.instance [None req-247c941c-7e95-4297-9247-f35b645a15da tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] Lazy-loading 'flavor' on Instance uuid 2a3f81e3-b84f-4370-a598-2f6a607c3d47 {{(pid=61806) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 837.059688] env[61806]: DEBUG oslo_vmware.api [None req-0ea3a38f-3871-4412-a6bc-d21c2086d262 tempest-AttachInterfacesV270Test-1974652806 tempest-AttachInterfacesV270Test-1974652806-project-member] Task: {'id': task-1294682, 'name': CopyVirtualDisk_Task} progress is 25%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 837.072456] env[61806]: DEBUG oslo_vmware.api [None req-4e0c8443-0fd4-459b-a4c5-7cf373aa1089 tempest-ServerAddressesNegativeTestJSON-1526246390 tempest-ServerAddressesNegativeTestJSON-1526246390-project-member] Task: {'id': session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]52476bc5-cdc4-32c7-3fa5-7a8153e34fa5, 'name': SearchDatastore_Task, 'duration_secs': 0.015925} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 837.073653] env[61806]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-8ea412c3-8a06-4364-b6e1-a587fc875280 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 837.077214] env[61806]: DEBUG nova.network.neutron [None req-c9a1ae7a-8313-4bdf-8de4-77929ce2d9bc tempest-ServerMetadataNegativeTestJSON-456453902 tempest-ServerMetadataNegativeTestJSON-456453902-project-member] [instance: 7f4907d0-178d-452a-8149-030becde8779] Updating instance_info_cache with network_info: [{"id": "718d0c6d-5962-4b66-a13f-963aa96b2985", "address": "fa:16:3e:ca:61:8b", "network": {"id": "19ea3feb-a5e7-4f23-a2b4-b853d49f878d", "bridge": "br-int", "label": "tempest-ServerMetadataNegativeTestJSON-568051735-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "42d6862cb4734bd9a4ffcc64070c92a3", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "030ecc21-dc1c-4283-854e-88e623b3970a", "external-id": "nsx-vlan-transportzone-577", "segmentation_id": 577, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap718d0c6d-59", "ovs_interfaceid": "718d0c6d-5962-4b66-a13f-963aa96b2985", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61806) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 837.084929] env[61806]: DEBUG oslo_vmware.api [None req-4e0c8443-0fd4-459b-a4c5-7cf373aa1089 tempest-ServerAddressesNegativeTestJSON-1526246390 tempest-ServerAddressesNegativeTestJSON-1526246390-project-member] Waiting for the task: (returnval){ [ 837.084929] env[61806]: value = "session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]5231083a-7cd0-2a7b-1ad7-861a0010882b" [ 837.084929] env[61806]: _type = "Task" [ 837.084929] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 837.095068] env[61806]: DEBUG oslo_vmware.api [None req-4e0c8443-0fd4-459b-a4c5-7cf373aa1089 tempest-ServerAddressesNegativeTestJSON-1526246390 tempest-ServerAddressesNegativeTestJSON-1526246390-project-member] Task: {'id': session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]5231083a-7cd0-2a7b-1ad7-861a0010882b, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 837.107243] env[61806]: DEBUG oslo_service.periodic_task [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=61806) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 837.107243] env[61806]: DEBUG oslo_service.periodic_task [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Running periodic task ComputeManager._sync_scheduler_instance_info {{(pid=61806) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 837.203523] env[61806]: DEBUG oslo_concurrency.lockutils [req-b856ac6a-7846-4611-a003-a75556e76da7 req-475d8cec-205e-4d72-8c36-2570184072e4 service nova] Releasing lock "refresh_cache-b41531f2-f28c-4d82-9682-0b557bbaa491" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 837.222258] env[61806]: DEBUG nova.compute.utils [None req-334c6338-dc9d-4c32-ad77-bd3fe007216c tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Using /dev/sd instead of None {{(pid=61806) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 837.226909] env[61806]: DEBUG nova.compute.manager [None req-334c6338-dc9d-4c32-ad77-bd3fe007216c tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] [instance: be19fc44-ecbe-489c-9b6e-be8957bb1dd0] Allocating IP information in the background. {{(pid=61806) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 837.227129] env[61806]: DEBUG nova.network.neutron [None req-334c6338-dc9d-4c32-ad77-bd3fe007216c tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] [instance: be19fc44-ecbe-489c-9b6e-be8957bb1dd0] allocate_for_instance() {{(pid=61806) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 837.241423] env[61806]: DEBUG oslo_vmware.api [None req-cbb90e6d-7dac-4c30-b860-1f0db7dd091e tempest-MultipleCreateTestJSON-1589015462 tempest-MultipleCreateTestJSON-1589015462-project-member] Task: {'id': task-1294681, 'name': PowerOnVM_Task} progress is 100%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 837.271251] env[61806]: DEBUG nova.policy [None req-334c6338-dc9d-4c32-ad77-bd3fe007216c tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '2b86d596e89645c2ab31332afdcbc3c2', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '7ab95fb72b5d46c3b6c7bebbccf897cd', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61806) authorize /opt/stack/nova/nova/policy.py:201}} [ 837.282781] env[61806]: DEBUG oslo_vmware.api [None req-fd7f690e-61c7-4f8c-a053-2b904c459e5b tempest-SecurityGroupsTestJSON-1213485581 tempest-SecurityGroupsTestJSON-1213485581-project-member] Task: {'id': task-1294683, 'name': PowerOffVM_Task} progress is 100%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 837.289443] env[61806]: DEBUG oslo_concurrency.lockutils [req-dccfed4b-c313-4c44-bcf9-f1f8979c828c req-6a7a8dc0-1edb-43d7-bbd6-a0ecc58ac377 service nova] Releasing lock "refresh_cache-c8033551-6591-4e37-ae78-4efe7145b10b" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 837.432104] env[61806]: DEBUG nova.objects.instance [None req-247c941c-7e95-4297-9247-f35b645a15da tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] Lazy-loading 'pci_requests' on Instance uuid 2a3f81e3-b84f-4370-a598-2f6a607c3d47 {{(pid=61806) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 837.548685] env[61806]: DEBUG nova.network.neutron [None req-334c6338-dc9d-4c32-ad77-bd3fe007216c tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] [instance: be19fc44-ecbe-489c-9b6e-be8957bb1dd0] Successfully created port: 6b41a0ca-4d2e-4ab1-9893-180a2f0febca {{(pid=61806) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 837.561922] env[61806]: DEBUG oslo_vmware.api [None req-0ea3a38f-3871-4412-a6bc-d21c2086d262 tempest-AttachInterfacesV270Test-1974652806 tempest-AttachInterfacesV270Test-1974652806-project-member] Task: {'id': task-1294682, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.707948} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 837.561922] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [None req-0ea3a38f-3871-4412-a6bc-d21c2086d262 tempest-AttachInterfacesV270Test-1974652806 tempest-AttachInterfacesV270Test-1974652806-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa.vmdk to [datastore2] 117139cd-4ed4-4bea-8ad3-f4ac32c692e4/117139cd-4ed4-4bea-8ad3-f4ac32c692e4.vmdk {{(pid=61806) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 837.562305] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-0ea3a38f-3871-4412-a6bc-d21c2086d262 tempest-AttachInterfacesV270Test-1974652806 tempest-AttachInterfacesV270Test-1974652806-project-member] [instance: 117139cd-4ed4-4bea-8ad3-f4ac32c692e4] Extending root virtual disk to 1048576 {{(pid=61806) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 837.562613] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-bee99c74-210c-4892-868d-d8baad2ac141 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 837.570517] env[61806]: DEBUG oslo_vmware.api [None req-0ea3a38f-3871-4412-a6bc-d21c2086d262 tempest-AttachInterfacesV270Test-1974652806 tempest-AttachInterfacesV270Test-1974652806-project-member] Waiting for the task: (returnval){ [ 837.570517] env[61806]: value = "task-1294684" [ 837.570517] env[61806]: _type = "Task" [ 837.570517] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 837.579795] env[61806]: DEBUG oslo_vmware.api [None req-0ea3a38f-3871-4412-a6bc-d21c2086d262 tempest-AttachInterfacesV270Test-1974652806 tempest-AttachInterfacesV270Test-1974652806-project-member] Task: {'id': task-1294684, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 837.580622] env[61806]: DEBUG oslo_concurrency.lockutils [None req-c9a1ae7a-8313-4bdf-8de4-77929ce2d9bc tempest-ServerMetadataNegativeTestJSON-456453902 tempest-ServerMetadataNegativeTestJSON-456453902-project-member] Releasing lock "refresh_cache-7f4907d0-178d-452a-8149-030becde8779" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 837.580622] env[61806]: DEBUG nova.compute.manager [None req-c9a1ae7a-8313-4bdf-8de4-77929ce2d9bc tempest-ServerMetadataNegativeTestJSON-456453902 tempest-ServerMetadataNegativeTestJSON-456453902-project-member] [instance: 7f4907d0-178d-452a-8149-030becde8779] Instance network_info: |[{"id": "718d0c6d-5962-4b66-a13f-963aa96b2985", "address": "fa:16:3e:ca:61:8b", "network": {"id": "19ea3feb-a5e7-4f23-a2b4-b853d49f878d", "bridge": "br-int", "label": "tempest-ServerMetadataNegativeTestJSON-568051735-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "42d6862cb4734bd9a4ffcc64070c92a3", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "030ecc21-dc1c-4283-854e-88e623b3970a", "external-id": "nsx-vlan-transportzone-577", "segmentation_id": 577, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap718d0c6d-59", "ovs_interfaceid": "718d0c6d-5962-4b66-a13f-963aa96b2985", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61806) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 837.580873] env[61806]: DEBUG oslo_concurrency.lockutils [req-ed667bab-0d2f-4f1b-8d09-0a10dd9f9ed6 req-900021cf-01b4-4cd2-9fec-d2be18885ed0 service nova] Acquired lock "refresh_cache-7f4907d0-178d-452a-8149-030becde8779" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 837.581131] env[61806]: DEBUG nova.network.neutron [req-ed667bab-0d2f-4f1b-8d09-0a10dd9f9ed6 req-900021cf-01b4-4cd2-9fec-d2be18885ed0 service nova] [instance: 7f4907d0-178d-452a-8149-030becde8779] Refreshing network info cache for port 718d0c6d-5962-4b66-a13f-963aa96b2985 {{(pid=61806) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 837.582806] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-c9a1ae7a-8313-4bdf-8de4-77929ce2d9bc tempest-ServerMetadataNegativeTestJSON-456453902 tempest-ServerMetadataNegativeTestJSON-456453902-project-member] [instance: 7f4907d0-178d-452a-8149-030becde8779] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:ca:61:8b', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '030ecc21-dc1c-4283-854e-88e623b3970a', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '718d0c6d-5962-4b66-a13f-963aa96b2985', 'vif_model': 'vmxnet3'}] {{(pid=61806) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 837.591129] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [None req-c9a1ae7a-8313-4bdf-8de4-77929ce2d9bc tempest-ServerMetadataNegativeTestJSON-456453902 tempest-ServerMetadataNegativeTestJSON-456453902-project-member] Creating folder: Project (42d6862cb4734bd9a4ffcc64070c92a3). Parent ref: group-v277609. {{(pid=61806) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 837.594456] env[61806]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-2d97cbc5-a937-4088-b873-02436033bcaf {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 837.607935] env[61806]: DEBUG oslo_vmware.api [None req-4e0c8443-0fd4-459b-a4c5-7cf373aa1089 tempest-ServerAddressesNegativeTestJSON-1526246390 tempest-ServerAddressesNegativeTestJSON-1526246390-project-member] Task: {'id': session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]5231083a-7cd0-2a7b-1ad7-861a0010882b, 'name': SearchDatastore_Task, 'duration_secs': 0.06011} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 837.609249] env[61806]: DEBUG oslo_concurrency.lockutils [None req-4e0c8443-0fd4-459b-a4c5-7cf373aa1089 tempest-ServerAddressesNegativeTestJSON-1526246390 tempest-ServerAddressesNegativeTestJSON-1526246390-project-member] Releasing lock "[datastore2] devstack-image-cache_base/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa.vmdk" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 837.609516] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [None req-4e0c8443-0fd4-459b-a4c5-7cf373aa1089 tempest-ServerAddressesNegativeTestJSON-1526246390 tempest-ServerAddressesNegativeTestJSON-1526246390-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa.vmdk to [datastore2] c8033551-6591-4e37-ae78-4efe7145b10b/c8033551-6591-4e37-ae78-4efe7145b10b.vmdk {{(pid=61806) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 837.614373] env[61806]: INFO nova.virt.vmwareapi.vm_util [None req-c9a1ae7a-8313-4bdf-8de4-77929ce2d9bc tempest-ServerMetadataNegativeTestJSON-456453902 tempest-ServerMetadataNegativeTestJSON-456453902-project-member] Created folder: Project (42d6862cb4734bd9a4ffcc64070c92a3) in parent group-v277609. [ 837.614735] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [None req-c9a1ae7a-8313-4bdf-8de4-77929ce2d9bc tempest-ServerMetadataNegativeTestJSON-456453902 tempest-ServerMetadataNegativeTestJSON-456453902-project-member] Creating folder: Instances. Parent ref: group-v277677. {{(pid=61806) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 837.614994] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-09acde0f-34a8-4b69-982c-e5f199fe9a7d {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 837.616997] env[61806]: DEBUG oslo_service.periodic_task [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=61806) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 837.617198] env[61806]: DEBUG nova.compute.manager [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Starting heal instance info cache {{(pid=61806) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9926}} [ 837.617268] env[61806]: DEBUG nova.compute.manager [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Rebuilding the list of instances to heal {{(pid=61806) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9930}} [ 837.620626] env[61806]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-8ebf02ed-5b84-4e93-8f97-9ee692fe05e6 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 837.628405] env[61806]: DEBUG oslo_vmware.api [None req-4e0c8443-0fd4-459b-a4c5-7cf373aa1089 tempest-ServerAddressesNegativeTestJSON-1526246390 tempest-ServerAddressesNegativeTestJSON-1526246390-project-member] Waiting for the task: (returnval){ [ 837.628405] env[61806]: value = "task-1294686" [ 837.628405] env[61806]: _type = "Task" [ 837.628405] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 837.633844] env[61806]: INFO nova.virt.vmwareapi.vm_util [None req-c9a1ae7a-8313-4bdf-8de4-77929ce2d9bc tempest-ServerMetadataNegativeTestJSON-456453902 tempest-ServerMetadataNegativeTestJSON-456453902-project-member] Created folder: Instances in parent group-v277677. [ 837.634110] env[61806]: DEBUG oslo.service.loopingcall [None req-c9a1ae7a-8313-4bdf-8de4-77929ce2d9bc tempest-ServerMetadataNegativeTestJSON-456453902 tempest-ServerMetadataNegativeTestJSON-456453902-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61806) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 837.637020] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 7f4907d0-178d-452a-8149-030becde8779] Creating VM on the ESX host {{(pid=61806) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 837.637283] env[61806]: DEBUG oslo_vmware.api [None req-4e0c8443-0fd4-459b-a4c5-7cf373aa1089 tempest-ServerAddressesNegativeTestJSON-1526246390 tempest-ServerAddressesNegativeTestJSON-1526246390-project-member] Task: {'id': task-1294686, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 837.637490] env[61806]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-34f77fdb-a664-45fc-b487-668dc60fa988 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 837.661078] env[61806]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 837.661078] env[61806]: value = "task-1294688" [ 837.661078] env[61806]: _type = "Task" [ 837.661078] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 837.669299] env[61806]: DEBUG oslo_vmware.api [-] Task: {'id': task-1294688, 'name': CreateVM_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 837.731280] env[61806]: DEBUG oslo_concurrency.lockutils [None req-4e219b6d-f08a-4088-99ff-2fcd194df36a tempest-ServerShowV257Test-1159790492 tempest-ServerShowV257Test-1159790492-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: held 1.014s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 837.733174] env[61806]: DEBUG nova.compute.manager [None req-334c6338-dc9d-4c32-ad77-bd3fe007216c tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] [instance: be19fc44-ecbe-489c-9b6e-be8957bb1dd0] Start building block device mappings for instance. {{(pid=61806) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 837.741035] env[61806]: DEBUG oslo_concurrency.lockutils [None req-b326e3ef-f103-433e-9385-2b5e433b613a tempest-ServerShowV247Test-1218443059 tempest-ServerShowV247Test-1218443059-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 13.124s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 837.741349] env[61806]: DEBUG nova.objects.instance [None req-b326e3ef-f103-433e-9385-2b5e433b613a tempest-ServerShowV247Test-1218443059 tempest-ServerShowV247Test-1218443059-project-member] Lazy-loading 'resources' on Instance uuid 4febd093-9f2b-494e-b175-e4693b0e3e0d {{(pid=61806) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 837.752363] env[61806]: DEBUG oslo_vmware.api [None req-cbb90e6d-7dac-4c30-b860-1f0db7dd091e tempest-MultipleCreateTestJSON-1589015462 tempest-MultipleCreateTestJSON-1589015462-project-member] Task: {'id': task-1294681, 'name': PowerOnVM_Task, 'duration_secs': 1.186794} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 837.752729] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [None req-cbb90e6d-7dac-4c30-b860-1f0db7dd091e tempest-MultipleCreateTestJSON-1589015462 tempest-MultipleCreateTestJSON-1589015462-project-member] [instance: ef502a7d-7dc0-4319-8ce3-6a8701288237] Powered on the VM {{(pid=61806) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 837.752970] env[61806]: INFO nova.compute.manager [None req-cbb90e6d-7dac-4c30-b860-1f0db7dd091e tempest-MultipleCreateTestJSON-1589015462 tempest-MultipleCreateTestJSON-1589015462-project-member] [instance: ef502a7d-7dc0-4319-8ce3-6a8701288237] Took 9.73 seconds to spawn the instance on the hypervisor. [ 837.753169] env[61806]: DEBUG nova.compute.manager [None req-cbb90e6d-7dac-4c30-b860-1f0db7dd091e tempest-MultipleCreateTestJSON-1589015462 tempest-MultipleCreateTestJSON-1589015462-project-member] [instance: ef502a7d-7dc0-4319-8ce3-6a8701288237] Checking state {{(pid=61806) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 837.754012] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7b3571eb-fd82-4b1e-bc2e-95dc79c919a4 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 837.788021] env[61806]: DEBUG oslo_vmware.api [None req-fd7f690e-61c7-4f8c-a053-2b904c459e5b tempest-SecurityGroupsTestJSON-1213485581 tempest-SecurityGroupsTestJSON-1213485581-project-member] Task: {'id': task-1294683, 'name': PowerOffVM_Task, 'duration_secs': 0.70863} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 837.788423] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [None req-fd7f690e-61c7-4f8c-a053-2b904c459e5b tempest-SecurityGroupsTestJSON-1213485581 tempest-SecurityGroupsTestJSON-1213485581-project-member] [instance: b41531f2-f28c-4d82-9682-0b557bbaa491] Powered off the VM {{(pid=61806) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 837.788695] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-fd7f690e-61c7-4f8c-a053-2b904c459e5b tempest-SecurityGroupsTestJSON-1213485581 tempest-SecurityGroupsTestJSON-1213485581-project-member] [instance: b41531f2-f28c-4d82-9682-0b557bbaa491] Unregistering the VM {{(pid=61806) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 837.789052] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-80a8214b-0f85-404b-a354-8d6bf0f02e72 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 837.862166] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-fd7f690e-61c7-4f8c-a053-2b904c459e5b tempest-SecurityGroupsTestJSON-1213485581 tempest-SecurityGroupsTestJSON-1213485581-project-member] [instance: b41531f2-f28c-4d82-9682-0b557bbaa491] Unregistered the VM {{(pid=61806) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 837.862496] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-fd7f690e-61c7-4f8c-a053-2b904c459e5b tempest-SecurityGroupsTestJSON-1213485581 tempest-SecurityGroupsTestJSON-1213485581-project-member] [instance: b41531f2-f28c-4d82-9682-0b557bbaa491] Deleting contents of the VM from datastore datastore2 {{(pid=61806) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 837.862749] env[61806]: DEBUG nova.virt.vmwareapi.ds_util [None req-fd7f690e-61c7-4f8c-a053-2b904c459e5b tempest-SecurityGroupsTestJSON-1213485581 tempest-SecurityGroupsTestJSON-1213485581-project-member] Deleting the datastore file [datastore2] b41531f2-f28c-4d82-9682-0b557bbaa491 {{(pid=61806) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 837.863050] env[61806]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-b1777c30-c24a-4659-ba53-ff5772a4b11f {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 837.869365] env[61806]: DEBUG oslo_vmware.api [None req-fd7f690e-61c7-4f8c-a053-2b904c459e5b tempest-SecurityGroupsTestJSON-1213485581 tempest-SecurityGroupsTestJSON-1213485581-project-member] Waiting for the task: (returnval){ [ 837.869365] env[61806]: value = "task-1294690" [ 837.869365] env[61806]: _type = "Task" [ 837.869365] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 837.879422] env[61806]: DEBUG oslo_vmware.api [None req-fd7f690e-61c7-4f8c-a053-2b904c459e5b tempest-SecurityGroupsTestJSON-1213485581 tempest-SecurityGroupsTestJSON-1213485581-project-member] Task: {'id': task-1294690, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 837.920209] env[61806]: DEBUG nova.network.neutron [req-ed667bab-0d2f-4f1b-8d09-0a10dd9f9ed6 req-900021cf-01b4-4cd2-9fec-d2be18885ed0 service nova] [instance: 7f4907d0-178d-452a-8149-030becde8779] Updated VIF entry in instance network info cache for port 718d0c6d-5962-4b66-a13f-963aa96b2985. {{(pid=61806) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 837.920604] env[61806]: DEBUG nova.network.neutron [req-ed667bab-0d2f-4f1b-8d09-0a10dd9f9ed6 req-900021cf-01b4-4cd2-9fec-d2be18885ed0 service nova] [instance: 7f4907d0-178d-452a-8149-030becde8779] Updating instance_info_cache with network_info: [{"id": "718d0c6d-5962-4b66-a13f-963aa96b2985", "address": "fa:16:3e:ca:61:8b", "network": {"id": "19ea3feb-a5e7-4f23-a2b4-b853d49f878d", "bridge": "br-int", "label": "tempest-ServerMetadataNegativeTestJSON-568051735-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "42d6862cb4734bd9a4ffcc64070c92a3", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "030ecc21-dc1c-4283-854e-88e623b3970a", "external-id": "nsx-vlan-transportzone-577", "segmentation_id": 577, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap718d0c6d-59", "ovs_interfaceid": "718d0c6d-5962-4b66-a13f-963aa96b2985", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61806) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 837.934236] env[61806]: DEBUG nova.objects.base [None req-247c941c-7e95-4297-9247-f35b645a15da tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] Object Instance<2a3f81e3-b84f-4370-a598-2f6a607c3d47> lazy-loaded attributes: flavor,pci_requests {{(pid=61806) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 837.934486] env[61806]: DEBUG nova.network.neutron [None req-247c941c-7e95-4297-9247-f35b645a15da tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] [instance: 2a3f81e3-b84f-4370-a598-2f6a607c3d47] allocate_for_instance() {{(pid=61806) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 838.015599] env[61806]: DEBUG oslo_concurrency.lockutils [None req-247c941c-7e95-4297-9247-f35b645a15da tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] Lock "interface-2a3f81e3-b84f-4370-a598-2f6a607c3d47-None" "released" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: held 1.088s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 838.080893] env[61806]: DEBUG oslo_vmware.api [None req-0ea3a38f-3871-4412-a6bc-d21c2086d262 tempest-AttachInterfacesV270Test-1974652806 tempest-AttachInterfacesV270Test-1974652806-project-member] Task: {'id': task-1294684, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.140804} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 838.081210] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-0ea3a38f-3871-4412-a6bc-d21c2086d262 tempest-AttachInterfacesV270Test-1974652806 tempest-AttachInterfacesV270Test-1974652806-project-member] [instance: 117139cd-4ed4-4bea-8ad3-f4ac32c692e4] Extended root virtual disk {{(pid=61806) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 838.081976] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-568ff8ba-d66d-44bb-8a82-181bbddad24c {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 838.106159] env[61806]: DEBUG nova.virt.vmwareapi.volumeops [None req-0ea3a38f-3871-4412-a6bc-d21c2086d262 tempest-AttachInterfacesV270Test-1974652806 tempest-AttachInterfacesV270Test-1974652806-project-member] [instance: 117139cd-4ed4-4bea-8ad3-f4ac32c692e4] Reconfiguring VM instance instance-00000044 to attach disk [datastore2] 117139cd-4ed4-4bea-8ad3-f4ac32c692e4/117139cd-4ed4-4bea-8ad3-f4ac32c692e4.vmdk or device None with type sparse {{(pid=61806) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 838.106159] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-b9835813-f83f-4d99-a7d4-806164d10c1f {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 838.124674] env[61806]: DEBUG nova.compute.manager [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] [instance: b41531f2-f28c-4d82-9682-0b557bbaa491] Skipping network cache update for instance because it is being deleted. {{(pid=61806) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9943}} [ 838.124900] env[61806]: DEBUG nova.compute.manager [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] [instance: ef502a7d-7dc0-4319-8ce3-6a8701288237] Skipping network cache update for instance because it is Building. {{(pid=61806) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9939}} [ 838.125103] env[61806]: DEBUG nova.compute.manager [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] [instance: 117139cd-4ed4-4bea-8ad3-f4ac32c692e4] Skipping network cache update for instance because it is Building. {{(pid=61806) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9939}} [ 838.125288] env[61806]: DEBUG nova.compute.manager [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] [instance: c8033551-6591-4e37-ae78-4efe7145b10b] Skipping network cache update for instance because it is Building. {{(pid=61806) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9939}} [ 838.125465] env[61806]: DEBUG nova.compute.manager [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] [instance: 7f4907d0-178d-452a-8149-030becde8779] Skipping network cache update for instance because it is Building. {{(pid=61806) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9939}} [ 838.125642] env[61806]: DEBUG nova.compute.manager [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] [instance: be19fc44-ecbe-489c-9b6e-be8957bb1dd0] Skipping network cache update for instance because it is Building. {{(pid=61806) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9939}} [ 838.137688] env[61806]: DEBUG oslo_vmware.api [None req-0ea3a38f-3871-4412-a6bc-d21c2086d262 tempest-AttachInterfacesV270Test-1974652806 tempest-AttachInterfacesV270Test-1974652806-project-member] Waiting for the task: (returnval){ [ 838.137688] env[61806]: value = "task-1294691" [ 838.137688] env[61806]: _type = "Task" [ 838.137688] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 838.143255] env[61806]: DEBUG oslo_vmware.api [None req-4e0c8443-0fd4-459b-a4c5-7cf373aa1089 tempest-ServerAddressesNegativeTestJSON-1526246390 tempest-ServerAddressesNegativeTestJSON-1526246390-project-member] Task: {'id': task-1294686, 'name': CopyVirtualDisk_Task} progress is 4%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 838.150042] env[61806]: DEBUG oslo_vmware.api [None req-0ea3a38f-3871-4412-a6bc-d21c2086d262 tempest-AttachInterfacesV270Test-1974652806 tempest-AttachInterfacesV270Test-1974652806-project-member] Task: {'id': task-1294691, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 838.173120] env[61806]: DEBUG oslo_vmware.api [-] Task: {'id': task-1294688, 'name': CreateVM_Task} progress is 25%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 838.178612] env[61806]: DEBUG oslo_concurrency.lockutils [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Acquiring lock "refresh_cache-e0ef0a35-82a5-495b-9d5c-5805e8306390" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 838.178773] env[61806]: DEBUG oslo_concurrency.lockutils [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Acquired lock "refresh_cache-e0ef0a35-82a5-495b-9d5c-5805e8306390" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 838.178923] env[61806]: DEBUG nova.network.neutron [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] [instance: e0ef0a35-82a5-495b-9d5c-5805e8306390] Forcefully refreshing network info cache for instance {{(pid=61806) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2004}} [ 838.179104] env[61806]: DEBUG nova.objects.instance [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Lazy-loading 'info_cache' on Instance uuid e0ef0a35-82a5-495b-9d5c-5805e8306390 {{(pid=61806) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 838.277938] env[61806]: INFO nova.compute.manager [None req-cbb90e6d-7dac-4c30-b860-1f0db7dd091e tempest-MultipleCreateTestJSON-1589015462 tempest-MultipleCreateTestJSON-1589015462-project-member] [instance: ef502a7d-7dc0-4319-8ce3-6a8701288237] Took 28.80 seconds to build instance. [ 838.384334] env[61806]: DEBUG oslo_vmware.api [None req-fd7f690e-61c7-4f8c-a053-2b904c459e5b tempest-SecurityGroupsTestJSON-1213485581 tempest-SecurityGroupsTestJSON-1213485581-project-member] Task: {'id': task-1294690, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 838.423384] env[61806]: DEBUG oslo_concurrency.lockutils [req-ed667bab-0d2f-4f1b-8d09-0a10dd9f9ed6 req-900021cf-01b4-4cd2-9fec-d2be18885ed0 service nova] Releasing lock "refresh_cache-7f4907d0-178d-452a-8149-030becde8779" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 838.549527] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-09be053c-5a49-486f-a2e1-f5c267e401e7 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 838.557558] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-73ef8ff0-48d0-4b56-9aa0-0eb2cbb94a32 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 838.587648] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bff4d1c1-53ac-44e9-9025-618dad42025c {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 838.595434] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-545b386a-b75e-4630-bc1c-628e1317991f {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 838.608454] env[61806]: DEBUG nova.compute.provider_tree [None req-b326e3ef-f103-433e-9385-2b5e433b613a tempest-ServerShowV247Test-1218443059 tempest-ServerShowV247Test-1218443059-project-member] Inventory has not changed in ProviderTree for provider: a2858be1-fd22-4e08-979e-87ad25293407 {{(pid=61806) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 838.640422] env[61806]: DEBUG oslo_vmware.api [None req-4e0c8443-0fd4-459b-a4c5-7cf373aa1089 tempest-ServerAddressesNegativeTestJSON-1526246390 tempest-ServerAddressesNegativeTestJSON-1526246390-project-member] Task: {'id': task-1294686, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.883945} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 838.643331] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [None req-4e0c8443-0fd4-459b-a4c5-7cf373aa1089 tempest-ServerAddressesNegativeTestJSON-1526246390 tempest-ServerAddressesNegativeTestJSON-1526246390-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa.vmdk to [datastore2] c8033551-6591-4e37-ae78-4efe7145b10b/c8033551-6591-4e37-ae78-4efe7145b10b.vmdk {{(pid=61806) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 838.643584] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-4e0c8443-0fd4-459b-a4c5-7cf373aa1089 tempest-ServerAddressesNegativeTestJSON-1526246390 tempest-ServerAddressesNegativeTestJSON-1526246390-project-member] [instance: c8033551-6591-4e37-ae78-4efe7145b10b] Extending root virtual disk to 1048576 {{(pid=61806) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 838.643809] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-7ffc4aa0-3515-4b62-a94d-342f834f47eb {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 838.650507] env[61806]: DEBUG oslo_vmware.api [None req-0ea3a38f-3871-4412-a6bc-d21c2086d262 tempest-AttachInterfacesV270Test-1974652806 tempest-AttachInterfacesV270Test-1974652806-project-member] Task: {'id': task-1294691, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 838.651895] env[61806]: DEBUG oslo_vmware.api [None req-4e0c8443-0fd4-459b-a4c5-7cf373aa1089 tempest-ServerAddressesNegativeTestJSON-1526246390 tempest-ServerAddressesNegativeTestJSON-1526246390-project-member] Waiting for the task: (returnval){ [ 838.651895] env[61806]: value = "task-1294692" [ 838.651895] env[61806]: _type = "Task" [ 838.651895] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 838.658889] env[61806]: DEBUG oslo_vmware.api [None req-4e0c8443-0fd4-459b-a4c5-7cf373aa1089 tempest-ServerAddressesNegativeTestJSON-1526246390 tempest-ServerAddressesNegativeTestJSON-1526246390-project-member] Task: {'id': task-1294692, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 838.670676] env[61806]: DEBUG oslo_vmware.api [-] Task: {'id': task-1294688, 'name': CreateVM_Task, 'duration_secs': 0.681168} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 838.670842] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 7f4907d0-178d-452a-8149-030becde8779] Created VM on the ESX host {{(pid=61806) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 838.671567] env[61806]: DEBUG oslo_concurrency.lockutils [None req-c9a1ae7a-8313-4bdf-8de4-77929ce2d9bc tempest-ServerMetadataNegativeTestJSON-456453902 tempest-ServerMetadataNegativeTestJSON-456453902-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 838.671744] env[61806]: DEBUG oslo_concurrency.lockutils [None req-c9a1ae7a-8313-4bdf-8de4-77929ce2d9bc tempest-ServerMetadataNegativeTestJSON-456453902 tempest-ServerMetadataNegativeTestJSON-456453902-project-member] Acquired lock "[datastore2] devstack-image-cache_base/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 838.672095] env[61806]: DEBUG oslo_concurrency.lockutils [None req-c9a1ae7a-8313-4bdf-8de4-77929ce2d9bc tempest-ServerMetadataNegativeTestJSON-456453902 tempest-ServerMetadataNegativeTestJSON-456453902-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 838.672354] env[61806]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-59a7cfad-cada-4eb2-a41d-32daabbf7f19 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 838.676558] env[61806]: DEBUG oslo_vmware.api [None req-c9a1ae7a-8313-4bdf-8de4-77929ce2d9bc tempest-ServerMetadataNegativeTestJSON-456453902 tempest-ServerMetadataNegativeTestJSON-456453902-project-member] Waiting for the task: (returnval){ [ 838.676558] env[61806]: value = "session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]52e8915e-55d0-dd27-1c65-d970cb49f48d" [ 838.676558] env[61806]: _type = "Task" [ 838.676558] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 838.686354] env[61806]: DEBUG oslo_vmware.api [None req-c9a1ae7a-8313-4bdf-8de4-77929ce2d9bc tempest-ServerMetadataNegativeTestJSON-456453902 tempest-ServerMetadataNegativeTestJSON-456453902-project-member] Task: {'id': session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]52e8915e-55d0-dd27-1c65-d970cb49f48d, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 838.751926] env[61806]: DEBUG nova.compute.manager [None req-334c6338-dc9d-4c32-ad77-bd3fe007216c tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] [instance: be19fc44-ecbe-489c-9b6e-be8957bb1dd0] Start spawning the instance on the hypervisor. {{(pid=61806) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 838.776013] env[61806]: DEBUG nova.virt.hardware [None req-334c6338-dc9d-4c32-ad77-bd3fe007216c tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-15T18:05:10Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-15T18:04:52Z,direct_url=,disk_format='vmdk',id=ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='84ba83fce2c349a6988173c5d6fc3b07',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-15T18:04:53Z,virtual_size=,visibility=), allow threads: False {{(pid=61806) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 838.776344] env[61806]: DEBUG nova.virt.hardware [None req-334c6338-dc9d-4c32-ad77-bd3fe007216c tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Flavor limits 0:0:0 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 838.776555] env[61806]: DEBUG nova.virt.hardware [None req-334c6338-dc9d-4c32-ad77-bd3fe007216c tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Image limits 0:0:0 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 838.776829] env[61806]: DEBUG nova.virt.hardware [None req-334c6338-dc9d-4c32-ad77-bd3fe007216c tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Flavor pref 0:0:0 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 838.777037] env[61806]: DEBUG nova.virt.hardware [None req-334c6338-dc9d-4c32-ad77-bd3fe007216c tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Image pref 0:0:0 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 838.777238] env[61806]: DEBUG nova.virt.hardware [None req-334c6338-dc9d-4c32-ad77-bd3fe007216c tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 838.777487] env[61806]: DEBUG nova.virt.hardware [None req-334c6338-dc9d-4c32-ad77-bd3fe007216c tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61806) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 838.777689] env[61806]: DEBUG nova.virt.hardware [None req-334c6338-dc9d-4c32-ad77-bd3fe007216c tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61806) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 838.777905] env[61806]: DEBUG nova.virt.hardware [None req-334c6338-dc9d-4c32-ad77-bd3fe007216c tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Got 1 possible topologies {{(pid=61806) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 838.778138] env[61806]: DEBUG nova.virt.hardware [None req-334c6338-dc9d-4c32-ad77-bd3fe007216c tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61806) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 838.778374] env[61806]: DEBUG nova.virt.hardware [None req-334c6338-dc9d-4c32-ad77-bd3fe007216c tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61806) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 838.779233] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0c2c1d37-d943-46c7-ba2c-6653d1d98a95 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 838.786229] env[61806]: DEBUG oslo_concurrency.lockutils [None req-cbb90e6d-7dac-4c30-b860-1f0db7dd091e tempest-MultipleCreateTestJSON-1589015462 tempest-MultipleCreateTestJSON-1589015462-project-member] Lock "ef502a7d-7dc0-4319-8ce3-6a8701288237" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 81.478s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 838.787525] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9083c63f-212f-4cfc-bae2-04663ed38528 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 838.880467] env[61806]: DEBUG oslo_vmware.api [None req-fd7f690e-61c7-4f8c-a053-2b904c459e5b tempest-SecurityGroupsTestJSON-1213485581 tempest-SecurityGroupsTestJSON-1213485581-project-member] Task: {'id': task-1294690, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.660399} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 838.880735] env[61806]: DEBUG nova.virt.vmwareapi.ds_util [None req-fd7f690e-61c7-4f8c-a053-2b904c459e5b tempest-SecurityGroupsTestJSON-1213485581 tempest-SecurityGroupsTestJSON-1213485581-project-member] Deleted the datastore file {{(pid=61806) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 838.880933] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-fd7f690e-61c7-4f8c-a053-2b904c459e5b tempest-SecurityGroupsTestJSON-1213485581 tempest-SecurityGroupsTestJSON-1213485581-project-member] [instance: b41531f2-f28c-4d82-9682-0b557bbaa491] Deleted contents of the VM from datastore datastore2 {{(pid=61806) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 838.881129] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-fd7f690e-61c7-4f8c-a053-2b904c459e5b tempest-SecurityGroupsTestJSON-1213485581 tempest-SecurityGroupsTestJSON-1213485581-project-member] [instance: b41531f2-f28c-4d82-9682-0b557bbaa491] Instance destroyed {{(pid=61806) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 838.881306] env[61806]: INFO nova.compute.manager [None req-fd7f690e-61c7-4f8c-a053-2b904c459e5b tempest-SecurityGroupsTestJSON-1213485581 tempest-SecurityGroupsTestJSON-1213485581-project-member] [instance: b41531f2-f28c-4d82-9682-0b557bbaa491] Took 2.12 seconds to destroy the instance on the hypervisor. [ 838.881548] env[61806]: DEBUG oslo.service.loopingcall [None req-fd7f690e-61c7-4f8c-a053-2b904c459e5b tempest-SecurityGroupsTestJSON-1213485581 tempest-SecurityGroupsTestJSON-1213485581-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61806) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 838.882054] env[61806]: DEBUG nova.compute.manager [-] [instance: b41531f2-f28c-4d82-9682-0b557bbaa491] Deallocating network for instance {{(pid=61806) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 838.882548] env[61806]: DEBUG nova.network.neutron [-] [instance: b41531f2-f28c-4d82-9682-0b557bbaa491] deallocate_for_instance() {{(pid=61806) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 839.110905] env[61806]: DEBUG nova.scheduler.client.report [None req-b326e3ef-f103-433e-9385-2b5e433b613a tempest-ServerShowV247Test-1218443059 tempest-ServerShowV247Test-1218443059-project-member] Inventory has not changed for provider a2858be1-fd22-4e08-979e-87ad25293407 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 138, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61806) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 839.152990] env[61806]: DEBUG oslo_vmware.api [None req-0ea3a38f-3871-4412-a6bc-d21c2086d262 tempest-AttachInterfacesV270Test-1974652806 tempest-AttachInterfacesV270Test-1974652806-project-member] Task: {'id': task-1294691, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 839.162159] env[61806]: DEBUG oslo_vmware.api [None req-4e0c8443-0fd4-459b-a4c5-7cf373aa1089 tempest-ServerAddressesNegativeTestJSON-1526246390 tempest-ServerAddressesNegativeTestJSON-1526246390-project-member] Task: {'id': task-1294692, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.073246} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 839.163050] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-4e0c8443-0fd4-459b-a4c5-7cf373aa1089 tempest-ServerAddressesNegativeTestJSON-1526246390 tempest-ServerAddressesNegativeTestJSON-1526246390-project-member] [instance: c8033551-6591-4e37-ae78-4efe7145b10b] Extended root virtual disk {{(pid=61806) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 839.163250] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-148f0096-6a59-4242-be80-ee2f67ce3bc9 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 839.185953] env[61806]: DEBUG nova.virt.vmwareapi.volumeops [None req-4e0c8443-0fd4-459b-a4c5-7cf373aa1089 tempest-ServerAddressesNegativeTestJSON-1526246390 tempest-ServerAddressesNegativeTestJSON-1526246390-project-member] [instance: c8033551-6591-4e37-ae78-4efe7145b10b] Reconfiguring VM instance instance-00000045 to attach disk [datastore2] c8033551-6591-4e37-ae78-4efe7145b10b/c8033551-6591-4e37-ae78-4efe7145b10b.vmdk or device None with type sparse {{(pid=61806) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 839.190166] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-cab9afcb-30f3-4f6b-b1f4-342247b0b17c {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 839.205503] env[61806]: DEBUG nova.compute.manager [req-48b551dc-06bc-423b-b78f-7015b99b0fb4 req-5392084c-8e7b-4e2a-a125-85b54c3a4def service nova] [instance: be19fc44-ecbe-489c-9b6e-be8957bb1dd0] Received event network-vif-plugged-6b41a0ca-4d2e-4ab1-9893-180a2f0febca {{(pid=61806) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 839.205702] env[61806]: DEBUG oslo_concurrency.lockutils [req-48b551dc-06bc-423b-b78f-7015b99b0fb4 req-5392084c-8e7b-4e2a-a125-85b54c3a4def service nova] Acquiring lock "be19fc44-ecbe-489c-9b6e-be8957bb1dd0-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 839.205911] env[61806]: DEBUG oslo_concurrency.lockutils [req-48b551dc-06bc-423b-b78f-7015b99b0fb4 req-5392084c-8e7b-4e2a-a125-85b54c3a4def service nova] Lock "be19fc44-ecbe-489c-9b6e-be8957bb1dd0-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 839.206094] env[61806]: DEBUG oslo_concurrency.lockutils [req-48b551dc-06bc-423b-b78f-7015b99b0fb4 req-5392084c-8e7b-4e2a-a125-85b54c3a4def service nova] Lock "be19fc44-ecbe-489c-9b6e-be8957bb1dd0-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 839.206266] env[61806]: DEBUG nova.compute.manager [req-48b551dc-06bc-423b-b78f-7015b99b0fb4 req-5392084c-8e7b-4e2a-a125-85b54c3a4def service nova] [instance: be19fc44-ecbe-489c-9b6e-be8957bb1dd0] No waiting events found dispatching network-vif-plugged-6b41a0ca-4d2e-4ab1-9893-180a2f0febca {{(pid=61806) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 839.206434] env[61806]: WARNING nova.compute.manager [req-48b551dc-06bc-423b-b78f-7015b99b0fb4 req-5392084c-8e7b-4e2a-a125-85b54c3a4def service nova] [instance: be19fc44-ecbe-489c-9b6e-be8957bb1dd0] Received unexpected event network-vif-plugged-6b41a0ca-4d2e-4ab1-9893-180a2f0febca for instance with vm_state building and task_state spawning. [ 839.214659] env[61806]: DEBUG oslo_vmware.api [None req-c9a1ae7a-8313-4bdf-8de4-77929ce2d9bc tempest-ServerMetadataNegativeTestJSON-456453902 tempest-ServerMetadataNegativeTestJSON-456453902-project-member] Task: {'id': session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]52e8915e-55d0-dd27-1c65-d970cb49f48d, 'name': SearchDatastore_Task, 'duration_secs': 0.011448} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 839.215852] env[61806]: DEBUG oslo_concurrency.lockutils [None req-c9a1ae7a-8313-4bdf-8de4-77929ce2d9bc tempest-ServerMetadataNegativeTestJSON-456453902 tempest-ServerMetadataNegativeTestJSON-456453902-project-member] Releasing lock "[datastore2] devstack-image-cache_base/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 839.216090] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-c9a1ae7a-8313-4bdf-8de4-77929ce2d9bc tempest-ServerMetadataNegativeTestJSON-456453902 tempest-ServerMetadataNegativeTestJSON-456453902-project-member] [instance: 7f4907d0-178d-452a-8149-030becde8779] Processing image ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa {{(pid=61806) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 839.216318] env[61806]: DEBUG oslo_concurrency.lockutils [None req-c9a1ae7a-8313-4bdf-8de4-77929ce2d9bc tempest-ServerMetadataNegativeTestJSON-456453902 tempest-ServerMetadataNegativeTestJSON-456453902-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa.vmdk" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 839.216469] env[61806]: DEBUG oslo_concurrency.lockutils [None req-c9a1ae7a-8313-4bdf-8de4-77929ce2d9bc tempest-ServerMetadataNegativeTestJSON-456453902 tempest-ServerMetadataNegativeTestJSON-456453902-project-member] Acquired lock "[datastore2] devstack-image-cache_base/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa.vmdk" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 839.216654] env[61806]: DEBUG nova.virt.vmwareapi.ds_util [None req-c9a1ae7a-8313-4bdf-8de4-77929ce2d9bc tempest-ServerMetadataNegativeTestJSON-456453902 tempest-ServerMetadataNegativeTestJSON-456453902-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61806) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 839.216958] env[61806]: DEBUG oslo_vmware.api [None req-4e0c8443-0fd4-459b-a4c5-7cf373aa1089 tempest-ServerAddressesNegativeTestJSON-1526246390 tempest-ServerAddressesNegativeTestJSON-1526246390-project-member] Waiting for the task: (returnval){ [ 839.216958] env[61806]: value = "task-1294693" [ 839.216958] env[61806]: _type = "Task" [ 839.216958] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 839.217284] env[61806]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-78c5c03f-37dd-4963-a304-47ae58728a98 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 839.229797] env[61806]: DEBUG oslo_vmware.api [None req-4e0c8443-0fd4-459b-a4c5-7cf373aa1089 tempest-ServerAddressesNegativeTestJSON-1526246390 tempest-ServerAddressesNegativeTestJSON-1526246390-project-member] Task: {'id': task-1294693, 'name': ReconfigVM_Task} progress is 10%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 839.231068] env[61806]: DEBUG nova.virt.vmwareapi.ds_util [None req-c9a1ae7a-8313-4bdf-8de4-77929ce2d9bc tempest-ServerMetadataNegativeTestJSON-456453902 tempest-ServerMetadataNegativeTestJSON-456453902-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61806) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 839.231358] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-c9a1ae7a-8313-4bdf-8de4-77929ce2d9bc tempest-ServerMetadataNegativeTestJSON-456453902 tempest-ServerMetadataNegativeTestJSON-456453902-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61806) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 839.232320] env[61806]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-49e21772-bd00-4d38-9a6a-b386a03d041f {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 839.237753] env[61806]: DEBUG oslo_vmware.api [None req-c9a1ae7a-8313-4bdf-8de4-77929ce2d9bc tempest-ServerMetadataNegativeTestJSON-456453902 tempest-ServerMetadataNegativeTestJSON-456453902-project-member] Waiting for the task: (returnval){ [ 839.237753] env[61806]: value = "session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]52af72b1-d4bb-d95f-7b99-1e2a6ce25c9b" [ 839.237753] env[61806]: _type = "Task" [ 839.237753] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 839.241334] env[61806]: DEBUG oslo_concurrency.lockutils [None req-3bd56cbd-68df-4151-a169-676f8bdda9ab tempest-MultipleCreateTestJSON-1589015462 tempest-MultipleCreateTestJSON-1589015462-project-member] Acquiring lock "25218e83-c1ab-4b97-ade1-7c32b6f99305" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 839.241618] env[61806]: DEBUG oslo_concurrency.lockutils [None req-3bd56cbd-68df-4151-a169-676f8bdda9ab tempest-MultipleCreateTestJSON-1589015462 tempest-MultipleCreateTestJSON-1589015462-project-member] Lock "25218e83-c1ab-4b97-ade1-7c32b6f99305" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 839.241768] env[61806]: DEBUG oslo_concurrency.lockutils [None req-3bd56cbd-68df-4151-a169-676f8bdda9ab tempest-MultipleCreateTestJSON-1589015462 tempest-MultipleCreateTestJSON-1589015462-project-member] Acquiring lock "25218e83-c1ab-4b97-ade1-7c32b6f99305-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 839.242357] env[61806]: DEBUG oslo_concurrency.lockutils [None req-3bd56cbd-68df-4151-a169-676f8bdda9ab tempest-MultipleCreateTestJSON-1589015462 tempest-MultipleCreateTestJSON-1589015462-project-member] Lock "25218e83-c1ab-4b97-ade1-7c32b6f99305-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.001s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 839.242551] env[61806]: DEBUG oslo_concurrency.lockutils [None req-3bd56cbd-68df-4151-a169-676f8bdda9ab tempest-MultipleCreateTestJSON-1589015462 tempest-MultipleCreateTestJSON-1589015462-project-member] Lock "25218e83-c1ab-4b97-ade1-7c32b6f99305-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 839.244676] env[61806]: INFO nova.compute.manager [None req-3bd56cbd-68df-4151-a169-676f8bdda9ab tempest-MultipleCreateTestJSON-1589015462 tempest-MultipleCreateTestJSON-1589015462-project-member] [instance: 25218e83-c1ab-4b97-ade1-7c32b6f99305] Terminating instance [ 839.246442] env[61806]: DEBUG nova.compute.manager [None req-3bd56cbd-68df-4151-a169-676f8bdda9ab tempest-MultipleCreateTestJSON-1589015462 tempest-MultipleCreateTestJSON-1589015462-project-member] [instance: 25218e83-c1ab-4b97-ade1-7c32b6f99305] Start destroying the instance on the hypervisor. {{(pid=61806) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 839.246662] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-3bd56cbd-68df-4151-a169-676f8bdda9ab tempest-MultipleCreateTestJSON-1589015462 tempest-MultipleCreateTestJSON-1589015462-project-member] [instance: 25218e83-c1ab-4b97-ade1-7c32b6f99305] Destroying instance {{(pid=61806) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 839.249178] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8f0f047f-03a0-41b9-a499-a8746312505f {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 839.257950] env[61806]: DEBUG oslo_vmware.api [None req-c9a1ae7a-8313-4bdf-8de4-77929ce2d9bc tempest-ServerMetadataNegativeTestJSON-456453902 tempest-ServerMetadataNegativeTestJSON-456453902-project-member] Task: {'id': session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]52af72b1-d4bb-d95f-7b99-1e2a6ce25c9b, 'name': SearchDatastore_Task, 'duration_secs': 0.014135} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 839.261412] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [None req-3bd56cbd-68df-4151-a169-676f8bdda9ab tempest-MultipleCreateTestJSON-1589015462 tempest-MultipleCreateTestJSON-1589015462-project-member] [instance: 25218e83-c1ab-4b97-ade1-7c32b6f99305] Powering off the VM {{(pid=61806) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 839.261664] env[61806]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-1cf42050-9341-4203-b818-f78e2c78b391 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 839.264217] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-964ba7ec-33a3-464e-b655-adaa0f657b9a {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 839.269512] env[61806]: DEBUG oslo_vmware.api [None req-c9a1ae7a-8313-4bdf-8de4-77929ce2d9bc tempest-ServerMetadataNegativeTestJSON-456453902 tempest-ServerMetadataNegativeTestJSON-456453902-project-member] Waiting for the task: (returnval){ [ 839.269512] env[61806]: value = "session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]52f8adf4-9b7c-a76a-d7cd-08041becb336" [ 839.269512] env[61806]: _type = "Task" [ 839.269512] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 839.273730] env[61806]: DEBUG oslo_vmware.api [None req-3bd56cbd-68df-4151-a169-676f8bdda9ab tempest-MultipleCreateTestJSON-1589015462 tempest-MultipleCreateTestJSON-1589015462-project-member] Waiting for the task: (returnval){ [ 839.273730] env[61806]: value = "task-1294694" [ 839.273730] env[61806]: _type = "Task" [ 839.273730] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 839.281052] env[61806]: DEBUG oslo_vmware.api [None req-c9a1ae7a-8313-4bdf-8de4-77929ce2d9bc tempest-ServerMetadataNegativeTestJSON-456453902 tempest-ServerMetadataNegativeTestJSON-456453902-project-member] Task: {'id': session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]52f8adf4-9b7c-a76a-d7cd-08041becb336, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 839.284179] env[61806]: DEBUG oslo_vmware.api [None req-3bd56cbd-68df-4151-a169-676f8bdda9ab tempest-MultipleCreateTestJSON-1589015462 tempest-MultipleCreateTestJSON-1589015462-project-member] Task: {'id': task-1294694, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 839.435994] env[61806]: DEBUG nova.compute.manager [req-df79f25b-0653-4e77-adbc-0a1400e0bed4 req-8b5baad3-0914-4550-a472-aa72efb1d4dc service nova] [instance: b41531f2-f28c-4d82-9682-0b557bbaa491] Received event network-vif-deleted-513ca4b5-b62c-486a-9860-cf267a15c3e5 {{(pid=61806) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 839.436250] env[61806]: INFO nova.compute.manager [req-df79f25b-0653-4e77-adbc-0a1400e0bed4 req-8b5baad3-0914-4550-a472-aa72efb1d4dc service nova] [instance: b41531f2-f28c-4d82-9682-0b557bbaa491] Neutron deleted interface 513ca4b5-b62c-486a-9860-cf267a15c3e5; detaching it from the instance and deleting it from the info cache [ 839.436480] env[61806]: DEBUG nova.network.neutron [req-df79f25b-0653-4e77-adbc-0a1400e0bed4 req-8b5baad3-0914-4550-a472-aa72efb1d4dc service nova] [instance: b41531f2-f28c-4d82-9682-0b557bbaa491] Updating instance_info_cache with network_info: [] {{(pid=61806) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 839.511611] env[61806]: DEBUG nova.network.neutron [None req-334c6338-dc9d-4c32-ad77-bd3fe007216c tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] [instance: be19fc44-ecbe-489c-9b6e-be8957bb1dd0] Successfully updated port: 6b41a0ca-4d2e-4ab1-9893-180a2f0febca {{(pid=61806) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 839.597647] env[61806]: DEBUG oslo_concurrency.lockutils [None req-182a714a-9a7d-40bc-ad59-30e74e8a3863 tempest-MultipleCreateTestJSON-1589015462 tempest-MultipleCreateTestJSON-1589015462-project-member] Acquiring lock "ef502a7d-7dc0-4319-8ce3-6a8701288237" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 839.597647] env[61806]: DEBUG oslo_concurrency.lockutils [None req-182a714a-9a7d-40bc-ad59-30e74e8a3863 tempest-MultipleCreateTestJSON-1589015462 tempest-MultipleCreateTestJSON-1589015462-project-member] Lock "ef502a7d-7dc0-4319-8ce3-6a8701288237" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 839.597647] env[61806]: DEBUG oslo_concurrency.lockutils [None req-182a714a-9a7d-40bc-ad59-30e74e8a3863 tempest-MultipleCreateTestJSON-1589015462 tempest-MultipleCreateTestJSON-1589015462-project-member] Acquiring lock "ef502a7d-7dc0-4319-8ce3-6a8701288237-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 839.597869] env[61806]: DEBUG oslo_concurrency.lockutils [None req-182a714a-9a7d-40bc-ad59-30e74e8a3863 tempest-MultipleCreateTestJSON-1589015462 tempest-MultipleCreateTestJSON-1589015462-project-member] Lock "ef502a7d-7dc0-4319-8ce3-6a8701288237-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 839.598150] env[61806]: DEBUG oslo_concurrency.lockutils [None req-182a714a-9a7d-40bc-ad59-30e74e8a3863 tempest-MultipleCreateTestJSON-1589015462 tempest-MultipleCreateTestJSON-1589015462-project-member] Lock "ef502a7d-7dc0-4319-8ce3-6a8701288237-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 839.600938] env[61806]: INFO nova.compute.manager [None req-182a714a-9a7d-40bc-ad59-30e74e8a3863 tempest-MultipleCreateTestJSON-1589015462 tempest-MultipleCreateTestJSON-1589015462-project-member] [instance: ef502a7d-7dc0-4319-8ce3-6a8701288237] Terminating instance [ 839.602856] env[61806]: DEBUG nova.compute.manager [None req-182a714a-9a7d-40bc-ad59-30e74e8a3863 tempest-MultipleCreateTestJSON-1589015462 tempest-MultipleCreateTestJSON-1589015462-project-member] [instance: ef502a7d-7dc0-4319-8ce3-6a8701288237] Start destroying the instance on the hypervisor. {{(pid=61806) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 839.603136] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-182a714a-9a7d-40bc-ad59-30e74e8a3863 tempest-MultipleCreateTestJSON-1589015462 tempest-MultipleCreateTestJSON-1589015462-project-member] [instance: ef502a7d-7dc0-4319-8ce3-6a8701288237] Destroying instance {{(pid=61806) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 839.603995] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9f0ee143-4ec2-4a06-8036-1f55f2735d50 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 839.611816] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [None req-182a714a-9a7d-40bc-ad59-30e74e8a3863 tempest-MultipleCreateTestJSON-1589015462 tempest-MultipleCreateTestJSON-1589015462-project-member] [instance: ef502a7d-7dc0-4319-8ce3-6a8701288237] Powering off the VM {{(pid=61806) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 839.612088] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-6fa94101-b67c-49c3-ae47-a1fb0fd1f4d1 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 839.616443] env[61806]: DEBUG oslo_concurrency.lockutils [None req-b326e3ef-f103-433e-9385-2b5e433b613a tempest-ServerShowV247Test-1218443059 tempest-ServerShowV247Test-1218443059-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.876s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 839.619593] env[61806]: DEBUG oslo_concurrency.lockutils [None req-4b919e4f-1f0b-40e0-b482-0547e22d55ff tempest-ImagesTestJSON-618852556 tempest-ImagesTestJSON-618852556-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 14.104s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 839.621185] env[61806]: INFO nova.compute.claims [None req-4b919e4f-1f0b-40e0-b482-0547e22d55ff tempest-ImagesTestJSON-618852556 tempest-ImagesTestJSON-618852556-project-member] [instance: 28e0baab-8516-42e3-acc2-9b8eb5192f57] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 839.623815] env[61806]: DEBUG oslo_vmware.api [None req-182a714a-9a7d-40bc-ad59-30e74e8a3863 tempest-MultipleCreateTestJSON-1589015462 tempest-MultipleCreateTestJSON-1589015462-project-member] Waiting for the task: (returnval){ [ 839.623815] env[61806]: value = "task-1294695" [ 839.623815] env[61806]: _type = "Task" [ 839.623815] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 839.632493] env[61806]: DEBUG oslo_vmware.api [None req-182a714a-9a7d-40bc-ad59-30e74e8a3863 tempest-MultipleCreateTestJSON-1589015462 tempest-MultipleCreateTestJSON-1589015462-project-member] Task: {'id': task-1294695, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 839.635052] env[61806]: DEBUG nova.network.neutron [-] [instance: b41531f2-f28c-4d82-9682-0b557bbaa491] Updating instance_info_cache with network_info: [] {{(pid=61806) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 839.646069] env[61806]: INFO nova.scheduler.client.report [None req-b326e3ef-f103-433e-9385-2b5e433b613a tempest-ServerShowV247Test-1218443059 tempest-ServerShowV247Test-1218443059-project-member] Deleted allocations for instance 4febd093-9f2b-494e-b175-e4693b0e3e0d [ 839.653746] env[61806]: DEBUG oslo_vmware.api [None req-0ea3a38f-3871-4412-a6bc-d21c2086d262 tempest-AttachInterfacesV270Test-1974652806 tempest-AttachInterfacesV270Test-1974652806-project-member] Task: {'id': task-1294691, 'name': ReconfigVM_Task, 'duration_secs': 1.149213} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 839.654270] env[61806]: DEBUG nova.virt.vmwareapi.volumeops [None req-0ea3a38f-3871-4412-a6bc-d21c2086d262 tempest-AttachInterfacesV270Test-1974652806 tempest-AttachInterfacesV270Test-1974652806-project-member] [instance: 117139cd-4ed4-4bea-8ad3-f4ac32c692e4] Reconfigured VM instance instance-00000044 to attach disk [datastore2] 117139cd-4ed4-4bea-8ad3-f4ac32c692e4/117139cd-4ed4-4bea-8ad3-f4ac32c692e4.vmdk or device None with type sparse {{(pid=61806) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 839.654860] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-1fc3275d-fd99-44fc-9600-be4259ee663c {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 839.660904] env[61806]: DEBUG oslo_vmware.api [None req-0ea3a38f-3871-4412-a6bc-d21c2086d262 tempest-AttachInterfacesV270Test-1974652806 tempest-AttachInterfacesV270Test-1974652806-project-member] Waiting for the task: (returnval){ [ 839.660904] env[61806]: value = "task-1294696" [ 839.660904] env[61806]: _type = "Task" [ 839.660904] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 839.669343] env[61806]: DEBUG oslo_vmware.api [None req-0ea3a38f-3871-4412-a6bc-d21c2086d262 tempest-AttachInterfacesV270Test-1974652806 tempest-AttachInterfacesV270Test-1974652806-project-member] Task: {'id': task-1294696, 'name': Rename_Task} progress is 5%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 839.674375] env[61806]: DEBUG nova.compute.manager [req-479b5f30-2427-45dd-b202-9ce1881e7e0c req-ba76e964-0672-4dba-ac99-8feb8ca36a2d service nova] [instance: be19fc44-ecbe-489c-9b6e-be8957bb1dd0] Received event network-changed-6b41a0ca-4d2e-4ab1-9893-180a2f0febca {{(pid=61806) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 839.674375] env[61806]: DEBUG nova.compute.manager [req-479b5f30-2427-45dd-b202-9ce1881e7e0c req-ba76e964-0672-4dba-ac99-8feb8ca36a2d service nova] [instance: be19fc44-ecbe-489c-9b6e-be8957bb1dd0] Refreshing instance network info cache due to event network-changed-6b41a0ca-4d2e-4ab1-9893-180a2f0febca. {{(pid=61806) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 839.674661] env[61806]: DEBUG oslo_concurrency.lockutils [req-479b5f30-2427-45dd-b202-9ce1881e7e0c req-ba76e964-0672-4dba-ac99-8feb8ca36a2d service nova] Acquiring lock "refresh_cache-be19fc44-ecbe-489c-9b6e-be8957bb1dd0" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 839.674661] env[61806]: DEBUG oslo_concurrency.lockutils [req-479b5f30-2427-45dd-b202-9ce1881e7e0c req-ba76e964-0672-4dba-ac99-8feb8ca36a2d service nova] Acquired lock "refresh_cache-be19fc44-ecbe-489c-9b6e-be8957bb1dd0" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 839.674742] env[61806]: DEBUG nova.network.neutron [req-479b5f30-2427-45dd-b202-9ce1881e7e0c req-ba76e964-0672-4dba-ac99-8feb8ca36a2d service nova] [instance: be19fc44-ecbe-489c-9b6e-be8957bb1dd0] Refreshing network info cache for port 6b41a0ca-4d2e-4ab1-9893-180a2f0febca {{(pid=61806) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 839.729610] env[61806]: DEBUG oslo_vmware.api [None req-4e0c8443-0fd4-459b-a4c5-7cf373aa1089 tempest-ServerAddressesNegativeTestJSON-1526246390 tempest-ServerAddressesNegativeTestJSON-1526246390-project-member] Task: {'id': task-1294693, 'name': ReconfigVM_Task, 'duration_secs': 0.39408} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 839.729938] env[61806]: DEBUG nova.virt.vmwareapi.volumeops [None req-4e0c8443-0fd4-459b-a4c5-7cf373aa1089 tempest-ServerAddressesNegativeTestJSON-1526246390 tempest-ServerAddressesNegativeTestJSON-1526246390-project-member] [instance: c8033551-6591-4e37-ae78-4efe7145b10b] Reconfigured VM instance instance-00000045 to attach disk [datastore2] c8033551-6591-4e37-ae78-4efe7145b10b/c8033551-6591-4e37-ae78-4efe7145b10b.vmdk or device None with type sparse {{(pid=61806) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 839.730962] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-0c75f199-97ae-4339-8cb5-9f5ee01fde9f {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 839.737270] env[61806]: DEBUG oslo_vmware.api [None req-4e0c8443-0fd4-459b-a4c5-7cf373aa1089 tempest-ServerAddressesNegativeTestJSON-1526246390 tempest-ServerAddressesNegativeTestJSON-1526246390-project-member] Waiting for the task: (returnval){ [ 839.737270] env[61806]: value = "task-1294697" [ 839.737270] env[61806]: _type = "Task" [ 839.737270] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 839.746614] env[61806]: DEBUG oslo_vmware.api [None req-4e0c8443-0fd4-459b-a4c5-7cf373aa1089 tempest-ServerAddressesNegativeTestJSON-1526246390 tempest-ServerAddressesNegativeTestJSON-1526246390-project-member] Task: {'id': task-1294697, 'name': Rename_Task} progress is 5%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 839.783449] env[61806]: DEBUG oslo_vmware.api [None req-c9a1ae7a-8313-4bdf-8de4-77929ce2d9bc tempest-ServerMetadataNegativeTestJSON-456453902 tempest-ServerMetadataNegativeTestJSON-456453902-project-member] Task: {'id': session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]52f8adf4-9b7c-a76a-d7cd-08041becb336, 'name': SearchDatastore_Task, 'duration_secs': 0.014886} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 839.786990] env[61806]: DEBUG oslo_concurrency.lockutils [None req-c9a1ae7a-8313-4bdf-8de4-77929ce2d9bc tempest-ServerMetadataNegativeTestJSON-456453902 tempest-ServerMetadataNegativeTestJSON-456453902-project-member] Releasing lock "[datastore2] devstack-image-cache_base/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa.vmdk" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 839.787345] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [None req-c9a1ae7a-8313-4bdf-8de4-77929ce2d9bc tempest-ServerMetadataNegativeTestJSON-456453902 tempest-ServerMetadataNegativeTestJSON-456453902-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa.vmdk to [datastore2] 7f4907d0-178d-452a-8149-030becde8779/7f4907d0-178d-452a-8149-030becde8779.vmdk {{(pid=61806) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 839.787681] env[61806]: DEBUG oslo_vmware.api [None req-3bd56cbd-68df-4151-a169-676f8bdda9ab tempest-MultipleCreateTestJSON-1589015462 tempest-MultipleCreateTestJSON-1589015462-project-member] Task: {'id': task-1294694, 'name': PowerOffVM_Task, 'duration_secs': 0.296121} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 839.787926] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-ef8ca321-58e3-4cc5-a6b7-e544a930c46c {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 839.790037] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [None req-3bd56cbd-68df-4151-a169-676f8bdda9ab tempest-MultipleCreateTestJSON-1589015462 tempest-MultipleCreateTestJSON-1589015462-project-member] [instance: 25218e83-c1ab-4b97-ade1-7c32b6f99305] Powered off the VM {{(pid=61806) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 839.790226] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-3bd56cbd-68df-4151-a169-676f8bdda9ab tempest-MultipleCreateTestJSON-1589015462 tempest-MultipleCreateTestJSON-1589015462-project-member] [instance: 25218e83-c1ab-4b97-ade1-7c32b6f99305] Unregistering the VM {{(pid=61806) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 839.790522] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-158c60d2-38be-45fb-b64f-92531a4fa555 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 839.797614] env[61806]: DEBUG oslo_vmware.api [None req-c9a1ae7a-8313-4bdf-8de4-77929ce2d9bc tempest-ServerMetadataNegativeTestJSON-456453902 tempest-ServerMetadataNegativeTestJSON-456453902-project-member] Waiting for the task: (returnval){ [ 839.797614] env[61806]: value = "task-1294698" [ 839.797614] env[61806]: _type = "Task" [ 839.797614] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 839.805097] env[61806]: DEBUG oslo_vmware.api [None req-c9a1ae7a-8313-4bdf-8de4-77929ce2d9bc tempest-ServerMetadataNegativeTestJSON-456453902 tempest-ServerMetadataNegativeTestJSON-456453902-project-member] Task: {'id': task-1294698, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 839.939061] env[61806]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-68fba25e-ecce-42b2-9e8c-a1e99a3a4aa8 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 839.948435] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-074af227-414f-4c00-8106-17cbc2109029 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 839.977107] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-3bd56cbd-68df-4151-a169-676f8bdda9ab tempest-MultipleCreateTestJSON-1589015462 tempest-MultipleCreateTestJSON-1589015462-project-member] [instance: 25218e83-c1ab-4b97-ade1-7c32b6f99305] Unregistered the VM {{(pid=61806) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 839.977415] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-3bd56cbd-68df-4151-a169-676f8bdda9ab tempest-MultipleCreateTestJSON-1589015462 tempest-MultipleCreateTestJSON-1589015462-project-member] [instance: 25218e83-c1ab-4b97-ade1-7c32b6f99305] Deleting contents of the VM from datastore datastore2 {{(pid=61806) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 839.977623] env[61806]: DEBUG nova.virt.vmwareapi.ds_util [None req-3bd56cbd-68df-4151-a169-676f8bdda9ab tempest-MultipleCreateTestJSON-1589015462 tempest-MultipleCreateTestJSON-1589015462-project-member] Deleting the datastore file [datastore2] 25218e83-c1ab-4b97-ade1-7c32b6f99305 {{(pid=61806) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 839.988775] env[61806]: DEBUG oslo_concurrency.lockutils [None req-f9d7b4b0-fa22-4fc5-86c2-688ab650f83e tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] Acquiring lock "interface-2a3f81e3-b84f-4370-a598-2f6a607c3d47-None" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 839.989051] env[61806]: DEBUG oslo_concurrency.lockutils [None req-f9d7b4b0-fa22-4fc5-86c2-688ab650f83e tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] Lock "interface-2a3f81e3-b84f-4370-a598-2f6a607c3d47-None" acquired by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: waited 0.000s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 839.989378] env[61806]: DEBUG nova.objects.instance [None req-f9d7b4b0-fa22-4fc5-86c2-688ab650f83e tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] Lazy-loading 'flavor' on Instance uuid 2a3f81e3-b84f-4370-a598-2f6a607c3d47 {{(pid=61806) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 839.990888] env[61806]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-28113028-e66c-43d2-a205-d30a70500760 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 839.992683] env[61806]: DEBUG nova.compute.manager [req-df79f25b-0653-4e77-adbc-0a1400e0bed4 req-8b5baad3-0914-4550-a472-aa72efb1d4dc service nova] [instance: b41531f2-f28c-4d82-9682-0b557bbaa491] Detach interface failed, port_id=513ca4b5-b62c-486a-9860-cf267a15c3e5, reason: Instance b41531f2-f28c-4d82-9682-0b557bbaa491 could not be found. {{(pid=61806) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10963}} [ 839.998327] env[61806]: DEBUG oslo_vmware.api [None req-3bd56cbd-68df-4151-a169-676f8bdda9ab tempest-MultipleCreateTestJSON-1589015462 tempest-MultipleCreateTestJSON-1589015462-project-member] Waiting for the task: (returnval){ [ 839.998327] env[61806]: value = "task-1294700" [ 839.998327] env[61806]: _type = "Task" [ 839.998327] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 840.007125] env[61806]: DEBUG oslo_vmware.api [None req-3bd56cbd-68df-4151-a169-676f8bdda9ab tempest-MultipleCreateTestJSON-1589015462 tempest-MultipleCreateTestJSON-1589015462-project-member] Task: {'id': task-1294700, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 840.012524] env[61806]: DEBUG nova.network.neutron [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] [instance: e0ef0a35-82a5-495b-9d5c-5805e8306390] Updating instance_info_cache with network_info: [{"id": "c42059da-e422-46df-be37-f9a0816b97c8", "address": "fa:16:3e:60:a9:54", "network": {"id": "288ebc87-a30f-439b-9885-9cdda68a9591", "bridge": "br-int", "label": "tempest-ImagesTestJSON-914411285-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4b9aa91aff1d4008ac5096902b77f852", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8cbc9b8f-ce19-4262-bf4d-88cd4f259a1c", "external-id": "nsx-vlan-transportzone-630", "segmentation_id": 630, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc42059da-e4", "ovs_interfaceid": "c42059da-e422-46df-be37-f9a0816b97c8", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61806) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 840.014036] env[61806]: DEBUG oslo_concurrency.lockutils [None req-334c6338-dc9d-4c32-ad77-bd3fe007216c tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Acquiring lock "refresh_cache-be19fc44-ecbe-489c-9b6e-be8957bb1dd0" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 840.139065] env[61806]: INFO nova.compute.manager [-] [instance: b41531f2-f28c-4d82-9682-0b557bbaa491] Took 1.26 seconds to deallocate network for instance. [ 840.139660] env[61806]: DEBUG oslo_vmware.api [None req-182a714a-9a7d-40bc-ad59-30e74e8a3863 tempest-MultipleCreateTestJSON-1589015462 tempest-MultipleCreateTestJSON-1589015462-project-member] Task: {'id': task-1294695, 'name': PowerOffVM_Task, 'duration_secs': 0.234993} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 840.141284] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [None req-182a714a-9a7d-40bc-ad59-30e74e8a3863 tempest-MultipleCreateTestJSON-1589015462 tempest-MultipleCreateTestJSON-1589015462-project-member] [instance: ef502a7d-7dc0-4319-8ce3-6a8701288237] Powered off the VM {{(pid=61806) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 840.141463] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-182a714a-9a7d-40bc-ad59-30e74e8a3863 tempest-MultipleCreateTestJSON-1589015462 tempest-MultipleCreateTestJSON-1589015462-project-member] [instance: ef502a7d-7dc0-4319-8ce3-6a8701288237] Unregistering the VM {{(pid=61806) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 840.141973] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-dabbb904-9f87-4625-82a4-511125054d7e {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 840.157677] env[61806]: DEBUG oslo_concurrency.lockutils [None req-b326e3ef-f103-433e-9385-2b5e433b613a tempest-ServerShowV247Test-1218443059 tempest-ServerShowV247Test-1218443059-project-member] Lock "4febd093-9f2b-494e-b175-e4693b0e3e0d" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 19.318s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 840.174753] env[61806]: DEBUG oslo_vmware.api [None req-0ea3a38f-3871-4412-a6bc-d21c2086d262 tempest-AttachInterfacesV270Test-1974652806 tempest-AttachInterfacesV270Test-1974652806-project-member] Task: {'id': task-1294696, 'name': Rename_Task, 'duration_secs': 0.258217} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 840.175255] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [None req-0ea3a38f-3871-4412-a6bc-d21c2086d262 tempest-AttachInterfacesV270Test-1974652806 tempest-AttachInterfacesV270Test-1974652806-project-member] [instance: 117139cd-4ed4-4bea-8ad3-f4ac32c692e4] Powering on the VM {{(pid=61806) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 840.176341] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-49ccc6b0-5487-476a-8763-09cc8286b301 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 840.187034] env[61806]: DEBUG oslo_vmware.api [None req-0ea3a38f-3871-4412-a6bc-d21c2086d262 tempest-AttachInterfacesV270Test-1974652806 tempest-AttachInterfacesV270Test-1974652806-project-member] Waiting for the task: (returnval){ [ 840.187034] env[61806]: value = "task-1294702" [ 840.187034] env[61806]: _type = "Task" [ 840.187034] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 840.198488] env[61806]: DEBUG oslo_vmware.api [None req-0ea3a38f-3871-4412-a6bc-d21c2086d262 tempest-AttachInterfacesV270Test-1974652806 tempest-AttachInterfacesV270Test-1974652806-project-member] Task: {'id': task-1294702, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 840.225141] env[61806]: DEBUG nova.network.neutron [req-479b5f30-2427-45dd-b202-9ce1881e7e0c req-ba76e964-0672-4dba-ac99-8feb8ca36a2d service nova] [instance: be19fc44-ecbe-489c-9b6e-be8957bb1dd0] Instance cache missing network info. {{(pid=61806) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 840.228598] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-182a714a-9a7d-40bc-ad59-30e74e8a3863 tempest-MultipleCreateTestJSON-1589015462 tempest-MultipleCreateTestJSON-1589015462-project-member] [instance: ef502a7d-7dc0-4319-8ce3-6a8701288237] Unregistered the VM {{(pid=61806) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 840.228824] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-182a714a-9a7d-40bc-ad59-30e74e8a3863 tempest-MultipleCreateTestJSON-1589015462 tempest-MultipleCreateTestJSON-1589015462-project-member] [instance: ef502a7d-7dc0-4319-8ce3-6a8701288237] Deleting contents of the VM from datastore datastore2 {{(pid=61806) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 840.229021] env[61806]: DEBUG nova.virt.vmwareapi.ds_util [None req-182a714a-9a7d-40bc-ad59-30e74e8a3863 tempest-MultipleCreateTestJSON-1589015462 tempest-MultipleCreateTestJSON-1589015462-project-member] Deleting the datastore file [datastore2] ef502a7d-7dc0-4319-8ce3-6a8701288237 {{(pid=61806) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 840.229297] env[61806]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-20ec9a8d-8660-4ec3-9484-e17963f921a9 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 840.235777] env[61806]: DEBUG oslo_vmware.api [None req-182a714a-9a7d-40bc-ad59-30e74e8a3863 tempest-MultipleCreateTestJSON-1589015462 tempest-MultipleCreateTestJSON-1589015462-project-member] Waiting for the task: (returnval){ [ 840.235777] env[61806]: value = "task-1294703" [ 840.235777] env[61806]: _type = "Task" [ 840.235777] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 840.248448] env[61806]: DEBUG oslo_vmware.api [None req-182a714a-9a7d-40bc-ad59-30e74e8a3863 tempest-MultipleCreateTestJSON-1589015462 tempest-MultipleCreateTestJSON-1589015462-project-member] Task: {'id': task-1294703, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 840.251915] env[61806]: DEBUG oslo_vmware.api [None req-4e0c8443-0fd4-459b-a4c5-7cf373aa1089 tempest-ServerAddressesNegativeTestJSON-1526246390 tempest-ServerAddressesNegativeTestJSON-1526246390-project-member] Task: {'id': task-1294697, 'name': Rename_Task, 'duration_secs': 0.178869} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 840.252225] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [None req-4e0c8443-0fd4-459b-a4c5-7cf373aa1089 tempest-ServerAddressesNegativeTestJSON-1526246390 tempest-ServerAddressesNegativeTestJSON-1526246390-project-member] [instance: c8033551-6591-4e37-ae78-4efe7145b10b] Powering on the VM {{(pid=61806) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 840.252493] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-99436ea6-2b8d-490a-bff2-5dd425d8966e {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 840.260447] env[61806]: DEBUG oslo_vmware.api [None req-4e0c8443-0fd4-459b-a4c5-7cf373aa1089 tempest-ServerAddressesNegativeTestJSON-1526246390 tempest-ServerAddressesNegativeTestJSON-1526246390-project-member] Waiting for the task: (returnval){ [ 840.260447] env[61806]: value = "task-1294704" [ 840.260447] env[61806]: _type = "Task" [ 840.260447] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 840.269919] env[61806]: DEBUG oslo_vmware.api [None req-4e0c8443-0fd4-459b-a4c5-7cf373aa1089 tempest-ServerAddressesNegativeTestJSON-1526246390 tempest-ServerAddressesNegativeTestJSON-1526246390-project-member] Task: {'id': task-1294704, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 840.312592] env[61806]: DEBUG oslo_vmware.api [None req-c9a1ae7a-8313-4bdf-8de4-77929ce2d9bc tempest-ServerMetadataNegativeTestJSON-456453902 tempest-ServerMetadataNegativeTestJSON-456453902-project-member] Task: {'id': task-1294698, 'name': CopyVirtualDisk_Task} progress is 25%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 840.359884] env[61806]: DEBUG nova.network.neutron [req-479b5f30-2427-45dd-b202-9ce1881e7e0c req-ba76e964-0672-4dba-ac99-8feb8ca36a2d service nova] [instance: be19fc44-ecbe-489c-9b6e-be8957bb1dd0] Updating instance_info_cache with network_info: [] {{(pid=61806) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 840.508049] env[61806]: DEBUG oslo_vmware.api [None req-3bd56cbd-68df-4151-a169-676f8bdda9ab tempest-MultipleCreateTestJSON-1589015462 tempest-MultipleCreateTestJSON-1589015462-project-member] Task: {'id': task-1294700, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.499435} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 840.508049] env[61806]: DEBUG nova.virt.vmwareapi.ds_util [None req-3bd56cbd-68df-4151-a169-676f8bdda9ab tempest-MultipleCreateTestJSON-1589015462 tempest-MultipleCreateTestJSON-1589015462-project-member] Deleted the datastore file {{(pid=61806) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 840.508250] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-3bd56cbd-68df-4151-a169-676f8bdda9ab tempest-MultipleCreateTestJSON-1589015462 tempest-MultipleCreateTestJSON-1589015462-project-member] [instance: 25218e83-c1ab-4b97-ade1-7c32b6f99305] Deleted contents of the VM from datastore datastore2 {{(pid=61806) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 840.508352] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-3bd56cbd-68df-4151-a169-676f8bdda9ab tempest-MultipleCreateTestJSON-1589015462 tempest-MultipleCreateTestJSON-1589015462-project-member] [instance: 25218e83-c1ab-4b97-ade1-7c32b6f99305] Instance destroyed {{(pid=61806) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 840.508569] env[61806]: INFO nova.compute.manager [None req-3bd56cbd-68df-4151-a169-676f8bdda9ab tempest-MultipleCreateTestJSON-1589015462 tempest-MultipleCreateTestJSON-1589015462-project-member] [instance: 25218e83-c1ab-4b97-ade1-7c32b6f99305] Took 1.26 seconds to destroy the instance on the hypervisor. [ 840.508852] env[61806]: DEBUG oslo.service.loopingcall [None req-3bd56cbd-68df-4151-a169-676f8bdda9ab tempest-MultipleCreateTestJSON-1589015462 tempest-MultipleCreateTestJSON-1589015462-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61806) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 840.509103] env[61806]: DEBUG nova.compute.manager [-] [instance: 25218e83-c1ab-4b97-ade1-7c32b6f99305] Deallocating network for instance {{(pid=61806) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 840.509210] env[61806]: DEBUG nova.network.neutron [-] [instance: 25218e83-c1ab-4b97-ade1-7c32b6f99305] deallocate_for_instance() {{(pid=61806) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 840.515622] env[61806]: DEBUG oslo_concurrency.lockutils [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Releasing lock "refresh_cache-e0ef0a35-82a5-495b-9d5c-5805e8306390" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 840.515817] env[61806]: DEBUG nova.compute.manager [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] [instance: e0ef0a35-82a5-495b-9d5c-5805e8306390] Updated the network info_cache for instance {{(pid=61806) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9997}} [ 840.516034] env[61806]: DEBUG oslo_service.periodic_task [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=61806) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 840.516225] env[61806]: DEBUG oslo_service.periodic_task [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=61806) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 840.516409] env[61806]: DEBUG oslo_service.periodic_task [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=61806) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 840.516569] env[61806]: DEBUG oslo_service.periodic_task [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=61806) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 840.516718] env[61806]: DEBUG oslo_service.periodic_task [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=61806) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 840.516888] env[61806]: DEBUG oslo_service.periodic_task [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=61806) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 840.517034] env[61806]: DEBUG nova.compute.manager [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=61806) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10545}} [ 840.517203] env[61806]: DEBUG oslo_service.periodic_task [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Running periodic task ComputeManager.update_available_resource {{(pid=61806) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 840.615810] env[61806]: DEBUG nova.objects.instance [None req-f9d7b4b0-fa22-4fc5-86c2-688ab650f83e tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] Lazy-loading 'pci_requests' on Instance uuid 2a3f81e3-b84f-4370-a598-2f6a607c3d47 {{(pid=61806) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 840.622231] env[61806]: DEBUG oslo_concurrency.lockutils [None req-81aa2ada-298c-4e33-a594-b57f3ac492a8 tempest-ServerShowV247Test-1218443059 tempest-ServerShowV247Test-1218443059-project-member] Acquiring lock "7c3fd2ff-ebd5-454f-a743-7fbae1088941" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 840.622494] env[61806]: DEBUG oslo_concurrency.lockutils [None req-81aa2ada-298c-4e33-a594-b57f3ac492a8 tempest-ServerShowV247Test-1218443059 tempest-ServerShowV247Test-1218443059-project-member] Lock "7c3fd2ff-ebd5-454f-a743-7fbae1088941" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 840.622791] env[61806]: DEBUG oslo_concurrency.lockutils [None req-81aa2ada-298c-4e33-a594-b57f3ac492a8 tempest-ServerShowV247Test-1218443059 tempest-ServerShowV247Test-1218443059-project-member] Acquiring lock "7c3fd2ff-ebd5-454f-a743-7fbae1088941-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 840.622997] env[61806]: DEBUG oslo_concurrency.lockutils [None req-81aa2ada-298c-4e33-a594-b57f3ac492a8 tempest-ServerShowV247Test-1218443059 tempest-ServerShowV247Test-1218443059-project-member] Lock "7c3fd2ff-ebd5-454f-a743-7fbae1088941-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 840.623189] env[61806]: DEBUG oslo_concurrency.lockutils [None req-81aa2ada-298c-4e33-a594-b57f3ac492a8 tempest-ServerShowV247Test-1218443059 tempest-ServerShowV247Test-1218443059-project-member] Lock "7c3fd2ff-ebd5-454f-a743-7fbae1088941-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 840.625982] env[61806]: INFO nova.compute.manager [None req-81aa2ada-298c-4e33-a594-b57f3ac492a8 tempest-ServerShowV247Test-1218443059 tempest-ServerShowV247Test-1218443059-project-member] [instance: 7c3fd2ff-ebd5-454f-a743-7fbae1088941] Terminating instance [ 840.631126] env[61806]: DEBUG oslo_concurrency.lockutils [None req-81aa2ada-298c-4e33-a594-b57f3ac492a8 tempest-ServerShowV247Test-1218443059 tempest-ServerShowV247Test-1218443059-project-member] Acquiring lock "refresh_cache-7c3fd2ff-ebd5-454f-a743-7fbae1088941" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 840.631368] env[61806]: DEBUG oslo_concurrency.lockutils [None req-81aa2ada-298c-4e33-a594-b57f3ac492a8 tempest-ServerShowV247Test-1218443059 tempest-ServerShowV247Test-1218443059-project-member] Acquired lock "refresh_cache-7c3fd2ff-ebd5-454f-a743-7fbae1088941" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 840.631736] env[61806]: DEBUG nova.network.neutron [None req-81aa2ada-298c-4e33-a594-b57f3ac492a8 tempest-ServerShowV247Test-1218443059 tempest-ServerShowV247Test-1218443059-project-member] [instance: 7c3fd2ff-ebd5-454f-a743-7fbae1088941] Building network info cache for instance {{(pid=61806) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 840.648508] env[61806]: DEBUG oslo_concurrency.lockutils [None req-fd7f690e-61c7-4f8c-a053-2b904c459e5b tempest-SecurityGroupsTestJSON-1213485581 tempest-SecurityGroupsTestJSON-1213485581-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 840.700667] env[61806]: DEBUG oslo_vmware.api [None req-0ea3a38f-3871-4412-a6bc-d21c2086d262 tempest-AttachInterfacesV270Test-1974652806 tempest-AttachInterfacesV270Test-1974652806-project-member] Task: {'id': task-1294702, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 840.750121] env[61806]: DEBUG oslo_vmware.api [None req-182a714a-9a7d-40bc-ad59-30e74e8a3863 tempest-MultipleCreateTestJSON-1589015462 tempest-MultipleCreateTestJSON-1589015462-project-member] Task: {'id': task-1294703, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.407002} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 840.750121] env[61806]: DEBUG nova.virt.vmwareapi.ds_util [None req-182a714a-9a7d-40bc-ad59-30e74e8a3863 tempest-MultipleCreateTestJSON-1589015462 tempest-MultipleCreateTestJSON-1589015462-project-member] Deleted the datastore file {{(pid=61806) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 840.750200] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-182a714a-9a7d-40bc-ad59-30e74e8a3863 tempest-MultipleCreateTestJSON-1589015462 tempest-MultipleCreateTestJSON-1589015462-project-member] [instance: ef502a7d-7dc0-4319-8ce3-6a8701288237] Deleted contents of the VM from datastore datastore2 {{(pid=61806) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 840.751805] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-182a714a-9a7d-40bc-ad59-30e74e8a3863 tempest-MultipleCreateTestJSON-1589015462 tempest-MultipleCreateTestJSON-1589015462-project-member] [instance: ef502a7d-7dc0-4319-8ce3-6a8701288237] Instance destroyed {{(pid=61806) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 840.751805] env[61806]: INFO nova.compute.manager [None req-182a714a-9a7d-40bc-ad59-30e74e8a3863 tempest-MultipleCreateTestJSON-1589015462 tempest-MultipleCreateTestJSON-1589015462-project-member] [instance: ef502a7d-7dc0-4319-8ce3-6a8701288237] Took 1.15 seconds to destroy the instance on the hypervisor. [ 840.751805] env[61806]: DEBUG oslo.service.loopingcall [None req-182a714a-9a7d-40bc-ad59-30e74e8a3863 tempest-MultipleCreateTestJSON-1589015462 tempest-MultipleCreateTestJSON-1589015462-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61806) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 840.755246] env[61806]: DEBUG nova.compute.manager [-] [instance: ef502a7d-7dc0-4319-8ce3-6a8701288237] Deallocating network for instance {{(pid=61806) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 840.755882] env[61806]: DEBUG nova.network.neutron [-] [instance: ef502a7d-7dc0-4319-8ce3-6a8701288237] deallocate_for_instance() {{(pid=61806) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 840.772585] env[61806]: DEBUG oslo_vmware.api [None req-4e0c8443-0fd4-459b-a4c5-7cf373aa1089 tempest-ServerAddressesNegativeTestJSON-1526246390 tempest-ServerAddressesNegativeTestJSON-1526246390-project-member] Task: {'id': task-1294704, 'name': PowerOnVM_Task} progress is 89%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 840.811799] env[61806]: DEBUG oslo_vmware.api [None req-c9a1ae7a-8313-4bdf-8de4-77929ce2d9bc tempest-ServerMetadataNegativeTestJSON-456453902 tempest-ServerMetadataNegativeTestJSON-456453902-project-member] Task: {'id': task-1294698, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.694155} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 840.814502] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [None req-c9a1ae7a-8313-4bdf-8de4-77929ce2d9bc tempest-ServerMetadataNegativeTestJSON-456453902 tempest-ServerMetadataNegativeTestJSON-456453902-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa.vmdk to [datastore2] 7f4907d0-178d-452a-8149-030becde8779/7f4907d0-178d-452a-8149-030becde8779.vmdk {{(pid=61806) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 840.814748] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-c9a1ae7a-8313-4bdf-8de4-77929ce2d9bc tempest-ServerMetadataNegativeTestJSON-456453902 tempest-ServerMetadataNegativeTestJSON-456453902-project-member] [instance: 7f4907d0-178d-452a-8149-030becde8779] Extending root virtual disk to 1048576 {{(pid=61806) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 840.815953] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-2621e3fa-3b40-4f3a-9b68-48464b2d7981 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 840.822881] env[61806]: DEBUG oslo_vmware.api [None req-c9a1ae7a-8313-4bdf-8de4-77929ce2d9bc tempest-ServerMetadataNegativeTestJSON-456453902 tempest-ServerMetadataNegativeTestJSON-456453902-project-member] Waiting for the task: (returnval){ [ 840.822881] env[61806]: value = "task-1294705" [ 840.822881] env[61806]: _type = "Task" [ 840.822881] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 840.835683] env[61806]: DEBUG oslo_vmware.api [None req-c9a1ae7a-8313-4bdf-8de4-77929ce2d9bc tempest-ServerMetadataNegativeTestJSON-456453902 tempest-ServerMetadataNegativeTestJSON-456453902-project-member] Task: {'id': task-1294705, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 840.863499] env[61806]: DEBUG oslo_concurrency.lockutils [req-479b5f30-2427-45dd-b202-9ce1881e7e0c req-ba76e964-0672-4dba-ac99-8feb8ca36a2d service nova] Releasing lock "refresh_cache-be19fc44-ecbe-489c-9b6e-be8957bb1dd0" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 840.864090] env[61806]: DEBUG oslo_concurrency.lockutils [None req-334c6338-dc9d-4c32-ad77-bd3fe007216c tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Acquired lock "refresh_cache-be19fc44-ecbe-489c-9b6e-be8957bb1dd0" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 840.864286] env[61806]: DEBUG nova.network.neutron [None req-334c6338-dc9d-4c32-ad77-bd3fe007216c tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] [instance: be19fc44-ecbe-489c-9b6e-be8957bb1dd0] Building network info cache for instance {{(pid=61806) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 840.947920] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9c62c79a-3551-4e48-b960-1df4dc252053 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 840.956112] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ef29c6b9-36e3-45f8-a109-70d12b6eeb5f {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 840.989878] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2b2880e7-96f6-4cad-958d-6615c29a24b8 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 840.997728] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dfe6ee46-5a7d-469b-84c6-5ba3135cd8ea {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 841.015608] env[61806]: DEBUG nova.compute.provider_tree [None req-4b919e4f-1f0b-40e0-b482-0547e22d55ff tempest-ImagesTestJSON-618852556 tempest-ImagesTestJSON-618852556-project-member] Inventory has not changed in ProviderTree for provider: a2858be1-fd22-4e08-979e-87ad25293407 {{(pid=61806) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 841.022541] env[61806]: DEBUG oslo_concurrency.lockutils [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 841.118632] env[61806]: DEBUG nova.objects.base [None req-f9d7b4b0-fa22-4fc5-86c2-688ab650f83e tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] Object Instance<2a3f81e3-b84f-4370-a598-2f6a607c3d47> lazy-loaded attributes: flavor,pci_requests {{(pid=61806) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 841.118907] env[61806]: DEBUG nova.network.neutron [None req-f9d7b4b0-fa22-4fc5-86c2-688ab650f83e tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] [instance: 2a3f81e3-b84f-4370-a598-2f6a607c3d47] allocate_for_instance() {{(pid=61806) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 841.150617] env[61806]: DEBUG nova.network.neutron [None req-81aa2ada-298c-4e33-a594-b57f3ac492a8 tempest-ServerShowV247Test-1218443059 tempest-ServerShowV247Test-1218443059-project-member] [instance: 7c3fd2ff-ebd5-454f-a743-7fbae1088941] Instance cache missing network info. {{(pid=61806) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 841.156444] env[61806]: DEBUG nova.policy [None req-f9d7b4b0-fa22-4fc5-86c2-688ab650f83e tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'b1eac9072ccb4b3bbd23f240e941a76e', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '6e391e1267014fafbb8f5a3211299819', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61806) authorize /opt/stack/nova/nova/policy.py:201}} [ 841.199046] env[61806]: DEBUG oslo_vmware.api [None req-0ea3a38f-3871-4412-a6bc-d21c2086d262 tempest-AttachInterfacesV270Test-1974652806 tempest-AttachInterfacesV270Test-1974652806-project-member] Task: {'id': task-1294702, 'name': PowerOnVM_Task} progress is 100%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 841.200052] env[61806]: DEBUG nova.network.neutron [None req-81aa2ada-298c-4e33-a594-b57f3ac492a8 tempest-ServerShowV247Test-1218443059 tempest-ServerShowV247Test-1218443059-project-member] [instance: 7c3fd2ff-ebd5-454f-a743-7fbae1088941] Updating instance_info_cache with network_info: [] {{(pid=61806) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 841.261789] env[61806]: DEBUG nova.network.neutron [-] [instance: 25218e83-c1ab-4b97-ade1-7c32b6f99305] Updating instance_info_cache with network_info: [] {{(pid=61806) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 841.272205] env[61806]: DEBUG oslo_vmware.api [None req-4e0c8443-0fd4-459b-a4c5-7cf373aa1089 tempest-ServerAddressesNegativeTestJSON-1526246390 tempest-ServerAddressesNegativeTestJSON-1526246390-project-member] Task: {'id': task-1294704, 'name': PowerOnVM_Task, 'duration_secs': 0.637985} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 841.273473] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [None req-4e0c8443-0fd4-459b-a4c5-7cf373aa1089 tempest-ServerAddressesNegativeTestJSON-1526246390 tempest-ServerAddressesNegativeTestJSON-1526246390-project-member] [instance: c8033551-6591-4e37-ae78-4efe7145b10b] Powered on the VM {{(pid=61806) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 841.273726] env[61806]: INFO nova.compute.manager [None req-4e0c8443-0fd4-459b-a4c5-7cf373aa1089 tempest-ServerAddressesNegativeTestJSON-1526246390 tempest-ServerAddressesNegativeTestJSON-1526246390-project-member] [instance: c8033551-6591-4e37-ae78-4efe7145b10b] Took 8.37 seconds to spawn the instance on the hypervisor. [ 841.273923] env[61806]: DEBUG nova.compute.manager [None req-4e0c8443-0fd4-459b-a4c5-7cf373aa1089 tempest-ServerAddressesNegativeTestJSON-1526246390 tempest-ServerAddressesNegativeTestJSON-1526246390-project-member] [instance: c8033551-6591-4e37-ae78-4efe7145b10b] Checking state {{(pid=61806) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 841.274974] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-229ce5f6-ec88-42d5-9c0f-159474d4ab95 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 841.334710] env[61806]: DEBUG oslo_vmware.api [None req-c9a1ae7a-8313-4bdf-8de4-77929ce2d9bc tempest-ServerMetadataNegativeTestJSON-456453902 tempest-ServerMetadataNegativeTestJSON-456453902-project-member] Task: {'id': task-1294705, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.090503} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 841.338022] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-c9a1ae7a-8313-4bdf-8de4-77929ce2d9bc tempest-ServerMetadataNegativeTestJSON-456453902 tempest-ServerMetadataNegativeTestJSON-456453902-project-member] [instance: 7f4907d0-178d-452a-8149-030becde8779] Extended root virtual disk {{(pid=61806) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 841.338022] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-478492d8-39cb-499b-ac7f-261c7c6d0777 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 841.362457] env[61806]: DEBUG nova.virt.vmwareapi.volumeops [None req-c9a1ae7a-8313-4bdf-8de4-77929ce2d9bc tempest-ServerMetadataNegativeTestJSON-456453902 tempest-ServerMetadataNegativeTestJSON-456453902-project-member] [instance: 7f4907d0-178d-452a-8149-030becde8779] Reconfiguring VM instance instance-00000046 to attach disk [datastore2] 7f4907d0-178d-452a-8149-030becde8779/7f4907d0-178d-452a-8149-030becde8779.vmdk or device None with type sparse {{(pid=61806) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 841.363226] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-0276d3ee-3f92-472c-828f-67a9eedd93af {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 841.386355] env[61806]: DEBUG oslo_vmware.api [None req-c9a1ae7a-8313-4bdf-8de4-77929ce2d9bc tempest-ServerMetadataNegativeTestJSON-456453902 tempest-ServerMetadataNegativeTestJSON-456453902-project-member] Waiting for the task: (returnval){ [ 841.386355] env[61806]: value = "task-1294706" [ 841.386355] env[61806]: _type = "Task" [ 841.386355] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 841.398343] env[61806]: DEBUG oslo_vmware.api [None req-c9a1ae7a-8313-4bdf-8de4-77929ce2d9bc tempest-ServerMetadataNegativeTestJSON-456453902 tempest-ServerMetadataNegativeTestJSON-456453902-project-member] Task: {'id': task-1294706, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 841.421231] env[61806]: DEBUG nova.network.neutron [None req-334c6338-dc9d-4c32-ad77-bd3fe007216c tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] [instance: be19fc44-ecbe-489c-9b6e-be8957bb1dd0] Instance cache missing network info. {{(pid=61806) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 841.474973] env[61806]: DEBUG nova.network.neutron [None req-f9d7b4b0-fa22-4fc5-86c2-688ab650f83e tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] [instance: 2a3f81e3-b84f-4370-a598-2f6a607c3d47] Successfully created port: 1f80e0d2-192f-4177-8212-acd6ca475851 {{(pid=61806) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 841.518622] env[61806]: DEBUG nova.scheduler.client.report [None req-4b919e4f-1f0b-40e0-b482-0547e22d55ff tempest-ImagesTestJSON-618852556 tempest-ImagesTestJSON-618852556-project-member] Inventory has not changed for provider a2858be1-fd22-4e08-979e-87ad25293407 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 138, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61806) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 841.551794] env[61806]: DEBUG nova.network.neutron [-] [instance: ef502a7d-7dc0-4319-8ce3-6a8701288237] Updating instance_info_cache with network_info: [] {{(pid=61806) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 841.570958] env[61806]: DEBUG nova.network.neutron [None req-334c6338-dc9d-4c32-ad77-bd3fe007216c tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] [instance: be19fc44-ecbe-489c-9b6e-be8957bb1dd0] Updating instance_info_cache with network_info: [{"id": "6b41a0ca-4d2e-4ab1-9893-180a2f0febca", "address": "fa:16:3e:50:8e:f1", "network": {"id": "b7e78559-d885-4b59-9c7f-9d966b424d04", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-1456721354-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7ab95fb72b5d46c3b6c7bebbccf897cd", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d891a662-7da0-4be5-8d0d-01b4ff70552a", "external-id": "nsx-vlan-transportzone-815", "segmentation_id": 815, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap6b41a0ca-4d", "ovs_interfaceid": "6b41a0ca-4d2e-4ab1-9893-180a2f0febca", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61806) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 841.596023] env[61806]: DEBUG nova.compute.manager [req-12574d22-cf24-4c1e-a3e8-d2682bdcdd1e req-81cb1d8b-fd4a-4bc7-ba57-6fe1f33064cf service nova] [instance: 25218e83-c1ab-4b97-ade1-7c32b6f99305] Received event network-vif-deleted-c8809e29-ea3c-4ea0-bee1-0731daee293d {{(pid=61806) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 841.699880] env[61806]: DEBUG oslo_vmware.api [None req-0ea3a38f-3871-4412-a6bc-d21c2086d262 tempest-AttachInterfacesV270Test-1974652806 tempest-AttachInterfacesV270Test-1974652806-project-member] Task: {'id': task-1294702, 'name': PowerOnVM_Task, 'duration_secs': 1.221018} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 841.700320] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [None req-0ea3a38f-3871-4412-a6bc-d21c2086d262 tempest-AttachInterfacesV270Test-1974652806 tempest-AttachInterfacesV270Test-1974652806-project-member] [instance: 117139cd-4ed4-4bea-8ad3-f4ac32c692e4] Powered on the VM {{(pid=61806) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 841.700626] env[61806]: INFO nova.compute.manager [None req-0ea3a38f-3871-4412-a6bc-d21c2086d262 tempest-AttachInterfacesV270Test-1974652806 tempest-AttachInterfacesV270Test-1974652806-project-member] [instance: 117139cd-4ed4-4bea-8ad3-f4ac32c692e4] Took 11.19 seconds to spawn the instance on the hypervisor. [ 841.700956] env[61806]: DEBUG nova.compute.manager [None req-0ea3a38f-3871-4412-a6bc-d21c2086d262 tempest-AttachInterfacesV270Test-1974652806 tempest-AttachInterfacesV270Test-1974652806-project-member] [instance: 117139cd-4ed4-4bea-8ad3-f4ac32c692e4] Checking state {{(pid=61806) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 841.701891] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7e958adc-125f-48ff-81e0-0316e499ded5 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 841.704800] env[61806]: DEBUG oslo_concurrency.lockutils [None req-81aa2ada-298c-4e33-a594-b57f3ac492a8 tempest-ServerShowV247Test-1218443059 tempest-ServerShowV247Test-1218443059-project-member] Releasing lock "refresh_cache-7c3fd2ff-ebd5-454f-a743-7fbae1088941" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 841.705410] env[61806]: DEBUG nova.compute.manager [None req-81aa2ada-298c-4e33-a594-b57f3ac492a8 tempest-ServerShowV247Test-1218443059 tempest-ServerShowV247Test-1218443059-project-member] [instance: 7c3fd2ff-ebd5-454f-a743-7fbae1088941] Start destroying the instance on the hypervisor. {{(pid=61806) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 841.705738] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-81aa2ada-298c-4e33-a594-b57f3ac492a8 tempest-ServerShowV247Test-1218443059 tempest-ServerShowV247Test-1218443059-project-member] [instance: 7c3fd2ff-ebd5-454f-a743-7fbae1088941] Destroying instance {{(pid=61806) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 841.706523] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-35ba5373-7230-42ad-8525-50edd04770f2 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 841.718719] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [None req-81aa2ada-298c-4e33-a594-b57f3ac492a8 tempest-ServerShowV247Test-1218443059 tempest-ServerShowV247Test-1218443059-project-member] [instance: 7c3fd2ff-ebd5-454f-a743-7fbae1088941] Powering off the VM {{(pid=61806) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 841.719541] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-6a111bf8-d782-4e3a-a46f-b5d67d6e9bde {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 841.726451] env[61806]: DEBUG oslo_vmware.api [None req-81aa2ada-298c-4e33-a594-b57f3ac492a8 tempest-ServerShowV247Test-1218443059 tempest-ServerShowV247Test-1218443059-project-member] Waiting for the task: (returnval){ [ 841.726451] env[61806]: value = "task-1294707" [ 841.726451] env[61806]: _type = "Task" [ 841.726451] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 841.734989] env[61806]: DEBUG nova.compute.manager [req-89c64881-2849-4d57-8ec1-18c1ed69cc80 req-81cd1502-a8fb-4e0a-abc5-237ac0fcea2a service nova] [instance: ef502a7d-7dc0-4319-8ce3-6a8701288237] Received event network-vif-deleted-9e500f22-fafc-4f40-ab22-09b4330b6d4f {{(pid=61806) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 841.739057] env[61806]: DEBUG oslo_vmware.api [None req-81aa2ada-298c-4e33-a594-b57f3ac492a8 tempest-ServerShowV247Test-1218443059 tempest-ServerShowV247Test-1218443059-project-member] Task: {'id': task-1294707, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 841.767389] env[61806]: INFO nova.compute.manager [-] [instance: 25218e83-c1ab-4b97-ade1-7c32b6f99305] Took 1.26 seconds to deallocate network for instance. [ 841.793012] env[61806]: INFO nova.compute.manager [None req-4e0c8443-0fd4-459b-a4c5-7cf373aa1089 tempest-ServerAddressesNegativeTestJSON-1526246390 tempest-ServerAddressesNegativeTestJSON-1526246390-project-member] [instance: c8033551-6591-4e37-ae78-4efe7145b10b] Took 27.45 seconds to build instance. [ 841.896630] env[61806]: DEBUG oslo_vmware.api [None req-c9a1ae7a-8313-4bdf-8de4-77929ce2d9bc tempest-ServerMetadataNegativeTestJSON-456453902 tempest-ServerMetadataNegativeTestJSON-456453902-project-member] Task: {'id': task-1294706, 'name': ReconfigVM_Task, 'duration_secs': 0.30113} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 841.896923] env[61806]: DEBUG nova.virt.vmwareapi.volumeops [None req-c9a1ae7a-8313-4bdf-8de4-77929ce2d9bc tempest-ServerMetadataNegativeTestJSON-456453902 tempest-ServerMetadataNegativeTestJSON-456453902-project-member] [instance: 7f4907d0-178d-452a-8149-030becde8779] Reconfigured VM instance instance-00000046 to attach disk [datastore2] 7f4907d0-178d-452a-8149-030becde8779/7f4907d0-178d-452a-8149-030becde8779.vmdk or device None with type sparse {{(pid=61806) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 841.897571] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-0b3345d4-fbd5-41e6-8ae8-6fb49f069e40 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 841.904547] env[61806]: DEBUG oslo_vmware.api [None req-c9a1ae7a-8313-4bdf-8de4-77929ce2d9bc tempest-ServerMetadataNegativeTestJSON-456453902 tempest-ServerMetadataNegativeTestJSON-456453902-project-member] Waiting for the task: (returnval){ [ 841.904547] env[61806]: value = "task-1294708" [ 841.904547] env[61806]: _type = "Task" [ 841.904547] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 841.912887] env[61806]: DEBUG oslo_vmware.api [None req-c9a1ae7a-8313-4bdf-8de4-77929ce2d9bc tempest-ServerMetadataNegativeTestJSON-456453902 tempest-ServerMetadataNegativeTestJSON-456453902-project-member] Task: {'id': task-1294708, 'name': Rename_Task} progress is 5%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 842.023597] env[61806]: DEBUG oslo_concurrency.lockutils [None req-4b919e4f-1f0b-40e0-b482-0547e22d55ff tempest-ImagesTestJSON-618852556 tempest-ImagesTestJSON-618852556-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.404s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 842.024324] env[61806]: DEBUG nova.compute.manager [None req-4b919e4f-1f0b-40e0-b482-0547e22d55ff tempest-ImagesTestJSON-618852556 tempest-ImagesTestJSON-618852556-project-member] [instance: 28e0baab-8516-42e3-acc2-9b8eb5192f57] Start building networks asynchronously for instance. {{(pid=61806) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 842.027133] env[61806]: DEBUG oslo_concurrency.lockutils [None req-81ee2b3f-4b48-4fda-b966-c7fd3249b12e tempest-ServerDiagnosticsV248Test-463927155 tempest-ServerDiagnosticsV248Test-463927155-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 15.522s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 842.027398] env[61806]: DEBUG nova.objects.instance [None req-81ee2b3f-4b48-4fda-b966-c7fd3249b12e tempest-ServerDiagnosticsV248Test-463927155 tempest-ServerDiagnosticsV248Test-463927155-project-member] Lazy-loading 'resources' on Instance uuid a934d02d-26aa-4900-b473-a58489e5629e {{(pid=61806) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 842.055262] env[61806]: INFO nova.compute.manager [-] [instance: ef502a7d-7dc0-4319-8ce3-6a8701288237] Took 1.30 seconds to deallocate network for instance. [ 842.073306] env[61806]: DEBUG oslo_concurrency.lockutils [None req-334c6338-dc9d-4c32-ad77-bd3fe007216c tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Releasing lock "refresh_cache-be19fc44-ecbe-489c-9b6e-be8957bb1dd0" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 842.073664] env[61806]: DEBUG nova.compute.manager [None req-334c6338-dc9d-4c32-ad77-bd3fe007216c tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] [instance: be19fc44-ecbe-489c-9b6e-be8957bb1dd0] Instance network_info: |[{"id": "6b41a0ca-4d2e-4ab1-9893-180a2f0febca", "address": "fa:16:3e:50:8e:f1", "network": {"id": "b7e78559-d885-4b59-9c7f-9d966b424d04", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-1456721354-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7ab95fb72b5d46c3b6c7bebbccf897cd", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d891a662-7da0-4be5-8d0d-01b4ff70552a", "external-id": "nsx-vlan-transportzone-815", "segmentation_id": 815, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap6b41a0ca-4d", "ovs_interfaceid": "6b41a0ca-4d2e-4ab1-9893-180a2f0febca", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61806) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 842.074139] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-334c6338-dc9d-4c32-ad77-bd3fe007216c tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] [instance: be19fc44-ecbe-489c-9b6e-be8957bb1dd0] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:50:8e:f1', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'd891a662-7da0-4be5-8d0d-01b4ff70552a', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '6b41a0ca-4d2e-4ab1-9893-180a2f0febca', 'vif_model': 'vmxnet3'}] {{(pid=61806) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 842.084622] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [None req-334c6338-dc9d-4c32-ad77-bd3fe007216c tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Creating folder: Project (7ab95fb72b5d46c3b6c7bebbccf897cd). Parent ref: group-v277609. {{(pid=61806) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 842.085496] env[61806]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-20b6a17a-b681-49c0-b252-a2b2d9d492f5 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 842.097338] env[61806]: INFO nova.virt.vmwareapi.vm_util [None req-334c6338-dc9d-4c32-ad77-bd3fe007216c tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Created folder: Project (7ab95fb72b5d46c3b6c7bebbccf897cd) in parent group-v277609. [ 842.097462] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [None req-334c6338-dc9d-4c32-ad77-bd3fe007216c tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Creating folder: Instances. Parent ref: group-v277680. {{(pid=61806) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 842.097948] env[61806]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-da133254-3212-488a-85e2-50c314a83d1c {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 842.108990] env[61806]: INFO nova.virt.vmwareapi.vm_util [None req-334c6338-dc9d-4c32-ad77-bd3fe007216c tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Created folder: Instances in parent group-v277680. [ 842.109281] env[61806]: DEBUG oslo.service.loopingcall [None req-334c6338-dc9d-4c32-ad77-bd3fe007216c tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61806) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 842.109516] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: be19fc44-ecbe-489c-9b6e-be8957bb1dd0] Creating VM on the ESX host {{(pid=61806) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 842.109791] env[61806]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-59ff498c-74c2-4b72-b5ee-2c0ff77aed5e {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 842.129248] env[61806]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 842.129248] env[61806]: value = "task-1294711" [ 842.129248] env[61806]: _type = "Task" [ 842.129248] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 842.137216] env[61806]: DEBUG oslo_vmware.api [-] Task: {'id': task-1294711, 'name': CreateVM_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 842.230640] env[61806]: INFO nova.compute.manager [None req-0ea3a38f-3871-4412-a6bc-d21c2086d262 tempest-AttachInterfacesV270Test-1974652806 tempest-AttachInterfacesV270Test-1974652806-project-member] [instance: 117139cd-4ed4-4bea-8ad3-f4ac32c692e4] Took 29.79 seconds to build instance. [ 842.237213] env[61806]: DEBUG oslo_vmware.api [None req-81aa2ada-298c-4e33-a594-b57f3ac492a8 tempest-ServerShowV247Test-1218443059 tempest-ServerShowV247Test-1218443059-project-member] Task: {'id': task-1294707, 'name': PowerOffVM_Task, 'duration_secs': 0.432855} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 842.237490] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [None req-81aa2ada-298c-4e33-a594-b57f3ac492a8 tempest-ServerShowV247Test-1218443059 tempest-ServerShowV247Test-1218443059-project-member] [instance: 7c3fd2ff-ebd5-454f-a743-7fbae1088941] Powered off the VM {{(pid=61806) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 842.237665] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-81aa2ada-298c-4e33-a594-b57f3ac492a8 tempest-ServerShowV247Test-1218443059 tempest-ServerShowV247Test-1218443059-project-member] [instance: 7c3fd2ff-ebd5-454f-a743-7fbae1088941] Unregistering the VM {{(pid=61806) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 842.237919] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-dbadf42e-74cb-4830-8bad-061f1bc61a01 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 842.267972] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-81aa2ada-298c-4e33-a594-b57f3ac492a8 tempest-ServerShowV247Test-1218443059 tempest-ServerShowV247Test-1218443059-project-member] [instance: 7c3fd2ff-ebd5-454f-a743-7fbae1088941] Unregistered the VM {{(pid=61806) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 842.268226] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-81aa2ada-298c-4e33-a594-b57f3ac492a8 tempest-ServerShowV247Test-1218443059 tempest-ServerShowV247Test-1218443059-project-member] [instance: 7c3fd2ff-ebd5-454f-a743-7fbae1088941] Deleting contents of the VM from datastore datastore2 {{(pid=61806) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 842.268458] env[61806]: DEBUG nova.virt.vmwareapi.ds_util [None req-81aa2ada-298c-4e33-a594-b57f3ac492a8 tempest-ServerShowV247Test-1218443059 tempest-ServerShowV247Test-1218443059-project-member] Deleting the datastore file [datastore2] 7c3fd2ff-ebd5-454f-a743-7fbae1088941 {{(pid=61806) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 842.268793] env[61806]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-3e94791c-f3a3-4481-9318-c7644eff65fb {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 842.274042] env[61806]: DEBUG oslo_concurrency.lockutils [None req-3bd56cbd-68df-4151-a169-676f8bdda9ab tempest-MultipleCreateTestJSON-1589015462 tempest-MultipleCreateTestJSON-1589015462-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 842.278791] env[61806]: DEBUG oslo_vmware.api [None req-81aa2ada-298c-4e33-a594-b57f3ac492a8 tempest-ServerShowV247Test-1218443059 tempest-ServerShowV247Test-1218443059-project-member] Waiting for the task: (returnval){ [ 842.278791] env[61806]: value = "task-1294713" [ 842.278791] env[61806]: _type = "Task" [ 842.278791] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 842.288903] env[61806]: DEBUG oslo_vmware.api [None req-81aa2ada-298c-4e33-a594-b57f3ac492a8 tempest-ServerShowV247Test-1218443059 tempest-ServerShowV247Test-1218443059-project-member] Task: {'id': task-1294713, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 842.294528] env[61806]: DEBUG oslo_concurrency.lockutils [None req-4e0c8443-0fd4-459b-a4c5-7cf373aa1089 tempest-ServerAddressesNegativeTestJSON-1526246390 tempest-ServerAddressesNegativeTestJSON-1526246390-project-member] Lock "c8033551-6591-4e37-ae78-4efe7145b10b" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 52.814s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 842.417710] env[61806]: DEBUG oslo_vmware.api [None req-c9a1ae7a-8313-4bdf-8de4-77929ce2d9bc tempest-ServerMetadataNegativeTestJSON-456453902 tempest-ServerMetadataNegativeTestJSON-456453902-project-member] Task: {'id': task-1294708, 'name': Rename_Task, 'duration_secs': 0.150049} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 842.417710] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [None req-c9a1ae7a-8313-4bdf-8de4-77929ce2d9bc tempest-ServerMetadataNegativeTestJSON-456453902 tempest-ServerMetadataNegativeTestJSON-456453902-project-member] [instance: 7f4907d0-178d-452a-8149-030becde8779] Powering on the VM {{(pid=61806) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 842.417710] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-00ed3e3a-7580-4452-b57d-f8e164cbea58 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 842.430073] env[61806]: DEBUG oslo_vmware.api [None req-c9a1ae7a-8313-4bdf-8de4-77929ce2d9bc tempest-ServerMetadataNegativeTestJSON-456453902 tempest-ServerMetadataNegativeTestJSON-456453902-project-member] Waiting for the task: (returnval){ [ 842.430073] env[61806]: value = "task-1294714" [ 842.430073] env[61806]: _type = "Task" [ 842.430073] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 842.439777] env[61806]: DEBUG oslo_vmware.api [None req-c9a1ae7a-8313-4bdf-8de4-77929ce2d9bc tempest-ServerMetadataNegativeTestJSON-456453902 tempest-ServerMetadataNegativeTestJSON-456453902-project-member] Task: {'id': task-1294714, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 842.533670] env[61806]: DEBUG nova.compute.utils [None req-4b919e4f-1f0b-40e0-b482-0547e22d55ff tempest-ImagesTestJSON-618852556 tempest-ImagesTestJSON-618852556-project-member] Using /dev/sd instead of None {{(pid=61806) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 842.535614] env[61806]: DEBUG nova.compute.manager [None req-4b919e4f-1f0b-40e0-b482-0547e22d55ff tempest-ImagesTestJSON-618852556 tempest-ImagesTestJSON-618852556-project-member] [instance: 28e0baab-8516-42e3-acc2-9b8eb5192f57] Allocating IP information in the background. {{(pid=61806) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 842.535783] env[61806]: DEBUG nova.network.neutron [None req-4b919e4f-1f0b-40e0-b482-0547e22d55ff tempest-ImagesTestJSON-618852556 tempest-ImagesTestJSON-618852556-project-member] [instance: 28e0baab-8516-42e3-acc2-9b8eb5192f57] allocate_for_instance() {{(pid=61806) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 842.562281] env[61806]: DEBUG oslo_concurrency.lockutils [None req-182a714a-9a7d-40bc-ad59-30e74e8a3863 tempest-MultipleCreateTestJSON-1589015462 tempest-MultipleCreateTestJSON-1589015462-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 842.599104] env[61806]: DEBUG nova.policy [None req-4b919e4f-1f0b-40e0-b482-0547e22d55ff tempest-ImagesTestJSON-618852556 tempest-ImagesTestJSON-618852556-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'e6fa11a9b63f4cd6b04baf3115431167', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '4b9aa91aff1d4008ac5096902b77f852', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61806) authorize /opt/stack/nova/nova/policy.py:201}} [ 842.639170] env[61806]: DEBUG oslo_vmware.api [-] Task: {'id': task-1294711, 'name': CreateVM_Task, 'duration_secs': 0.440437} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 842.641598] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: be19fc44-ecbe-489c-9b6e-be8957bb1dd0] Created VM on the ESX host {{(pid=61806) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 842.642539] env[61806]: DEBUG oslo_concurrency.lockutils [None req-334c6338-dc9d-4c32-ad77-bd3fe007216c tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 842.642952] env[61806]: DEBUG oslo_concurrency.lockutils [None req-334c6338-dc9d-4c32-ad77-bd3fe007216c tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Acquired lock "[datastore2] devstack-image-cache_base/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 842.643042] env[61806]: DEBUG oslo_concurrency.lockutils [None req-334c6338-dc9d-4c32-ad77-bd3fe007216c tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 842.643280] env[61806]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-905f4aca-7b67-460b-ade7-3af99c9e7d6c {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 842.648250] env[61806]: DEBUG oslo_vmware.api [None req-334c6338-dc9d-4c32-ad77-bd3fe007216c tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Waiting for the task: (returnval){ [ 842.648250] env[61806]: value = "session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]52a0ebe9-d1c4-89c5-6188-55569efbfac7" [ 842.648250] env[61806]: _type = "Task" [ 842.648250] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 842.658628] env[61806]: DEBUG oslo_vmware.api [None req-334c6338-dc9d-4c32-ad77-bd3fe007216c tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Task: {'id': session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]52a0ebe9-d1c4-89c5-6188-55569efbfac7, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 842.733540] env[61806]: DEBUG oslo_concurrency.lockutils [None req-0ea3a38f-3871-4412-a6bc-d21c2086d262 tempest-AttachInterfacesV270Test-1974652806 tempest-AttachInterfacesV270Test-1974652806-project-member] Lock "117139cd-4ed4-4bea-8ad3-f4ac32c692e4" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 73.924s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 842.791308] env[61806]: DEBUG oslo_vmware.api [None req-81aa2ada-298c-4e33-a594-b57f3ac492a8 tempest-ServerShowV247Test-1218443059 tempest-ServerShowV247Test-1218443059-project-member] Task: {'id': task-1294713, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.127809} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 842.791740] env[61806]: DEBUG nova.virt.vmwareapi.ds_util [None req-81aa2ada-298c-4e33-a594-b57f3ac492a8 tempest-ServerShowV247Test-1218443059 tempest-ServerShowV247Test-1218443059-project-member] Deleted the datastore file {{(pid=61806) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 842.792087] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-81aa2ada-298c-4e33-a594-b57f3ac492a8 tempest-ServerShowV247Test-1218443059 tempest-ServerShowV247Test-1218443059-project-member] [instance: 7c3fd2ff-ebd5-454f-a743-7fbae1088941] Deleted contents of the VM from datastore datastore2 {{(pid=61806) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 842.792282] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-81aa2ada-298c-4e33-a594-b57f3ac492a8 tempest-ServerShowV247Test-1218443059 tempest-ServerShowV247Test-1218443059-project-member] [instance: 7c3fd2ff-ebd5-454f-a743-7fbae1088941] Instance destroyed {{(pid=61806) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 842.792539] env[61806]: INFO nova.compute.manager [None req-81aa2ada-298c-4e33-a594-b57f3ac492a8 tempest-ServerShowV247Test-1218443059 tempest-ServerShowV247Test-1218443059-project-member] [instance: 7c3fd2ff-ebd5-454f-a743-7fbae1088941] Took 1.09 seconds to destroy the instance on the hypervisor. [ 842.792851] env[61806]: DEBUG oslo.service.loopingcall [None req-81aa2ada-298c-4e33-a594-b57f3ac492a8 tempest-ServerShowV247Test-1218443059 tempest-ServerShowV247Test-1218443059-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61806) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 842.793040] env[61806]: DEBUG nova.compute.manager [-] [instance: 7c3fd2ff-ebd5-454f-a743-7fbae1088941] Deallocating network for instance {{(pid=61806) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 842.794043] env[61806]: DEBUG nova.network.neutron [-] [instance: 7c3fd2ff-ebd5-454f-a743-7fbae1088941] deallocate_for_instance() {{(pid=61806) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 842.821361] env[61806]: DEBUG nova.network.neutron [-] [instance: 7c3fd2ff-ebd5-454f-a743-7fbae1088941] Instance cache missing network info. {{(pid=61806) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 842.839995] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f37b9c6d-9c37-46a7-9e8a-35e9cd842523 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 842.850870] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b3254084-50fc-4890-9dcc-9086052653ac {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 842.897792] env[61806]: DEBUG nova.network.neutron [None req-4b919e4f-1f0b-40e0-b482-0547e22d55ff tempest-ImagesTestJSON-618852556 tempest-ImagesTestJSON-618852556-project-member] [instance: 28e0baab-8516-42e3-acc2-9b8eb5192f57] Successfully created port: 2be8aa59-fbf0-4081-955d-9e031781b864 {{(pid=61806) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 842.900919] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2adc89dc-0fde-4fa1-8646-77a0e012a438 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 842.909539] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bcd72241-cc3d-45d6-b652-502d63fef89a {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 842.926377] env[61806]: DEBUG nova.compute.provider_tree [None req-81ee2b3f-4b48-4fda-b966-c7fd3249b12e tempest-ServerDiagnosticsV248Test-463927155 tempest-ServerDiagnosticsV248Test-463927155-project-member] Inventory has not changed in ProviderTree for provider: a2858be1-fd22-4e08-979e-87ad25293407 {{(pid=61806) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 842.940201] env[61806]: DEBUG oslo_vmware.api [None req-c9a1ae7a-8313-4bdf-8de4-77929ce2d9bc tempest-ServerMetadataNegativeTestJSON-456453902 tempest-ServerMetadataNegativeTestJSON-456453902-project-member] Task: {'id': task-1294714, 'name': PowerOnVM_Task} progress is 89%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 842.968728] env[61806]: DEBUG oslo_concurrency.lockutils [None req-40a350df-b9c4-4f7d-968f-0160bcbc0d06 tempest-ServerAddressesNegativeTestJSON-1526246390 tempest-ServerAddressesNegativeTestJSON-1526246390-project-member] Acquiring lock "c8033551-6591-4e37-ae78-4efe7145b10b" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 842.969115] env[61806]: DEBUG oslo_concurrency.lockutils [None req-40a350df-b9c4-4f7d-968f-0160bcbc0d06 tempest-ServerAddressesNegativeTestJSON-1526246390 tempest-ServerAddressesNegativeTestJSON-1526246390-project-member] Lock "c8033551-6591-4e37-ae78-4efe7145b10b" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 842.969287] env[61806]: DEBUG oslo_concurrency.lockutils [None req-40a350df-b9c4-4f7d-968f-0160bcbc0d06 tempest-ServerAddressesNegativeTestJSON-1526246390 tempest-ServerAddressesNegativeTestJSON-1526246390-project-member] Acquiring lock "c8033551-6591-4e37-ae78-4efe7145b10b-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 842.969547] env[61806]: DEBUG oslo_concurrency.lockutils [None req-40a350df-b9c4-4f7d-968f-0160bcbc0d06 tempest-ServerAddressesNegativeTestJSON-1526246390 tempest-ServerAddressesNegativeTestJSON-1526246390-project-member] Lock "c8033551-6591-4e37-ae78-4efe7145b10b-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 842.969900] env[61806]: DEBUG oslo_concurrency.lockutils [None req-40a350df-b9c4-4f7d-968f-0160bcbc0d06 tempest-ServerAddressesNegativeTestJSON-1526246390 tempest-ServerAddressesNegativeTestJSON-1526246390-project-member] Lock "c8033551-6591-4e37-ae78-4efe7145b10b-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 842.972887] env[61806]: INFO nova.compute.manager [None req-40a350df-b9c4-4f7d-968f-0160bcbc0d06 tempest-ServerAddressesNegativeTestJSON-1526246390 tempest-ServerAddressesNegativeTestJSON-1526246390-project-member] [instance: c8033551-6591-4e37-ae78-4efe7145b10b] Terminating instance [ 842.974795] env[61806]: DEBUG nova.compute.manager [None req-40a350df-b9c4-4f7d-968f-0160bcbc0d06 tempest-ServerAddressesNegativeTestJSON-1526246390 tempest-ServerAddressesNegativeTestJSON-1526246390-project-member] [instance: c8033551-6591-4e37-ae78-4efe7145b10b] Start destroying the instance on the hypervisor. {{(pid=61806) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 842.975077] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-40a350df-b9c4-4f7d-968f-0160bcbc0d06 tempest-ServerAddressesNegativeTestJSON-1526246390 tempest-ServerAddressesNegativeTestJSON-1526246390-project-member] [instance: c8033551-6591-4e37-ae78-4efe7145b10b] Destroying instance {{(pid=61806) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 842.975927] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b170fc3b-f1be-4964-ae03-af93a0274725 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 842.984395] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [None req-40a350df-b9c4-4f7d-968f-0160bcbc0d06 tempest-ServerAddressesNegativeTestJSON-1526246390 tempest-ServerAddressesNegativeTestJSON-1526246390-project-member] [instance: c8033551-6591-4e37-ae78-4efe7145b10b] Powering off the VM {{(pid=61806) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 842.984705] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-1cbd759b-4ef2-47a3-87c7-b1d9a74a78bb {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 842.992888] env[61806]: DEBUG oslo_vmware.api [None req-40a350df-b9c4-4f7d-968f-0160bcbc0d06 tempest-ServerAddressesNegativeTestJSON-1526246390 tempest-ServerAddressesNegativeTestJSON-1526246390-project-member] Waiting for the task: (returnval){ [ 842.992888] env[61806]: value = "task-1294715" [ 842.992888] env[61806]: _type = "Task" [ 842.992888] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 843.002496] env[61806]: DEBUG oslo_vmware.api [None req-40a350df-b9c4-4f7d-968f-0160bcbc0d06 tempest-ServerAddressesNegativeTestJSON-1526246390 tempest-ServerAddressesNegativeTestJSON-1526246390-project-member] Task: {'id': task-1294715, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 843.039574] env[61806]: DEBUG nova.compute.manager [None req-4b919e4f-1f0b-40e0-b482-0547e22d55ff tempest-ImagesTestJSON-618852556 tempest-ImagesTestJSON-618852556-project-member] [instance: 28e0baab-8516-42e3-acc2-9b8eb5192f57] Start building block device mappings for instance. {{(pid=61806) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 843.049405] env[61806]: DEBUG nova.network.neutron [None req-f9d7b4b0-fa22-4fc5-86c2-688ab650f83e tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] [instance: 2a3f81e3-b84f-4370-a598-2f6a607c3d47] Successfully updated port: 1f80e0d2-192f-4177-8212-acd6ca475851 {{(pid=61806) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 843.159769] env[61806]: DEBUG oslo_vmware.api [None req-334c6338-dc9d-4c32-ad77-bd3fe007216c tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Task: {'id': session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]52a0ebe9-d1c4-89c5-6188-55569efbfac7, 'name': SearchDatastore_Task, 'duration_secs': 0.011429} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 843.160114] env[61806]: DEBUG oslo_concurrency.lockutils [None req-334c6338-dc9d-4c32-ad77-bd3fe007216c tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Releasing lock "[datastore2] devstack-image-cache_base/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 843.160352] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-334c6338-dc9d-4c32-ad77-bd3fe007216c tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] [instance: be19fc44-ecbe-489c-9b6e-be8957bb1dd0] Processing image ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa {{(pid=61806) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 843.160742] env[61806]: DEBUG oslo_concurrency.lockutils [None req-334c6338-dc9d-4c32-ad77-bd3fe007216c tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa.vmdk" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 843.160980] env[61806]: DEBUG oslo_concurrency.lockutils [None req-334c6338-dc9d-4c32-ad77-bd3fe007216c tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Acquired lock "[datastore2] devstack-image-cache_base/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa.vmdk" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 843.161208] env[61806]: DEBUG nova.virt.vmwareapi.ds_util [None req-334c6338-dc9d-4c32-ad77-bd3fe007216c tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61806) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 843.161486] env[61806]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-6fc960c0-231a-4fad-adca-d554323dc9d0 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 843.172130] env[61806]: DEBUG nova.virt.vmwareapi.ds_util [None req-334c6338-dc9d-4c32-ad77-bd3fe007216c tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61806) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 843.172318] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-334c6338-dc9d-4c32-ad77-bd3fe007216c tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61806) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 843.173190] env[61806]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-baac2a57-8bf5-46fb-a5b4-519e2a59b2e0 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 843.178355] env[61806]: DEBUG oslo_vmware.api [None req-334c6338-dc9d-4c32-ad77-bd3fe007216c tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Waiting for the task: (returnval){ [ 843.178355] env[61806]: value = "session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]52d65a26-21d0-68cc-6992-52b94485cebd" [ 843.178355] env[61806]: _type = "Task" [ 843.178355] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 843.187190] env[61806]: DEBUG oslo_vmware.api [None req-334c6338-dc9d-4c32-ad77-bd3fe007216c tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Task: {'id': session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]52d65a26-21d0-68cc-6992-52b94485cebd, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 843.324368] env[61806]: DEBUG nova.network.neutron [-] [instance: 7c3fd2ff-ebd5-454f-a743-7fbae1088941] Updating instance_info_cache with network_info: [] {{(pid=61806) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 843.350689] env[61806]: DEBUG oslo_concurrency.lockutils [None req-0563170c-7d39-409f-a6dd-2eda4e3f4539 tempest-AttachInterfacesV270Test-1974652806 tempest-AttachInterfacesV270Test-1974652806-project-member] Acquiring lock "interface-117139cd-4ed4-4bea-8ad3-f4ac32c692e4-None" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 843.351044] env[61806]: DEBUG oslo_concurrency.lockutils [None req-0563170c-7d39-409f-a6dd-2eda4e3f4539 tempest-AttachInterfacesV270Test-1974652806 tempest-AttachInterfacesV270Test-1974652806-project-member] Lock "interface-117139cd-4ed4-4bea-8ad3-f4ac32c692e4-None" acquired by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: waited 0.001s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 843.351514] env[61806]: DEBUG nova.objects.instance [None req-0563170c-7d39-409f-a6dd-2eda4e3f4539 tempest-AttachInterfacesV270Test-1974652806 tempest-AttachInterfacesV270Test-1974652806-project-member] Lazy-loading 'flavor' on Instance uuid 117139cd-4ed4-4bea-8ad3-f4ac32c692e4 {{(pid=61806) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 843.429487] env[61806]: DEBUG nova.scheduler.client.report [None req-81ee2b3f-4b48-4fda-b966-c7fd3249b12e tempest-ServerDiagnosticsV248Test-463927155 tempest-ServerDiagnosticsV248Test-463927155-project-member] Inventory has not changed for provider a2858be1-fd22-4e08-979e-87ad25293407 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 138, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61806) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 843.443744] env[61806]: DEBUG oslo_vmware.api [None req-c9a1ae7a-8313-4bdf-8de4-77929ce2d9bc tempest-ServerMetadataNegativeTestJSON-456453902 tempest-ServerMetadataNegativeTestJSON-456453902-project-member] Task: {'id': task-1294714, 'name': PowerOnVM_Task, 'duration_secs': 0.6073} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 843.444751] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [None req-c9a1ae7a-8313-4bdf-8de4-77929ce2d9bc tempest-ServerMetadataNegativeTestJSON-456453902 tempest-ServerMetadataNegativeTestJSON-456453902-project-member] [instance: 7f4907d0-178d-452a-8149-030becde8779] Powered on the VM {{(pid=61806) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 843.444886] env[61806]: INFO nova.compute.manager [None req-c9a1ae7a-8313-4bdf-8de4-77929ce2d9bc tempest-ServerMetadataNegativeTestJSON-456453902 tempest-ServerMetadataNegativeTestJSON-456453902-project-member] [instance: 7f4907d0-178d-452a-8149-030becde8779] Took 8.16 seconds to spawn the instance on the hypervisor. [ 843.445080] env[61806]: DEBUG nova.compute.manager [None req-c9a1ae7a-8313-4bdf-8de4-77929ce2d9bc tempest-ServerMetadataNegativeTestJSON-456453902 tempest-ServerMetadataNegativeTestJSON-456453902-project-member] [instance: 7f4907d0-178d-452a-8149-030becde8779] Checking state {{(pid=61806) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 843.445867] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f15175bd-0f8d-4176-af70-24b0637c5ad6 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 843.502018] env[61806]: DEBUG oslo_vmware.api [None req-40a350df-b9c4-4f7d-968f-0160bcbc0d06 tempest-ServerAddressesNegativeTestJSON-1526246390 tempest-ServerAddressesNegativeTestJSON-1526246390-project-member] Task: {'id': task-1294715, 'name': PowerOffVM_Task, 'duration_secs': 0.396882} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 843.502309] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [None req-40a350df-b9c4-4f7d-968f-0160bcbc0d06 tempest-ServerAddressesNegativeTestJSON-1526246390 tempest-ServerAddressesNegativeTestJSON-1526246390-project-member] [instance: c8033551-6591-4e37-ae78-4efe7145b10b] Powered off the VM {{(pid=61806) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 843.502480] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-40a350df-b9c4-4f7d-968f-0160bcbc0d06 tempest-ServerAddressesNegativeTestJSON-1526246390 tempest-ServerAddressesNegativeTestJSON-1526246390-project-member] [instance: c8033551-6591-4e37-ae78-4efe7145b10b] Unregistering the VM {{(pid=61806) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 843.502761] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-735d3ddf-ff53-4b56-b790-17dce681dd48 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 843.552090] env[61806]: DEBUG oslo_concurrency.lockutils [None req-f9d7b4b0-fa22-4fc5-86c2-688ab650f83e tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] Acquiring lock "refresh_cache-2a3f81e3-b84f-4370-a598-2f6a607c3d47" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 843.552311] env[61806]: DEBUG oslo_concurrency.lockutils [None req-f9d7b4b0-fa22-4fc5-86c2-688ab650f83e tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] Acquired lock "refresh_cache-2a3f81e3-b84f-4370-a598-2f6a607c3d47" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 843.552513] env[61806]: DEBUG nova.network.neutron [None req-f9d7b4b0-fa22-4fc5-86c2-688ab650f83e tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] [instance: 2a3f81e3-b84f-4370-a598-2f6a607c3d47] Building network info cache for instance {{(pid=61806) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 843.580879] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-40a350df-b9c4-4f7d-968f-0160bcbc0d06 tempest-ServerAddressesNegativeTestJSON-1526246390 tempest-ServerAddressesNegativeTestJSON-1526246390-project-member] [instance: c8033551-6591-4e37-ae78-4efe7145b10b] Unregistered the VM {{(pid=61806) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 843.581069] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-40a350df-b9c4-4f7d-968f-0160bcbc0d06 tempest-ServerAddressesNegativeTestJSON-1526246390 tempest-ServerAddressesNegativeTestJSON-1526246390-project-member] [instance: c8033551-6591-4e37-ae78-4efe7145b10b] Deleting contents of the VM from datastore datastore2 {{(pid=61806) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 843.581123] env[61806]: DEBUG nova.virt.vmwareapi.ds_util [None req-40a350df-b9c4-4f7d-968f-0160bcbc0d06 tempest-ServerAddressesNegativeTestJSON-1526246390 tempest-ServerAddressesNegativeTestJSON-1526246390-project-member] Deleting the datastore file [datastore2] c8033551-6591-4e37-ae78-4efe7145b10b {{(pid=61806) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 843.582132] env[61806]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-ea30a454-3b21-4bf6-9e65-8233bd97bee7 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 843.589342] env[61806]: DEBUG oslo_vmware.api [None req-40a350df-b9c4-4f7d-968f-0160bcbc0d06 tempest-ServerAddressesNegativeTestJSON-1526246390 tempest-ServerAddressesNegativeTestJSON-1526246390-project-member] Waiting for the task: (returnval){ [ 843.589342] env[61806]: value = "task-1294717" [ 843.589342] env[61806]: _type = "Task" [ 843.589342] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 843.599651] env[61806]: DEBUG oslo_vmware.api [None req-40a350df-b9c4-4f7d-968f-0160bcbc0d06 tempest-ServerAddressesNegativeTestJSON-1526246390 tempest-ServerAddressesNegativeTestJSON-1526246390-project-member] Task: {'id': task-1294717, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 843.692457] env[61806]: DEBUG oslo_vmware.api [None req-334c6338-dc9d-4c32-ad77-bd3fe007216c tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Task: {'id': session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]52d65a26-21d0-68cc-6992-52b94485cebd, 'name': SearchDatastore_Task, 'duration_secs': 0.017057} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 843.693583] env[61806]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-50df34a3-1058-43fb-9c06-0b27e9437ff9 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 843.700035] env[61806]: DEBUG oslo_vmware.api [None req-334c6338-dc9d-4c32-ad77-bd3fe007216c tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Waiting for the task: (returnval){ [ 843.700035] env[61806]: value = "session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]522e2405-d9f1-2c15-2cdd-70e68b688bae" [ 843.700035] env[61806]: _type = "Task" [ 843.700035] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 843.710422] env[61806]: DEBUG oslo_vmware.api [None req-334c6338-dc9d-4c32-ad77-bd3fe007216c tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Task: {'id': session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]522e2405-d9f1-2c15-2cdd-70e68b688bae, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 843.826946] env[61806]: INFO nova.compute.manager [-] [instance: 7c3fd2ff-ebd5-454f-a743-7fbae1088941] Took 1.03 seconds to deallocate network for instance. [ 843.856044] env[61806]: DEBUG nova.objects.instance [None req-0563170c-7d39-409f-a6dd-2eda4e3f4539 tempest-AttachInterfacesV270Test-1974652806 tempest-AttachInterfacesV270Test-1974652806-project-member] Lazy-loading 'pci_requests' on Instance uuid 117139cd-4ed4-4bea-8ad3-f4ac32c692e4 {{(pid=61806) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 843.939835] env[61806]: DEBUG oslo_concurrency.lockutils [None req-81ee2b3f-4b48-4fda-b966-c7fd3249b12e tempest-ServerDiagnosticsV248Test-463927155 tempest-ServerDiagnosticsV248Test-463927155-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.913s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 843.943372] env[61806]: DEBUG oslo_concurrency.lockutils [None req-f4c6ec76-6b50-4a2c-bb86-09b71e318ce8 tempest-ServerShowV257Test-1159790492 tempest-ServerShowV257Test-1159790492-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 16.290s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 843.943731] env[61806]: DEBUG nova.objects.instance [None req-f4c6ec76-6b50-4a2c-bb86-09b71e318ce8 tempest-ServerShowV257Test-1159790492 tempest-ServerShowV257Test-1159790492-project-member] Lazy-loading 'resources' on Instance uuid 616ec206-9804-469e-ab5c-41aea7f048aa {{(pid=61806) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 843.962870] env[61806]: INFO nova.compute.manager [None req-c9a1ae7a-8313-4bdf-8de4-77929ce2d9bc tempest-ServerMetadataNegativeTestJSON-456453902 tempest-ServerMetadataNegativeTestJSON-456453902-project-member] [instance: 7f4907d0-178d-452a-8149-030becde8779] Took 25.98 seconds to build instance. [ 843.965238] env[61806]: INFO nova.scheduler.client.report [None req-81ee2b3f-4b48-4fda-b966-c7fd3249b12e tempest-ServerDiagnosticsV248Test-463927155 tempest-ServerDiagnosticsV248Test-463927155-project-member] Deleted allocations for instance a934d02d-26aa-4900-b473-a58489e5629e [ 844.049600] env[61806]: DEBUG nova.compute.manager [None req-4b919e4f-1f0b-40e0-b482-0547e22d55ff tempest-ImagesTestJSON-618852556 tempest-ImagesTestJSON-618852556-project-member] [instance: 28e0baab-8516-42e3-acc2-9b8eb5192f57] Start spawning the instance on the hypervisor. {{(pid=61806) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 844.076694] env[61806]: DEBUG nova.virt.hardware [None req-4b919e4f-1f0b-40e0-b482-0547e22d55ff tempest-ImagesTestJSON-618852556 tempest-ImagesTestJSON-618852556-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-15T18:05:10Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='81d3e1db94dce2ecd8cb59052af5f2d9',container_format='bare',created_at=2024-10-15T18:11:34Z,direct_url=,disk_format='vmdk',id=e1cfdcc0-a515-4a37-bf5d-903e61ffb7f2,min_disk=1,min_ram=0,name='tempest-test-snap-530130087',owner='4b9aa91aff1d4008ac5096902b77f852',properties=ImageMetaProps,protected=,size=21334016,status='active',tags=,updated_at=2024-10-15T18:11:48Z,virtual_size=,visibility=), allow threads: False {{(pid=61806) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 844.077025] env[61806]: DEBUG nova.virt.hardware [None req-4b919e4f-1f0b-40e0-b482-0547e22d55ff tempest-ImagesTestJSON-618852556 tempest-ImagesTestJSON-618852556-project-member] Flavor limits 0:0:0 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 844.077241] env[61806]: DEBUG nova.virt.hardware [None req-4b919e4f-1f0b-40e0-b482-0547e22d55ff tempest-ImagesTestJSON-618852556 tempest-ImagesTestJSON-618852556-project-member] Image limits 0:0:0 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 844.077476] env[61806]: DEBUG nova.virt.hardware [None req-4b919e4f-1f0b-40e0-b482-0547e22d55ff tempest-ImagesTestJSON-618852556 tempest-ImagesTestJSON-618852556-project-member] Flavor pref 0:0:0 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 844.077670] env[61806]: DEBUG nova.virt.hardware [None req-4b919e4f-1f0b-40e0-b482-0547e22d55ff tempest-ImagesTestJSON-618852556 tempest-ImagesTestJSON-618852556-project-member] Image pref 0:0:0 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 844.077876] env[61806]: DEBUG nova.virt.hardware [None req-4b919e4f-1f0b-40e0-b482-0547e22d55ff tempest-ImagesTestJSON-618852556 tempest-ImagesTestJSON-618852556-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 844.078151] env[61806]: DEBUG nova.virt.hardware [None req-4b919e4f-1f0b-40e0-b482-0547e22d55ff tempest-ImagesTestJSON-618852556 tempest-ImagesTestJSON-618852556-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61806) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 844.078383] env[61806]: DEBUG nova.virt.hardware [None req-4b919e4f-1f0b-40e0-b482-0547e22d55ff tempest-ImagesTestJSON-618852556 tempest-ImagesTestJSON-618852556-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61806) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 844.078605] env[61806]: DEBUG nova.virt.hardware [None req-4b919e4f-1f0b-40e0-b482-0547e22d55ff tempest-ImagesTestJSON-618852556 tempest-ImagesTestJSON-618852556-project-member] Got 1 possible topologies {{(pid=61806) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 844.078814] env[61806]: DEBUG nova.virt.hardware [None req-4b919e4f-1f0b-40e0-b482-0547e22d55ff tempest-ImagesTestJSON-618852556 tempest-ImagesTestJSON-618852556-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61806) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 844.079046] env[61806]: DEBUG nova.virt.hardware [None req-4b919e4f-1f0b-40e0-b482-0547e22d55ff tempest-ImagesTestJSON-618852556 tempest-ImagesTestJSON-618852556-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61806) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 844.079975] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ffc60537-09d4-447c-8444-0a0caab8f369 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 844.088784] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8235ba65-b360-4f01-8142-7d00d8a161dc {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 844.096955] env[61806]: WARNING nova.network.neutron [None req-f9d7b4b0-fa22-4fc5-86c2-688ab650f83e tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] [instance: 2a3f81e3-b84f-4370-a598-2f6a607c3d47] a7ffda72-e57c-41fa-a7ed-d7b12084bc9d already exists in list: networks containing: ['a7ffda72-e57c-41fa-a7ed-d7b12084bc9d']. ignoring it [ 844.111251] env[61806]: DEBUG oslo_vmware.api [None req-40a350df-b9c4-4f7d-968f-0160bcbc0d06 tempest-ServerAddressesNegativeTestJSON-1526246390 tempest-ServerAddressesNegativeTestJSON-1526246390-project-member] Task: {'id': task-1294717, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.369088} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 844.111762] env[61806]: DEBUG nova.virt.vmwareapi.ds_util [None req-40a350df-b9c4-4f7d-968f-0160bcbc0d06 tempest-ServerAddressesNegativeTestJSON-1526246390 tempest-ServerAddressesNegativeTestJSON-1526246390-project-member] Deleted the datastore file {{(pid=61806) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 844.111965] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-40a350df-b9c4-4f7d-968f-0160bcbc0d06 tempest-ServerAddressesNegativeTestJSON-1526246390 tempest-ServerAddressesNegativeTestJSON-1526246390-project-member] [instance: c8033551-6591-4e37-ae78-4efe7145b10b] Deleted contents of the VM from datastore datastore2 {{(pid=61806) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 844.112155] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-40a350df-b9c4-4f7d-968f-0160bcbc0d06 tempest-ServerAddressesNegativeTestJSON-1526246390 tempest-ServerAddressesNegativeTestJSON-1526246390-project-member] [instance: c8033551-6591-4e37-ae78-4efe7145b10b] Instance destroyed {{(pid=61806) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 844.112336] env[61806]: INFO nova.compute.manager [None req-40a350df-b9c4-4f7d-968f-0160bcbc0d06 tempest-ServerAddressesNegativeTestJSON-1526246390 tempest-ServerAddressesNegativeTestJSON-1526246390-project-member] [instance: c8033551-6591-4e37-ae78-4efe7145b10b] Took 1.14 seconds to destroy the instance on the hypervisor. [ 844.112581] env[61806]: DEBUG oslo.service.loopingcall [None req-40a350df-b9c4-4f7d-968f-0160bcbc0d06 tempest-ServerAddressesNegativeTestJSON-1526246390 tempest-ServerAddressesNegativeTestJSON-1526246390-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61806) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 844.112861] env[61806]: DEBUG nova.compute.manager [-] [instance: c8033551-6591-4e37-ae78-4efe7145b10b] Deallocating network for instance {{(pid=61806) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 844.112973] env[61806]: DEBUG nova.network.neutron [-] [instance: c8033551-6591-4e37-ae78-4efe7145b10b] deallocate_for_instance() {{(pid=61806) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 844.213857] env[61806]: DEBUG oslo_vmware.api [None req-334c6338-dc9d-4c32-ad77-bd3fe007216c tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Task: {'id': session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]522e2405-d9f1-2c15-2cdd-70e68b688bae, 'name': SearchDatastore_Task, 'duration_secs': 0.010534} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 844.213857] env[61806]: DEBUG oslo_concurrency.lockutils [None req-334c6338-dc9d-4c32-ad77-bd3fe007216c tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Releasing lock "[datastore2] devstack-image-cache_base/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa.vmdk" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 844.213857] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [None req-334c6338-dc9d-4c32-ad77-bd3fe007216c tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa.vmdk to [datastore2] be19fc44-ecbe-489c-9b6e-be8957bb1dd0/be19fc44-ecbe-489c-9b6e-be8957bb1dd0.vmdk {{(pid=61806) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 844.213857] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-01ce1193-6122-43fc-96ea-647bb7dbf020 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 844.221022] env[61806]: DEBUG oslo_vmware.api [None req-334c6338-dc9d-4c32-ad77-bd3fe007216c tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Waiting for the task: (returnval){ [ 844.221022] env[61806]: value = "task-1294718" [ 844.221022] env[61806]: _type = "Task" [ 844.221022] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 844.227571] env[61806]: DEBUG oslo_vmware.api [None req-334c6338-dc9d-4c32-ad77-bd3fe007216c tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Task: {'id': task-1294718, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 844.334332] env[61806]: DEBUG oslo_concurrency.lockutils [None req-81aa2ada-298c-4e33-a594-b57f3ac492a8 tempest-ServerShowV247Test-1218443059 tempest-ServerShowV247Test-1218443059-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 844.361800] env[61806]: DEBUG nova.objects.base [None req-0563170c-7d39-409f-a6dd-2eda4e3f4539 tempest-AttachInterfacesV270Test-1974652806 tempest-AttachInterfacesV270Test-1974652806-project-member] Object Instance<117139cd-4ed4-4bea-8ad3-f4ac32c692e4> lazy-loaded attributes: flavor,pci_requests {{(pid=61806) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 844.362441] env[61806]: DEBUG nova.network.neutron [None req-0563170c-7d39-409f-a6dd-2eda4e3f4539 tempest-AttachInterfacesV270Test-1974652806 tempest-AttachInterfacesV270Test-1974652806-project-member] [instance: 117139cd-4ed4-4bea-8ad3-f4ac32c692e4] allocate_for_instance() {{(pid=61806) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 844.467464] env[61806]: DEBUG oslo_concurrency.lockutils [None req-c9a1ae7a-8313-4bdf-8de4-77929ce2d9bc tempest-ServerMetadataNegativeTestJSON-456453902 tempest-ServerMetadataNegativeTestJSON-456453902-project-member] Lock "7f4907d0-178d-452a-8149-030becde8779" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 44.229s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 844.477144] env[61806]: DEBUG oslo_concurrency.lockutils [None req-81ee2b3f-4b48-4fda-b966-c7fd3249b12e tempest-ServerDiagnosticsV248Test-463927155 tempest-ServerDiagnosticsV248Test-463927155-project-member] Lock "a934d02d-26aa-4900-b473-a58489e5629e" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 21.777s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 844.510977] env[61806]: DEBUG oslo_concurrency.lockutils [None req-0563170c-7d39-409f-a6dd-2eda4e3f4539 tempest-AttachInterfacesV270Test-1974652806 tempest-AttachInterfacesV270Test-1974652806-project-member] Lock "interface-117139cd-4ed4-4bea-8ad3-f4ac32c692e4-None" "released" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: held 1.160s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 844.578360] env[61806]: DEBUG nova.network.neutron [None req-4b919e4f-1f0b-40e0-b482-0547e22d55ff tempest-ImagesTestJSON-618852556 tempest-ImagesTestJSON-618852556-project-member] [instance: 28e0baab-8516-42e3-acc2-9b8eb5192f57] Successfully updated port: 2be8aa59-fbf0-4081-955d-9e031781b864 {{(pid=61806) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 844.691179] env[61806]: DEBUG nova.network.neutron [None req-f9d7b4b0-fa22-4fc5-86c2-688ab650f83e tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] [instance: 2a3f81e3-b84f-4370-a598-2f6a607c3d47] Updating instance_info_cache with network_info: [{"id": "a839bcaa-9990-4bad-be1f-1f06eb1c978e", "address": "fa:16:3e:0b:8a:a6", "network": {"id": "a7ffda72-e57c-41fa-a7ed-d7b12084bc9d", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-62794865-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.193", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "6e391e1267014fafbb8f5a3211299819", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8b3cfeb1-f262-4fd9-b506-8e9c0733e2d8", "external-id": "nsx-vlan-transportzone-119", "segmentation_id": 119, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa839bcaa-99", "ovs_interfaceid": "a839bcaa-9990-4bad-be1f-1f06eb1c978e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "1f80e0d2-192f-4177-8212-acd6ca475851", "address": "fa:16:3e:aa:aa:6f", "network": {"id": "a7ffda72-e57c-41fa-a7ed-d7b12084bc9d", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-62794865-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "6e391e1267014fafbb8f5a3211299819", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8b3cfeb1-f262-4fd9-b506-8e9c0733e2d8", "external-id": "nsx-vlan-transportzone-119", "segmentation_id": 119, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1f80e0d2-19", "ovs_interfaceid": "1f80e0d2-192f-4177-8212-acd6ca475851", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61806) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 844.729950] env[61806]: DEBUG oslo_vmware.api [None req-334c6338-dc9d-4c32-ad77-bd3fe007216c tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Task: {'id': task-1294718, 'name': CopyVirtualDisk_Task} progress is 4%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 844.731641] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-45a5ea5e-bfad-4a0d-a4f8-f7d5a3ba3e7f {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 844.739083] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2d15213d-f8b2-4291-928f-70c790f2a52e {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 844.768983] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-771188bf-7ac1-4b9e-a544-960c7d3179ee {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 844.776816] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-db16a1a0-65a1-4a0b-ad2f-66bbc28ef250 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 844.793957] env[61806]: DEBUG nova.compute.provider_tree [None req-f4c6ec76-6b50-4a2c-bb86-09b71e318ce8 tempest-ServerShowV257Test-1159790492 tempest-ServerShowV257Test-1159790492-project-member] Inventory has not changed in ProviderTree for provider: a2858be1-fd22-4e08-979e-87ad25293407 {{(pid=61806) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 844.993069] env[61806]: DEBUG nova.network.neutron [-] [instance: c8033551-6591-4e37-ae78-4efe7145b10b] Updating instance_info_cache with network_info: [] {{(pid=61806) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 845.081348] env[61806]: DEBUG oslo_concurrency.lockutils [None req-4b919e4f-1f0b-40e0-b482-0547e22d55ff tempest-ImagesTestJSON-618852556 tempest-ImagesTestJSON-618852556-project-member] Acquiring lock "refresh_cache-28e0baab-8516-42e3-acc2-9b8eb5192f57" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 845.081762] env[61806]: DEBUG oslo_concurrency.lockutils [None req-4b919e4f-1f0b-40e0-b482-0547e22d55ff tempest-ImagesTestJSON-618852556 tempest-ImagesTestJSON-618852556-project-member] Acquired lock "refresh_cache-28e0baab-8516-42e3-acc2-9b8eb5192f57" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 845.081762] env[61806]: DEBUG nova.network.neutron [None req-4b919e4f-1f0b-40e0-b482-0547e22d55ff tempest-ImagesTestJSON-618852556 tempest-ImagesTestJSON-618852556-project-member] [instance: 28e0baab-8516-42e3-acc2-9b8eb5192f57] Building network info cache for instance {{(pid=61806) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 845.196245] env[61806]: DEBUG oslo_concurrency.lockutils [None req-f9d7b4b0-fa22-4fc5-86c2-688ab650f83e tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] Releasing lock "refresh_cache-2a3f81e3-b84f-4370-a598-2f6a607c3d47" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 845.196917] env[61806]: DEBUG oslo_concurrency.lockutils [None req-f9d7b4b0-fa22-4fc5-86c2-688ab650f83e tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] Acquiring lock "2a3f81e3-b84f-4370-a598-2f6a607c3d47" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 845.197101] env[61806]: DEBUG oslo_concurrency.lockutils [None req-f9d7b4b0-fa22-4fc5-86c2-688ab650f83e tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] Acquired lock "2a3f81e3-b84f-4370-a598-2f6a607c3d47" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 845.197968] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cb217fc5-08dc-4eb0-b862-407c4463a4c9 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 845.215177] env[61806]: DEBUG nova.virt.hardware [None req-f9d7b4b0-fa22-4fc5-86c2-688ab650f83e tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-15T18:05:10Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format=,created_at=,direct_url=,disk_format=,id=,min_disk=,min_ram=,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=61806) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 845.215451] env[61806]: DEBUG nova.virt.hardware [None req-f9d7b4b0-fa22-4fc5-86c2-688ab650f83e tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] Flavor limits 0:0:0 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 845.215687] env[61806]: DEBUG nova.virt.hardware [None req-f9d7b4b0-fa22-4fc5-86c2-688ab650f83e tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] Image limits 0:0:0 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 845.215958] env[61806]: DEBUG nova.virt.hardware [None req-f9d7b4b0-fa22-4fc5-86c2-688ab650f83e tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] Flavor pref 0:0:0 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 845.216183] env[61806]: DEBUG nova.virt.hardware [None req-f9d7b4b0-fa22-4fc5-86c2-688ab650f83e tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] Image pref 0:0:0 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 845.216450] env[61806]: DEBUG nova.virt.hardware [None req-f9d7b4b0-fa22-4fc5-86c2-688ab650f83e tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 845.216817] env[61806]: DEBUG nova.virt.hardware [None req-f9d7b4b0-fa22-4fc5-86c2-688ab650f83e tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61806) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 845.217056] env[61806]: DEBUG nova.virt.hardware [None req-f9d7b4b0-fa22-4fc5-86c2-688ab650f83e tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61806) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 845.217285] env[61806]: DEBUG nova.virt.hardware [None req-f9d7b4b0-fa22-4fc5-86c2-688ab650f83e tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] Got 1 possible topologies {{(pid=61806) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 845.217432] env[61806]: DEBUG nova.virt.hardware [None req-f9d7b4b0-fa22-4fc5-86c2-688ab650f83e tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61806) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 845.217638] env[61806]: DEBUG nova.virt.hardware [None req-f9d7b4b0-fa22-4fc5-86c2-688ab650f83e tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61806) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 845.223884] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-f9d7b4b0-fa22-4fc5-86c2-688ab650f83e tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] [instance: 2a3f81e3-b84f-4370-a598-2f6a607c3d47] Reconfiguring VM to attach interface {{(pid=61806) attach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1929}} [ 845.224544] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-2372b664-67ce-47b7-b1b8-c4643fb50067 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 845.243970] env[61806]: DEBUG oslo_vmware.api [None req-334c6338-dc9d-4c32-ad77-bd3fe007216c tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Task: {'id': task-1294718, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.987245} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 845.245076] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [None req-334c6338-dc9d-4c32-ad77-bd3fe007216c tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa.vmdk to [datastore2] be19fc44-ecbe-489c-9b6e-be8957bb1dd0/be19fc44-ecbe-489c-9b6e-be8957bb1dd0.vmdk {{(pid=61806) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 845.245299] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-334c6338-dc9d-4c32-ad77-bd3fe007216c tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] [instance: be19fc44-ecbe-489c-9b6e-be8957bb1dd0] Extending root virtual disk to 1048576 {{(pid=61806) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 845.245602] env[61806]: DEBUG oslo_vmware.api [None req-f9d7b4b0-fa22-4fc5-86c2-688ab650f83e tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] Waiting for the task: (returnval){ [ 845.245602] env[61806]: value = "task-1294719" [ 845.245602] env[61806]: _type = "Task" [ 845.245602] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 845.245791] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-b01b44de-9c38-40f0-a6bb-a7a829f8dbdb {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 845.254417] env[61806]: DEBUG oslo_vmware.api [None req-f9d7b4b0-fa22-4fc5-86c2-688ab650f83e tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] Task: {'id': task-1294719, 'name': ReconfigVM_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 845.255527] env[61806]: DEBUG oslo_vmware.api [None req-334c6338-dc9d-4c32-ad77-bd3fe007216c tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Waiting for the task: (returnval){ [ 845.255527] env[61806]: value = "task-1294720" [ 845.255527] env[61806]: _type = "Task" [ 845.255527] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 845.264064] env[61806]: DEBUG oslo_vmware.api [None req-334c6338-dc9d-4c32-ad77-bd3fe007216c tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Task: {'id': task-1294720, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 845.297353] env[61806]: DEBUG nova.scheduler.client.report [None req-f4c6ec76-6b50-4a2c-bb86-09b71e318ce8 tempest-ServerShowV257Test-1159790492 tempest-ServerShowV257Test-1159790492-project-member] Inventory has not changed for provider a2858be1-fd22-4e08-979e-87ad25293407 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 138, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61806) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 845.495607] env[61806]: INFO nova.compute.manager [-] [instance: c8033551-6591-4e37-ae78-4efe7145b10b] Took 1.38 seconds to deallocate network for instance. [ 845.613250] env[61806]: DEBUG nova.network.neutron [None req-4b919e4f-1f0b-40e0-b482-0547e22d55ff tempest-ImagesTestJSON-618852556 tempest-ImagesTestJSON-618852556-project-member] [instance: 28e0baab-8516-42e3-acc2-9b8eb5192f57] Instance cache missing network info. {{(pid=61806) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 845.753017] env[61806]: DEBUG nova.network.neutron [None req-4b919e4f-1f0b-40e0-b482-0547e22d55ff tempest-ImagesTestJSON-618852556 tempest-ImagesTestJSON-618852556-project-member] [instance: 28e0baab-8516-42e3-acc2-9b8eb5192f57] Updating instance_info_cache with network_info: [{"id": "2be8aa59-fbf0-4081-955d-9e031781b864", "address": "fa:16:3e:a9:79:bd", "network": {"id": "288ebc87-a30f-439b-9885-9cdda68a9591", "bridge": "br-int", "label": "tempest-ImagesTestJSON-914411285-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4b9aa91aff1d4008ac5096902b77f852", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8cbc9b8f-ce19-4262-bf4d-88cd4f259a1c", "external-id": "nsx-vlan-transportzone-630", "segmentation_id": 630, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2be8aa59-fb", "ovs_interfaceid": "2be8aa59-fbf0-4081-955d-9e031781b864", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61806) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 845.763051] env[61806]: DEBUG oslo_vmware.api [None req-f9d7b4b0-fa22-4fc5-86c2-688ab650f83e tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] Task: {'id': task-1294719, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 845.768036] env[61806]: DEBUG oslo_vmware.api [None req-334c6338-dc9d-4c32-ad77-bd3fe007216c tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Task: {'id': task-1294720, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.065527} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 845.768327] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-334c6338-dc9d-4c32-ad77-bd3fe007216c tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] [instance: be19fc44-ecbe-489c-9b6e-be8957bb1dd0] Extended root virtual disk {{(pid=61806) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 845.769189] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4507028e-c3d7-43e6-b520-b71bfdb5df07 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 845.792255] env[61806]: DEBUG nova.virt.vmwareapi.volumeops [None req-334c6338-dc9d-4c32-ad77-bd3fe007216c tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] [instance: be19fc44-ecbe-489c-9b6e-be8957bb1dd0] Reconfiguring VM instance instance-00000047 to attach disk [datastore2] be19fc44-ecbe-489c-9b6e-be8957bb1dd0/be19fc44-ecbe-489c-9b6e-be8957bb1dd0.vmdk or device None with type sparse {{(pid=61806) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 845.792835] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-17a9e7c7-ecde-4732-af04-698c59a1597e {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 845.807715] env[61806]: DEBUG oslo_concurrency.lockutils [None req-f4c6ec76-6b50-4a2c-bb86-09b71e318ce8 tempest-ServerShowV257Test-1159790492 tempest-ServerShowV257Test-1159790492-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.866s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 845.809875] env[61806]: DEBUG oslo_concurrency.lockutils [None req-0b05357b-24cd-44f2-8a41-86cdceeb41ee tempest-VolumesAdminNegativeTest-1387177911 tempest-VolumesAdminNegativeTest-1387177911-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 17.863s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 845.810128] env[61806]: DEBUG nova.objects.instance [None req-0b05357b-24cd-44f2-8a41-86cdceeb41ee tempest-VolumesAdminNegativeTest-1387177911 tempest-VolumesAdminNegativeTest-1387177911-project-member] Lazy-loading 'resources' on Instance uuid 4373b735-31cf-4b53-b655-38555cf212a5 {{(pid=61806) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 845.818274] env[61806]: DEBUG oslo_vmware.api [None req-334c6338-dc9d-4c32-ad77-bd3fe007216c tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Waiting for the task: (returnval){ [ 845.818274] env[61806]: value = "task-1294721" [ 845.818274] env[61806]: _type = "Task" [ 845.818274] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 845.827639] env[61806]: DEBUG oslo_vmware.api [None req-334c6338-dc9d-4c32-ad77-bd3fe007216c tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Task: {'id': task-1294721, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 845.828649] env[61806]: INFO nova.scheduler.client.report [None req-f4c6ec76-6b50-4a2c-bb86-09b71e318ce8 tempest-ServerShowV257Test-1159790492 tempest-ServerShowV257Test-1159790492-project-member] Deleted allocations for instance 616ec206-9804-469e-ab5c-41aea7f048aa [ 846.004480] env[61806]: DEBUG oslo_concurrency.lockutils [None req-40a350df-b9c4-4f7d-968f-0160bcbc0d06 tempest-ServerAddressesNegativeTestJSON-1526246390 tempest-ServerAddressesNegativeTestJSON-1526246390-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 846.257715] env[61806]: DEBUG oslo_concurrency.lockutils [None req-4b919e4f-1f0b-40e0-b482-0547e22d55ff tempest-ImagesTestJSON-618852556 tempest-ImagesTestJSON-618852556-project-member] Releasing lock "refresh_cache-28e0baab-8516-42e3-acc2-9b8eb5192f57" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 846.258054] env[61806]: DEBUG nova.compute.manager [None req-4b919e4f-1f0b-40e0-b482-0547e22d55ff tempest-ImagesTestJSON-618852556 tempest-ImagesTestJSON-618852556-project-member] [instance: 28e0baab-8516-42e3-acc2-9b8eb5192f57] Instance network_info: |[{"id": "2be8aa59-fbf0-4081-955d-9e031781b864", "address": "fa:16:3e:a9:79:bd", "network": {"id": "288ebc87-a30f-439b-9885-9cdda68a9591", "bridge": "br-int", "label": "tempest-ImagesTestJSON-914411285-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4b9aa91aff1d4008ac5096902b77f852", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8cbc9b8f-ce19-4262-bf4d-88cd4f259a1c", "external-id": "nsx-vlan-transportzone-630", "segmentation_id": 630, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2be8aa59-fb", "ovs_interfaceid": "2be8aa59-fbf0-4081-955d-9e031781b864", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61806) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 846.258308] env[61806]: DEBUG oslo_vmware.api [None req-f9d7b4b0-fa22-4fc5-86c2-688ab650f83e tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] Task: {'id': task-1294719, 'name': ReconfigVM_Task, 'duration_secs': 0.804802} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 846.258670] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-4b919e4f-1f0b-40e0-b482-0547e22d55ff tempest-ImagesTestJSON-618852556 tempest-ImagesTestJSON-618852556-project-member] [instance: 28e0baab-8516-42e3-acc2-9b8eb5192f57] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:a9:79:bd', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '8cbc9b8f-ce19-4262-bf4d-88cd4f259a1c', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '2be8aa59-fbf0-4081-955d-9e031781b864', 'vif_model': 'vmxnet3'}] {{(pid=61806) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 846.266506] env[61806]: DEBUG oslo.service.loopingcall [None req-4b919e4f-1f0b-40e0-b482-0547e22d55ff tempest-ImagesTestJSON-618852556 tempest-ImagesTestJSON-618852556-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61806) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 846.267032] env[61806]: DEBUG oslo_concurrency.lockutils [None req-f9d7b4b0-fa22-4fc5-86c2-688ab650f83e tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] Releasing lock "2a3f81e3-b84f-4370-a598-2f6a607c3d47" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 846.267242] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-f9d7b4b0-fa22-4fc5-86c2-688ab650f83e tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] [instance: 2a3f81e3-b84f-4370-a598-2f6a607c3d47] Reconfigured VM to attach interface {{(pid=61806) attach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1943}} [ 846.269671] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 28e0baab-8516-42e3-acc2-9b8eb5192f57] Creating VM on the ESX host {{(pid=61806) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 846.269904] env[61806]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-c0643f1d-b88a-4ebb-8711-533c7b4fd330 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 846.290817] env[61806]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 846.290817] env[61806]: value = "task-1294722" [ 846.290817] env[61806]: _type = "Task" [ 846.290817] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 846.298687] env[61806]: DEBUG oslo_vmware.api [-] Task: {'id': task-1294722, 'name': CreateVM_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 846.328543] env[61806]: DEBUG oslo_vmware.api [None req-334c6338-dc9d-4c32-ad77-bd3fe007216c tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Task: {'id': task-1294721, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 846.335898] env[61806]: DEBUG oslo_concurrency.lockutils [None req-f4c6ec76-6b50-4a2c-bb86-09b71e318ce8 tempest-ServerShowV257Test-1159790492 tempest-ServerShowV257Test-1159790492-project-member] Lock "616ec206-9804-469e-ab5c-41aea7f048aa" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 23.002s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 846.492447] env[61806]: DEBUG nova.compute.manager [req-7bf50f99-c0fe-4fed-9f5a-1c1e83a293e9 req-437d3a97-597e-4f7a-a495-9c3660c807ce service nova] [instance: 2a3f81e3-b84f-4370-a598-2f6a607c3d47] Received event network-vif-plugged-1f80e0d2-192f-4177-8212-acd6ca475851 {{(pid=61806) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 846.492447] env[61806]: DEBUG oslo_concurrency.lockutils [req-7bf50f99-c0fe-4fed-9f5a-1c1e83a293e9 req-437d3a97-597e-4f7a-a495-9c3660c807ce service nova] Acquiring lock "2a3f81e3-b84f-4370-a598-2f6a607c3d47-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 846.492447] env[61806]: DEBUG oslo_concurrency.lockutils [req-7bf50f99-c0fe-4fed-9f5a-1c1e83a293e9 req-437d3a97-597e-4f7a-a495-9c3660c807ce service nova] Lock "2a3f81e3-b84f-4370-a598-2f6a607c3d47-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 846.492612] env[61806]: DEBUG oslo_concurrency.lockutils [req-7bf50f99-c0fe-4fed-9f5a-1c1e83a293e9 req-437d3a97-597e-4f7a-a495-9c3660c807ce service nova] Lock "2a3f81e3-b84f-4370-a598-2f6a607c3d47-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 846.492910] env[61806]: DEBUG nova.compute.manager [req-7bf50f99-c0fe-4fed-9f5a-1c1e83a293e9 req-437d3a97-597e-4f7a-a495-9c3660c807ce service nova] [instance: 2a3f81e3-b84f-4370-a598-2f6a607c3d47] No waiting events found dispatching network-vif-plugged-1f80e0d2-192f-4177-8212-acd6ca475851 {{(pid=61806) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 846.493020] env[61806]: WARNING nova.compute.manager [req-7bf50f99-c0fe-4fed-9f5a-1c1e83a293e9 req-437d3a97-597e-4f7a-a495-9c3660c807ce service nova] [instance: 2a3f81e3-b84f-4370-a598-2f6a607c3d47] Received unexpected event network-vif-plugged-1f80e0d2-192f-4177-8212-acd6ca475851 for instance with vm_state active and task_state None. [ 846.493153] env[61806]: DEBUG nova.compute.manager [req-7bf50f99-c0fe-4fed-9f5a-1c1e83a293e9 req-437d3a97-597e-4f7a-a495-9c3660c807ce service nova] [instance: 2a3f81e3-b84f-4370-a598-2f6a607c3d47] Received event network-changed-1f80e0d2-192f-4177-8212-acd6ca475851 {{(pid=61806) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 846.493278] env[61806]: DEBUG nova.compute.manager [req-7bf50f99-c0fe-4fed-9f5a-1c1e83a293e9 req-437d3a97-597e-4f7a-a495-9c3660c807ce service nova] [instance: 2a3f81e3-b84f-4370-a598-2f6a607c3d47] Refreshing instance network info cache due to event network-changed-1f80e0d2-192f-4177-8212-acd6ca475851. {{(pid=61806) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 846.493463] env[61806]: DEBUG oslo_concurrency.lockutils [req-7bf50f99-c0fe-4fed-9f5a-1c1e83a293e9 req-437d3a97-597e-4f7a-a495-9c3660c807ce service nova] Acquiring lock "refresh_cache-2a3f81e3-b84f-4370-a598-2f6a607c3d47" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 846.493615] env[61806]: DEBUG oslo_concurrency.lockutils [req-7bf50f99-c0fe-4fed-9f5a-1c1e83a293e9 req-437d3a97-597e-4f7a-a495-9c3660c807ce service nova] Acquired lock "refresh_cache-2a3f81e3-b84f-4370-a598-2f6a607c3d47" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 846.493777] env[61806]: DEBUG nova.network.neutron [req-7bf50f99-c0fe-4fed-9f5a-1c1e83a293e9 req-437d3a97-597e-4f7a-a495-9c3660c807ce service nova] [instance: 2a3f81e3-b84f-4370-a598-2f6a607c3d47] Refreshing network info cache for port 1f80e0d2-192f-4177-8212-acd6ca475851 {{(pid=61806) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 846.550842] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-22387d79-013a-4e19-8ef2-ee76d4051d1b {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 846.559853] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-66693cc8-e3c3-45ab-9433-0701e7928df2 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 846.596536] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4f9cede0-602f-48b1-b964-dbdb06b027ff {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 846.604765] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-af44511f-0602-4482-bade-8699195f36a1 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 846.618716] env[61806]: DEBUG nova.compute.provider_tree [None req-0b05357b-24cd-44f2-8a41-86cdceeb41ee tempest-VolumesAdminNegativeTest-1387177911 tempest-VolumesAdminNegativeTest-1387177911-project-member] Inventory has not changed in ProviderTree for provider: a2858be1-fd22-4e08-979e-87ad25293407 {{(pid=61806) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 846.637677] env[61806]: DEBUG nova.compute.manager [req-9a5ff409-4bec-4d44-9c97-1ad20a32abe2 req-257ebf74-630c-4dd7-88ac-d509d4ce1dad service nova] [instance: 28e0baab-8516-42e3-acc2-9b8eb5192f57] Received event network-vif-plugged-2be8aa59-fbf0-4081-955d-9e031781b864 {{(pid=61806) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 846.637887] env[61806]: DEBUG oslo_concurrency.lockutils [req-9a5ff409-4bec-4d44-9c97-1ad20a32abe2 req-257ebf74-630c-4dd7-88ac-d509d4ce1dad service nova] Acquiring lock "28e0baab-8516-42e3-acc2-9b8eb5192f57-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 846.638109] env[61806]: DEBUG oslo_concurrency.lockutils [req-9a5ff409-4bec-4d44-9c97-1ad20a32abe2 req-257ebf74-630c-4dd7-88ac-d509d4ce1dad service nova] Lock "28e0baab-8516-42e3-acc2-9b8eb5192f57-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 846.638316] env[61806]: DEBUG oslo_concurrency.lockutils [req-9a5ff409-4bec-4d44-9c97-1ad20a32abe2 req-257ebf74-630c-4dd7-88ac-d509d4ce1dad service nova] Lock "28e0baab-8516-42e3-acc2-9b8eb5192f57-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 846.638454] env[61806]: DEBUG nova.compute.manager [req-9a5ff409-4bec-4d44-9c97-1ad20a32abe2 req-257ebf74-630c-4dd7-88ac-d509d4ce1dad service nova] [instance: 28e0baab-8516-42e3-acc2-9b8eb5192f57] No waiting events found dispatching network-vif-plugged-2be8aa59-fbf0-4081-955d-9e031781b864 {{(pid=61806) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 846.638945] env[61806]: WARNING nova.compute.manager [req-9a5ff409-4bec-4d44-9c97-1ad20a32abe2 req-257ebf74-630c-4dd7-88ac-d509d4ce1dad service nova] [instance: 28e0baab-8516-42e3-acc2-9b8eb5192f57] Received unexpected event network-vif-plugged-2be8aa59-fbf0-4081-955d-9e031781b864 for instance with vm_state building and task_state spawning. [ 846.786946] env[61806]: DEBUG oslo_concurrency.lockutils [None req-f9d7b4b0-fa22-4fc5-86c2-688ab650f83e tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] Lock "interface-2a3f81e3-b84f-4370-a598-2f6a607c3d47-None" "released" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: held 6.798s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 846.804176] env[61806]: DEBUG oslo_vmware.api [-] Task: {'id': task-1294722, 'name': CreateVM_Task, 'duration_secs': 0.471533} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 846.804398] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 28e0baab-8516-42e3-acc2-9b8eb5192f57] Created VM on the ESX host {{(pid=61806) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 846.805435] env[61806]: DEBUG oslo_concurrency.lockutils [None req-4b919e4f-1f0b-40e0-b482-0547e22d55ff tempest-ImagesTestJSON-618852556 tempest-ImagesTestJSON-618852556-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/e1cfdcc0-a515-4a37-bf5d-903e61ffb7f2" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 846.805863] env[61806]: DEBUG oslo_concurrency.lockutils [None req-4b919e4f-1f0b-40e0-b482-0547e22d55ff tempest-ImagesTestJSON-618852556 tempest-ImagesTestJSON-618852556-project-member] Acquired lock "[datastore2] devstack-image-cache_base/e1cfdcc0-a515-4a37-bf5d-903e61ffb7f2" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 846.806311] env[61806]: DEBUG oslo_concurrency.lockutils [None req-4b919e4f-1f0b-40e0-b482-0547e22d55ff tempest-ImagesTestJSON-618852556 tempest-ImagesTestJSON-618852556-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/e1cfdcc0-a515-4a37-bf5d-903e61ffb7f2" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 846.807468] env[61806]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-0ab10ad6-5e46-4e4d-95ff-e482875d6708 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 846.816589] env[61806]: DEBUG oslo_vmware.api [None req-4b919e4f-1f0b-40e0-b482-0547e22d55ff tempest-ImagesTestJSON-618852556 tempest-ImagesTestJSON-618852556-project-member] Waiting for the task: (returnval){ [ 846.816589] env[61806]: value = "session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]525ac78c-6af6-eb3e-e178-6319d7045bad" [ 846.816589] env[61806]: _type = "Task" [ 846.816589] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 846.823243] env[61806]: DEBUG oslo_vmware.api [None req-4b919e4f-1f0b-40e0-b482-0547e22d55ff tempest-ImagesTestJSON-618852556 tempest-ImagesTestJSON-618852556-project-member] Task: {'id': session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]525ac78c-6af6-eb3e-e178-6319d7045bad, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 846.831552] env[61806]: DEBUG oslo_vmware.api [None req-334c6338-dc9d-4c32-ad77-bd3fe007216c tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Task: {'id': task-1294721, 'name': ReconfigVM_Task, 'duration_secs': 0.563621} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 846.831552] env[61806]: DEBUG nova.virt.vmwareapi.volumeops [None req-334c6338-dc9d-4c32-ad77-bd3fe007216c tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] [instance: be19fc44-ecbe-489c-9b6e-be8957bb1dd0] Reconfigured VM instance instance-00000047 to attach disk [datastore2] be19fc44-ecbe-489c-9b6e-be8957bb1dd0/be19fc44-ecbe-489c-9b6e-be8957bb1dd0.vmdk or device None with type sparse {{(pid=61806) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 846.831711] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-a19f9f9f-fb1c-4f54-8f9d-f4237750d392 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 846.837927] env[61806]: DEBUG oslo_vmware.api [None req-334c6338-dc9d-4c32-ad77-bd3fe007216c tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Waiting for the task: (returnval){ [ 846.837927] env[61806]: value = "task-1294723" [ 846.837927] env[61806]: _type = "Task" [ 846.837927] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 846.846935] env[61806]: DEBUG oslo_vmware.api [None req-334c6338-dc9d-4c32-ad77-bd3fe007216c tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Task: {'id': task-1294723, 'name': Rename_Task} progress is 5%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 847.122145] env[61806]: DEBUG nova.scheduler.client.report [None req-0b05357b-24cd-44f2-8a41-86cdceeb41ee tempest-VolumesAdminNegativeTest-1387177911 tempest-VolumesAdminNegativeTest-1387177911-project-member] Inventory has not changed for provider a2858be1-fd22-4e08-979e-87ad25293407 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 138, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61806) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 847.209746] env[61806]: DEBUG oslo_concurrency.lockutils [None req-f19a9d0d-15e4-4e05-ab2c-e194a287f3a9 tempest-AttachInterfacesV270Test-1974652806 tempest-AttachInterfacesV270Test-1974652806-project-member] Acquiring lock "117139cd-4ed4-4bea-8ad3-f4ac32c692e4" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 847.209950] env[61806]: DEBUG oslo_concurrency.lockutils [None req-f19a9d0d-15e4-4e05-ab2c-e194a287f3a9 tempest-AttachInterfacesV270Test-1974652806 tempest-AttachInterfacesV270Test-1974652806-project-member] Lock "117139cd-4ed4-4bea-8ad3-f4ac32c692e4" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 847.210182] env[61806]: DEBUG oslo_concurrency.lockutils [None req-f19a9d0d-15e4-4e05-ab2c-e194a287f3a9 tempest-AttachInterfacesV270Test-1974652806 tempest-AttachInterfacesV270Test-1974652806-project-member] Acquiring lock "117139cd-4ed4-4bea-8ad3-f4ac32c692e4-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 847.213088] env[61806]: DEBUG oslo_concurrency.lockutils [None req-f19a9d0d-15e4-4e05-ab2c-e194a287f3a9 tempest-AttachInterfacesV270Test-1974652806 tempest-AttachInterfacesV270Test-1974652806-project-member] Lock "117139cd-4ed4-4bea-8ad3-f4ac32c692e4-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 847.213088] env[61806]: DEBUG oslo_concurrency.lockutils [None req-f19a9d0d-15e4-4e05-ab2c-e194a287f3a9 tempest-AttachInterfacesV270Test-1974652806 tempest-AttachInterfacesV270Test-1974652806-project-member] Lock "117139cd-4ed4-4bea-8ad3-f4ac32c692e4-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 847.213088] env[61806]: INFO nova.compute.manager [None req-f19a9d0d-15e4-4e05-ab2c-e194a287f3a9 tempest-AttachInterfacesV270Test-1974652806 tempest-AttachInterfacesV270Test-1974652806-project-member] [instance: 117139cd-4ed4-4bea-8ad3-f4ac32c692e4] Terminating instance [ 847.214632] env[61806]: DEBUG nova.compute.manager [None req-f19a9d0d-15e4-4e05-ab2c-e194a287f3a9 tempest-AttachInterfacesV270Test-1974652806 tempest-AttachInterfacesV270Test-1974652806-project-member] [instance: 117139cd-4ed4-4bea-8ad3-f4ac32c692e4] Start destroying the instance on the hypervisor. {{(pid=61806) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 847.214859] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-f19a9d0d-15e4-4e05-ab2c-e194a287f3a9 tempest-AttachInterfacesV270Test-1974652806 tempest-AttachInterfacesV270Test-1974652806-project-member] [instance: 117139cd-4ed4-4bea-8ad3-f4ac32c692e4] Destroying instance {{(pid=61806) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 847.215745] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-639cb800-7556-48b8-ac9f-7f3fed65a3ef {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 847.225550] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [None req-f19a9d0d-15e4-4e05-ab2c-e194a287f3a9 tempest-AttachInterfacesV270Test-1974652806 tempest-AttachInterfacesV270Test-1974652806-project-member] [instance: 117139cd-4ed4-4bea-8ad3-f4ac32c692e4] Powering off the VM {{(pid=61806) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 847.225853] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-e791283e-2f2a-4d6e-964a-10a68a06fef9 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 847.232243] env[61806]: DEBUG oslo_vmware.api [None req-f19a9d0d-15e4-4e05-ab2c-e194a287f3a9 tempest-AttachInterfacesV270Test-1974652806 tempest-AttachInterfacesV270Test-1974652806-project-member] Waiting for the task: (returnval){ [ 847.232243] env[61806]: value = "task-1294724" [ 847.232243] env[61806]: _type = "Task" [ 847.232243] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 847.241739] env[61806]: DEBUG oslo_vmware.api [None req-f19a9d0d-15e4-4e05-ab2c-e194a287f3a9 tempest-AttachInterfacesV270Test-1974652806 tempest-AttachInterfacesV270Test-1974652806-project-member] Task: {'id': task-1294724, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 847.271645] env[61806]: DEBUG nova.network.neutron [req-7bf50f99-c0fe-4fed-9f5a-1c1e83a293e9 req-437d3a97-597e-4f7a-a495-9c3660c807ce service nova] [instance: 2a3f81e3-b84f-4370-a598-2f6a607c3d47] Updated VIF entry in instance network info cache for port 1f80e0d2-192f-4177-8212-acd6ca475851. {{(pid=61806) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 847.272175] env[61806]: DEBUG nova.network.neutron [req-7bf50f99-c0fe-4fed-9f5a-1c1e83a293e9 req-437d3a97-597e-4f7a-a495-9c3660c807ce service nova] [instance: 2a3f81e3-b84f-4370-a598-2f6a607c3d47] Updating instance_info_cache with network_info: [{"id": "a839bcaa-9990-4bad-be1f-1f06eb1c978e", "address": "fa:16:3e:0b:8a:a6", "network": {"id": "a7ffda72-e57c-41fa-a7ed-d7b12084bc9d", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-62794865-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.193", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "6e391e1267014fafbb8f5a3211299819", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8b3cfeb1-f262-4fd9-b506-8e9c0733e2d8", "external-id": "nsx-vlan-transportzone-119", "segmentation_id": 119, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa839bcaa-99", "ovs_interfaceid": "a839bcaa-9990-4bad-be1f-1f06eb1c978e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "1f80e0d2-192f-4177-8212-acd6ca475851", "address": "fa:16:3e:aa:aa:6f", "network": {"id": "a7ffda72-e57c-41fa-a7ed-d7b12084bc9d", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-62794865-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "6e391e1267014fafbb8f5a3211299819", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8b3cfeb1-f262-4fd9-b506-8e9c0733e2d8", "external-id": "nsx-vlan-transportzone-119", "segmentation_id": 119, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1f80e0d2-19", "ovs_interfaceid": "1f80e0d2-192f-4177-8212-acd6ca475851", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61806) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 847.325029] env[61806]: DEBUG oslo_concurrency.lockutils [None req-4b919e4f-1f0b-40e0-b482-0547e22d55ff tempest-ImagesTestJSON-618852556 tempest-ImagesTestJSON-618852556-project-member] Releasing lock "[datastore2] devstack-image-cache_base/e1cfdcc0-a515-4a37-bf5d-903e61ffb7f2" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 847.325206] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-4b919e4f-1f0b-40e0-b482-0547e22d55ff tempest-ImagesTestJSON-618852556 tempest-ImagesTestJSON-618852556-project-member] [instance: 28e0baab-8516-42e3-acc2-9b8eb5192f57] Processing image e1cfdcc0-a515-4a37-bf5d-903e61ffb7f2 {{(pid=61806) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 847.325780] env[61806]: DEBUG oslo_concurrency.lockutils [None req-4b919e4f-1f0b-40e0-b482-0547e22d55ff tempest-ImagesTestJSON-618852556 tempest-ImagesTestJSON-618852556-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/e1cfdcc0-a515-4a37-bf5d-903e61ffb7f2/e1cfdcc0-a515-4a37-bf5d-903e61ffb7f2.vmdk" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 847.325780] env[61806]: DEBUG oslo_concurrency.lockutils [None req-4b919e4f-1f0b-40e0-b482-0547e22d55ff tempest-ImagesTestJSON-618852556 tempest-ImagesTestJSON-618852556-project-member] Acquired lock "[datastore2] devstack-image-cache_base/e1cfdcc0-a515-4a37-bf5d-903e61ffb7f2/e1cfdcc0-a515-4a37-bf5d-903e61ffb7f2.vmdk" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 847.325780] env[61806]: DEBUG nova.virt.vmwareapi.ds_util [None req-4b919e4f-1f0b-40e0-b482-0547e22d55ff tempest-ImagesTestJSON-618852556 tempest-ImagesTestJSON-618852556-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61806) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 847.326086] env[61806]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-5c31b654-7262-4911-873a-604cffda2480 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 847.337445] env[61806]: DEBUG nova.virt.vmwareapi.ds_util [None req-4b919e4f-1f0b-40e0-b482-0547e22d55ff tempest-ImagesTestJSON-618852556 tempest-ImagesTestJSON-618852556-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61806) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 847.337677] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-4b919e4f-1f0b-40e0-b482-0547e22d55ff tempest-ImagesTestJSON-618852556 tempest-ImagesTestJSON-618852556-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61806) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 847.338500] env[61806]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-6c2338f2-08ad-4fd0-b9d1-b8925b4cc24d {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 847.350275] env[61806]: DEBUG oslo_vmware.api [None req-4b919e4f-1f0b-40e0-b482-0547e22d55ff tempest-ImagesTestJSON-618852556 tempest-ImagesTestJSON-618852556-project-member] Waiting for the task: (returnval){ [ 847.350275] env[61806]: value = "session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]52a0e17f-9fd2-dee8-aeb9-c6364c18101f" [ 847.350275] env[61806]: _type = "Task" [ 847.350275] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 847.350275] env[61806]: DEBUG oslo_vmware.api [None req-334c6338-dc9d-4c32-ad77-bd3fe007216c tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Task: {'id': task-1294723, 'name': Rename_Task, 'duration_secs': 0.350014} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 847.352949] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [None req-334c6338-dc9d-4c32-ad77-bd3fe007216c tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] [instance: be19fc44-ecbe-489c-9b6e-be8957bb1dd0] Powering on the VM {{(pid=61806) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 847.353608] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-cf33533c-24d8-4f7c-bc5c-8fad55ba8490 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 847.359967] env[61806]: DEBUG oslo_vmware.api [None req-4b919e4f-1f0b-40e0-b482-0547e22d55ff tempest-ImagesTestJSON-618852556 tempest-ImagesTestJSON-618852556-project-member] Task: {'id': session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]52a0e17f-9fd2-dee8-aeb9-c6364c18101f, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 847.361532] env[61806]: DEBUG oslo_vmware.api [None req-334c6338-dc9d-4c32-ad77-bd3fe007216c tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Waiting for the task: (returnval){ [ 847.361532] env[61806]: value = "task-1294725" [ 847.361532] env[61806]: _type = "Task" [ 847.361532] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 847.368963] env[61806]: DEBUG oslo_vmware.api [None req-334c6338-dc9d-4c32-ad77-bd3fe007216c tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Task: {'id': task-1294725, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 847.627621] env[61806]: DEBUG oslo_concurrency.lockutils [None req-0b05357b-24cd-44f2-8a41-86cdceeb41ee tempest-VolumesAdminNegativeTest-1387177911 tempest-VolumesAdminNegativeTest-1387177911-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.818s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 847.630465] env[61806]: DEBUG oslo_concurrency.lockutils [None req-90367f73-05d7-4a77-94a7-20cc6ea5d878 tempest-ServerAddressesTestJSON-661730579 tempest-ServerAddressesTestJSON-661730579-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 18.570s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 847.630465] env[61806]: DEBUG nova.objects.instance [None req-90367f73-05d7-4a77-94a7-20cc6ea5d878 tempest-ServerAddressesTestJSON-661730579 tempest-ServerAddressesTestJSON-661730579-project-member] Lazy-loading 'resources' on Instance uuid a842df44-d8a9-4376-b9fc-5ca19a68a4b7 {{(pid=61806) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 847.656758] env[61806]: INFO nova.scheduler.client.report [None req-0b05357b-24cd-44f2-8a41-86cdceeb41ee tempest-VolumesAdminNegativeTest-1387177911 tempest-VolumesAdminNegativeTest-1387177911-project-member] Deleted allocations for instance 4373b735-31cf-4b53-b655-38555cf212a5 [ 847.742827] env[61806]: DEBUG oslo_vmware.api [None req-f19a9d0d-15e4-4e05-ab2c-e194a287f3a9 tempest-AttachInterfacesV270Test-1974652806 tempest-AttachInterfacesV270Test-1974652806-project-member] Task: {'id': task-1294724, 'name': PowerOffVM_Task, 'duration_secs': 0.345603} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 847.743147] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [None req-f19a9d0d-15e4-4e05-ab2c-e194a287f3a9 tempest-AttachInterfacesV270Test-1974652806 tempest-AttachInterfacesV270Test-1974652806-project-member] [instance: 117139cd-4ed4-4bea-8ad3-f4ac32c692e4] Powered off the VM {{(pid=61806) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 847.743331] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-f19a9d0d-15e4-4e05-ab2c-e194a287f3a9 tempest-AttachInterfacesV270Test-1974652806 tempest-AttachInterfacesV270Test-1974652806-project-member] [instance: 117139cd-4ed4-4bea-8ad3-f4ac32c692e4] Unregistering the VM {{(pid=61806) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 847.743613] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-77e0a473-b588-478b-affd-a82ada219686 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 847.774579] env[61806]: DEBUG oslo_concurrency.lockutils [req-7bf50f99-c0fe-4fed-9f5a-1c1e83a293e9 req-437d3a97-597e-4f7a-a495-9c3660c807ce service nova] Releasing lock "refresh_cache-2a3f81e3-b84f-4370-a598-2f6a607c3d47" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 847.813018] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-f19a9d0d-15e4-4e05-ab2c-e194a287f3a9 tempest-AttachInterfacesV270Test-1974652806 tempest-AttachInterfacesV270Test-1974652806-project-member] [instance: 117139cd-4ed4-4bea-8ad3-f4ac32c692e4] Unregistered the VM {{(pid=61806) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 847.813279] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-f19a9d0d-15e4-4e05-ab2c-e194a287f3a9 tempest-AttachInterfacesV270Test-1974652806 tempest-AttachInterfacesV270Test-1974652806-project-member] [instance: 117139cd-4ed4-4bea-8ad3-f4ac32c692e4] Deleting contents of the VM from datastore datastore2 {{(pid=61806) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 847.813532] env[61806]: DEBUG nova.virt.vmwareapi.ds_util [None req-f19a9d0d-15e4-4e05-ab2c-e194a287f3a9 tempest-AttachInterfacesV270Test-1974652806 tempest-AttachInterfacesV270Test-1974652806-project-member] Deleting the datastore file [datastore2] 117139cd-4ed4-4bea-8ad3-f4ac32c692e4 {{(pid=61806) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 847.813915] env[61806]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-f5d9875c-f688-481a-8b0e-9ddf14a703f7 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 847.820320] env[61806]: DEBUG oslo_vmware.api [None req-f19a9d0d-15e4-4e05-ab2c-e194a287f3a9 tempest-AttachInterfacesV270Test-1974652806 tempest-AttachInterfacesV270Test-1974652806-project-member] Waiting for the task: (returnval){ [ 847.820320] env[61806]: value = "task-1294727" [ 847.820320] env[61806]: _type = "Task" [ 847.820320] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 847.828836] env[61806]: DEBUG oslo_vmware.api [None req-f19a9d0d-15e4-4e05-ab2c-e194a287f3a9 tempest-AttachInterfacesV270Test-1974652806 tempest-AttachInterfacesV270Test-1974652806-project-member] Task: {'id': task-1294727, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 847.860784] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-4b919e4f-1f0b-40e0-b482-0547e22d55ff tempest-ImagesTestJSON-618852556 tempest-ImagesTestJSON-618852556-project-member] [instance: 28e0baab-8516-42e3-acc2-9b8eb5192f57] Preparing fetch location {{(pid=61806) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:633}} [ 847.861102] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-4b919e4f-1f0b-40e0-b482-0547e22d55ff tempest-ImagesTestJSON-618852556 tempest-ImagesTestJSON-618852556-project-member] [instance: 28e0baab-8516-42e3-acc2-9b8eb5192f57] Fetch image to [datastore2] OSTACK_IMG_cdc60310-6efe-499a-8b76-fbc7cd9f40d4/OSTACK_IMG_cdc60310-6efe-499a-8b76-fbc7cd9f40d4.vmdk {{(pid=61806) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:635}} [ 847.861342] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-4b919e4f-1f0b-40e0-b482-0547e22d55ff tempest-ImagesTestJSON-618852556 tempest-ImagesTestJSON-618852556-project-member] [instance: 28e0baab-8516-42e3-acc2-9b8eb5192f57] Downloading stream optimized image e1cfdcc0-a515-4a37-bf5d-903e61ffb7f2 to [datastore2] OSTACK_IMG_cdc60310-6efe-499a-8b76-fbc7cd9f40d4/OSTACK_IMG_cdc60310-6efe-499a-8b76-fbc7cd9f40d4.vmdk on the data store datastore2 as vApp {{(pid=61806) _fetch_image_as_vapp /opt/stack/nova/nova/virt/vmwareapi/vmops.py:437}} [ 847.861780] env[61806]: DEBUG nova.virt.vmwareapi.images [None req-4b919e4f-1f0b-40e0-b482-0547e22d55ff tempest-ImagesTestJSON-618852556 tempest-ImagesTestJSON-618852556-project-member] [instance: 28e0baab-8516-42e3-acc2-9b8eb5192f57] Downloading image file data e1cfdcc0-a515-4a37-bf5d-903e61ffb7f2 to the ESX as VM named 'OSTACK_IMG_cdc60310-6efe-499a-8b76-fbc7cd9f40d4' {{(pid=61806) fetch_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:336}} [ 847.872025] env[61806]: DEBUG oslo_vmware.api [None req-334c6338-dc9d-4c32-ad77-bd3fe007216c tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Task: {'id': task-1294725, 'name': PowerOnVM_Task} progress is 88%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 847.947753] env[61806]: DEBUG oslo_vmware.rw_handles [None req-4b919e4f-1f0b-40e0-b482-0547e22d55ff tempest-ImagesTestJSON-618852556 tempest-ImagesTestJSON-618852556-project-member] Creating HttpNfcLease lease for vApp import into resource pool: (val){ [ 847.947753] env[61806]: value = "resgroup-9" [ 847.947753] env[61806]: _type = "ResourcePool" [ 847.947753] env[61806]: }. {{(pid=61806) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:453}} [ 847.948411] env[61806]: DEBUG oslo_vmware.service [-] Invoking ResourcePool.ImportVApp with opID=oslo.vmware-ce881490-6dba-4d26-a340-d4a3410e7751 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 847.975512] env[61806]: DEBUG oslo_vmware.rw_handles [None req-4b919e4f-1f0b-40e0-b482-0547e22d55ff tempest-ImagesTestJSON-618852556 tempest-ImagesTestJSON-618852556-project-member] Lease: (returnval){ [ 847.975512] env[61806]: value = "session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]523ae947-d9da-4c70-c594-a9dc1f5a4e38" [ 847.975512] env[61806]: _type = "HttpNfcLease" [ 847.975512] env[61806]: } obtained for vApp import into resource pool (val){ [ 847.975512] env[61806]: value = "resgroup-9" [ 847.975512] env[61806]: _type = "ResourcePool" [ 847.975512] env[61806]: }. {{(pid=61806) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:461}} [ 847.975812] env[61806]: DEBUG oslo_vmware.api [None req-4b919e4f-1f0b-40e0-b482-0547e22d55ff tempest-ImagesTestJSON-618852556 tempest-ImagesTestJSON-618852556-project-member] Waiting for the lease: (returnval){ [ 847.975812] env[61806]: value = "session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]523ae947-d9da-4c70-c594-a9dc1f5a4e38" [ 847.975812] env[61806]: _type = "HttpNfcLease" [ 847.975812] env[61806]: } to be ready. {{(pid=61806) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 847.984818] env[61806]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 847.984818] env[61806]: value = "session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]523ae947-d9da-4c70-c594-a9dc1f5a4e38" [ 847.984818] env[61806]: _type = "HttpNfcLease" [ 847.984818] env[61806]: } is initializing. {{(pid=61806) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 848.167696] env[61806]: DEBUG oslo_concurrency.lockutils [None req-0b05357b-24cd-44f2-8a41-86cdceeb41ee tempest-VolumesAdminNegativeTest-1387177911 tempest-VolumesAdminNegativeTest-1387177911-project-member] Lock "4373b735-31cf-4b53-b655-38555cf212a5" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 23.438s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 848.333021] env[61806]: DEBUG oslo_vmware.api [None req-f19a9d0d-15e4-4e05-ab2c-e194a287f3a9 tempest-AttachInterfacesV270Test-1974652806 tempest-AttachInterfacesV270Test-1974652806-project-member] Task: {'id': task-1294727, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.187722} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 848.333120] env[61806]: DEBUG nova.virt.vmwareapi.ds_util [None req-f19a9d0d-15e4-4e05-ab2c-e194a287f3a9 tempest-AttachInterfacesV270Test-1974652806 tempest-AttachInterfacesV270Test-1974652806-project-member] Deleted the datastore file {{(pid=61806) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 848.333284] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-f19a9d0d-15e4-4e05-ab2c-e194a287f3a9 tempest-AttachInterfacesV270Test-1974652806 tempest-AttachInterfacesV270Test-1974652806-project-member] [instance: 117139cd-4ed4-4bea-8ad3-f4ac32c692e4] Deleted contents of the VM from datastore datastore2 {{(pid=61806) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 848.333987] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-f19a9d0d-15e4-4e05-ab2c-e194a287f3a9 tempest-AttachInterfacesV270Test-1974652806 tempest-AttachInterfacesV270Test-1974652806-project-member] [instance: 117139cd-4ed4-4bea-8ad3-f4ac32c692e4] Instance destroyed {{(pid=61806) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 848.333987] env[61806]: INFO nova.compute.manager [None req-f19a9d0d-15e4-4e05-ab2c-e194a287f3a9 tempest-AttachInterfacesV270Test-1974652806 tempest-AttachInterfacesV270Test-1974652806-project-member] [instance: 117139cd-4ed4-4bea-8ad3-f4ac32c692e4] Took 1.12 seconds to destroy the instance on the hypervisor. [ 848.333987] env[61806]: DEBUG oslo.service.loopingcall [None req-f19a9d0d-15e4-4e05-ab2c-e194a287f3a9 tempest-AttachInterfacesV270Test-1974652806 tempest-AttachInterfacesV270Test-1974652806-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61806) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 848.334238] env[61806]: DEBUG nova.compute.manager [-] [instance: 117139cd-4ed4-4bea-8ad3-f4ac32c692e4] Deallocating network for instance {{(pid=61806) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 848.334280] env[61806]: DEBUG nova.network.neutron [-] [instance: 117139cd-4ed4-4bea-8ad3-f4ac32c692e4] deallocate_for_instance() {{(pid=61806) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 848.337368] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-428606a7-6147-4f7d-97a4-5114ebaeca01 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 848.344038] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-80361167-0d22-4312-9d57-ca4046349501 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 848.379760] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-51307cb6-c414-4e92-a6c9-591515f382d9 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 848.387806] env[61806]: DEBUG oslo_vmware.api [None req-334c6338-dc9d-4c32-ad77-bd3fe007216c tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Task: {'id': task-1294725, 'name': PowerOnVM_Task, 'duration_secs': 0.667184} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 848.390145] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [None req-334c6338-dc9d-4c32-ad77-bd3fe007216c tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] [instance: be19fc44-ecbe-489c-9b6e-be8957bb1dd0] Powered on the VM {{(pid=61806) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 848.390389] env[61806]: INFO nova.compute.manager [None req-334c6338-dc9d-4c32-ad77-bd3fe007216c tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] [instance: be19fc44-ecbe-489c-9b6e-be8957bb1dd0] Took 9.64 seconds to spawn the instance on the hypervisor. [ 848.390582] env[61806]: DEBUG nova.compute.manager [None req-334c6338-dc9d-4c32-ad77-bd3fe007216c tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] [instance: be19fc44-ecbe-489c-9b6e-be8957bb1dd0] Checking state {{(pid=61806) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 848.391633] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-75f0d5fb-375b-4217-b0ac-e09dad019812 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 848.395305] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aa266607-0ea2-4b3e-a5ad-5ef1639bb87c {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 848.416142] env[61806]: DEBUG nova.compute.provider_tree [None req-90367f73-05d7-4a77-94a7-20cc6ea5d878 tempest-ServerAddressesTestJSON-661730579 tempest-ServerAddressesTestJSON-661730579-project-member] Inventory has not changed in ProviderTree for provider: a2858be1-fd22-4e08-979e-87ad25293407 {{(pid=61806) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 848.471742] env[61806]: DEBUG oslo_concurrency.lockutils [None req-d93d1274-43bb-49c2-a2cd-82e2c5e1853d tempest-ServerMetadataNegativeTestJSON-456453902 tempest-ServerMetadataNegativeTestJSON-456453902-project-member] Acquiring lock "7f4907d0-178d-452a-8149-030becde8779" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 848.472027] env[61806]: DEBUG oslo_concurrency.lockutils [None req-d93d1274-43bb-49c2-a2cd-82e2c5e1853d tempest-ServerMetadataNegativeTestJSON-456453902 tempest-ServerMetadataNegativeTestJSON-456453902-project-member] Lock "7f4907d0-178d-452a-8149-030becde8779" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 848.472274] env[61806]: DEBUG oslo_concurrency.lockutils [None req-d93d1274-43bb-49c2-a2cd-82e2c5e1853d tempest-ServerMetadataNegativeTestJSON-456453902 tempest-ServerMetadataNegativeTestJSON-456453902-project-member] Acquiring lock "7f4907d0-178d-452a-8149-030becde8779-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 848.472470] env[61806]: DEBUG oslo_concurrency.lockutils [None req-d93d1274-43bb-49c2-a2cd-82e2c5e1853d tempest-ServerMetadataNegativeTestJSON-456453902 tempest-ServerMetadataNegativeTestJSON-456453902-project-member] Lock "7f4907d0-178d-452a-8149-030becde8779-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 848.472711] env[61806]: DEBUG oslo_concurrency.lockutils [None req-d93d1274-43bb-49c2-a2cd-82e2c5e1853d tempest-ServerMetadataNegativeTestJSON-456453902 tempest-ServerMetadataNegativeTestJSON-456453902-project-member] Lock "7f4907d0-178d-452a-8149-030becde8779-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 848.475835] env[61806]: INFO nova.compute.manager [None req-d93d1274-43bb-49c2-a2cd-82e2c5e1853d tempest-ServerMetadataNegativeTestJSON-456453902 tempest-ServerMetadataNegativeTestJSON-456453902-project-member] [instance: 7f4907d0-178d-452a-8149-030becde8779] Terminating instance [ 848.481548] env[61806]: DEBUG nova.compute.manager [None req-d93d1274-43bb-49c2-a2cd-82e2c5e1853d tempest-ServerMetadataNegativeTestJSON-456453902 tempest-ServerMetadataNegativeTestJSON-456453902-project-member] [instance: 7f4907d0-178d-452a-8149-030becde8779] Start destroying the instance on the hypervisor. {{(pid=61806) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 848.481758] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-d93d1274-43bb-49c2-a2cd-82e2c5e1853d tempest-ServerMetadataNegativeTestJSON-456453902 tempest-ServerMetadataNegativeTestJSON-456453902-project-member] [instance: 7f4907d0-178d-452a-8149-030becde8779] Destroying instance {{(pid=61806) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 848.482742] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f1a96831-19c2-40f1-bb9d-2a440ea35dc5 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 848.496191] env[61806]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 848.496191] env[61806]: value = "session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]523ae947-d9da-4c70-c594-a9dc1f5a4e38" [ 848.496191] env[61806]: _type = "HttpNfcLease" [ 848.496191] env[61806]: } is initializing. {{(pid=61806) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 848.498705] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [None req-d93d1274-43bb-49c2-a2cd-82e2c5e1853d tempest-ServerMetadataNegativeTestJSON-456453902 tempest-ServerMetadataNegativeTestJSON-456453902-project-member] [instance: 7f4907d0-178d-452a-8149-030becde8779] Powering off the VM {{(pid=61806) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 848.499031] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-faf28f8e-7b58-4c0b-95a9-1aae7eca6599 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 848.505131] env[61806]: DEBUG oslo_vmware.api [None req-d93d1274-43bb-49c2-a2cd-82e2c5e1853d tempest-ServerMetadataNegativeTestJSON-456453902 tempest-ServerMetadataNegativeTestJSON-456453902-project-member] Waiting for the task: (returnval){ [ 848.505131] env[61806]: value = "task-1294729" [ 848.505131] env[61806]: _type = "Task" [ 848.505131] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 848.515146] env[61806]: DEBUG oslo_vmware.api [None req-d93d1274-43bb-49c2-a2cd-82e2c5e1853d tempest-ServerMetadataNegativeTestJSON-456453902 tempest-ServerMetadataNegativeTestJSON-456453902-project-member] Task: {'id': task-1294729, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 848.672182] env[61806]: DEBUG oslo_concurrency.lockutils [None req-fbc8cbed-1b00-40c8-a788-e6e317b3c0cb tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] Acquiring lock "interface-2a3f81e3-b84f-4370-a598-2f6a607c3d47-64b56aa1-c5e3-4093-acee-b75f7ead331e" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 848.672889] env[61806]: DEBUG oslo_concurrency.lockutils [None req-fbc8cbed-1b00-40c8-a788-e6e317b3c0cb tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] Lock "interface-2a3f81e3-b84f-4370-a598-2f6a607c3d47-64b56aa1-c5e3-4093-acee-b75f7ead331e" acquired by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: waited 0.001s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 848.673489] env[61806]: DEBUG nova.objects.instance [None req-fbc8cbed-1b00-40c8-a788-e6e317b3c0cb tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] Lazy-loading 'flavor' on Instance uuid 2a3f81e3-b84f-4370-a598-2f6a607c3d47 {{(pid=61806) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 848.927941] env[61806]: DEBUG nova.scheduler.client.report [None req-90367f73-05d7-4a77-94a7-20cc6ea5d878 tempest-ServerAddressesTestJSON-661730579 tempest-ServerAddressesTestJSON-661730579-project-member] Inventory has not changed for provider a2858be1-fd22-4e08-979e-87ad25293407 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 138, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61806) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 848.934060] env[61806]: INFO nova.compute.manager [None req-334c6338-dc9d-4c32-ad77-bd3fe007216c tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] [instance: be19fc44-ecbe-489c-9b6e-be8957bb1dd0] Took 26.48 seconds to build instance. [ 848.986195] env[61806]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 848.986195] env[61806]: value = "session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]523ae947-d9da-4c70-c594-a9dc1f5a4e38" [ 848.986195] env[61806]: _type = "HttpNfcLease" [ 848.986195] env[61806]: } is ready. {{(pid=61806) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 848.986897] env[61806]: DEBUG oslo_vmware.rw_handles [None req-4b919e4f-1f0b-40e0-b482-0547e22d55ff tempest-ImagesTestJSON-618852556 tempest-ImagesTestJSON-618852556-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 848.986897] env[61806]: value = "session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]523ae947-d9da-4c70-c594-a9dc1f5a4e38" [ 848.986897] env[61806]: _type = "HttpNfcLease" [ 848.986897] env[61806]: }. {{(pid=61806) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:467}} [ 848.987926] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4602721b-3f58-4f2a-9ab6-f14b68ea7c05 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 848.996732] env[61806]: DEBUG oslo_vmware.rw_handles [None req-4b919e4f-1f0b-40e0-b482-0547e22d55ff tempest-ImagesTestJSON-618852556 tempest-ImagesTestJSON-618852556-project-member] Found VMDK URL: https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52c8ccf8-2d4a-aaf7-d0cd-1cb8021f4093/disk-0.vmdk from lease info. {{(pid=61806) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 848.997140] env[61806]: DEBUG oslo_vmware.rw_handles [None req-4b919e4f-1f0b-40e0-b482-0547e22d55ff tempest-ImagesTestJSON-618852556 tempest-ImagesTestJSON-618852556-project-member] Creating HTTP connection to write to file with size = 21334016 and URL = https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52c8ccf8-2d4a-aaf7-d0cd-1cb8021f4093/disk-0.vmdk. {{(pid=61806) _create_write_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:122}} [ 849.070812] env[61806]: DEBUG nova.compute.manager [req-694e8208-f309-43ea-a8a9-5486d89a55dc req-dcaa7ce0-7757-4f52-aa4a-97efd1d9b94d service nova] [instance: c8033551-6591-4e37-ae78-4efe7145b10b] Received event network-vif-deleted-e8bfbf94-495b-4d13-8d2c-7a02e7a90387 {{(pid=61806) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 849.071061] env[61806]: DEBUG nova.compute.manager [req-694e8208-f309-43ea-a8a9-5486d89a55dc req-dcaa7ce0-7757-4f52-aa4a-97efd1d9b94d service nova] [instance: 28e0baab-8516-42e3-acc2-9b8eb5192f57] Received event network-changed-2be8aa59-fbf0-4081-955d-9e031781b864 {{(pid=61806) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 849.071247] env[61806]: DEBUG nova.compute.manager [req-694e8208-f309-43ea-a8a9-5486d89a55dc req-dcaa7ce0-7757-4f52-aa4a-97efd1d9b94d service nova] [instance: 28e0baab-8516-42e3-acc2-9b8eb5192f57] Refreshing instance network info cache due to event network-changed-2be8aa59-fbf0-4081-955d-9e031781b864. {{(pid=61806) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 849.071463] env[61806]: DEBUG oslo_concurrency.lockutils [req-694e8208-f309-43ea-a8a9-5486d89a55dc req-dcaa7ce0-7757-4f52-aa4a-97efd1d9b94d service nova] Acquiring lock "refresh_cache-28e0baab-8516-42e3-acc2-9b8eb5192f57" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 849.071601] env[61806]: DEBUG oslo_concurrency.lockutils [req-694e8208-f309-43ea-a8a9-5486d89a55dc req-dcaa7ce0-7757-4f52-aa4a-97efd1d9b94d service nova] Acquired lock "refresh_cache-28e0baab-8516-42e3-acc2-9b8eb5192f57" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 849.071761] env[61806]: DEBUG nova.network.neutron [req-694e8208-f309-43ea-a8a9-5486d89a55dc req-dcaa7ce0-7757-4f52-aa4a-97efd1d9b94d service nova] [instance: 28e0baab-8516-42e3-acc2-9b8eb5192f57] Refreshing network info cache for port 2be8aa59-fbf0-4081-955d-9e031781b864 {{(pid=61806) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 849.074204] env[61806]: DEBUG nova.compute.manager [req-0e9fc532-b997-40c0-98ac-029997d6a7d4 req-9f5130ee-5f7a-45ce-8688-6bd78b8b0e28 service nova] [instance: 117139cd-4ed4-4bea-8ad3-f4ac32c692e4] Received event network-vif-deleted-f984af48-8d6f-48b1-a3b3-1f988c36b872 {{(pid=61806) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 849.074398] env[61806]: INFO nova.compute.manager [req-0e9fc532-b997-40c0-98ac-029997d6a7d4 req-9f5130ee-5f7a-45ce-8688-6bd78b8b0e28 service nova] [instance: 117139cd-4ed4-4bea-8ad3-f4ac32c692e4] Neutron deleted interface f984af48-8d6f-48b1-a3b3-1f988c36b872; detaching it from the instance and deleting it from the info cache [ 849.074574] env[61806]: DEBUG nova.network.neutron [req-0e9fc532-b997-40c0-98ac-029997d6a7d4 req-9f5130ee-5f7a-45ce-8688-6bd78b8b0e28 service nova] [instance: 117139cd-4ed4-4bea-8ad3-f4ac32c692e4] Updating instance_info_cache with network_info: [] {{(pid=61806) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 849.085418] env[61806]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-6c01fffe-f6ee-4511-b255-e727cc23baf5 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 849.088559] env[61806]: DEBUG oslo_vmware.api [None req-d93d1274-43bb-49c2-a2cd-82e2c5e1853d tempest-ServerMetadataNegativeTestJSON-456453902 tempest-ServerMetadataNegativeTestJSON-456453902-project-member] Task: {'id': task-1294729, 'name': PowerOffVM_Task, 'duration_secs': 0.323316} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 849.090594] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [None req-d93d1274-43bb-49c2-a2cd-82e2c5e1853d tempest-ServerMetadataNegativeTestJSON-456453902 tempest-ServerMetadataNegativeTestJSON-456453902-project-member] [instance: 7f4907d0-178d-452a-8149-030becde8779] Powered off the VM {{(pid=61806) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 849.090792] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-d93d1274-43bb-49c2-a2cd-82e2c5e1853d tempest-ServerMetadataNegativeTestJSON-456453902 tempest-ServerMetadataNegativeTestJSON-456453902-project-member] [instance: 7f4907d0-178d-452a-8149-030becde8779] Unregistering the VM {{(pid=61806) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 849.092233] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-01b65db0-4ac1-40a4-9d50-f473f2fab099 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 849.164016] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-d93d1274-43bb-49c2-a2cd-82e2c5e1853d tempest-ServerMetadataNegativeTestJSON-456453902 tempest-ServerMetadataNegativeTestJSON-456453902-project-member] [instance: 7f4907d0-178d-452a-8149-030becde8779] Unregistered the VM {{(pid=61806) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 849.164248] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-d93d1274-43bb-49c2-a2cd-82e2c5e1853d tempest-ServerMetadataNegativeTestJSON-456453902 tempest-ServerMetadataNegativeTestJSON-456453902-project-member] [instance: 7f4907d0-178d-452a-8149-030becde8779] Deleting contents of the VM from datastore datastore2 {{(pid=61806) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 849.164444] env[61806]: DEBUG nova.virt.vmwareapi.ds_util [None req-d93d1274-43bb-49c2-a2cd-82e2c5e1853d tempest-ServerMetadataNegativeTestJSON-456453902 tempest-ServerMetadataNegativeTestJSON-456453902-project-member] Deleting the datastore file [datastore2] 7f4907d0-178d-452a-8149-030becde8779 {{(pid=61806) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 849.164714] env[61806]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-28e4f6b0-1e94-4c3e-b36f-4afc2e9d0088 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 849.168485] env[61806]: DEBUG nova.network.neutron [-] [instance: 117139cd-4ed4-4bea-8ad3-f4ac32c692e4] Updating instance_info_cache with network_info: [] {{(pid=61806) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 849.183224] env[61806]: DEBUG oslo_vmware.api [None req-d93d1274-43bb-49c2-a2cd-82e2c5e1853d tempest-ServerMetadataNegativeTestJSON-456453902 tempest-ServerMetadataNegativeTestJSON-456453902-project-member] Waiting for the task: (returnval){ [ 849.183224] env[61806]: value = "task-1294731" [ 849.183224] env[61806]: _type = "Task" [ 849.183224] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 849.198217] env[61806]: DEBUG oslo_vmware.api [None req-d93d1274-43bb-49c2-a2cd-82e2c5e1853d tempest-ServerMetadataNegativeTestJSON-456453902 tempest-ServerMetadataNegativeTestJSON-456453902-project-member] Task: {'id': task-1294731, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 849.437054] env[61806]: DEBUG oslo_concurrency.lockutils [None req-90367f73-05d7-4a77-94a7-20cc6ea5d878 tempest-ServerAddressesTestJSON-661730579 tempest-ServerAddressesTestJSON-661730579-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.805s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 849.438058] env[61806]: DEBUG oslo_concurrency.lockutils [None req-334c6338-dc9d-4c32-ad77-bd3fe007216c tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Lock "be19fc44-ecbe-489c-9b6e-be8957bb1dd0" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 42.467s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 849.438868] env[61806]: DEBUG oslo_concurrency.lockutils [None req-fd7f690e-61c7-4f8c-a053-2b904c459e5b tempest-SecurityGroupsTestJSON-1213485581 tempest-SecurityGroupsTestJSON-1213485581-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 8.791s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 849.439313] env[61806]: DEBUG nova.objects.instance [None req-fd7f690e-61c7-4f8c-a053-2b904c459e5b tempest-SecurityGroupsTestJSON-1213485581 tempest-SecurityGroupsTestJSON-1213485581-project-member] Lazy-loading 'resources' on Instance uuid b41531f2-f28c-4d82-9682-0b557bbaa491 {{(pid=61806) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 849.445441] env[61806]: DEBUG nova.objects.instance [None req-fbc8cbed-1b00-40c8-a788-e6e317b3c0cb tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] Lazy-loading 'pci_requests' on Instance uuid 2a3f81e3-b84f-4370-a598-2f6a607c3d47 {{(pid=61806) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 849.472786] env[61806]: INFO nova.scheduler.client.report [None req-90367f73-05d7-4a77-94a7-20cc6ea5d878 tempest-ServerAddressesTestJSON-661730579 tempest-ServerAddressesTestJSON-661730579-project-member] Deleted allocations for instance a842df44-d8a9-4376-b9fc-5ca19a68a4b7 [ 849.582788] env[61806]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-1dc78012-08ed-4ece-8aab-6e4449a84d5e {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 849.589936] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e9ceb196-6234-42bf-827a-c564ebb0510b {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 849.620634] env[61806]: DEBUG nova.compute.manager [req-0e9fc532-b997-40c0-98ac-029997d6a7d4 req-9f5130ee-5f7a-45ce-8688-6bd78b8b0e28 service nova] [instance: 117139cd-4ed4-4bea-8ad3-f4ac32c692e4] Detach interface failed, port_id=f984af48-8d6f-48b1-a3b3-1f988c36b872, reason: Instance 117139cd-4ed4-4bea-8ad3-f4ac32c692e4 could not be found. {{(pid=61806) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10963}} [ 849.672944] env[61806]: INFO nova.compute.manager [-] [instance: 117139cd-4ed4-4bea-8ad3-f4ac32c692e4] Took 1.34 seconds to deallocate network for instance. [ 849.677198] env[61806]: DEBUG oslo_concurrency.lockutils [None req-64647cce-deb3-4f3d-927e-8bb03c4dc9ee tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Acquiring lock "be19fc44-ecbe-489c-9b6e-be8957bb1dd0" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 849.677198] env[61806]: DEBUG oslo_concurrency.lockutils [None req-64647cce-deb3-4f3d-927e-8bb03c4dc9ee tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Lock "be19fc44-ecbe-489c-9b6e-be8957bb1dd0" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 849.677198] env[61806]: DEBUG oslo_concurrency.lockutils [None req-64647cce-deb3-4f3d-927e-8bb03c4dc9ee tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Acquiring lock "be19fc44-ecbe-489c-9b6e-be8957bb1dd0-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 849.677198] env[61806]: DEBUG oslo_concurrency.lockutils [None req-64647cce-deb3-4f3d-927e-8bb03c4dc9ee tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Lock "be19fc44-ecbe-489c-9b6e-be8957bb1dd0-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 849.677198] env[61806]: DEBUG oslo_concurrency.lockutils [None req-64647cce-deb3-4f3d-927e-8bb03c4dc9ee tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Lock "be19fc44-ecbe-489c-9b6e-be8957bb1dd0-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 849.682444] env[61806]: INFO nova.compute.manager [None req-64647cce-deb3-4f3d-927e-8bb03c4dc9ee tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] [instance: be19fc44-ecbe-489c-9b6e-be8957bb1dd0] Terminating instance [ 849.689609] env[61806]: DEBUG nova.compute.manager [None req-64647cce-deb3-4f3d-927e-8bb03c4dc9ee tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] [instance: be19fc44-ecbe-489c-9b6e-be8957bb1dd0] Start destroying the instance on the hypervisor. {{(pid=61806) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 849.690078] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-64647cce-deb3-4f3d-927e-8bb03c4dc9ee tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] [instance: be19fc44-ecbe-489c-9b6e-be8957bb1dd0] Destroying instance {{(pid=61806) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 849.691127] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-49b0862a-b394-4c85-a991-72b97e8959c1 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 849.701559] env[61806]: DEBUG oslo_vmware.api [None req-d93d1274-43bb-49c2-a2cd-82e2c5e1853d tempest-ServerMetadataNegativeTestJSON-456453902 tempest-ServerMetadataNegativeTestJSON-456453902-project-member] Task: {'id': task-1294731, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.260089} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 849.703999] env[61806]: DEBUG nova.virt.vmwareapi.ds_util [None req-d93d1274-43bb-49c2-a2cd-82e2c5e1853d tempest-ServerMetadataNegativeTestJSON-456453902 tempest-ServerMetadataNegativeTestJSON-456453902-project-member] Deleted the datastore file {{(pid=61806) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 849.704354] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-d93d1274-43bb-49c2-a2cd-82e2c5e1853d tempest-ServerMetadataNegativeTestJSON-456453902 tempest-ServerMetadataNegativeTestJSON-456453902-project-member] [instance: 7f4907d0-178d-452a-8149-030becde8779] Deleted contents of the VM from datastore datastore2 {{(pid=61806) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 849.704674] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-d93d1274-43bb-49c2-a2cd-82e2c5e1853d tempest-ServerMetadataNegativeTestJSON-456453902 tempest-ServerMetadataNegativeTestJSON-456453902-project-member] [instance: 7f4907d0-178d-452a-8149-030becde8779] Instance destroyed {{(pid=61806) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 849.705091] env[61806]: INFO nova.compute.manager [None req-d93d1274-43bb-49c2-a2cd-82e2c5e1853d tempest-ServerMetadataNegativeTestJSON-456453902 tempest-ServerMetadataNegativeTestJSON-456453902-project-member] [instance: 7f4907d0-178d-452a-8149-030becde8779] Took 1.22 seconds to destroy the instance on the hypervisor. [ 849.705541] env[61806]: DEBUG oslo.service.loopingcall [None req-d93d1274-43bb-49c2-a2cd-82e2c5e1853d tempest-ServerMetadataNegativeTestJSON-456453902 tempest-ServerMetadataNegativeTestJSON-456453902-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61806) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 849.706024] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [None req-64647cce-deb3-4f3d-927e-8bb03c4dc9ee tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] [instance: be19fc44-ecbe-489c-9b6e-be8957bb1dd0] Powering off the VM {{(pid=61806) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 849.706353] env[61806]: DEBUG nova.compute.manager [-] [instance: 7f4907d0-178d-452a-8149-030becde8779] Deallocating network for instance {{(pid=61806) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 849.706548] env[61806]: DEBUG nova.network.neutron [-] [instance: 7f4907d0-178d-452a-8149-030becde8779] deallocate_for_instance() {{(pid=61806) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 849.708374] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-622c3d70-2a74-4bda-8568-9bb81ee3f783 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 849.717257] env[61806]: DEBUG oslo_vmware.api [None req-64647cce-deb3-4f3d-927e-8bb03c4dc9ee tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Waiting for the task: (returnval){ [ 849.717257] env[61806]: value = "task-1294732" [ 849.717257] env[61806]: _type = "Task" [ 849.717257] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 849.725197] env[61806]: DEBUG oslo_vmware.api [None req-64647cce-deb3-4f3d-927e-8bb03c4dc9ee tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Task: {'id': task-1294732, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 849.955413] env[61806]: DEBUG nova.objects.base [None req-fbc8cbed-1b00-40c8-a788-e6e317b3c0cb tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] Object Instance<2a3f81e3-b84f-4370-a598-2f6a607c3d47> lazy-loaded attributes: flavor,pci_requests {{(pid=61806) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 849.955612] env[61806]: DEBUG nova.network.neutron [None req-fbc8cbed-1b00-40c8-a788-e6e317b3c0cb tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] [instance: 2a3f81e3-b84f-4370-a598-2f6a607c3d47] allocate_for_instance() {{(pid=61806) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 849.985281] env[61806]: DEBUG oslo_concurrency.lockutils [None req-90367f73-05d7-4a77-94a7-20cc6ea5d878 tempest-ServerAddressesTestJSON-661730579 tempest-ServerAddressesTestJSON-661730579-project-member] Lock "a842df44-d8a9-4376-b9fc-5ca19a68a4b7" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 24.759s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 850.179582] env[61806]: DEBUG oslo_vmware.rw_handles [None req-4b919e4f-1f0b-40e0-b482-0547e22d55ff tempest-ImagesTestJSON-618852556 tempest-ImagesTestJSON-618852556-project-member] Completed reading data from the image iterator. {{(pid=61806) read /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:765}} [ 850.179582] env[61806]: DEBUG oslo_vmware.rw_handles [None req-4b919e4f-1f0b-40e0-b482-0547e22d55ff tempest-ImagesTestJSON-618852556 tempest-ImagesTestJSON-618852556-project-member] Getting lease state for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52c8ccf8-2d4a-aaf7-d0cd-1cb8021f4093/disk-0.vmdk. {{(pid=61806) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 850.180353] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b4abac41-f211-4774-8e9f-9a6c559339d0 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 850.189926] env[61806]: DEBUG nova.network.neutron [req-694e8208-f309-43ea-a8a9-5486d89a55dc req-dcaa7ce0-7757-4f52-aa4a-97efd1d9b94d service nova] [instance: 28e0baab-8516-42e3-acc2-9b8eb5192f57] Updated VIF entry in instance network info cache for port 2be8aa59-fbf0-4081-955d-9e031781b864. {{(pid=61806) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 850.190493] env[61806]: DEBUG nova.network.neutron [req-694e8208-f309-43ea-a8a9-5486d89a55dc req-dcaa7ce0-7757-4f52-aa4a-97efd1d9b94d service nova] [instance: 28e0baab-8516-42e3-acc2-9b8eb5192f57] Updating instance_info_cache with network_info: [{"id": "2be8aa59-fbf0-4081-955d-9e031781b864", "address": "fa:16:3e:a9:79:bd", "network": {"id": "288ebc87-a30f-439b-9885-9cdda68a9591", "bridge": "br-int", "label": "tempest-ImagesTestJSON-914411285-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4b9aa91aff1d4008ac5096902b77f852", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8cbc9b8f-ce19-4262-bf4d-88cd4f259a1c", "external-id": "nsx-vlan-transportzone-630", "segmentation_id": 630, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2be8aa59-fb", "ovs_interfaceid": "2be8aa59-fbf0-4081-955d-9e031781b864", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61806) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 850.193226] env[61806]: DEBUG oslo_concurrency.lockutils [None req-f19a9d0d-15e4-4e05-ab2c-e194a287f3a9 tempest-AttachInterfacesV270Test-1974652806 tempest-AttachInterfacesV270Test-1974652806-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 850.199376] env[61806]: DEBUG oslo_vmware.rw_handles [None req-4b919e4f-1f0b-40e0-b482-0547e22d55ff tempest-ImagesTestJSON-618852556 tempest-ImagesTestJSON-618852556-project-member] Lease for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52c8ccf8-2d4a-aaf7-d0cd-1cb8021f4093/disk-0.vmdk is in state: ready. {{(pid=61806) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 850.199376] env[61806]: DEBUG oslo_vmware.rw_handles [None req-4b919e4f-1f0b-40e0-b482-0547e22d55ff tempest-ImagesTestJSON-618852556 tempest-ImagesTestJSON-618852556-project-member] Releasing lease for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52c8ccf8-2d4a-aaf7-d0cd-1cb8021f4093/disk-0.vmdk. {{(pid=61806) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:440}} [ 850.199376] env[61806]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseComplete with opID=oslo.vmware-4b37d549-ee2e-42d4-83ea-469f186c17a9 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 850.229637] env[61806]: DEBUG oslo_vmware.api [None req-64647cce-deb3-4f3d-927e-8bb03c4dc9ee tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Task: {'id': task-1294732, 'name': PowerOffVM_Task, 'duration_secs': 0.29212} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 850.232586] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [None req-64647cce-deb3-4f3d-927e-8bb03c4dc9ee tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] [instance: be19fc44-ecbe-489c-9b6e-be8957bb1dd0] Powered off the VM {{(pid=61806) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 850.232850] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-64647cce-deb3-4f3d-927e-8bb03c4dc9ee tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] [instance: be19fc44-ecbe-489c-9b6e-be8957bb1dd0] Unregistering the VM {{(pid=61806) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 850.236314] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-bdcea928-2f2c-49bd-b3e9-dd0b90324be9 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 850.282930] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7e213508-9a48-4351-802f-a7ccefd9df8d {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 850.293610] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d1bef650-5d13-4e4d-8c47-8412278f538c {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 850.330931] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-236dc15f-6861-4142-8bf6-05914fbc9402 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 850.333764] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-64647cce-deb3-4f3d-927e-8bb03c4dc9ee tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] [instance: be19fc44-ecbe-489c-9b6e-be8957bb1dd0] Unregistered the VM {{(pid=61806) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 850.334027] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-64647cce-deb3-4f3d-927e-8bb03c4dc9ee tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] [instance: be19fc44-ecbe-489c-9b6e-be8957bb1dd0] Deleting contents of the VM from datastore datastore2 {{(pid=61806) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 850.334227] env[61806]: DEBUG nova.virt.vmwareapi.ds_util [None req-64647cce-deb3-4f3d-927e-8bb03c4dc9ee tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Deleting the datastore file [datastore2] be19fc44-ecbe-489c-9b6e-be8957bb1dd0 {{(pid=61806) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 850.334499] env[61806]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-eac6d4b1-b427-4d9c-9558-fc435618c862 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 850.341789] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4138e806-9775-4d0e-99b9-5a8e9e193fc8 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 850.347176] env[61806]: DEBUG oslo_vmware.api [None req-64647cce-deb3-4f3d-927e-8bb03c4dc9ee tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Waiting for the task: (returnval){ [ 850.347176] env[61806]: value = "task-1294734" [ 850.347176] env[61806]: _type = "Task" [ 850.347176] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 850.359478] env[61806]: DEBUG nova.compute.provider_tree [None req-fd7f690e-61c7-4f8c-a053-2b904c459e5b tempest-SecurityGroupsTestJSON-1213485581 tempest-SecurityGroupsTestJSON-1213485581-project-member] Inventory has not changed in ProviderTree for provider: a2858be1-fd22-4e08-979e-87ad25293407 {{(pid=61806) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 850.368772] env[61806]: DEBUG oslo_vmware.api [None req-64647cce-deb3-4f3d-927e-8bb03c4dc9ee tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Task: {'id': task-1294734, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 850.537797] env[61806]: DEBUG nova.network.neutron [-] [instance: 7f4907d0-178d-452a-8149-030becde8779] Updating instance_info_cache with network_info: [] {{(pid=61806) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 850.617415] env[61806]: DEBUG oslo_vmware.rw_handles [None req-4b919e4f-1f0b-40e0-b482-0547e22d55ff tempest-ImagesTestJSON-618852556 tempest-ImagesTestJSON-618852556-project-member] Closed VMDK write handle for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52c8ccf8-2d4a-aaf7-d0cd-1cb8021f4093/disk-0.vmdk. {{(pid=61806) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:646}} [ 850.621025] env[61806]: INFO nova.virt.vmwareapi.images [None req-4b919e4f-1f0b-40e0-b482-0547e22d55ff tempest-ImagesTestJSON-618852556 tempest-ImagesTestJSON-618852556-project-member] [instance: 28e0baab-8516-42e3-acc2-9b8eb5192f57] Downloaded image file data e1cfdcc0-a515-4a37-bf5d-903e61ffb7f2 [ 850.621025] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ed8b89df-9baa-459a-94e3-d8ad7232ad59 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 850.642788] env[61806]: DEBUG nova.policy [None req-fbc8cbed-1b00-40c8-a788-e6e317b3c0cb tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'b1eac9072ccb4b3bbd23f240e941a76e', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '6e391e1267014fafbb8f5a3211299819', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61806) authorize /opt/stack/nova/nova/policy.py:201}} [ 850.644820] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-46e3ba68-4ca9-45da-a12f-4b9d3946c583 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 850.669980] env[61806]: INFO nova.virt.vmwareapi.images [None req-4b919e4f-1f0b-40e0-b482-0547e22d55ff tempest-ImagesTestJSON-618852556 tempest-ImagesTestJSON-618852556-project-member] [instance: 28e0baab-8516-42e3-acc2-9b8eb5192f57] The imported VM was unregistered [ 850.672463] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-4b919e4f-1f0b-40e0-b482-0547e22d55ff tempest-ImagesTestJSON-618852556 tempest-ImagesTestJSON-618852556-project-member] [instance: 28e0baab-8516-42e3-acc2-9b8eb5192f57] Caching image {{(pid=61806) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:638}} [ 850.672698] env[61806]: DEBUG nova.virt.vmwareapi.ds_util [None req-4b919e4f-1f0b-40e0-b482-0547e22d55ff tempest-ImagesTestJSON-618852556 tempest-ImagesTestJSON-618852556-project-member] Creating directory with path [datastore2] devstack-image-cache_base/e1cfdcc0-a515-4a37-bf5d-903e61ffb7f2 {{(pid=61806) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 850.673141] env[61806]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-c952365d-0fb2-408d-8e8a-792075394614 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 850.694859] env[61806]: DEBUG oslo_concurrency.lockutils [req-694e8208-f309-43ea-a8a9-5486d89a55dc req-dcaa7ce0-7757-4f52-aa4a-97efd1d9b94d service nova] Releasing lock "refresh_cache-28e0baab-8516-42e3-acc2-9b8eb5192f57" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 850.705710] env[61806]: DEBUG nova.virt.vmwareapi.ds_util [None req-4b919e4f-1f0b-40e0-b482-0547e22d55ff tempest-ImagesTestJSON-618852556 tempest-ImagesTestJSON-618852556-project-member] Created directory with path [datastore2] devstack-image-cache_base/e1cfdcc0-a515-4a37-bf5d-903e61ffb7f2 {{(pid=61806) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 850.705919] env[61806]: DEBUG nova.virt.vmwareapi.ds_util [None req-4b919e4f-1f0b-40e0-b482-0547e22d55ff tempest-ImagesTestJSON-618852556 tempest-ImagesTestJSON-618852556-project-member] Moving virtual disk from [datastore2] OSTACK_IMG_cdc60310-6efe-499a-8b76-fbc7cd9f40d4/OSTACK_IMG_cdc60310-6efe-499a-8b76-fbc7cd9f40d4.vmdk to [datastore2] devstack-image-cache_base/e1cfdcc0-a515-4a37-bf5d-903e61ffb7f2/e1cfdcc0-a515-4a37-bf5d-903e61ffb7f2.vmdk. {{(pid=61806) disk_move /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:263}} [ 850.706208] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.MoveVirtualDisk_Task with opID=oslo.vmware-b9296051-d834-4fcd-906d-cb9e4f107c8c {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 850.714470] env[61806]: DEBUG oslo_vmware.api [None req-4b919e4f-1f0b-40e0-b482-0547e22d55ff tempest-ImagesTestJSON-618852556 tempest-ImagesTestJSON-618852556-project-member] Waiting for the task: (returnval){ [ 850.714470] env[61806]: value = "task-1294736" [ 850.714470] env[61806]: _type = "Task" [ 850.714470] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 850.723660] env[61806]: DEBUG oslo_vmware.api [None req-4b919e4f-1f0b-40e0-b482-0547e22d55ff tempest-ImagesTestJSON-618852556 tempest-ImagesTestJSON-618852556-project-member] Task: {'id': task-1294736, 'name': MoveVirtualDisk_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 850.858251] env[61806]: DEBUG oslo_vmware.api [None req-64647cce-deb3-4f3d-927e-8bb03c4dc9ee tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Task: {'id': task-1294734, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.29246} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 850.858532] env[61806]: DEBUG nova.virt.vmwareapi.ds_util [None req-64647cce-deb3-4f3d-927e-8bb03c4dc9ee tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Deleted the datastore file {{(pid=61806) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 850.858724] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-64647cce-deb3-4f3d-927e-8bb03c4dc9ee tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] [instance: be19fc44-ecbe-489c-9b6e-be8957bb1dd0] Deleted contents of the VM from datastore datastore2 {{(pid=61806) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 850.858909] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-64647cce-deb3-4f3d-927e-8bb03c4dc9ee tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] [instance: be19fc44-ecbe-489c-9b6e-be8957bb1dd0] Instance destroyed {{(pid=61806) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 850.859105] env[61806]: INFO nova.compute.manager [None req-64647cce-deb3-4f3d-927e-8bb03c4dc9ee tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] [instance: be19fc44-ecbe-489c-9b6e-be8957bb1dd0] Took 1.17 seconds to destroy the instance on the hypervisor. [ 850.859354] env[61806]: DEBUG oslo.service.loopingcall [None req-64647cce-deb3-4f3d-927e-8bb03c4dc9ee tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61806) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 850.859554] env[61806]: DEBUG nova.compute.manager [-] [instance: be19fc44-ecbe-489c-9b6e-be8957bb1dd0] Deallocating network for instance {{(pid=61806) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 850.859642] env[61806]: DEBUG nova.network.neutron [-] [instance: be19fc44-ecbe-489c-9b6e-be8957bb1dd0] deallocate_for_instance() {{(pid=61806) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 850.866340] env[61806]: DEBUG nova.scheduler.client.report [None req-fd7f690e-61c7-4f8c-a053-2b904c459e5b tempest-SecurityGroupsTestJSON-1213485581 tempest-SecurityGroupsTestJSON-1213485581-project-member] Inventory has not changed for provider a2858be1-fd22-4e08-979e-87ad25293407 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 138, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61806) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 851.042041] env[61806]: INFO nova.compute.manager [-] [instance: 7f4907d0-178d-452a-8149-030becde8779] Took 1.33 seconds to deallocate network for instance. [ 851.230522] env[61806]: DEBUG oslo_vmware.api [None req-4b919e4f-1f0b-40e0-b482-0547e22d55ff tempest-ImagesTestJSON-618852556 tempest-ImagesTestJSON-618852556-project-member] Task: {'id': task-1294736, 'name': MoveVirtualDisk_Task} progress is 9%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 851.371169] env[61806]: DEBUG oslo_concurrency.lockutils [None req-fd7f690e-61c7-4f8c-a053-2b904c459e5b tempest-SecurityGroupsTestJSON-1213485581 tempest-SecurityGroupsTestJSON-1213485581-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.932s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 851.376577] env[61806]: DEBUG oslo_concurrency.lockutils [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 10.354s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 851.376887] env[61806]: DEBUG oslo_concurrency.lockutils [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 851.376967] env[61806]: DEBUG nova.compute.resource_tracker [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=61806) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 851.377336] env[61806]: DEBUG oslo_concurrency.lockutils [None req-3bd56cbd-68df-4151-a169-676f8bdda9ab tempest-MultipleCreateTestJSON-1589015462 tempest-MultipleCreateTestJSON-1589015462-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 9.103s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 851.377577] env[61806]: DEBUG nova.objects.instance [None req-3bd56cbd-68df-4151-a169-676f8bdda9ab tempest-MultipleCreateTestJSON-1589015462 tempest-MultipleCreateTestJSON-1589015462-project-member] Lazy-loading 'resources' on Instance uuid 25218e83-c1ab-4b97-ade1-7c32b6f99305 {{(pid=61806) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 851.383785] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a32793af-75b2-420a-8f58-d53616c3b998 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 851.395746] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ab2c9ff5-0aa7-4f9c-a237-1e98dd8286a7 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 851.420983] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cef49d29-b64b-4ad8-830d-0cfddfa34d14 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 851.426312] env[61806]: INFO nova.scheduler.client.report [None req-fd7f690e-61c7-4f8c-a053-2b904c459e5b tempest-SecurityGroupsTestJSON-1213485581 tempest-SecurityGroupsTestJSON-1213485581-project-member] Deleted allocations for instance b41531f2-f28c-4d82-9682-0b557bbaa491 [ 851.438782] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bbf840dd-1c9f-48bf-b2f3-a31bae0b9d5d {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 851.478995] env[61806]: DEBUG nova.compute.resource_tracker [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=181506MB free_disk=138GB free_vcpus=48 pci_devices=None {{(pid=61806) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 851.479179] env[61806]: DEBUG oslo_concurrency.lockutils [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 851.482706] env[61806]: DEBUG nova.compute.manager [req-8c4505bf-869b-41de-9b99-0a6ad97a2251 req-5e9c93ac-af8b-4949-b594-16db1e05f0b6 service nova] [instance: 7f4907d0-178d-452a-8149-030becde8779] Received event network-vif-deleted-718d0c6d-5962-4b66-a13f-963aa96b2985 {{(pid=61806) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 851.550387] env[61806]: DEBUG oslo_concurrency.lockutils [None req-d93d1274-43bb-49c2-a2cd-82e2c5e1853d tempest-ServerMetadataNegativeTestJSON-456453902 tempest-ServerMetadataNegativeTestJSON-456453902-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 851.713666] env[61806]: DEBUG nova.compute.manager [req-66e2a2a2-58e6-48f1-beea-45f6bf80f0be req-0751fce0-a23b-44e9-8e6c-72d7e45dd60c service nova] [instance: be19fc44-ecbe-489c-9b6e-be8957bb1dd0] Received event network-vif-deleted-6b41a0ca-4d2e-4ab1-9893-180a2f0febca {{(pid=61806) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 851.713718] env[61806]: INFO nova.compute.manager [req-66e2a2a2-58e6-48f1-beea-45f6bf80f0be req-0751fce0-a23b-44e9-8e6c-72d7e45dd60c service nova] [instance: be19fc44-ecbe-489c-9b6e-be8957bb1dd0] Neutron deleted interface 6b41a0ca-4d2e-4ab1-9893-180a2f0febca; detaching it from the instance and deleting it from the info cache [ 851.713952] env[61806]: DEBUG nova.network.neutron [req-66e2a2a2-58e6-48f1-beea-45f6bf80f0be req-0751fce0-a23b-44e9-8e6c-72d7e45dd60c service nova] [instance: be19fc44-ecbe-489c-9b6e-be8957bb1dd0] Updating instance_info_cache with network_info: [] {{(pid=61806) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 851.726846] env[61806]: DEBUG oslo_vmware.api [None req-4b919e4f-1f0b-40e0-b482-0547e22d55ff tempest-ImagesTestJSON-618852556 tempest-ImagesTestJSON-618852556-project-member] Task: {'id': task-1294736, 'name': MoveVirtualDisk_Task} progress is 26%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 851.938605] env[61806]: DEBUG oslo_concurrency.lockutils [None req-fd7f690e-61c7-4f8c-a053-2b904c459e5b tempest-SecurityGroupsTestJSON-1213485581 tempest-SecurityGroupsTestJSON-1213485581-project-member] Lock "b41531f2-f28c-4d82-9682-0b557bbaa491" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 15.189s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 851.990703] env[61806]: DEBUG nova.network.neutron [-] [instance: be19fc44-ecbe-489c-9b6e-be8957bb1dd0] Updating instance_info_cache with network_info: [] {{(pid=61806) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 852.099119] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d8c80105-5931-410e-ac50-97070bbb58db {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 852.108326] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-24bc4e7f-7e4a-4693-af00-588b09f1e274 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 852.143989] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3c36fbd1-1b01-40df-96ae-ecd0b7474574 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 852.152797] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1c9b45f9-3dae-4d64-b7b9-a04d7826911f {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 852.169504] env[61806]: DEBUG nova.compute.provider_tree [None req-3bd56cbd-68df-4151-a169-676f8bdda9ab tempest-MultipleCreateTestJSON-1589015462 tempest-MultipleCreateTestJSON-1589015462-project-member] Inventory has not changed in ProviderTree for provider: a2858be1-fd22-4e08-979e-87ad25293407 {{(pid=61806) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 852.216967] env[61806]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-376c7d61-d979-4c1b-89b9-48f257b2c23a {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 852.230065] env[61806]: DEBUG oslo_vmware.api [None req-4b919e4f-1f0b-40e0-b482-0547e22d55ff tempest-ImagesTestJSON-618852556 tempest-ImagesTestJSON-618852556-project-member] Task: {'id': task-1294736, 'name': MoveVirtualDisk_Task} progress is 46%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 852.235989] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-27c367fd-9d4f-4fda-8b70-6f623859b835 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 852.267022] env[61806]: DEBUG nova.compute.manager [req-66e2a2a2-58e6-48f1-beea-45f6bf80f0be req-0751fce0-a23b-44e9-8e6c-72d7e45dd60c service nova] [instance: be19fc44-ecbe-489c-9b6e-be8957bb1dd0] Detach interface failed, port_id=6b41a0ca-4d2e-4ab1-9893-180a2f0febca, reason: Instance be19fc44-ecbe-489c-9b6e-be8957bb1dd0 could not be found. {{(pid=61806) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10963}} [ 852.495910] env[61806]: INFO nova.compute.manager [-] [instance: be19fc44-ecbe-489c-9b6e-be8957bb1dd0] Took 1.63 seconds to deallocate network for instance. [ 852.673114] env[61806]: DEBUG nova.scheduler.client.report [None req-3bd56cbd-68df-4151-a169-676f8bdda9ab tempest-MultipleCreateTestJSON-1589015462 tempest-MultipleCreateTestJSON-1589015462-project-member] Inventory has not changed for provider a2858be1-fd22-4e08-979e-87ad25293407 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 138, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61806) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 852.729818] env[61806]: DEBUG oslo_vmware.api [None req-4b919e4f-1f0b-40e0-b482-0547e22d55ff tempest-ImagesTestJSON-618852556 tempest-ImagesTestJSON-618852556-project-member] Task: {'id': task-1294736, 'name': MoveVirtualDisk_Task} progress is 69%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 853.001922] env[61806]: DEBUG oslo_concurrency.lockutils [None req-64647cce-deb3-4f3d-927e-8bb03c4dc9ee tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 853.082841] env[61806]: DEBUG nova.network.neutron [None req-fbc8cbed-1b00-40c8-a788-e6e317b3c0cb tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] [instance: 2a3f81e3-b84f-4370-a598-2f6a607c3d47] Successfully updated port: 64b56aa1-c5e3-4093-acee-b75f7ead331e {{(pid=61806) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 853.184830] env[61806]: DEBUG oslo_concurrency.lockutils [None req-3bd56cbd-68df-4151-a169-676f8bdda9ab tempest-MultipleCreateTestJSON-1589015462 tempest-MultipleCreateTestJSON-1589015462-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.807s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 853.187912] env[61806]: DEBUG oslo_concurrency.lockutils [None req-182a714a-9a7d-40bc-ad59-30e74e8a3863 tempest-MultipleCreateTestJSON-1589015462 tempest-MultipleCreateTestJSON-1589015462-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 10.626s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 853.189021] env[61806]: DEBUG nova.objects.instance [None req-182a714a-9a7d-40bc-ad59-30e74e8a3863 tempest-MultipleCreateTestJSON-1589015462 tempest-MultipleCreateTestJSON-1589015462-project-member] Lazy-loading 'resources' on Instance uuid ef502a7d-7dc0-4319-8ce3-6a8701288237 {{(pid=61806) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 853.223828] env[61806]: INFO nova.scheduler.client.report [None req-3bd56cbd-68df-4151-a169-676f8bdda9ab tempest-MultipleCreateTestJSON-1589015462 tempest-MultipleCreateTestJSON-1589015462-project-member] Deleted allocations for instance 25218e83-c1ab-4b97-ade1-7c32b6f99305 [ 853.244454] env[61806]: DEBUG oslo_vmware.api [None req-4b919e4f-1f0b-40e0-b482-0547e22d55ff tempest-ImagesTestJSON-618852556 tempest-ImagesTestJSON-618852556-project-member] Task: {'id': task-1294736, 'name': MoveVirtualDisk_Task} progress is 88%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 853.588227] env[61806]: DEBUG oslo_concurrency.lockutils [None req-fbc8cbed-1b00-40c8-a788-e6e317b3c0cb tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] Acquiring lock "refresh_cache-2a3f81e3-b84f-4370-a598-2f6a607c3d47" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 853.588360] env[61806]: DEBUG oslo_concurrency.lockutils [None req-fbc8cbed-1b00-40c8-a788-e6e317b3c0cb tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] Acquired lock "refresh_cache-2a3f81e3-b84f-4370-a598-2f6a607c3d47" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 853.588548] env[61806]: DEBUG nova.network.neutron [None req-fbc8cbed-1b00-40c8-a788-e6e317b3c0cb tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] [instance: 2a3f81e3-b84f-4370-a598-2f6a607c3d47] Building network info cache for instance {{(pid=61806) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 853.739635] env[61806]: DEBUG oslo_vmware.api [None req-4b919e4f-1f0b-40e0-b482-0547e22d55ff tempest-ImagesTestJSON-618852556 tempest-ImagesTestJSON-618852556-project-member] Task: {'id': task-1294736, 'name': MoveVirtualDisk_Task, 'duration_secs': 2.818181} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 853.739938] env[61806]: INFO nova.virt.vmwareapi.ds_util [None req-4b919e4f-1f0b-40e0-b482-0547e22d55ff tempest-ImagesTestJSON-618852556 tempest-ImagesTestJSON-618852556-project-member] Moved virtual disk from [datastore2] OSTACK_IMG_cdc60310-6efe-499a-8b76-fbc7cd9f40d4/OSTACK_IMG_cdc60310-6efe-499a-8b76-fbc7cd9f40d4.vmdk to [datastore2] devstack-image-cache_base/e1cfdcc0-a515-4a37-bf5d-903e61ffb7f2/e1cfdcc0-a515-4a37-bf5d-903e61ffb7f2.vmdk. [ 853.740167] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-4b919e4f-1f0b-40e0-b482-0547e22d55ff tempest-ImagesTestJSON-618852556 tempest-ImagesTestJSON-618852556-project-member] [instance: 28e0baab-8516-42e3-acc2-9b8eb5192f57] Cleaning up location [datastore2] OSTACK_IMG_cdc60310-6efe-499a-8b76-fbc7cd9f40d4 {{(pid=61806) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:640}} [ 853.741012] env[61806]: DEBUG nova.virt.vmwareapi.ds_util [None req-4b919e4f-1f0b-40e0-b482-0547e22d55ff tempest-ImagesTestJSON-618852556 tempest-ImagesTestJSON-618852556-project-member] Deleting the datastore file [datastore2] OSTACK_IMG_cdc60310-6efe-499a-8b76-fbc7cd9f40d4 {{(pid=61806) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 853.741012] env[61806]: DEBUG oslo_concurrency.lockutils [None req-3bd56cbd-68df-4151-a169-676f8bdda9ab tempest-MultipleCreateTestJSON-1589015462 tempest-MultipleCreateTestJSON-1589015462-project-member] Lock "25218e83-c1ab-4b97-ade1-7c32b6f99305" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 14.499s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 853.741682] env[61806]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-9e2cdd37-d40b-420a-a3b6-a66b32bcf147 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 853.751963] env[61806]: DEBUG oslo_vmware.api [None req-4b919e4f-1f0b-40e0-b482-0547e22d55ff tempest-ImagesTestJSON-618852556 tempest-ImagesTestJSON-618852556-project-member] Waiting for the task: (returnval){ [ 853.751963] env[61806]: value = "task-1294737" [ 853.751963] env[61806]: _type = "Task" [ 853.751963] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 853.762356] env[61806]: DEBUG oslo_vmware.api [None req-4b919e4f-1f0b-40e0-b482-0547e22d55ff tempest-ImagesTestJSON-618852556 tempest-ImagesTestJSON-618852556-project-member] Task: {'id': task-1294737, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 853.936575] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-76cd7554-a883-4926-ac4a-ea7181f794f3 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 853.947024] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f63e3fe8-c169-42f1-b349-9d5035769e23 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 853.983927] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-70580a39-3373-4589-8c17-d7d87a6e2896 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 853.992666] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aaa8da4c-9f74-415d-a4de-9ea3570dab5b {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 854.007337] env[61806]: DEBUG nova.compute.provider_tree [None req-182a714a-9a7d-40bc-ad59-30e74e8a3863 tempest-MultipleCreateTestJSON-1589015462 tempest-MultipleCreateTestJSON-1589015462-project-member] Inventory has not changed in ProviderTree for provider: a2858be1-fd22-4e08-979e-87ad25293407 {{(pid=61806) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 854.141812] env[61806]: WARNING nova.network.neutron [None req-fbc8cbed-1b00-40c8-a788-e6e317b3c0cb tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] [instance: 2a3f81e3-b84f-4370-a598-2f6a607c3d47] a7ffda72-e57c-41fa-a7ed-d7b12084bc9d already exists in list: networks containing: ['a7ffda72-e57c-41fa-a7ed-d7b12084bc9d']. ignoring it [ 854.142058] env[61806]: WARNING nova.network.neutron [None req-fbc8cbed-1b00-40c8-a788-e6e317b3c0cb tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] [instance: 2a3f81e3-b84f-4370-a598-2f6a607c3d47] a7ffda72-e57c-41fa-a7ed-d7b12084bc9d already exists in list: networks containing: ['a7ffda72-e57c-41fa-a7ed-d7b12084bc9d']. ignoring it [ 854.265090] env[61806]: DEBUG oslo_vmware.api [None req-4b919e4f-1f0b-40e0-b482-0547e22d55ff tempest-ImagesTestJSON-618852556 tempest-ImagesTestJSON-618852556-project-member] Task: {'id': task-1294737, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.03427} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 854.266998] env[61806]: DEBUG nova.virt.vmwareapi.ds_util [None req-4b919e4f-1f0b-40e0-b482-0547e22d55ff tempest-ImagesTestJSON-618852556 tempest-ImagesTestJSON-618852556-project-member] Deleted the datastore file {{(pid=61806) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 854.266998] env[61806]: DEBUG oslo_concurrency.lockutils [None req-4b919e4f-1f0b-40e0-b482-0547e22d55ff tempest-ImagesTestJSON-618852556 tempest-ImagesTestJSON-618852556-project-member] Releasing lock "[datastore2] devstack-image-cache_base/e1cfdcc0-a515-4a37-bf5d-903e61ffb7f2/e1cfdcc0-a515-4a37-bf5d-903e61ffb7f2.vmdk" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 854.266998] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [None req-4b919e4f-1f0b-40e0-b482-0547e22d55ff tempest-ImagesTestJSON-618852556 tempest-ImagesTestJSON-618852556-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/e1cfdcc0-a515-4a37-bf5d-903e61ffb7f2/e1cfdcc0-a515-4a37-bf5d-903e61ffb7f2.vmdk to [datastore2] 28e0baab-8516-42e3-acc2-9b8eb5192f57/28e0baab-8516-42e3-acc2-9b8eb5192f57.vmdk {{(pid=61806) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 854.266998] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-2d9c7987-3243-4d89-a146-1176f499c5f5 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 854.275394] env[61806]: DEBUG oslo_vmware.api [None req-4b919e4f-1f0b-40e0-b482-0547e22d55ff tempest-ImagesTestJSON-618852556 tempest-ImagesTestJSON-618852556-project-member] Waiting for the task: (returnval){ [ 854.275394] env[61806]: value = "task-1294738" [ 854.275394] env[61806]: _type = "Task" [ 854.275394] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 854.283060] env[61806]: DEBUG oslo_vmware.api [None req-4b919e4f-1f0b-40e0-b482-0547e22d55ff tempest-ImagesTestJSON-618852556 tempest-ImagesTestJSON-618852556-project-member] Task: {'id': task-1294738, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 854.348118] env[61806]: DEBUG nova.compute.manager [req-8b9c4290-0c65-4a19-aa46-bb5d498fcf3c req-5c9744ab-1409-4334-bac1-8bae6f18d01b service nova] [instance: 2a3f81e3-b84f-4370-a598-2f6a607c3d47] Received event network-vif-plugged-64b56aa1-c5e3-4093-acee-b75f7ead331e {{(pid=61806) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 854.348118] env[61806]: DEBUG oslo_concurrency.lockutils [req-8b9c4290-0c65-4a19-aa46-bb5d498fcf3c req-5c9744ab-1409-4334-bac1-8bae6f18d01b service nova] Acquiring lock "2a3f81e3-b84f-4370-a598-2f6a607c3d47-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 854.348118] env[61806]: DEBUG oslo_concurrency.lockutils [req-8b9c4290-0c65-4a19-aa46-bb5d498fcf3c req-5c9744ab-1409-4334-bac1-8bae6f18d01b service nova] Lock "2a3f81e3-b84f-4370-a598-2f6a607c3d47-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 854.348118] env[61806]: DEBUG oslo_concurrency.lockutils [req-8b9c4290-0c65-4a19-aa46-bb5d498fcf3c req-5c9744ab-1409-4334-bac1-8bae6f18d01b service nova] Lock "2a3f81e3-b84f-4370-a598-2f6a607c3d47-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 854.348118] env[61806]: DEBUG nova.compute.manager [req-8b9c4290-0c65-4a19-aa46-bb5d498fcf3c req-5c9744ab-1409-4334-bac1-8bae6f18d01b service nova] [instance: 2a3f81e3-b84f-4370-a598-2f6a607c3d47] No waiting events found dispatching network-vif-plugged-64b56aa1-c5e3-4093-acee-b75f7ead331e {{(pid=61806) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 854.348118] env[61806]: WARNING nova.compute.manager [req-8b9c4290-0c65-4a19-aa46-bb5d498fcf3c req-5c9744ab-1409-4334-bac1-8bae6f18d01b service nova] [instance: 2a3f81e3-b84f-4370-a598-2f6a607c3d47] Received unexpected event network-vif-plugged-64b56aa1-c5e3-4093-acee-b75f7ead331e for instance with vm_state active and task_state None. [ 854.349248] env[61806]: DEBUG nova.compute.manager [req-8b9c4290-0c65-4a19-aa46-bb5d498fcf3c req-5c9744ab-1409-4334-bac1-8bae6f18d01b service nova] [instance: 2a3f81e3-b84f-4370-a598-2f6a607c3d47] Received event network-changed-64b56aa1-c5e3-4093-acee-b75f7ead331e {{(pid=61806) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 854.349973] env[61806]: DEBUG nova.compute.manager [req-8b9c4290-0c65-4a19-aa46-bb5d498fcf3c req-5c9744ab-1409-4334-bac1-8bae6f18d01b service nova] [instance: 2a3f81e3-b84f-4370-a598-2f6a607c3d47] Refreshing instance network info cache due to event network-changed-64b56aa1-c5e3-4093-acee-b75f7ead331e. {{(pid=61806) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 854.350907] env[61806]: DEBUG oslo_concurrency.lockutils [req-8b9c4290-0c65-4a19-aa46-bb5d498fcf3c req-5c9744ab-1409-4334-bac1-8bae6f18d01b service nova] Acquiring lock "refresh_cache-2a3f81e3-b84f-4370-a598-2f6a607c3d47" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 854.359438] env[61806]: INFO nova.compute.manager [None req-9d9c27a2-c2da-48b4-9d8f-5aa531ec53d7 tempest-ServerActionsV293TestJSON-906151111 tempest-ServerActionsV293TestJSON-906151111-project-member] [instance: 0758901a-7093-41d3-b0e2-5c519333abdd] Rebuilding instance [ 854.418019] env[61806]: DEBUG nova.compute.manager [None req-9d9c27a2-c2da-48b4-9d8f-5aa531ec53d7 tempest-ServerActionsV293TestJSON-906151111 tempest-ServerActionsV293TestJSON-906151111-project-member] [instance: 0758901a-7093-41d3-b0e2-5c519333abdd] Checking state {{(pid=61806) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 854.418019] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4ff40a83-a0b7-4685-8a91-2bed5d574e80 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 854.511587] env[61806]: DEBUG nova.scheduler.client.report [None req-182a714a-9a7d-40bc-ad59-30e74e8a3863 tempest-MultipleCreateTestJSON-1589015462 tempest-MultipleCreateTestJSON-1589015462-project-member] Inventory has not changed for provider a2858be1-fd22-4e08-979e-87ad25293407 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 138, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61806) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 854.786323] env[61806]: DEBUG oslo_vmware.api [None req-4b919e4f-1f0b-40e0-b482-0547e22d55ff tempest-ImagesTestJSON-618852556 tempest-ImagesTestJSON-618852556-project-member] Task: {'id': task-1294738, 'name': CopyVirtualDisk_Task} progress is 21%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 854.877953] env[61806]: DEBUG nova.network.neutron [None req-fbc8cbed-1b00-40c8-a788-e6e317b3c0cb tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] [instance: 2a3f81e3-b84f-4370-a598-2f6a607c3d47] Updating instance_info_cache with network_info: [{"id": "a839bcaa-9990-4bad-be1f-1f06eb1c978e", "address": "fa:16:3e:0b:8a:a6", "network": {"id": "a7ffda72-e57c-41fa-a7ed-d7b12084bc9d", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-62794865-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.193", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "6e391e1267014fafbb8f5a3211299819", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8b3cfeb1-f262-4fd9-b506-8e9c0733e2d8", "external-id": "nsx-vlan-transportzone-119", "segmentation_id": 119, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa839bcaa-99", "ovs_interfaceid": "a839bcaa-9990-4bad-be1f-1f06eb1c978e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "1f80e0d2-192f-4177-8212-acd6ca475851", "address": "fa:16:3e:aa:aa:6f", "network": {"id": "a7ffda72-e57c-41fa-a7ed-d7b12084bc9d", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-62794865-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "6e391e1267014fafbb8f5a3211299819", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8b3cfeb1-f262-4fd9-b506-8e9c0733e2d8", "external-id": "nsx-vlan-transportzone-119", "segmentation_id": 119, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1f80e0d2-19", "ovs_interfaceid": "1f80e0d2-192f-4177-8212-acd6ca475851", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "64b56aa1-c5e3-4093-acee-b75f7ead331e", "address": "fa:16:3e:0e:4f:49", "network": {"id": "a7ffda72-e57c-41fa-a7ed-d7b12084bc9d", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-62794865-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "6e391e1267014fafbb8f5a3211299819", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8b3cfeb1-f262-4fd9-b506-8e9c0733e2d8", "external-id": "nsx-vlan-transportzone-119", "segmentation_id": 119, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap64b56aa1-c5", "ovs_interfaceid": "64b56aa1-c5e3-4093-acee-b75f7ead331e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": true, "delegate_create": true, "meta": {}}] {{(pid=61806) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 854.928439] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [None req-9d9c27a2-c2da-48b4-9d8f-5aa531ec53d7 tempest-ServerActionsV293TestJSON-906151111 tempest-ServerActionsV293TestJSON-906151111-project-member] [instance: 0758901a-7093-41d3-b0e2-5c519333abdd] Powering off the VM {{(pid=61806) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 854.928691] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-ff5b5a08-3e97-4ede-ab36-2348efdce893 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 854.937079] env[61806]: DEBUG oslo_vmware.api [None req-9d9c27a2-c2da-48b4-9d8f-5aa531ec53d7 tempest-ServerActionsV293TestJSON-906151111 tempest-ServerActionsV293TestJSON-906151111-project-member] Waiting for the task: (returnval){ [ 854.937079] env[61806]: value = "task-1294739" [ 854.937079] env[61806]: _type = "Task" [ 854.937079] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 854.947983] env[61806]: DEBUG oslo_vmware.api [None req-9d9c27a2-c2da-48b4-9d8f-5aa531ec53d7 tempest-ServerActionsV293TestJSON-906151111 tempest-ServerActionsV293TestJSON-906151111-project-member] Task: {'id': task-1294739, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 855.018562] env[61806]: DEBUG oslo_concurrency.lockutils [None req-182a714a-9a7d-40bc-ad59-30e74e8a3863 tempest-MultipleCreateTestJSON-1589015462 tempest-MultipleCreateTestJSON-1589015462-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.831s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 855.023802] env[61806]: DEBUG oslo_concurrency.lockutils [None req-81aa2ada-298c-4e33-a594-b57f3ac492a8 tempest-ServerShowV247Test-1218443059 tempest-ServerShowV247Test-1218443059-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 10.687s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 855.024189] env[61806]: DEBUG nova.objects.instance [None req-81aa2ada-298c-4e33-a594-b57f3ac492a8 tempest-ServerShowV247Test-1218443059 tempest-ServerShowV247Test-1218443059-project-member] Lazy-loading 'resources' on Instance uuid 7c3fd2ff-ebd5-454f-a743-7fbae1088941 {{(pid=61806) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 855.062104] env[61806]: INFO nova.scheduler.client.report [None req-182a714a-9a7d-40bc-ad59-30e74e8a3863 tempest-MultipleCreateTestJSON-1589015462 tempest-MultipleCreateTestJSON-1589015462-project-member] Deleted allocations for instance ef502a7d-7dc0-4319-8ce3-6a8701288237 [ 855.286554] env[61806]: DEBUG oslo_vmware.api [None req-4b919e4f-1f0b-40e0-b482-0547e22d55ff tempest-ImagesTestJSON-618852556 tempest-ImagesTestJSON-618852556-project-member] Task: {'id': task-1294738, 'name': CopyVirtualDisk_Task} progress is 43%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 855.380565] env[61806]: DEBUG oslo_concurrency.lockutils [None req-fbc8cbed-1b00-40c8-a788-e6e317b3c0cb tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] Releasing lock "refresh_cache-2a3f81e3-b84f-4370-a598-2f6a607c3d47" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 855.381425] env[61806]: DEBUG oslo_concurrency.lockutils [None req-fbc8cbed-1b00-40c8-a788-e6e317b3c0cb tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] Acquiring lock "2a3f81e3-b84f-4370-a598-2f6a607c3d47" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 855.381591] env[61806]: DEBUG oslo_concurrency.lockutils [None req-fbc8cbed-1b00-40c8-a788-e6e317b3c0cb tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] Acquired lock "2a3f81e3-b84f-4370-a598-2f6a607c3d47" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 855.381893] env[61806]: DEBUG oslo_concurrency.lockutils [req-8b9c4290-0c65-4a19-aa46-bb5d498fcf3c req-5c9744ab-1409-4334-bac1-8bae6f18d01b service nova] Acquired lock "refresh_cache-2a3f81e3-b84f-4370-a598-2f6a607c3d47" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 855.382457] env[61806]: DEBUG nova.network.neutron [req-8b9c4290-0c65-4a19-aa46-bb5d498fcf3c req-5c9744ab-1409-4334-bac1-8bae6f18d01b service nova] [instance: 2a3f81e3-b84f-4370-a598-2f6a607c3d47] Refreshing network info cache for port 64b56aa1-c5e3-4093-acee-b75f7ead331e {{(pid=61806) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 855.384576] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9b5344f1-3e24-48a6-a03d-f7c6989d0402 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 855.420979] env[61806]: DEBUG nova.virt.hardware [None req-fbc8cbed-1b00-40c8-a788-e6e317b3c0cb tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-15T18:05:10Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format=,created_at=,direct_url=,disk_format=,id=,min_disk=,min_ram=,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=61806) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 855.422767] env[61806]: DEBUG nova.virt.hardware [None req-fbc8cbed-1b00-40c8-a788-e6e317b3c0cb tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] Flavor limits 0:0:0 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 855.422767] env[61806]: DEBUG nova.virt.hardware [None req-fbc8cbed-1b00-40c8-a788-e6e317b3c0cb tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] Image limits 0:0:0 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 855.422767] env[61806]: DEBUG nova.virt.hardware [None req-fbc8cbed-1b00-40c8-a788-e6e317b3c0cb tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] Flavor pref 0:0:0 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 855.422767] env[61806]: DEBUG nova.virt.hardware [None req-fbc8cbed-1b00-40c8-a788-e6e317b3c0cb tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] Image pref 0:0:0 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 855.422767] env[61806]: DEBUG nova.virt.hardware [None req-fbc8cbed-1b00-40c8-a788-e6e317b3c0cb tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 855.422767] env[61806]: DEBUG nova.virt.hardware [None req-fbc8cbed-1b00-40c8-a788-e6e317b3c0cb tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61806) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 855.422767] env[61806]: DEBUG nova.virt.hardware [None req-fbc8cbed-1b00-40c8-a788-e6e317b3c0cb tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61806) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 855.422767] env[61806]: DEBUG nova.virt.hardware [None req-fbc8cbed-1b00-40c8-a788-e6e317b3c0cb tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] Got 1 possible topologies {{(pid=61806) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 855.423123] env[61806]: DEBUG nova.virt.hardware [None req-fbc8cbed-1b00-40c8-a788-e6e317b3c0cb tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61806) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 855.424029] env[61806]: DEBUG nova.virt.hardware [None req-fbc8cbed-1b00-40c8-a788-e6e317b3c0cb tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61806) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 855.429942] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-fbc8cbed-1b00-40c8-a788-e6e317b3c0cb tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] [instance: 2a3f81e3-b84f-4370-a598-2f6a607c3d47] Reconfiguring VM to attach interface {{(pid=61806) attach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1929}} [ 855.432163] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-db30b1a1-1c7d-4628-ba3c-c81cf3c84714 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 855.457787] env[61806]: DEBUG oslo_vmware.api [None req-9d9c27a2-c2da-48b4-9d8f-5aa531ec53d7 tempest-ServerActionsV293TestJSON-906151111 tempest-ServerActionsV293TestJSON-906151111-project-member] Task: {'id': task-1294739, 'name': PowerOffVM_Task, 'duration_secs': 0.211475} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 855.459334] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [None req-9d9c27a2-c2da-48b4-9d8f-5aa531ec53d7 tempest-ServerActionsV293TestJSON-906151111 tempest-ServerActionsV293TestJSON-906151111-project-member] [instance: 0758901a-7093-41d3-b0e2-5c519333abdd] Powered off the VM {{(pid=61806) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 855.460317] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [None req-9d9c27a2-c2da-48b4-9d8f-5aa531ec53d7 tempest-ServerActionsV293TestJSON-906151111 tempest-ServerActionsV293TestJSON-906151111-project-member] [instance: 0758901a-7093-41d3-b0e2-5c519333abdd] Powering off the VM {{(pid=61806) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 855.460745] env[61806]: DEBUG oslo_vmware.api [None req-fbc8cbed-1b00-40c8-a788-e6e317b3c0cb tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] Waiting for the task: (returnval){ [ 855.460745] env[61806]: value = "task-1294740" [ 855.460745] env[61806]: _type = "Task" [ 855.460745] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 855.460959] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-198daed2-d7ee-4a17-aca0-47fc2a1a7776 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 855.472416] env[61806]: DEBUG oslo_vmware.api [None req-fbc8cbed-1b00-40c8-a788-e6e317b3c0cb tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] Task: {'id': task-1294740, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 855.473675] env[61806]: DEBUG oslo_vmware.api [None req-9d9c27a2-c2da-48b4-9d8f-5aa531ec53d7 tempest-ServerActionsV293TestJSON-906151111 tempest-ServerActionsV293TestJSON-906151111-project-member] Waiting for the task: (returnval){ [ 855.473675] env[61806]: value = "task-1294741" [ 855.473675] env[61806]: _type = "Task" [ 855.473675] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 855.482966] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [None req-9d9c27a2-c2da-48b4-9d8f-5aa531ec53d7 tempest-ServerActionsV293TestJSON-906151111 tempest-ServerActionsV293TestJSON-906151111-project-member] [instance: 0758901a-7093-41d3-b0e2-5c519333abdd] VM already powered off {{(pid=61806) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1509}} [ 855.483249] env[61806]: DEBUG nova.virt.vmwareapi.volumeops [None req-9d9c27a2-c2da-48b4-9d8f-5aa531ec53d7 tempest-ServerActionsV293TestJSON-906151111 tempest-ServerActionsV293TestJSON-906151111-project-member] [instance: 0758901a-7093-41d3-b0e2-5c519333abdd] Volume detach. Driver type: vmdk {{(pid=61806) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 855.483585] env[61806]: DEBUG nova.virt.vmwareapi.volumeops [None req-9d9c27a2-c2da-48b4-9d8f-5aa531ec53d7 tempest-ServerActionsV293TestJSON-906151111 tempest-ServerActionsV293TestJSON-906151111-project-member] [instance: 0758901a-7093-41d3-b0e2-5c519333abdd] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-277626', 'volume_id': '1b19f6ff-b694-445e-9bc1-d581e9268324', 'name': 'volume-1b19f6ff-b694-445e-9bc1-d581e9268324', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '0758901a-7093-41d3-b0e2-5c519333abdd', 'attached_at': '', 'detached_at': '', 'volume_id': '1b19f6ff-b694-445e-9bc1-d581e9268324', 'serial': '1b19f6ff-b694-445e-9bc1-d581e9268324'} {{(pid=61806) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 855.484803] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4f2e7e9a-7657-43df-8b6f-9987573f38ec {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 855.505679] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ef3c18fe-c541-488a-9f70-741225005bc9 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 855.513998] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ffb5a11e-4ef3-466c-88aa-d83f2f5b4105 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 855.541930] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3de0f278-5436-4442-951d-bf3378413566 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 855.562305] env[61806]: DEBUG nova.virt.vmwareapi.volumeops [None req-9d9c27a2-c2da-48b4-9d8f-5aa531ec53d7 tempest-ServerActionsV293TestJSON-906151111 tempest-ServerActionsV293TestJSON-906151111-project-member] The volume has not been displaced from its original location: [datastore1] volume-1b19f6ff-b694-445e-9bc1-d581e9268324/volume-1b19f6ff-b694-445e-9bc1-d581e9268324.vmdk. No consolidation needed. {{(pid=61806) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 855.568871] env[61806]: DEBUG nova.virt.vmwareapi.volumeops [None req-9d9c27a2-c2da-48b4-9d8f-5aa531ec53d7 tempest-ServerActionsV293TestJSON-906151111 tempest-ServerActionsV293TestJSON-906151111-project-member] [instance: 0758901a-7093-41d3-b0e2-5c519333abdd] Reconfiguring VM instance instance-0000003c to detach disk 2000 {{(pid=61806) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 855.576643] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-5269779a-bb92-4e63-a60e-7206579bf9d0 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 855.592449] env[61806]: DEBUG oslo_concurrency.lockutils [None req-182a714a-9a7d-40bc-ad59-30e74e8a3863 tempest-MultipleCreateTestJSON-1589015462 tempest-MultipleCreateTestJSON-1589015462-project-member] Lock "ef502a7d-7dc0-4319-8ce3-6a8701288237" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 15.995s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 855.599930] env[61806]: DEBUG oslo_vmware.api [None req-9d9c27a2-c2da-48b4-9d8f-5aa531ec53d7 tempest-ServerActionsV293TestJSON-906151111 tempest-ServerActionsV293TestJSON-906151111-project-member] Waiting for the task: (returnval){ [ 855.599930] env[61806]: value = "task-1294742" [ 855.599930] env[61806]: _type = "Task" [ 855.599930] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 855.610328] env[61806]: DEBUG oslo_vmware.api [None req-9d9c27a2-c2da-48b4-9d8f-5aa531ec53d7 tempest-ServerActionsV293TestJSON-906151111 tempest-ServerActionsV293TestJSON-906151111-project-member] Task: {'id': task-1294742, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 855.785666] env[61806]: DEBUG oslo_vmware.api [None req-4b919e4f-1f0b-40e0-b482-0547e22d55ff tempest-ImagesTestJSON-618852556 tempest-ImagesTestJSON-618852556-project-member] Task: {'id': task-1294738, 'name': CopyVirtualDisk_Task} progress is 63%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 855.794928] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aed4ef47-1d5d-48bd-a310-68f1268b172b {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 855.802957] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-650ca1e1-0b22-4ffc-b618-86020acb89ee {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 855.834034] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b3054ea3-cd82-4126-92d7-fd71af7524f0 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 855.841596] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-97d29119-c313-4fbb-8608-d3fb0af4b6f7 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 855.856958] env[61806]: DEBUG nova.compute.provider_tree [None req-81aa2ada-298c-4e33-a594-b57f3ac492a8 tempest-ServerShowV247Test-1218443059 tempest-ServerShowV247Test-1218443059-project-member] Inventory has not changed in ProviderTree for provider: a2858be1-fd22-4e08-979e-87ad25293407 {{(pid=61806) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 855.977257] env[61806]: DEBUG oslo_vmware.api [None req-fbc8cbed-1b00-40c8-a788-e6e317b3c0cb tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] Task: {'id': task-1294740, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 856.054864] env[61806]: DEBUG oslo_concurrency.lockutils [None req-ca966d32-e3ef-4c81-a661-22cd95d9e4ec tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Acquiring lock "efa19aac-3ad7-42d5-a6d8-859050de6e63" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 856.055178] env[61806]: DEBUG oslo_concurrency.lockutils [None req-ca966d32-e3ef-4c81-a661-22cd95d9e4ec tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Lock "efa19aac-3ad7-42d5-a6d8-859050de6e63" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 856.110633] env[61806]: DEBUG oslo_vmware.api [None req-9d9c27a2-c2da-48b4-9d8f-5aa531ec53d7 tempest-ServerActionsV293TestJSON-906151111 tempest-ServerActionsV293TestJSON-906151111-project-member] Task: {'id': task-1294742, 'name': ReconfigVM_Task, 'duration_secs': 0.204745} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 856.110930] env[61806]: DEBUG nova.virt.vmwareapi.volumeops [None req-9d9c27a2-c2da-48b4-9d8f-5aa531ec53d7 tempest-ServerActionsV293TestJSON-906151111 tempest-ServerActionsV293TestJSON-906151111-project-member] [instance: 0758901a-7093-41d3-b0e2-5c519333abdd] Reconfigured VM instance instance-0000003c to detach disk 2000 {{(pid=61806) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 856.116099] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-7ed802a1-4fe9-48c0-a4e9-7d16d66593f7 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 856.136789] env[61806]: DEBUG oslo_vmware.api [None req-9d9c27a2-c2da-48b4-9d8f-5aa531ec53d7 tempest-ServerActionsV293TestJSON-906151111 tempest-ServerActionsV293TestJSON-906151111-project-member] Waiting for the task: (returnval){ [ 856.136789] env[61806]: value = "task-1294743" [ 856.136789] env[61806]: _type = "Task" [ 856.136789] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 856.146137] env[61806]: DEBUG oslo_vmware.api [None req-9d9c27a2-c2da-48b4-9d8f-5aa531ec53d7 tempest-ServerActionsV293TestJSON-906151111 tempest-ServerActionsV293TestJSON-906151111-project-member] Task: {'id': task-1294743, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 856.289933] env[61806]: DEBUG oslo_vmware.api [None req-4b919e4f-1f0b-40e0-b482-0547e22d55ff tempest-ImagesTestJSON-618852556 tempest-ImagesTestJSON-618852556-project-member] Task: {'id': task-1294738, 'name': CopyVirtualDisk_Task} progress is 83%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 856.360019] env[61806]: DEBUG nova.scheduler.client.report [None req-81aa2ada-298c-4e33-a594-b57f3ac492a8 tempest-ServerShowV247Test-1218443059 tempest-ServerShowV247Test-1218443059-project-member] Inventory has not changed for provider a2858be1-fd22-4e08-979e-87ad25293407 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 138, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61806) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 856.443862] env[61806]: DEBUG nova.network.neutron [req-8b9c4290-0c65-4a19-aa46-bb5d498fcf3c req-5c9744ab-1409-4334-bac1-8bae6f18d01b service nova] [instance: 2a3f81e3-b84f-4370-a598-2f6a607c3d47] Updated VIF entry in instance network info cache for port 64b56aa1-c5e3-4093-acee-b75f7ead331e. {{(pid=61806) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 856.444505] env[61806]: DEBUG nova.network.neutron [req-8b9c4290-0c65-4a19-aa46-bb5d498fcf3c req-5c9744ab-1409-4334-bac1-8bae6f18d01b service nova] [instance: 2a3f81e3-b84f-4370-a598-2f6a607c3d47] Updating instance_info_cache with network_info: [{"id": "a839bcaa-9990-4bad-be1f-1f06eb1c978e", "address": "fa:16:3e:0b:8a:a6", "network": {"id": "a7ffda72-e57c-41fa-a7ed-d7b12084bc9d", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-62794865-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.193", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "6e391e1267014fafbb8f5a3211299819", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8b3cfeb1-f262-4fd9-b506-8e9c0733e2d8", "external-id": "nsx-vlan-transportzone-119", "segmentation_id": 119, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa839bcaa-99", "ovs_interfaceid": "a839bcaa-9990-4bad-be1f-1f06eb1c978e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "1f80e0d2-192f-4177-8212-acd6ca475851", "address": "fa:16:3e:aa:aa:6f", "network": {"id": "a7ffda72-e57c-41fa-a7ed-d7b12084bc9d", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-62794865-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "6e391e1267014fafbb8f5a3211299819", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8b3cfeb1-f262-4fd9-b506-8e9c0733e2d8", "external-id": "nsx-vlan-transportzone-119", "segmentation_id": 119, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1f80e0d2-19", "ovs_interfaceid": "1f80e0d2-192f-4177-8212-acd6ca475851", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "64b56aa1-c5e3-4093-acee-b75f7ead331e", "address": "fa:16:3e:0e:4f:49", "network": {"id": "a7ffda72-e57c-41fa-a7ed-d7b12084bc9d", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-62794865-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "6e391e1267014fafbb8f5a3211299819", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8b3cfeb1-f262-4fd9-b506-8e9c0733e2d8", "external-id": "nsx-vlan-transportzone-119", "segmentation_id": 119, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap64b56aa1-c5", "ovs_interfaceid": "64b56aa1-c5e3-4093-acee-b75f7ead331e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": true, "delegate_create": true, "meta": {}}] {{(pid=61806) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 856.474178] env[61806]: DEBUG oslo_vmware.api [None req-fbc8cbed-1b00-40c8-a788-e6e317b3c0cb tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] Task: {'id': task-1294740, 'name': ReconfigVM_Task} progress is 99%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 856.560651] env[61806]: DEBUG nova.compute.manager [None req-ca966d32-e3ef-4c81-a661-22cd95d9e4ec tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] [instance: efa19aac-3ad7-42d5-a6d8-859050de6e63] Starting instance... {{(pid=61806) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 856.646845] env[61806]: DEBUG oslo_vmware.api [None req-9d9c27a2-c2da-48b4-9d8f-5aa531ec53d7 tempest-ServerActionsV293TestJSON-906151111 tempest-ServerActionsV293TestJSON-906151111-project-member] Task: {'id': task-1294743, 'name': ReconfigVM_Task, 'duration_secs': 0.218904} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 856.647209] env[61806]: DEBUG nova.virt.vmwareapi.volumeops [None req-9d9c27a2-c2da-48b4-9d8f-5aa531ec53d7 tempest-ServerActionsV293TestJSON-906151111 tempest-ServerActionsV293TestJSON-906151111-project-member] [instance: 0758901a-7093-41d3-b0e2-5c519333abdd] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-277626', 'volume_id': '1b19f6ff-b694-445e-9bc1-d581e9268324', 'name': 'volume-1b19f6ff-b694-445e-9bc1-d581e9268324', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '0758901a-7093-41d3-b0e2-5c519333abdd', 'attached_at': '', 'detached_at': '', 'volume_id': '1b19f6ff-b694-445e-9bc1-d581e9268324', 'serial': '1b19f6ff-b694-445e-9bc1-d581e9268324'} {{(pid=61806) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 856.647512] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-9d9c27a2-c2da-48b4-9d8f-5aa531ec53d7 tempest-ServerActionsV293TestJSON-906151111 tempest-ServerActionsV293TestJSON-906151111-project-member] [instance: 0758901a-7093-41d3-b0e2-5c519333abdd] Destroying instance {{(pid=61806) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 856.648317] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8f1c627b-4511-4864-b1ca-13b1158ba409 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 856.654925] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-9d9c27a2-c2da-48b4-9d8f-5aa531ec53d7 tempest-ServerActionsV293TestJSON-906151111 tempest-ServerActionsV293TestJSON-906151111-project-member] [instance: 0758901a-7093-41d3-b0e2-5c519333abdd] Unregistering the VM {{(pid=61806) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 856.655189] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-2dcd9cdd-9424-457d-843c-4d9cb2683a88 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 856.722906] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-9d9c27a2-c2da-48b4-9d8f-5aa531ec53d7 tempest-ServerActionsV293TestJSON-906151111 tempest-ServerActionsV293TestJSON-906151111-project-member] [instance: 0758901a-7093-41d3-b0e2-5c519333abdd] Unregistered the VM {{(pid=61806) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 856.723199] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-9d9c27a2-c2da-48b4-9d8f-5aa531ec53d7 tempest-ServerActionsV293TestJSON-906151111 tempest-ServerActionsV293TestJSON-906151111-project-member] [instance: 0758901a-7093-41d3-b0e2-5c519333abdd] Deleting contents of the VM from datastore datastore1 {{(pid=61806) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 856.723392] env[61806]: DEBUG nova.virt.vmwareapi.ds_util [None req-9d9c27a2-c2da-48b4-9d8f-5aa531ec53d7 tempest-ServerActionsV293TestJSON-906151111 tempest-ServerActionsV293TestJSON-906151111-project-member] Deleting the datastore file [datastore1] 0758901a-7093-41d3-b0e2-5c519333abdd {{(pid=61806) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 856.723674] env[61806]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-79d246d5-4c90-4a99-b7dc-7499dc07ad93 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 856.730833] env[61806]: DEBUG oslo_vmware.api [None req-9d9c27a2-c2da-48b4-9d8f-5aa531ec53d7 tempest-ServerActionsV293TestJSON-906151111 tempest-ServerActionsV293TestJSON-906151111-project-member] Waiting for the task: (returnval){ [ 856.730833] env[61806]: value = "task-1294745" [ 856.730833] env[61806]: _type = "Task" [ 856.730833] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 856.738618] env[61806]: DEBUG oslo_vmware.api [None req-9d9c27a2-c2da-48b4-9d8f-5aa531ec53d7 tempest-ServerActionsV293TestJSON-906151111 tempest-ServerActionsV293TestJSON-906151111-project-member] Task: {'id': task-1294745, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 856.788733] env[61806]: DEBUG oslo_vmware.api [None req-4b919e4f-1f0b-40e0-b482-0547e22d55ff tempest-ImagesTestJSON-618852556 tempest-ImagesTestJSON-618852556-project-member] Task: {'id': task-1294738, 'name': CopyVirtualDisk_Task, 'duration_secs': 2.421151} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 856.788952] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [None req-4b919e4f-1f0b-40e0-b482-0547e22d55ff tempest-ImagesTestJSON-618852556 tempest-ImagesTestJSON-618852556-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/e1cfdcc0-a515-4a37-bf5d-903e61ffb7f2/e1cfdcc0-a515-4a37-bf5d-903e61ffb7f2.vmdk to [datastore2] 28e0baab-8516-42e3-acc2-9b8eb5192f57/28e0baab-8516-42e3-acc2-9b8eb5192f57.vmdk {{(pid=61806) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 856.789776] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-79f218c8-674f-4f27-aa69-177dba767723 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 856.811265] env[61806]: DEBUG nova.virt.vmwareapi.volumeops [None req-4b919e4f-1f0b-40e0-b482-0547e22d55ff tempest-ImagesTestJSON-618852556 tempest-ImagesTestJSON-618852556-project-member] [instance: 28e0baab-8516-42e3-acc2-9b8eb5192f57] Reconfiguring VM instance instance-00000048 to attach disk [datastore2] 28e0baab-8516-42e3-acc2-9b8eb5192f57/28e0baab-8516-42e3-acc2-9b8eb5192f57.vmdk or device None with type streamOptimized {{(pid=61806) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 856.811594] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-cd6186bb-37dc-4f35-b32a-ee7b826e37c5 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 856.832081] env[61806]: DEBUG oslo_vmware.api [None req-4b919e4f-1f0b-40e0-b482-0547e22d55ff tempest-ImagesTestJSON-618852556 tempest-ImagesTestJSON-618852556-project-member] Waiting for the task: (returnval){ [ 856.832081] env[61806]: value = "task-1294746" [ 856.832081] env[61806]: _type = "Task" [ 856.832081] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 856.840903] env[61806]: DEBUG oslo_vmware.api [None req-4b919e4f-1f0b-40e0-b482-0547e22d55ff tempest-ImagesTestJSON-618852556 tempest-ImagesTestJSON-618852556-project-member] Task: {'id': task-1294746, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 856.865433] env[61806]: DEBUG oslo_concurrency.lockutils [None req-81aa2ada-298c-4e33-a594-b57f3ac492a8 tempest-ServerShowV247Test-1218443059 tempest-ServerShowV247Test-1218443059-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.844s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 856.867855] env[61806]: DEBUG oslo_concurrency.lockutils [None req-40a350df-b9c4-4f7d-968f-0160bcbc0d06 tempest-ServerAddressesNegativeTestJSON-1526246390 tempest-ServerAddressesNegativeTestJSON-1526246390-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 10.864s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 856.868156] env[61806]: DEBUG nova.objects.instance [None req-40a350df-b9c4-4f7d-968f-0160bcbc0d06 tempest-ServerAddressesNegativeTestJSON-1526246390 tempest-ServerAddressesNegativeTestJSON-1526246390-project-member] Lazy-loading 'resources' on Instance uuid c8033551-6591-4e37-ae78-4efe7145b10b {{(pid=61806) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 856.897044] env[61806]: INFO nova.scheduler.client.report [None req-81aa2ada-298c-4e33-a594-b57f3ac492a8 tempest-ServerShowV247Test-1218443059 tempest-ServerShowV247Test-1218443059-project-member] Deleted allocations for instance 7c3fd2ff-ebd5-454f-a743-7fbae1088941 [ 856.947311] env[61806]: DEBUG oslo_concurrency.lockutils [req-8b9c4290-0c65-4a19-aa46-bb5d498fcf3c req-5c9744ab-1409-4334-bac1-8bae6f18d01b service nova] Releasing lock "refresh_cache-2a3f81e3-b84f-4370-a598-2f6a607c3d47" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 856.974128] env[61806]: DEBUG oslo_vmware.api [None req-fbc8cbed-1b00-40c8-a788-e6e317b3c0cb tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] Task: {'id': task-1294740, 'name': ReconfigVM_Task} progress is 99%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 857.081973] env[61806]: DEBUG oslo_concurrency.lockutils [None req-ca966d32-e3ef-4c81-a661-22cd95d9e4ec tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 857.245564] env[61806]: DEBUG oslo_vmware.api [None req-9d9c27a2-c2da-48b4-9d8f-5aa531ec53d7 tempest-ServerActionsV293TestJSON-906151111 tempest-ServerActionsV293TestJSON-906151111-project-member] Task: {'id': task-1294745, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.083156} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 857.245797] env[61806]: DEBUG nova.virt.vmwareapi.ds_util [None req-9d9c27a2-c2da-48b4-9d8f-5aa531ec53d7 tempest-ServerActionsV293TestJSON-906151111 tempest-ServerActionsV293TestJSON-906151111-project-member] Deleted the datastore file {{(pid=61806) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 857.245990] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-9d9c27a2-c2da-48b4-9d8f-5aa531ec53d7 tempest-ServerActionsV293TestJSON-906151111 tempest-ServerActionsV293TestJSON-906151111-project-member] [instance: 0758901a-7093-41d3-b0e2-5c519333abdd] Deleted contents of the VM from datastore datastore1 {{(pid=61806) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 857.246192] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-9d9c27a2-c2da-48b4-9d8f-5aa531ec53d7 tempest-ServerActionsV293TestJSON-906151111 tempest-ServerActionsV293TestJSON-906151111-project-member] [instance: 0758901a-7093-41d3-b0e2-5c519333abdd] Instance destroyed {{(pid=61806) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 857.315422] env[61806]: DEBUG nova.virt.vmwareapi.volumeops [None req-9d9c27a2-c2da-48b4-9d8f-5aa531ec53d7 tempest-ServerActionsV293TestJSON-906151111 tempest-ServerActionsV293TestJSON-906151111-project-member] [instance: 0758901a-7093-41d3-b0e2-5c519333abdd] Volume detach. Driver type: vmdk {{(pid=61806) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 857.315794] env[61806]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-0b9d1841-d621-47de-883b-ec0587f25e1d {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 857.324604] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5c074479-dd3f-435f-9553-37ab7447bd68 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 857.345692] env[61806]: DEBUG oslo_vmware.api [None req-4b919e4f-1f0b-40e0-b482-0547e22d55ff tempest-ImagesTestJSON-618852556 tempest-ImagesTestJSON-618852556-project-member] Task: {'id': task-1294746, 'name': ReconfigVM_Task, 'duration_secs': 0.419805} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 857.352570] env[61806]: DEBUG nova.virt.vmwareapi.volumeops [None req-4b919e4f-1f0b-40e0-b482-0547e22d55ff tempest-ImagesTestJSON-618852556 tempest-ImagesTestJSON-618852556-project-member] [instance: 28e0baab-8516-42e3-acc2-9b8eb5192f57] Reconfigured VM instance instance-00000048 to attach disk [datastore2] 28e0baab-8516-42e3-acc2-9b8eb5192f57/28e0baab-8516-42e3-acc2-9b8eb5192f57.vmdk or device None with type streamOptimized {{(pid=61806) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 857.354508] env[61806]: ERROR nova.compute.manager [None req-9d9c27a2-c2da-48b4-9d8f-5aa531ec53d7 tempest-ServerActionsV293TestJSON-906151111 tempest-ServerActionsV293TestJSON-906151111-project-member] [instance: 0758901a-7093-41d3-b0e2-5c519333abdd] Failed to detach volume 1b19f6ff-b694-445e-9bc1-d581e9268324 from /dev/sda: nova.exception.InstanceNotFound: Instance 0758901a-7093-41d3-b0e2-5c519333abdd could not be found. [ 857.354508] env[61806]: ERROR nova.compute.manager [instance: 0758901a-7093-41d3-b0e2-5c519333abdd] Traceback (most recent call last): [ 857.354508] env[61806]: ERROR nova.compute.manager [instance: 0758901a-7093-41d3-b0e2-5c519333abdd] File "/opt/stack/nova/nova/compute/manager.py", line 4142, in _do_rebuild_instance [ 857.354508] env[61806]: ERROR nova.compute.manager [instance: 0758901a-7093-41d3-b0e2-5c519333abdd] self.driver.rebuild(**kwargs) [ 857.354508] env[61806]: ERROR nova.compute.manager [instance: 0758901a-7093-41d3-b0e2-5c519333abdd] File "/opt/stack/nova/nova/virt/driver.py", line 493, in rebuild [ 857.354508] env[61806]: ERROR nova.compute.manager [instance: 0758901a-7093-41d3-b0e2-5c519333abdd] raise NotImplementedError() [ 857.354508] env[61806]: ERROR nova.compute.manager [instance: 0758901a-7093-41d3-b0e2-5c519333abdd] NotImplementedError [ 857.354508] env[61806]: ERROR nova.compute.manager [instance: 0758901a-7093-41d3-b0e2-5c519333abdd] [ 857.354508] env[61806]: ERROR nova.compute.manager [instance: 0758901a-7093-41d3-b0e2-5c519333abdd] During handling of the above exception, another exception occurred: [ 857.354508] env[61806]: ERROR nova.compute.manager [instance: 0758901a-7093-41d3-b0e2-5c519333abdd] [ 857.354508] env[61806]: ERROR nova.compute.manager [instance: 0758901a-7093-41d3-b0e2-5c519333abdd] Traceback (most recent call last): [ 857.354508] env[61806]: ERROR nova.compute.manager [instance: 0758901a-7093-41d3-b0e2-5c519333abdd] File "/opt/stack/nova/nova/compute/manager.py", line 3565, in _detach_root_volume [ 857.354508] env[61806]: ERROR nova.compute.manager [instance: 0758901a-7093-41d3-b0e2-5c519333abdd] self.driver.detach_volume(context, old_connection_info, [ 857.354508] env[61806]: ERROR nova.compute.manager [instance: 0758901a-7093-41d3-b0e2-5c519333abdd] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 553, in detach_volume [ 857.354508] env[61806]: ERROR nova.compute.manager [instance: 0758901a-7093-41d3-b0e2-5c519333abdd] return self._volumeops.detach_volume(connection_info, instance) [ 857.354508] env[61806]: ERROR nova.compute.manager [instance: 0758901a-7093-41d3-b0e2-5c519333abdd] File "/opt/stack/nova/nova/virt/vmwareapi/volumeops.py", line 649, in detach_volume [ 857.354508] env[61806]: ERROR nova.compute.manager [instance: 0758901a-7093-41d3-b0e2-5c519333abdd] self._detach_volume_vmdk(connection_info, instance) [ 857.354508] env[61806]: ERROR nova.compute.manager [instance: 0758901a-7093-41d3-b0e2-5c519333abdd] File "/opt/stack/nova/nova/virt/vmwareapi/volumeops.py", line 569, in _detach_volume_vmdk [ 857.354508] env[61806]: ERROR nova.compute.manager [instance: 0758901a-7093-41d3-b0e2-5c519333abdd] vm_ref = vm_util.get_vm_ref(self._session, instance) [ 857.354508] env[61806]: ERROR nova.compute.manager [instance: 0758901a-7093-41d3-b0e2-5c519333abdd] File "/opt/stack/nova/nova/virt/vmwareapi/vm_util.py", line 1135, in get_vm_ref [ 857.354508] env[61806]: ERROR nova.compute.manager [instance: 0758901a-7093-41d3-b0e2-5c519333abdd] stable_ref.fetch_moref(session) [ 857.354508] env[61806]: ERROR nova.compute.manager [instance: 0758901a-7093-41d3-b0e2-5c519333abdd] File "/opt/stack/nova/nova/virt/vmwareapi/vm_util.py", line 1126, in fetch_moref [ 857.354508] env[61806]: ERROR nova.compute.manager [instance: 0758901a-7093-41d3-b0e2-5c519333abdd] raise exception.InstanceNotFound(instance_id=self._uuid) [ 857.354508] env[61806]: ERROR nova.compute.manager [instance: 0758901a-7093-41d3-b0e2-5c519333abdd] nova.exception.InstanceNotFound: Instance 0758901a-7093-41d3-b0e2-5c519333abdd could not be found. [ 857.354508] env[61806]: ERROR nova.compute.manager [instance: 0758901a-7093-41d3-b0e2-5c519333abdd] [ 857.357399] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-595ecffa-70d6-42d4-9dbd-87a55a631fb8 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 857.364409] env[61806]: DEBUG oslo_vmware.api [None req-4b919e4f-1f0b-40e0-b482-0547e22d55ff tempest-ImagesTestJSON-618852556 tempest-ImagesTestJSON-618852556-project-member] Waiting for the task: (returnval){ [ 857.364409] env[61806]: value = "task-1294747" [ 857.364409] env[61806]: _type = "Task" [ 857.364409] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 857.375528] env[61806]: DEBUG oslo_vmware.api [None req-4b919e4f-1f0b-40e0-b482-0547e22d55ff tempest-ImagesTestJSON-618852556 tempest-ImagesTestJSON-618852556-project-member] Task: {'id': task-1294747, 'name': Rename_Task} progress is 5%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 857.410297] env[61806]: DEBUG oslo_concurrency.lockutils [None req-81aa2ada-298c-4e33-a594-b57f3ac492a8 tempest-ServerShowV247Test-1218443059 tempest-ServerShowV247Test-1218443059-project-member] Lock "7c3fd2ff-ebd5-454f-a743-7fbae1088941" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 16.787s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 857.487274] env[61806]: DEBUG oslo_vmware.api [None req-fbc8cbed-1b00-40c8-a788-e6e317b3c0cb tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] Task: {'id': task-1294740, 'name': ReconfigVM_Task, 'duration_secs': 1.623504} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 857.491205] env[61806]: DEBUG oslo_concurrency.lockutils [None req-fbc8cbed-1b00-40c8-a788-e6e317b3c0cb tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] Releasing lock "2a3f81e3-b84f-4370-a598-2f6a607c3d47" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 857.491720] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-fbc8cbed-1b00-40c8-a788-e6e317b3c0cb tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] [instance: 2a3f81e3-b84f-4370-a598-2f6a607c3d47] Reconfigured VM to attach interface {{(pid=61806) attach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1943}} [ 857.515239] env[61806]: DEBUG nova.compute.utils [None req-9d9c27a2-c2da-48b4-9d8f-5aa531ec53d7 tempest-ServerActionsV293TestJSON-906151111 tempest-ServerActionsV293TestJSON-906151111-project-member] [instance: 0758901a-7093-41d3-b0e2-5c519333abdd] Build of instance 0758901a-7093-41d3-b0e2-5c519333abdd aborted: Failed to rebuild volume backed instance. {{(pid=61806) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 857.518276] env[61806]: ERROR nova.compute.manager [None req-9d9c27a2-c2da-48b4-9d8f-5aa531ec53d7 tempest-ServerActionsV293TestJSON-906151111 tempest-ServerActionsV293TestJSON-906151111-project-member] [instance: 0758901a-7093-41d3-b0e2-5c519333abdd] Setting instance vm_state to ERROR: nova.exception.BuildAbortException: Build of instance 0758901a-7093-41d3-b0e2-5c519333abdd aborted: Failed to rebuild volume backed instance. [ 857.518276] env[61806]: ERROR nova.compute.manager [instance: 0758901a-7093-41d3-b0e2-5c519333abdd] Traceback (most recent call last): [ 857.518276] env[61806]: ERROR nova.compute.manager [instance: 0758901a-7093-41d3-b0e2-5c519333abdd] File "/opt/stack/nova/nova/compute/manager.py", line 4142, in _do_rebuild_instance [ 857.518276] env[61806]: ERROR nova.compute.manager [instance: 0758901a-7093-41d3-b0e2-5c519333abdd] self.driver.rebuild(**kwargs) [ 857.518276] env[61806]: ERROR nova.compute.manager [instance: 0758901a-7093-41d3-b0e2-5c519333abdd] File "/opt/stack/nova/nova/virt/driver.py", line 493, in rebuild [ 857.518276] env[61806]: ERROR nova.compute.manager [instance: 0758901a-7093-41d3-b0e2-5c519333abdd] raise NotImplementedError() [ 857.518276] env[61806]: ERROR nova.compute.manager [instance: 0758901a-7093-41d3-b0e2-5c519333abdd] NotImplementedError [ 857.518276] env[61806]: ERROR nova.compute.manager [instance: 0758901a-7093-41d3-b0e2-5c519333abdd] [ 857.518276] env[61806]: ERROR nova.compute.manager [instance: 0758901a-7093-41d3-b0e2-5c519333abdd] During handling of the above exception, another exception occurred: [ 857.518276] env[61806]: ERROR nova.compute.manager [instance: 0758901a-7093-41d3-b0e2-5c519333abdd] [ 857.518276] env[61806]: ERROR nova.compute.manager [instance: 0758901a-7093-41d3-b0e2-5c519333abdd] Traceback (most recent call last): [ 857.518276] env[61806]: ERROR nova.compute.manager [instance: 0758901a-7093-41d3-b0e2-5c519333abdd] File "/opt/stack/nova/nova/compute/manager.py", line 3600, in _rebuild_volume_backed_instance [ 857.518276] env[61806]: ERROR nova.compute.manager [instance: 0758901a-7093-41d3-b0e2-5c519333abdd] self._detach_root_volume(context, instance, root_bdm) [ 857.518276] env[61806]: ERROR nova.compute.manager [instance: 0758901a-7093-41d3-b0e2-5c519333abdd] File "/opt/stack/nova/nova/compute/manager.py", line 3579, in _detach_root_volume [ 857.518276] env[61806]: ERROR nova.compute.manager [instance: 0758901a-7093-41d3-b0e2-5c519333abdd] with excutils.save_and_reraise_exception(): [ 857.518276] env[61806]: ERROR nova.compute.manager [instance: 0758901a-7093-41d3-b0e2-5c519333abdd] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 857.518276] env[61806]: ERROR nova.compute.manager [instance: 0758901a-7093-41d3-b0e2-5c519333abdd] self.force_reraise() [ 857.518276] env[61806]: ERROR nova.compute.manager [instance: 0758901a-7093-41d3-b0e2-5c519333abdd] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 857.518276] env[61806]: ERROR nova.compute.manager [instance: 0758901a-7093-41d3-b0e2-5c519333abdd] raise self.value [ 857.518276] env[61806]: ERROR nova.compute.manager [instance: 0758901a-7093-41d3-b0e2-5c519333abdd] File "/opt/stack/nova/nova/compute/manager.py", line 3565, in _detach_root_volume [ 857.518276] env[61806]: ERROR nova.compute.manager [instance: 0758901a-7093-41d3-b0e2-5c519333abdd] self.driver.detach_volume(context, old_connection_info, [ 857.518276] env[61806]: ERROR nova.compute.manager [instance: 0758901a-7093-41d3-b0e2-5c519333abdd] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 553, in detach_volume [ 857.518276] env[61806]: ERROR nova.compute.manager [instance: 0758901a-7093-41d3-b0e2-5c519333abdd] return self._volumeops.detach_volume(connection_info, instance) [ 857.518276] env[61806]: ERROR nova.compute.manager [instance: 0758901a-7093-41d3-b0e2-5c519333abdd] File "/opt/stack/nova/nova/virt/vmwareapi/volumeops.py", line 649, in detach_volume [ 857.518276] env[61806]: ERROR nova.compute.manager [instance: 0758901a-7093-41d3-b0e2-5c519333abdd] self._detach_volume_vmdk(connection_info, instance) [ 857.518276] env[61806]: ERROR nova.compute.manager [instance: 0758901a-7093-41d3-b0e2-5c519333abdd] File "/opt/stack/nova/nova/virt/vmwareapi/volumeops.py", line 569, in _detach_volume_vmdk [ 857.518276] env[61806]: ERROR nova.compute.manager [instance: 0758901a-7093-41d3-b0e2-5c519333abdd] vm_ref = vm_util.get_vm_ref(self._session, instance) [ 857.518276] env[61806]: ERROR nova.compute.manager [instance: 0758901a-7093-41d3-b0e2-5c519333abdd] File "/opt/stack/nova/nova/virt/vmwareapi/vm_util.py", line 1135, in get_vm_ref [ 857.518276] env[61806]: ERROR nova.compute.manager [instance: 0758901a-7093-41d3-b0e2-5c519333abdd] stable_ref.fetch_moref(session) [ 857.518276] env[61806]: ERROR nova.compute.manager [instance: 0758901a-7093-41d3-b0e2-5c519333abdd] File "/opt/stack/nova/nova/virt/vmwareapi/vm_util.py", line 1126, in fetch_moref [ 857.518276] env[61806]: ERROR nova.compute.manager [instance: 0758901a-7093-41d3-b0e2-5c519333abdd] raise exception.InstanceNotFound(instance_id=self._uuid) [ 857.518276] env[61806]: ERROR nova.compute.manager [instance: 0758901a-7093-41d3-b0e2-5c519333abdd] nova.exception.InstanceNotFound: Instance 0758901a-7093-41d3-b0e2-5c519333abdd could not be found. [ 857.518276] env[61806]: ERROR nova.compute.manager [instance: 0758901a-7093-41d3-b0e2-5c519333abdd] [ 857.518276] env[61806]: ERROR nova.compute.manager [instance: 0758901a-7093-41d3-b0e2-5c519333abdd] During handling of the above exception, another exception occurred: [ 857.518276] env[61806]: ERROR nova.compute.manager [instance: 0758901a-7093-41d3-b0e2-5c519333abdd] [ 857.518276] env[61806]: ERROR nova.compute.manager [instance: 0758901a-7093-41d3-b0e2-5c519333abdd] Traceback (most recent call last): [ 857.518276] env[61806]: ERROR nova.compute.manager [instance: 0758901a-7093-41d3-b0e2-5c519333abdd] File "/opt/stack/nova/nova/compute/manager.py", line 10863, in _error_out_instance_on_exception [ 857.518276] env[61806]: ERROR nova.compute.manager [instance: 0758901a-7093-41d3-b0e2-5c519333abdd] yield [ 857.518276] env[61806]: ERROR nova.compute.manager [instance: 0758901a-7093-41d3-b0e2-5c519333abdd] File "/opt/stack/nova/nova/compute/manager.py", line 3868, in rebuild_instance [ 857.518276] env[61806]: ERROR nova.compute.manager [instance: 0758901a-7093-41d3-b0e2-5c519333abdd] self._do_rebuild_instance_with_claim( [ 857.519907] env[61806]: ERROR nova.compute.manager [instance: 0758901a-7093-41d3-b0e2-5c519333abdd] File "/opt/stack/nova/nova/compute/manager.py", line 3954, in _do_rebuild_instance_with_claim [ 857.519907] env[61806]: ERROR nova.compute.manager [instance: 0758901a-7093-41d3-b0e2-5c519333abdd] self._do_rebuild_instance( [ 857.519907] env[61806]: ERROR nova.compute.manager [instance: 0758901a-7093-41d3-b0e2-5c519333abdd] File "/opt/stack/nova/nova/compute/manager.py", line 4146, in _do_rebuild_instance [ 857.519907] env[61806]: ERROR nova.compute.manager [instance: 0758901a-7093-41d3-b0e2-5c519333abdd] self._rebuild_default_impl(**kwargs) [ 857.519907] env[61806]: ERROR nova.compute.manager [instance: 0758901a-7093-41d3-b0e2-5c519333abdd] File "/opt/stack/nova/nova/compute/manager.py", line 3723, in _rebuild_default_impl [ 857.519907] env[61806]: ERROR nova.compute.manager [instance: 0758901a-7093-41d3-b0e2-5c519333abdd] self._rebuild_volume_backed_instance( [ 857.519907] env[61806]: ERROR nova.compute.manager [instance: 0758901a-7093-41d3-b0e2-5c519333abdd] File "/opt/stack/nova/nova/compute/manager.py", line 3615, in _rebuild_volume_backed_instance [ 857.519907] env[61806]: ERROR nova.compute.manager [instance: 0758901a-7093-41d3-b0e2-5c519333abdd] raise exception.BuildAbortException( [ 857.519907] env[61806]: ERROR nova.compute.manager [instance: 0758901a-7093-41d3-b0e2-5c519333abdd] nova.exception.BuildAbortException: Build of instance 0758901a-7093-41d3-b0e2-5c519333abdd aborted: Failed to rebuild volume backed instance. [ 857.519907] env[61806]: ERROR nova.compute.manager [instance: 0758901a-7093-41d3-b0e2-5c519333abdd] [ 857.582867] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-470a6798-ca50-4aba-a915-7b527ddf6325 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 857.591176] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2f253b8c-f9ce-4bd5-bf0e-3ac6b3c21292 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 857.626022] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d405699b-0da9-4726-a1d3-c54b435c75b2 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 857.631911] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-38fb34b6-9df6-4966-8312-45ecd4d9b84f {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 857.645800] env[61806]: DEBUG nova.compute.provider_tree [None req-40a350df-b9c4-4f7d-968f-0160bcbc0d06 tempest-ServerAddressesNegativeTestJSON-1526246390 tempest-ServerAddressesNegativeTestJSON-1526246390-project-member] Inventory has not changed in ProviderTree for provider: a2858be1-fd22-4e08-979e-87ad25293407 {{(pid=61806) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 857.875246] env[61806]: DEBUG oslo_vmware.api [None req-4b919e4f-1f0b-40e0-b482-0547e22d55ff tempest-ImagesTestJSON-618852556 tempest-ImagesTestJSON-618852556-project-member] Task: {'id': task-1294747, 'name': Rename_Task, 'duration_secs': 0.13846} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 857.875645] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [None req-4b919e4f-1f0b-40e0-b482-0547e22d55ff tempest-ImagesTestJSON-618852556 tempest-ImagesTestJSON-618852556-project-member] [instance: 28e0baab-8516-42e3-acc2-9b8eb5192f57] Powering on the VM {{(pid=61806) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 857.875974] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-06660d8c-dcba-4b44-8a1e-07bf20119988 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 857.882526] env[61806]: DEBUG oslo_vmware.api [None req-4b919e4f-1f0b-40e0-b482-0547e22d55ff tempest-ImagesTestJSON-618852556 tempest-ImagesTestJSON-618852556-project-member] Waiting for the task: (returnval){ [ 857.882526] env[61806]: value = "task-1294748" [ 857.882526] env[61806]: _type = "Task" [ 857.882526] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 857.891047] env[61806]: DEBUG oslo_vmware.api [None req-4b919e4f-1f0b-40e0-b482-0547e22d55ff tempest-ImagesTestJSON-618852556 tempest-ImagesTestJSON-618852556-project-member] Task: {'id': task-1294748, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 857.999329] env[61806]: DEBUG oslo_concurrency.lockutils [None req-fbc8cbed-1b00-40c8-a788-e6e317b3c0cb tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] Lock "interface-2a3f81e3-b84f-4370-a598-2f6a607c3d47-64b56aa1-c5e3-4093-acee-b75f7ead331e" "released" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: held 9.326s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 858.151986] env[61806]: DEBUG nova.scheduler.client.report [None req-40a350df-b9c4-4f7d-968f-0160bcbc0d06 tempest-ServerAddressesNegativeTestJSON-1526246390 tempest-ServerAddressesNegativeTestJSON-1526246390-project-member] Inventory has not changed for provider a2858be1-fd22-4e08-979e-87ad25293407 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 138, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61806) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 858.393561] env[61806]: DEBUG oslo_vmware.api [None req-4b919e4f-1f0b-40e0-b482-0547e22d55ff tempest-ImagesTestJSON-618852556 tempest-ImagesTestJSON-618852556-project-member] Task: {'id': task-1294748, 'name': PowerOnVM_Task, 'duration_secs': 0.493272} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 858.394182] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [None req-4b919e4f-1f0b-40e0-b482-0547e22d55ff tempest-ImagesTestJSON-618852556 tempest-ImagesTestJSON-618852556-project-member] [instance: 28e0baab-8516-42e3-acc2-9b8eb5192f57] Powered on the VM {{(pid=61806) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 858.394182] env[61806]: INFO nova.compute.manager [None req-4b919e4f-1f0b-40e0-b482-0547e22d55ff tempest-ImagesTestJSON-618852556 tempest-ImagesTestJSON-618852556-project-member] [instance: 28e0baab-8516-42e3-acc2-9b8eb5192f57] Took 14.34 seconds to spawn the instance on the hypervisor. [ 858.394270] env[61806]: DEBUG nova.compute.manager [None req-4b919e4f-1f0b-40e0-b482-0547e22d55ff tempest-ImagesTestJSON-618852556 tempest-ImagesTestJSON-618852556-project-member] [instance: 28e0baab-8516-42e3-acc2-9b8eb5192f57] Checking state {{(pid=61806) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 858.395207] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-61ec55fd-3260-4d7e-9440-a20372d00ad7 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 858.656788] env[61806]: DEBUG oslo_concurrency.lockutils [None req-40a350df-b9c4-4f7d-968f-0160bcbc0d06 tempest-ServerAddressesNegativeTestJSON-1526246390 tempest-ServerAddressesNegativeTestJSON-1526246390-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.789s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 858.659552] env[61806]: DEBUG oslo_concurrency.lockutils [None req-f19a9d0d-15e4-4e05-ab2c-e194a287f3a9 tempest-AttachInterfacesV270Test-1974652806 tempest-AttachInterfacesV270Test-1974652806-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 8.467s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 858.659892] env[61806]: DEBUG nova.objects.instance [None req-f19a9d0d-15e4-4e05-ab2c-e194a287f3a9 tempest-AttachInterfacesV270Test-1974652806 tempest-AttachInterfacesV270Test-1974652806-project-member] Lazy-loading 'resources' on Instance uuid 117139cd-4ed4-4bea-8ad3-f4ac32c692e4 {{(pid=61806) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 858.691467] env[61806]: INFO nova.scheduler.client.report [None req-40a350df-b9c4-4f7d-968f-0160bcbc0d06 tempest-ServerAddressesNegativeTestJSON-1526246390 tempest-ServerAddressesNegativeTestJSON-1526246390-project-member] Deleted allocations for instance c8033551-6591-4e37-ae78-4efe7145b10b [ 858.916432] env[61806]: INFO nova.compute.manager [None req-4b919e4f-1f0b-40e0-b482-0547e22d55ff tempest-ImagesTestJSON-618852556 tempest-ImagesTestJSON-618852556-project-member] [instance: 28e0baab-8516-42e3-acc2-9b8eb5192f57] Took 33.43 seconds to build instance. [ 859.202353] env[61806]: DEBUG oslo_concurrency.lockutils [None req-40a350df-b9c4-4f7d-968f-0160bcbc0d06 tempest-ServerAddressesNegativeTestJSON-1526246390 tempest-ServerAddressesNegativeTestJSON-1526246390-project-member] Lock "c8033551-6591-4e37-ae78-4efe7145b10b" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 16.233s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 859.331167] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a64b20a4-57c6-4e47-b9c6-9e7371253fdd {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 859.339857] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ba589513-6e24-42aa-b354-74dba3fbe4a5 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 859.369681] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9ba9e2ac-4a62-498c-976d-433ad3943db5 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 859.377131] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e4aea0cb-118f-4690-a7b9-041f770fb55d {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 859.393470] env[61806]: DEBUG nova.compute.provider_tree [None req-f19a9d0d-15e4-4e05-ab2c-e194a287f3a9 tempest-AttachInterfacesV270Test-1974652806 tempest-AttachInterfacesV270Test-1974652806-project-member] Inventory has not changed in ProviderTree for provider: a2858be1-fd22-4e08-979e-87ad25293407 {{(pid=61806) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 859.419651] env[61806]: DEBUG oslo_concurrency.lockutils [None req-4b919e4f-1f0b-40e0-b482-0547e22d55ff tempest-ImagesTestJSON-618852556 tempest-ImagesTestJSON-618852556-project-member] Lock "28e0baab-8516-42e3-acc2-9b8eb5192f57" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 50.003s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 859.532342] env[61806]: DEBUG oslo_concurrency.lockutils [None req-9d9c27a2-c2da-48b4-9d8f-5aa531ec53d7 tempest-ServerActionsV293TestJSON-906151111 tempest-ServerActionsV293TestJSON-906151111-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 859.719699] env[61806]: DEBUG oslo_concurrency.lockutils [None req-13a0ddcb-4deb-45c0-93df-f8b645118f1e tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] Acquiring lock "interface-2a3f81e3-b84f-4370-a598-2f6a607c3d47-1f80e0d2-192f-4177-8212-acd6ca475851" by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 859.720071] env[61806]: DEBUG oslo_concurrency.lockutils [None req-13a0ddcb-4deb-45c0-93df-f8b645118f1e tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] Lock "interface-2a3f81e3-b84f-4370-a598-2f6a607c3d47-1f80e0d2-192f-4177-8212-acd6ca475851" acquired by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" :: waited 0.000s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 859.897127] env[61806]: DEBUG nova.scheduler.client.report [None req-f19a9d0d-15e4-4e05-ab2c-e194a287f3a9 tempest-AttachInterfacesV270Test-1974652806 tempest-AttachInterfacesV270Test-1974652806-project-member] Inventory has not changed for provider a2858be1-fd22-4e08-979e-87ad25293407 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 138, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61806) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 860.050928] env[61806]: DEBUG oslo_concurrency.lockutils [None req-20dfd5c3-f9cd-45b7-8782-70160298a47b tempest-ServerActionsV293TestJSON-906151111 tempest-ServerActionsV293TestJSON-906151111-project-member] Acquiring lock "0758901a-7093-41d3-b0e2-5c519333abdd" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 860.051296] env[61806]: DEBUG oslo_concurrency.lockutils [None req-20dfd5c3-f9cd-45b7-8782-70160298a47b tempest-ServerActionsV293TestJSON-906151111 tempest-ServerActionsV293TestJSON-906151111-project-member] Lock "0758901a-7093-41d3-b0e2-5c519333abdd" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 860.051540] env[61806]: DEBUG oslo_concurrency.lockutils [None req-20dfd5c3-f9cd-45b7-8782-70160298a47b tempest-ServerActionsV293TestJSON-906151111 tempest-ServerActionsV293TestJSON-906151111-project-member] Acquiring lock "0758901a-7093-41d3-b0e2-5c519333abdd-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 860.051735] env[61806]: DEBUG oslo_concurrency.lockutils [None req-20dfd5c3-f9cd-45b7-8782-70160298a47b tempest-ServerActionsV293TestJSON-906151111 tempest-ServerActionsV293TestJSON-906151111-project-member] Lock "0758901a-7093-41d3-b0e2-5c519333abdd-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 860.051911] env[61806]: DEBUG oslo_concurrency.lockutils [None req-20dfd5c3-f9cd-45b7-8782-70160298a47b tempest-ServerActionsV293TestJSON-906151111 tempest-ServerActionsV293TestJSON-906151111-project-member] Lock "0758901a-7093-41d3-b0e2-5c519333abdd-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 860.054315] env[61806]: INFO nova.compute.manager [None req-20dfd5c3-f9cd-45b7-8782-70160298a47b tempest-ServerActionsV293TestJSON-906151111 tempest-ServerActionsV293TestJSON-906151111-project-member] [instance: 0758901a-7093-41d3-b0e2-5c519333abdd] Terminating instance [ 860.056989] env[61806]: DEBUG nova.compute.manager [None req-20dfd5c3-f9cd-45b7-8782-70160298a47b tempest-ServerActionsV293TestJSON-906151111 tempest-ServerActionsV293TestJSON-906151111-project-member] [instance: 0758901a-7093-41d3-b0e2-5c519333abdd] Start destroying the instance on the hypervisor. {{(pid=61806) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 860.057293] env[61806]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-9e2c1391-ea2a-4c23-8b2d-ea962b5194ce {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 860.066443] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-19f1ffbf-b916-4aea-a584-91ce1c49f9a7 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 860.091105] env[61806]: WARNING nova.virt.vmwareapi.driver [None req-20dfd5c3-f9cd-45b7-8782-70160298a47b tempest-ServerActionsV293TestJSON-906151111 tempest-ServerActionsV293TestJSON-906151111-project-member] [instance: 0758901a-7093-41d3-b0e2-5c519333abdd] Instance does not exists. Proceeding to delete instance properties on datastore: nova.exception.InstanceNotFound: Instance 0758901a-7093-41d3-b0e2-5c519333abdd could not be found. [ 860.091333] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-20dfd5c3-f9cd-45b7-8782-70160298a47b tempest-ServerActionsV293TestJSON-906151111 tempest-ServerActionsV293TestJSON-906151111-project-member] [instance: 0758901a-7093-41d3-b0e2-5c519333abdd] Destroying instance {{(pid=61806) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 860.091987] env[61806]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-2df83440-805c-4b62-94f7-f3b6d6efe9cb {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 860.100981] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-65b0bd4f-0a81-4452-95ec-0fdc372c1c9a {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 860.125887] env[61806]: WARNING nova.virt.vmwareapi.vmops [None req-20dfd5c3-f9cd-45b7-8782-70160298a47b tempest-ServerActionsV293TestJSON-906151111 tempest-ServerActionsV293TestJSON-906151111-project-member] [instance: 0758901a-7093-41d3-b0e2-5c519333abdd] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 0758901a-7093-41d3-b0e2-5c519333abdd could not be found. [ 860.126107] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-20dfd5c3-f9cd-45b7-8782-70160298a47b tempest-ServerActionsV293TestJSON-906151111 tempest-ServerActionsV293TestJSON-906151111-project-member] [instance: 0758901a-7093-41d3-b0e2-5c519333abdd] Instance destroyed {{(pid=61806) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 860.126297] env[61806]: INFO nova.compute.manager [None req-20dfd5c3-f9cd-45b7-8782-70160298a47b tempest-ServerActionsV293TestJSON-906151111 tempest-ServerActionsV293TestJSON-906151111-project-member] [instance: 0758901a-7093-41d3-b0e2-5c519333abdd] Took 0.07 seconds to destroy the instance on the hypervisor. [ 860.126540] env[61806]: DEBUG oslo.service.loopingcall [None req-20dfd5c3-f9cd-45b7-8782-70160298a47b tempest-ServerActionsV293TestJSON-906151111 tempest-ServerActionsV293TestJSON-906151111-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61806) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 860.126885] env[61806]: DEBUG nova.compute.manager [-] [instance: 0758901a-7093-41d3-b0e2-5c519333abdd] Deallocating network for instance {{(pid=61806) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 860.127039] env[61806]: DEBUG nova.network.neutron [-] [instance: 0758901a-7093-41d3-b0e2-5c519333abdd] deallocate_for_instance() {{(pid=61806) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 860.185699] env[61806]: DEBUG oslo_concurrency.lockutils [None req-eec47054-9fa4-495a-837f-6d4ed5d258b3 tempest-ImagesTestJSON-618852556 tempest-ImagesTestJSON-618852556-project-member] Acquiring lock "28e0baab-8516-42e3-acc2-9b8eb5192f57" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 860.185929] env[61806]: DEBUG oslo_concurrency.lockutils [None req-eec47054-9fa4-495a-837f-6d4ed5d258b3 tempest-ImagesTestJSON-618852556 tempest-ImagesTestJSON-618852556-project-member] Lock "28e0baab-8516-42e3-acc2-9b8eb5192f57" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 860.186234] env[61806]: DEBUG oslo_concurrency.lockutils [None req-eec47054-9fa4-495a-837f-6d4ed5d258b3 tempest-ImagesTestJSON-618852556 tempest-ImagesTestJSON-618852556-project-member] Acquiring lock "28e0baab-8516-42e3-acc2-9b8eb5192f57-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 860.186325] env[61806]: DEBUG oslo_concurrency.lockutils [None req-eec47054-9fa4-495a-837f-6d4ed5d258b3 tempest-ImagesTestJSON-618852556 tempest-ImagesTestJSON-618852556-project-member] Lock "28e0baab-8516-42e3-acc2-9b8eb5192f57-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 860.186489] env[61806]: DEBUG oslo_concurrency.lockutils [None req-eec47054-9fa4-495a-837f-6d4ed5d258b3 tempest-ImagesTestJSON-618852556 tempest-ImagesTestJSON-618852556-project-member] Lock "28e0baab-8516-42e3-acc2-9b8eb5192f57-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 860.188547] env[61806]: INFO nova.compute.manager [None req-eec47054-9fa4-495a-837f-6d4ed5d258b3 tempest-ImagesTestJSON-618852556 tempest-ImagesTestJSON-618852556-project-member] [instance: 28e0baab-8516-42e3-acc2-9b8eb5192f57] Terminating instance [ 860.190696] env[61806]: DEBUG nova.compute.manager [None req-eec47054-9fa4-495a-837f-6d4ed5d258b3 tempest-ImagesTestJSON-618852556 tempest-ImagesTestJSON-618852556-project-member] [instance: 28e0baab-8516-42e3-acc2-9b8eb5192f57] Start destroying the instance on the hypervisor. {{(pid=61806) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 860.190857] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-eec47054-9fa4-495a-837f-6d4ed5d258b3 tempest-ImagesTestJSON-618852556 tempest-ImagesTestJSON-618852556-project-member] [instance: 28e0baab-8516-42e3-acc2-9b8eb5192f57] Destroying instance {{(pid=61806) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 860.191700] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ba788e71-9058-45fa-acab-aefd727383a9 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 860.198943] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [None req-eec47054-9fa4-495a-837f-6d4ed5d258b3 tempest-ImagesTestJSON-618852556 tempest-ImagesTestJSON-618852556-project-member] [instance: 28e0baab-8516-42e3-acc2-9b8eb5192f57] Powering off the VM {{(pid=61806) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 860.199231] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-6ba857e1-f79b-4e18-9df5-100eb6866445 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 860.204676] env[61806]: DEBUG oslo_vmware.api [None req-eec47054-9fa4-495a-837f-6d4ed5d258b3 tempest-ImagesTestJSON-618852556 tempest-ImagesTestJSON-618852556-project-member] Waiting for the task: (returnval){ [ 860.204676] env[61806]: value = "task-1294749" [ 860.204676] env[61806]: _type = "Task" [ 860.204676] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 860.213744] env[61806]: DEBUG oslo_vmware.api [None req-eec47054-9fa4-495a-837f-6d4ed5d258b3 tempest-ImagesTestJSON-618852556 tempest-ImagesTestJSON-618852556-project-member] Task: {'id': task-1294749, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 860.222442] env[61806]: DEBUG oslo_concurrency.lockutils [None req-13a0ddcb-4deb-45c0-93df-f8b645118f1e tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] Acquiring lock "2a3f81e3-b84f-4370-a598-2f6a607c3d47" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 860.222621] env[61806]: DEBUG oslo_concurrency.lockutils [None req-13a0ddcb-4deb-45c0-93df-f8b645118f1e tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] Acquired lock "2a3f81e3-b84f-4370-a598-2f6a607c3d47" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 860.223440] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ab35a210-f7ab-465c-849a-919d2ffcb587 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 860.243410] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6c597e18-ade0-436f-a456-a98d16ffb87d {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 860.272016] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-13a0ddcb-4deb-45c0-93df-f8b645118f1e tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] [instance: 2a3f81e3-b84f-4370-a598-2f6a607c3d47] Reconfiguring VM to detach interface {{(pid=61806) detach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1974}} [ 860.272723] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-18bd5ef5-8a21-4bb5-b1a6-a8eabdc00b98 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 860.291608] env[61806]: DEBUG oslo_vmware.api [None req-13a0ddcb-4deb-45c0-93df-f8b645118f1e tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] Waiting for the task: (returnval){ [ 860.291608] env[61806]: value = "task-1294750" [ 860.291608] env[61806]: _type = "Task" [ 860.291608] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 860.300466] env[61806]: DEBUG oslo_vmware.api [None req-13a0ddcb-4deb-45c0-93df-f8b645118f1e tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] Task: {'id': task-1294750, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 860.402421] env[61806]: DEBUG oslo_concurrency.lockutils [None req-f19a9d0d-15e4-4e05-ab2c-e194a287f3a9 tempest-AttachInterfacesV270Test-1974652806 tempest-AttachInterfacesV270Test-1974652806-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.743s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 860.407024] env[61806]: DEBUG oslo_concurrency.lockutils [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 8.926s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 860.435336] env[61806]: INFO nova.scheduler.client.report [None req-f19a9d0d-15e4-4e05-ab2c-e194a287f3a9 tempest-AttachInterfacesV270Test-1974652806 tempest-AttachInterfacesV270Test-1974652806-project-member] Deleted allocations for instance 117139cd-4ed4-4bea-8ad3-f4ac32c692e4 [ 860.715715] env[61806]: DEBUG oslo_vmware.api [None req-eec47054-9fa4-495a-837f-6d4ed5d258b3 tempest-ImagesTestJSON-618852556 tempest-ImagesTestJSON-618852556-project-member] Task: {'id': task-1294749, 'name': PowerOffVM_Task, 'duration_secs': 0.207736} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 860.715985] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [None req-eec47054-9fa4-495a-837f-6d4ed5d258b3 tempest-ImagesTestJSON-618852556 tempest-ImagesTestJSON-618852556-project-member] [instance: 28e0baab-8516-42e3-acc2-9b8eb5192f57] Powered off the VM {{(pid=61806) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 860.716169] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-eec47054-9fa4-495a-837f-6d4ed5d258b3 tempest-ImagesTestJSON-618852556 tempest-ImagesTestJSON-618852556-project-member] [instance: 28e0baab-8516-42e3-acc2-9b8eb5192f57] Unregistering the VM {{(pid=61806) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 860.716474] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-0c5058ae-d849-4944-9294-ca7a6130c932 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 860.730117] env[61806]: DEBUG nova.compute.manager [req-eb42afaf-8560-43f7-8cd9-503c705406c0 req-22743396-df1e-4e25-a9fc-5c7d11c1390c service nova] [instance: 0758901a-7093-41d3-b0e2-5c519333abdd] Received event network-vif-deleted-5ed8d8cc-fa88-438d-80e3-22e0c009373d {{(pid=61806) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 860.730117] env[61806]: INFO nova.compute.manager [req-eb42afaf-8560-43f7-8cd9-503c705406c0 req-22743396-df1e-4e25-a9fc-5c7d11c1390c service nova] [instance: 0758901a-7093-41d3-b0e2-5c519333abdd] Neutron deleted interface 5ed8d8cc-fa88-438d-80e3-22e0c009373d; detaching it from the instance and deleting it from the info cache [ 860.730117] env[61806]: DEBUG nova.network.neutron [req-eb42afaf-8560-43f7-8cd9-503c705406c0 req-22743396-df1e-4e25-a9fc-5c7d11c1390c service nova] [instance: 0758901a-7093-41d3-b0e2-5c519333abdd] Updating instance_info_cache with network_info: [] {{(pid=61806) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 860.779641] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-eec47054-9fa4-495a-837f-6d4ed5d258b3 tempest-ImagesTestJSON-618852556 tempest-ImagesTestJSON-618852556-project-member] [instance: 28e0baab-8516-42e3-acc2-9b8eb5192f57] Unregistered the VM {{(pid=61806) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 860.779972] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-eec47054-9fa4-495a-837f-6d4ed5d258b3 tempest-ImagesTestJSON-618852556 tempest-ImagesTestJSON-618852556-project-member] [instance: 28e0baab-8516-42e3-acc2-9b8eb5192f57] Deleting contents of the VM from datastore datastore2 {{(pid=61806) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 860.780241] env[61806]: DEBUG nova.virt.vmwareapi.ds_util [None req-eec47054-9fa4-495a-837f-6d4ed5d258b3 tempest-ImagesTestJSON-618852556 tempest-ImagesTestJSON-618852556-project-member] Deleting the datastore file [datastore2] 28e0baab-8516-42e3-acc2-9b8eb5192f57 {{(pid=61806) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 860.780573] env[61806]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-1329711a-b54d-4ec0-8b9c-da6bdfe185ef {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 860.787779] env[61806]: DEBUG oslo_vmware.api [None req-eec47054-9fa4-495a-837f-6d4ed5d258b3 tempest-ImagesTestJSON-618852556 tempest-ImagesTestJSON-618852556-project-member] Waiting for the task: (returnval){ [ 860.787779] env[61806]: value = "task-1294752" [ 860.787779] env[61806]: _type = "Task" [ 860.787779] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 860.800340] env[61806]: DEBUG oslo_vmware.api [None req-eec47054-9fa4-495a-837f-6d4ed5d258b3 tempest-ImagesTestJSON-618852556 tempest-ImagesTestJSON-618852556-project-member] Task: {'id': task-1294752, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 860.805809] env[61806]: DEBUG oslo_vmware.api [None req-13a0ddcb-4deb-45c0-93df-f8b645118f1e tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] Task: {'id': task-1294750, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 860.948635] env[61806]: DEBUG oslo_concurrency.lockutils [None req-f19a9d0d-15e4-4e05-ab2c-e194a287f3a9 tempest-AttachInterfacesV270Test-1974652806 tempest-AttachInterfacesV270Test-1974652806-project-member] Lock "117139cd-4ed4-4bea-8ad3-f4ac32c692e4" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 13.739s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 861.206695] env[61806]: DEBUG nova.network.neutron [-] [instance: 0758901a-7093-41d3-b0e2-5c519333abdd] Updating instance_info_cache with network_info: [] {{(pid=61806) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 861.235717] env[61806]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-d4ee87c7-9ea0-449a-9564-19c2234bc600 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 861.245766] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e8a11df7-70fa-4623-925a-9136aeec8028 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 861.270488] env[61806]: DEBUG nova.compute.manager [req-eb42afaf-8560-43f7-8cd9-503c705406c0 req-22743396-df1e-4e25-a9fc-5c7d11c1390c service nova] [instance: 0758901a-7093-41d3-b0e2-5c519333abdd] Detach interface failed, port_id=5ed8d8cc-fa88-438d-80e3-22e0c009373d, reason: Instance 0758901a-7093-41d3-b0e2-5c519333abdd could not be found. {{(pid=61806) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10963}} [ 861.302360] env[61806]: DEBUG oslo_vmware.api [None req-eec47054-9fa4-495a-837f-6d4ed5d258b3 tempest-ImagesTestJSON-618852556 tempest-ImagesTestJSON-618852556-project-member] Task: {'id': task-1294752, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.150413} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 861.305467] env[61806]: DEBUG nova.virt.vmwareapi.ds_util [None req-eec47054-9fa4-495a-837f-6d4ed5d258b3 tempest-ImagesTestJSON-618852556 tempest-ImagesTestJSON-618852556-project-member] Deleted the datastore file {{(pid=61806) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 861.305820] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-eec47054-9fa4-495a-837f-6d4ed5d258b3 tempest-ImagesTestJSON-618852556 tempest-ImagesTestJSON-618852556-project-member] [instance: 28e0baab-8516-42e3-acc2-9b8eb5192f57] Deleted contents of the VM from datastore datastore2 {{(pid=61806) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 861.306118] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-eec47054-9fa4-495a-837f-6d4ed5d258b3 tempest-ImagesTestJSON-618852556 tempest-ImagesTestJSON-618852556-project-member] [instance: 28e0baab-8516-42e3-acc2-9b8eb5192f57] Instance destroyed {{(pid=61806) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 861.306484] env[61806]: INFO nova.compute.manager [None req-eec47054-9fa4-495a-837f-6d4ed5d258b3 tempest-ImagesTestJSON-618852556 tempest-ImagesTestJSON-618852556-project-member] [instance: 28e0baab-8516-42e3-acc2-9b8eb5192f57] Took 1.12 seconds to destroy the instance on the hypervisor. [ 861.306834] env[61806]: DEBUG oslo.service.loopingcall [None req-eec47054-9fa4-495a-837f-6d4ed5d258b3 tempest-ImagesTestJSON-618852556 tempest-ImagesTestJSON-618852556-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61806) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 861.307175] env[61806]: DEBUG oslo_vmware.api [None req-13a0ddcb-4deb-45c0-93df-f8b645118f1e tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] Task: {'id': task-1294750, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 861.307488] env[61806]: DEBUG nova.compute.manager [-] [instance: 28e0baab-8516-42e3-acc2-9b8eb5192f57] Deallocating network for instance {{(pid=61806) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 861.307690] env[61806]: DEBUG nova.network.neutron [-] [instance: 28e0baab-8516-42e3-acc2-9b8eb5192f57] deallocate_for_instance() {{(pid=61806) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 861.454629] env[61806]: DEBUG nova.compute.resource_tracker [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Instance e0ef0a35-82a5-495b-9d5c-5805e8306390 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61806) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 861.454629] env[61806]: DEBUG nova.compute.resource_tracker [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Instance 2a3f81e3-b84f-4370-a598-2f6a607c3d47 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61806) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 861.454629] env[61806]: DEBUG nova.compute.resource_tracker [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Instance 43dcce0d-4fdb-4ee1-8309-92e92d2331a9 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61806) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 861.455288] env[61806]: DEBUG nova.compute.resource_tracker [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Instance 67d1bdd8-cf8f-4fb5-bd3c-42008e4d60e3 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61806) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 861.455288] env[61806]: WARNING nova.compute.resource_tracker [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Instance 7f4907d0-178d-452a-8149-030becde8779 is not being actively managed by this compute host but has allocations referencing this compute host: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. Skipping heal of allocation because we do not know what to do. [ 861.455288] env[61806]: WARNING nova.compute.resource_tracker [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Instance be19fc44-ecbe-489c-9b6e-be8957bb1dd0 is not being actively managed by this compute host but has allocations referencing this compute host: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. Skipping heal of allocation because we do not know what to do. [ 861.455288] env[61806]: DEBUG nova.compute.resource_tracker [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Instance 28e0baab-8516-42e3-acc2-9b8eb5192f57 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61806) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 861.712506] env[61806]: INFO nova.compute.manager [-] [instance: 0758901a-7093-41d3-b0e2-5c519333abdd] Took 1.59 seconds to deallocate network for instance. [ 861.805330] env[61806]: DEBUG oslo_vmware.api [None req-13a0ddcb-4deb-45c0-93df-f8b645118f1e tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] Task: {'id': task-1294750, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 862.155412] env[61806]: DEBUG nova.compute.resource_tracker [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Instance efa19aac-3ad7-42d5-a6d8-859050de6e63 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61806) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 862.155412] env[61806]: DEBUG nova.compute.resource_tracker [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Total usable vcpus: 48, total allocated vcpus: 6 {{(pid=61806) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 862.155412] env[61806]: DEBUG nova.compute.resource_tracker [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=1664MB phys_disk=200GB used_disk=5GB total_vcpus=48 used_vcpus=6 pci_stats=[] {{(pid=61806) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 862.162516] env[61806]: DEBUG nova.network.neutron [-] [instance: 28e0baab-8516-42e3-acc2-9b8eb5192f57] Updating instance_info_cache with network_info: [] {{(pid=61806) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 862.274904] env[61806]: INFO nova.compute.manager [None req-20dfd5c3-f9cd-45b7-8782-70160298a47b tempest-ServerActionsV293TestJSON-906151111 tempest-ServerActionsV293TestJSON-906151111-project-member] [instance: 0758901a-7093-41d3-b0e2-5c519333abdd] Took 0.56 seconds to detach 1 volumes for instance. [ 862.281279] env[61806]: DEBUG nova.compute.manager [None req-20dfd5c3-f9cd-45b7-8782-70160298a47b tempest-ServerActionsV293TestJSON-906151111 tempest-ServerActionsV293TestJSON-906151111-project-member] [instance: 0758901a-7093-41d3-b0e2-5c519333abdd] Deleting volume: 1b19f6ff-b694-445e-9bc1-d581e9268324 {{(pid=61806) _cleanup_volumes /opt/stack/nova/nova/compute/manager.py:3247}} [ 862.304708] env[61806]: DEBUG oslo_vmware.api [None req-13a0ddcb-4deb-45c0-93df-f8b645118f1e tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] Task: {'id': task-1294750, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 862.325931] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4a8ddc8c-a2d2-47ea-8a49-dabbdb23a09b {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 862.334814] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1aa0e0a0-3dad-43cc-a610-c46a69c42cfe {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 862.376035] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-02c5f5e3-ca21-4b49-9fa8-29115b742134 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 862.386386] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f60f264c-19c6-4ab9-8507-6dc6fb2b767d {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 862.404586] env[61806]: DEBUG nova.compute.provider_tree [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Inventory has not changed in ProviderTree for provider: a2858be1-fd22-4e08-979e-87ad25293407 {{(pid=61806) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 862.666563] env[61806]: INFO nova.compute.manager [-] [instance: 28e0baab-8516-42e3-acc2-9b8eb5192f57] Took 1.36 seconds to deallocate network for instance. [ 862.763770] env[61806]: DEBUG nova.compute.manager [req-a9ecfc16-659e-40e0-9c9f-8d73a9abe7e5 req-6dd37106-9b86-447a-ac70-5ef3dbdecd13 service nova] [instance: 28e0baab-8516-42e3-acc2-9b8eb5192f57] Received event network-vif-deleted-2be8aa59-fbf0-4081-955d-9e031781b864 {{(pid=61806) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 862.808615] env[61806]: DEBUG oslo_vmware.api [None req-13a0ddcb-4deb-45c0-93df-f8b645118f1e tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] Task: {'id': task-1294750, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 862.880664] env[61806]: DEBUG oslo_concurrency.lockutils [None req-20dfd5c3-f9cd-45b7-8782-70160298a47b tempest-ServerActionsV293TestJSON-906151111 tempest-ServerActionsV293TestJSON-906151111-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 862.908888] env[61806]: DEBUG nova.scheduler.client.report [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Inventory has not changed for provider a2858be1-fd22-4e08-979e-87ad25293407 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 138, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61806) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 863.175927] env[61806]: DEBUG oslo_concurrency.lockutils [None req-eec47054-9fa4-495a-837f-6d4ed5d258b3 tempest-ImagesTestJSON-618852556 tempest-ImagesTestJSON-618852556-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 863.306623] env[61806]: DEBUG oslo_vmware.api [None req-13a0ddcb-4deb-45c0-93df-f8b645118f1e tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] Task: {'id': task-1294750, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 863.415465] env[61806]: DEBUG nova.compute.resource_tracker [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=61806) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 863.415844] env[61806]: DEBUG oslo_concurrency.lockutils [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 3.010s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 863.416345] env[61806]: DEBUG oslo_concurrency.lockutils [None req-d93d1274-43bb-49c2-a2cd-82e2c5e1853d tempest-ServerMetadataNegativeTestJSON-456453902 tempest-ServerMetadataNegativeTestJSON-456453902-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 11.866s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 863.416682] env[61806]: DEBUG oslo_concurrency.lockutils [None req-d93d1274-43bb-49c2-a2cd-82e2c5e1853d tempest-ServerMetadataNegativeTestJSON-456453902 tempest-ServerMetadataNegativeTestJSON-456453902-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 863.421594] env[61806]: DEBUG oslo_concurrency.lockutils [None req-64647cce-deb3-4f3d-927e-8bb03c4dc9ee tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 10.419s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 863.422194] env[61806]: DEBUG oslo_concurrency.lockutils [None req-64647cce-deb3-4f3d-927e-8bb03c4dc9ee tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.001s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 863.427430] env[61806]: DEBUG oslo_concurrency.lockutils [None req-ca966d32-e3ef-4c81-a661-22cd95d9e4ec tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 6.345s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 863.431318] env[61806]: INFO nova.compute.claims [None req-ca966d32-e3ef-4c81-a661-22cd95d9e4ec tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] [instance: efa19aac-3ad7-42d5-a6d8-859050de6e63] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 863.474012] env[61806]: INFO nova.scheduler.client.report [None req-d93d1274-43bb-49c2-a2cd-82e2c5e1853d tempest-ServerMetadataNegativeTestJSON-456453902 tempest-ServerMetadataNegativeTestJSON-456453902-project-member] Deleted allocations for instance 7f4907d0-178d-452a-8149-030becde8779 [ 863.477127] env[61806]: INFO nova.scheduler.client.report [None req-64647cce-deb3-4f3d-927e-8bb03c4dc9ee tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Deleted allocations for instance be19fc44-ecbe-489c-9b6e-be8957bb1dd0 [ 863.630725] env[61806]: DEBUG oslo_concurrency.lockutils [None req-b996cb64-3154-4eef-9fcd-aa24605b9020 tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] Acquiring lock "d1802da4-a670-427d-84d7-0fcb4717e18f" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 863.632024] env[61806]: DEBUG oslo_concurrency.lockutils [None req-b996cb64-3154-4eef-9fcd-aa24605b9020 tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] Lock "d1802da4-a670-427d-84d7-0fcb4717e18f" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 863.808718] env[61806]: DEBUG oslo_vmware.api [None req-13a0ddcb-4deb-45c0-93df-f8b645118f1e tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] Task: {'id': task-1294750, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 863.992391] env[61806]: DEBUG oslo_concurrency.lockutils [None req-d93d1274-43bb-49c2-a2cd-82e2c5e1853d tempest-ServerMetadataNegativeTestJSON-456453902 tempest-ServerMetadataNegativeTestJSON-456453902-project-member] Lock "7f4907d0-178d-452a-8149-030becde8779" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 15.520s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 863.999666] env[61806]: DEBUG oslo_concurrency.lockutils [None req-64647cce-deb3-4f3d-927e-8bb03c4dc9ee tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Lock "be19fc44-ecbe-489c-9b6e-be8957bb1dd0" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 14.324s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 864.136024] env[61806]: DEBUG nova.compute.manager [None req-b996cb64-3154-4eef-9fcd-aa24605b9020 tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] [instance: d1802da4-a670-427d-84d7-0fcb4717e18f] Starting instance... {{(pid=61806) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 864.308054] env[61806]: DEBUG oslo_vmware.api [None req-13a0ddcb-4deb-45c0-93df-f8b645118f1e tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] Task: {'id': task-1294750, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 864.567704] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7334fee8-84f6-4172-be44-eed901955832 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 864.577172] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-86e33d27-7c62-4a02-966c-43e33c99e4b3 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 864.607999] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d571c08f-a1e2-4ca7-801b-9791a585dd27 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 864.616471] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e88c89e7-67bf-4875-ad52-783c2cf5484a {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 864.632274] env[61806]: DEBUG nova.compute.provider_tree [None req-ca966d32-e3ef-4c81-a661-22cd95d9e4ec tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Inventory has not changed in ProviderTree for provider: a2858be1-fd22-4e08-979e-87ad25293407 {{(pid=61806) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 864.655385] env[61806]: DEBUG oslo_concurrency.lockutils [None req-b996cb64-3154-4eef-9fcd-aa24605b9020 tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 864.808127] env[61806]: DEBUG oslo_vmware.api [None req-13a0ddcb-4deb-45c0-93df-f8b645118f1e tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] Task: {'id': task-1294750, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 865.138146] env[61806]: DEBUG nova.scheduler.client.report [None req-ca966d32-e3ef-4c81-a661-22cd95d9e4ec tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Inventory has not changed for provider a2858be1-fd22-4e08-979e-87ad25293407 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 138, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61806) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 865.309256] env[61806]: DEBUG oslo_vmware.api [None req-13a0ddcb-4deb-45c0-93df-f8b645118f1e tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] Task: {'id': task-1294750, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 865.437248] env[61806]: DEBUG oslo_concurrency.lockutils [None req-98f42353-addf-4bdc-9891-17066a23ed33 tempest-AttachVolumeNegativeTest-1683033704 tempest-AttachVolumeNegativeTest-1683033704-project-member] Acquiring lock "43dcce0d-4fdb-4ee1-8309-92e92d2331a9" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 865.437855] env[61806]: DEBUG oslo_concurrency.lockutils [None req-98f42353-addf-4bdc-9891-17066a23ed33 tempest-AttachVolumeNegativeTest-1683033704 tempest-AttachVolumeNegativeTest-1683033704-project-member] Lock "43dcce0d-4fdb-4ee1-8309-92e92d2331a9" acquired by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: waited 0.001s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 865.644420] env[61806]: DEBUG oslo_concurrency.lockutils [None req-ca966d32-e3ef-4c81-a661-22cd95d9e4ec tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.217s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 865.645190] env[61806]: DEBUG nova.compute.manager [None req-ca966d32-e3ef-4c81-a661-22cd95d9e4ec tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] [instance: efa19aac-3ad7-42d5-a6d8-859050de6e63] Start building networks asynchronously for instance. {{(pid=61806) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 865.650750] env[61806]: DEBUG oslo_concurrency.lockutils [None req-9d9c27a2-c2da-48b4-9d8f-5aa531ec53d7 tempest-ServerActionsV293TestJSON-906151111 tempest-ServerActionsV293TestJSON-906151111-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 6.117s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 865.711858] env[61806]: DEBUG oslo_concurrency.lockutils [None req-799f40df-a27b-4200-803a-166b5c1e29fd tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Acquiring lock "59c22c16-3d1c-4e64-bfac-7f8886a1927e" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 865.712361] env[61806]: DEBUG oslo_concurrency.lockutils [None req-799f40df-a27b-4200-803a-166b5c1e29fd tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Lock "59c22c16-3d1c-4e64-bfac-7f8886a1927e" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 865.806508] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-83b6595c-c69d-4c77-8fc1-591cac9e53a4 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 865.812197] env[61806]: DEBUG oslo_vmware.api [None req-13a0ddcb-4deb-45c0-93df-f8b645118f1e tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] Task: {'id': task-1294750, 'name': ReconfigVM_Task} progress is 18%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 865.817160] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-947675c6-c975-4fc5-b16f-d4a7b152b50c {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 865.849604] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e71faa52-5212-46ab-883a-d43344a269ea {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 865.858079] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6e13bfb3-ce38-4fc8-9d6b-83073e22f656 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 865.874656] env[61806]: DEBUG nova.compute.provider_tree [None req-9d9c27a2-c2da-48b4-9d8f-5aa531ec53d7 tempest-ServerActionsV293TestJSON-906151111 tempest-ServerActionsV293TestJSON-906151111-project-member] Inventory has not changed in ProviderTree for provider: a2858be1-fd22-4e08-979e-87ad25293407 {{(pid=61806) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 865.941643] env[61806]: DEBUG nova.compute.utils [None req-98f42353-addf-4bdc-9891-17066a23ed33 tempest-AttachVolumeNegativeTest-1683033704 tempest-AttachVolumeNegativeTest-1683033704-project-member] Using /dev/sd instead of None {{(pid=61806) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 866.154372] env[61806]: DEBUG nova.compute.utils [None req-ca966d32-e3ef-4c81-a661-22cd95d9e4ec tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Using /dev/sd instead of None {{(pid=61806) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 866.155820] env[61806]: DEBUG nova.compute.manager [None req-ca966d32-e3ef-4c81-a661-22cd95d9e4ec tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] [instance: efa19aac-3ad7-42d5-a6d8-859050de6e63] Allocating IP information in the background. {{(pid=61806) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 866.155992] env[61806]: DEBUG nova.network.neutron [None req-ca966d32-e3ef-4c81-a661-22cd95d9e4ec tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] [instance: efa19aac-3ad7-42d5-a6d8-859050de6e63] allocate_for_instance() {{(pid=61806) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 866.217617] env[61806]: DEBUG nova.compute.manager [None req-799f40df-a27b-4200-803a-166b5c1e29fd tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] [instance: 59c22c16-3d1c-4e64-bfac-7f8886a1927e] Starting instance... {{(pid=61806) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 866.224944] env[61806]: DEBUG nova.policy [None req-ca966d32-e3ef-4c81-a661-22cd95d9e4ec tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'b2631ce1f6dd40a4bfc5da7f90cde260', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '4b9d94c3d684474b99e305fb6cd9b95f', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61806) authorize /opt/stack/nova/nova/policy.py:201}} [ 866.310350] env[61806]: DEBUG oslo_vmware.api [None req-13a0ddcb-4deb-45c0-93df-f8b645118f1e tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] Task: {'id': task-1294750, 'name': ReconfigVM_Task, 'duration_secs': 5.785584} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 866.310653] env[61806]: DEBUG oslo_concurrency.lockutils [None req-13a0ddcb-4deb-45c0-93df-f8b645118f1e tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] Releasing lock "2a3f81e3-b84f-4370-a598-2f6a607c3d47" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 866.310869] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-13a0ddcb-4deb-45c0-93df-f8b645118f1e tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] [instance: 2a3f81e3-b84f-4370-a598-2f6a607c3d47] Reconfigured VM to detach interface {{(pid=61806) detach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1984}} [ 866.377649] env[61806]: DEBUG nova.scheduler.client.report [None req-9d9c27a2-c2da-48b4-9d8f-5aa531ec53d7 tempest-ServerActionsV293TestJSON-906151111 tempest-ServerActionsV293TestJSON-906151111-project-member] Inventory has not changed for provider a2858be1-fd22-4e08-979e-87ad25293407 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 138, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61806) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 866.444771] env[61806]: DEBUG oslo_concurrency.lockutils [None req-98f42353-addf-4bdc-9891-17066a23ed33 tempest-AttachVolumeNegativeTest-1683033704 tempest-AttachVolumeNegativeTest-1683033704-project-member] Lock "43dcce0d-4fdb-4ee1-8309-92e92d2331a9" "released" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: held 1.007s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 866.659546] env[61806]: DEBUG nova.compute.manager [None req-ca966d32-e3ef-4c81-a661-22cd95d9e4ec tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] [instance: efa19aac-3ad7-42d5-a6d8-859050de6e63] Start building block device mappings for instance. {{(pid=61806) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 866.744307] env[61806]: DEBUG oslo_concurrency.lockutils [None req-799f40df-a27b-4200-803a-166b5c1e29fd tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 866.884927] env[61806]: DEBUG oslo_concurrency.lockutils [None req-9d9c27a2-c2da-48b4-9d8f-5aa531ec53d7 tempest-ServerActionsV293TestJSON-906151111 tempest-ServerActionsV293TestJSON-906151111-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.233s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 866.884927] env[61806]: INFO nova.compute.manager [None req-9d9c27a2-c2da-48b4-9d8f-5aa531ec53d7 tempest-ServerActionsV293TestJSON-906151111 tempest-ServerActionsV293TestJSON-906151111-project-member] [instance: 0758901a-7093-41d3-b0e2-5c519333abdd] Successfully reverted task state from rebuilding on failure for instance. [ 866.889123] env[61806]: DEBUG oslo_concurrency.lockutils [None req-20dfd5c3-f9cd-45b7-8782-70160298a47b tempest-ServerActionsV293TestJSON-906151111 tempest-ServerActionsV293TestJSON-906151111-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 4.008s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 866.889123] env[61806]: DEBUG nova.objects.instance [None req-20dfd5c3-f9cd-45b7-8782-70160298a47b tempest-ServerActionsV293TestJSON-906151111 tempest-ServerActionsV293TestJSON-906151111-project-member] Lazy-loading 'resources' on Instance uuid 0758901a-7093-41d3-b0e2-5c519333abdd {{(pid=61806) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 867.075892] env[61806]: DEBUG nova.network.neutron [None req-ca966d32-e3ef-4c81-a661-22cd95d9e4ec tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] [instance: efa19aac-3ad7-42d5-a6d8-859050de6e63] Successfully created port: fea0c8d9-43ab-4462-9427-73d9ae4a47b2 {{(pid=61806) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 867.135771] env[61806]: DEBUG nova.compute.manager [req-cc465fbe-2cac-448e-9ba1-14895659ed54 req-9f011109-5090-497e-a4a6-1549890d17f2 service nova] [instance: 2a3f81e3-b84f-4370-a598-2f6a607c3d47] Received event network-vif-deleted-1f80e0d2-192f-4177-8212-acd6ca475851 {{(pid=61806) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 867.136018] env[61806]: INFO nova.compute.manager [req-cc465fbe-2cac-448e-9ba1-14895659ed54 req-9f011109-5090-497e-a4a6-1549890d17f2 service nova] [instance: 2a3f81e3-b84f-4370-a598-2f6a607c3d47] Neutron deleted interface 1f80e0d2-192f-4177-8212-acd6ca475851; detaching it from the instance and deleting it from the info cache [ 867.136331] env[61806]: DEBUG nova.network.neutron [req-cc465fbe-2cac-448e-9ba1-14895659ed54 req-9f011109-5090-497e-a4a6-1549890d17f2 service nova] [instance: 2a3f81e3-b84f-4370-a598-2f6a607c3d47] Updating instance_info_cache with network_info: [{"id": "a839bcaa-9990-4bad-be1f-1f06eb1c978e", "address": "fa:16:3e:0b:8a:a6", "network": {"id": "a7ffda72-e57c-41fa-a7ed-d7b12084bc9d", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-62794865-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.193", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "6e391e1267014fafbb8f5a3211299819", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8b3cfeb1-f262-4fd9-b506-8e9c0733e2d8", "external-id": "nsx-vlan-transportzone-119", "segmentation_id": 119, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa839bcaa-99", "ovs_interfaceid": "a839bcaa-9990-4bad-be1f-1f06eb1c978e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "64b56aa1-c5e3-4093-acee-b75f7ead331e", "address": "fa:16:3e:0e:4f:49", "network": {"id": "a7ffda72-e57c-41fa-a7ed-d7b12084bc9d", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-62794865-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "6e391e1267014fafbb8f5a3211299819", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8b3cfeb1-f262-4fd9-b506-8e9c0733e2d8", "external-id": "nsx-vlan-transportzone-119", "segmentation_id": 119, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap64b56aa1-c5", "ovs_interfaceid": "64b56aa1-c5e3-4093-acee-b75f7ead331e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": true, "delegate_create": true, "meta": {}}] {{(pid=61806) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 867.550865] env[61806]: DEBUG oslo_concurrency.lockutils [None req-98f42353-addf-4bdc-9891-17066a23ed33 tempest-AttachVolumeNegativeTest-1683033704 tempest-AttachVolumeNegativeTest-1683033704-project-member] Acquiring lock "43dcce0d-4fdb-4ee1-8309-92e92d2331a9" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 867.551165] env[61806]: DEBUG oslo_concurrency.lockutils [None req-98f42353-addf-4bdc-9891-17066a23ed33 tempest-AttachVolumeNegativeTest-1683033704 tempest-AttachVolumeNegativeTest-1683033704-project-member] Lock "43dcce0d-4fdb-4ee1-8309-92e92d2331a9" acquired by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: waited 0.000s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 867.551401] env[61806]: INFO nova.compute.manager [None req-98f42353-addf-4bdc-9891-17066a23ed33 tempest-AttachVolumeNegativeTest-1683033704 tempest-AttachVolumeNegativeTest-1683033704-project-member] [instance: 43dcce0d-4fdb-4ee1-8309-92e92d2331a9] Attaching volume 21f9e064-7d34-44c9-9759-32f3d73dec59 to /dev/sdb [ 867.587793] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4de3651a-8f47-471f-90d2-33b0424160da {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 867.599742] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-14ea704f-6c23-414e-a641-8f36f78e8469 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 867.638734] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c4c105c1-d8bc-46f4-b88c-e8c481ebb87b {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 867.641835] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4e328804-156f-467c-82e9-ca5763f7f6b8 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 867.645850] env[61806]: DEBUG oslo_concurrency.lockutils [req-cc465fbe-2cac-448e-9ba1-14895659ed54 req-9f011109-5090-497e-a4a6-1549890d17f2 service nova] Acquiring lock "2a3f81e3-b84f-4370-a598-2f6a607c3d47" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 867.645850] env[61806]: DEBUG oslo_concurrency.lockutils [req-cc465fbe-2cac-448e-9ba1-14895659ed54 req-9f011109-5090-497e-a4a6-1549890d17f2 service nova] Acquired lock "2a3f81e3-b84f-4370-a598-2f6a607c3d47" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 867.647321] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ee457fbc-1fe9-4740-8a9d-a6921eda5f02 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 867.657387] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e8b936d6-a2e8-4723-bf6e-6889f6fed2ed {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 867.677396] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c0a4d5c2-b939-4eb0-b4e0-0067404c272d {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 867.681518] env[61806]: DEBUG nova.compute.manager [None req-ca966d32-e3ef-4c81-a661-22cd95d9e4ec tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] [instance: efa19aac-3ad7-42d5-a6d8-859050de6e63] Start spawning the instance on the hypervisor. {{(pid=61806) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 867.683594] env[61806]: DEBUG oslo_concurrency.lockutils [req-cc465fbe-2cac-448e-9ba1-14895659ed54 req-9f011109-5090-497e-a4a6-1549890d17f2 service nova] Releasing lock "2a3f81e3-b84f-4370-a598-2f6a607c3d47" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 867.683849] env[61806]: WARNING nova.compute.manager [req-cc465fbe-2cac-448e-9ba1-14895659ed54 req-9f011109-5090-497e-a4a6-1549890d17f2 service nova] [instance: 2a3f81e3-b84f-4370-a598-2f6a607c3d47] Detach interface failed, port_id=1f80e0d2-192f-4177-8212-acd6ca475851, reason: No device with interface-id 1f80e0d2-192f-4177-8212-acd6ca475851 exists on VM: nova.exception.NotFound: No device with interface-id 1f80e0d2-192f-4177-8212-acd6ca475851 exists on VM [ 867.696898] env[61806]: DEBUG nova.compute.provider_tree [None req-20dfd5c3-f9cd-45b7-8782-70160298a47b tempest-ServerActionsV293TestJSON-906151111 tempest-ServerActionsV293TestJSON-906151111-project-member] Inventory has not changed in ProviderTree for provider: a2858be1-fd22-4e08-979e-87ad25293407 {{(pid=61806) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 867.702966] env[61806]: DEBUG nova.virt.block_device [None req-98f42353-addf-4bdc-9891-17066a23ed33 tempest-AttachVolumeNegativeTest-1683033704 tempest-AttachVolumeNegativeTest-1683033704-project-member] [instance: 43dcce0d-4fdb-4ee1-8309-92e92d2331a9] Updating existing volume attachment record: 0b5f9494-280a-4426-b072-a4b0d8a7395a {{(pid=61806) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 867.720631] env[61806]: DEBUG nova.virt.hardware [None req-ca966d32-e3ef-4c81-a661-22cd95d9e4ec tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-15T18:05:10Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-15T18:04:52Z,direct_url=,disk_format='vmdk',id=ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='84ba83fce2c349a6988173c5d6fc3b07',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-15T18:04:53Z,virtual_size=,visibility=), allow threads: False {{(pid=61806) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 867.720631] env[61806]: DEBUG nova.virt.hardware [None req-ca966d32-e3ef-4c81-a661-22cd95d9e4ec tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Flavor limits 0:0:0 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 867.720631] env[61806]: DEBUG nova.virt.hardware [None req-ca966d32-e3ef-4c81-a661-22cd95d9e4ec tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Image limits 0:0:0 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 867.720631] env[61806]: DEBUG nova.virt.hardware [None req-ca966d32-e3ef-4c81-a661-22cd95d9e4ec tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Flavor pref 0:0:0 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 867.720857] env[61806]: DEBUG nova.virt.hardware [None req-ca966d32-e3ef-4c81-a661-22cd95d9e4ec tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Image pref 0:0:0 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 867.721203] env[61806]: DEBUG nova.virt.hardware [None req-ca966d32-e3ef-4c81-a661-22cd95d9e4ec tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 867.721203] env[61806]: DEBUG nova.virt.hardware [None req-ca966d32-e3ef-4c81-a661-22cd95d9e4ec tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61806) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 867.722181] env[61806]: DEBUG nova.virt.hardware [None req-ca966d32-e3ef-4c81-a661-22cd95d9e4ec tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61806) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 867.722181] env[61806]: DEBUG nova.virt.hardware [None req-ca966d32-e3ef-4c81-a661-22cd95d9e4ec tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Got 1 possible topologies {{(pid=61806) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 867.722181] env[61806]: DEBUG nova.virt.hardware [None req-ca966d32-e3ef-4c81-a661-22cd95d9e4ec tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61806) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 867.722181] env[61806]: DEBUG nova.virt.hardware [None req-ca966d32-e3ef-4c81-a661-22cd95d9e4ec tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61806) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 867.723876] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3e2988dc-3b0a-4715-a063-882e743be102 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 867.735670] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e066b244-79ea-4a6f-9424-e04a35be2a38 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 868.108318] env[61806]: DEBUG oslo_concurrency.lockutils [None req-13a0ddcb-4deb-45c0-93df-f8b645118f1e tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] Acquiring lock "refresh_cache-2a3f81e3-b84f-4370-a598-2f6a607c3d47" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 868.108318] env[61806]: DEBUG oslo_concurrency.lockutils [None req-13a0ddcb-4deb-45c0-93df-f8b645118f1e tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] Acquired lock "refresh_cache-2a3f81e3-b84f-4370-a598-2f6a607c3d47" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 868.108318] env[61806]: DEBUG nova.network.neutron [None req-13a0ddcb-4deb-45c0-93df-f8b645118f1e tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] [instance: 2a3f81e3-b84f-4370-a598-2f6a607c3d47] Building network info cache for instance {{(pid=61806) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 868.202985] env[61806]: DEBUG nova.scheduler.client.report [None req-20dfd5c3-f9cd-45b7-8782-70160298a47b tempest-ServerActionsV293TestJSON-906151111 tempest-ServerActionsV293TestJSON-906151111-project-member] Inventory has not changed for provider a2858be1-fd22-4e08-979e-87ad25293407 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 138, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61806) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 868.264008] env[61806]: DEBUG oslo_concurrency.lockutils [None req-a7b134bb-7562-4442-8a02-daf53033761c tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] Acquiring lock "2a3f81e3-b84f-4370-a598-2f6a607c3d47" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 868.264367] env[61806]: DEBUG oslo_concurrency.lockutils [None req-a7b134bb-7562-4442-8a02-daf53033761c tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] Lock "2a3f81e3-b84f-4370-a598-2f6a607c3d47" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 868.264591] env[61806]: DEBUG oslo_concurrency.lockutils [None req-a7b134bb-7562-4442-8a02-daf53033761c tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] Acquiring lock "2a3f81e3-b84f-4370-a598-2f6a607c3d47-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 868.264868] env[61806]: DEBUG oslo_concurrency.lockutils [None req-a7b134bb-7562-4442-8a02-daf53033761c tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] Lock "2a3f81e3-b84f-4370-a598-2f6a607c3d47-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 868.264977] env[61806]: DEBUG oslo_concurrency.lockutils [None req-a7b134bb-7562-4442-8a02-daf53033761c tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] Lock "2a3f81e3-b84f-4370-a598-2f6a607c3d47-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 868.267279] env[61806]: INFO nova.compute.manager [None req-a7b134bb-7562-4442-8a02-daf53033761c tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] [instance: 2a3f81e3-b84f-4370-a598-2f6a607c3d47] Terminating instance [ 868.270521] env[61806]: DEBUG nova.compute.manager [None req-a7b134bb-7562-4442-8a02-daf53033761c tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] [instance: 2a3f81e3-b84f-4370-a598-2f6a607c3d47] Start destroying the instance on the hypervisor. {{(pid=61806) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 868.270521] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-a7b134bb-7562-4442-8a02-daf53033761c tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] [instance: 2a3f81e3-b84f-4370-a598-2f6a607c3d47] Destroying instance {{(pid=61806) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 868.272627] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-61f49450-bc78-4372-aa75-15454484a9b9 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 868.282285] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [None req-a7b134bb-7562-4442-8a02-daf53033761c tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] [instance: 2a3f81e3-b84f-4370-a598-2f6a607c3d47] Powering off the VM {{(pid=61806) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 868.282553] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-01ea5f22-cf8d-40eb-8d0a-371f6e19f7af {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 868.290761] env[61806]: DEBUG oslo_vmware.api [None req-a7b134bb-7562-4442-8a02-daf53033761c tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] Waiting for the task: (returnval){ [ 868.290761] env[61806]: value = "task-1294757" [ 868.290761] env[61806]: _type = "Task" [ 868.290761] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 868.300574] env[61806]: DEBUG oslo_vmware.api [None req-a7b134bb-7562-4442-8a02-daf53033761c tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] Task: {'id': task-1294757, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 868.346230] env[61806]: DEBUG oslo_concurrency.lockutils [None req-b85317d7-e79b-4ab3-83e4-2de701507449 tempest-ServerTagsTestJSON-1292926242 tempest-ServerTagsTestJSON-1292926242-project-member] Acquiring lock "9c20d337-b8f5-4965-b707-0f87c4ab6fc4" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 868.346230] env[61806]: DEBUG oslo_concurrency.lockutils [None req-b85317d7-e79b-4ab3-83e4-2de701507449 tempest-ServerTagsTestJSON-1292926242 tempest-ServerTagsTestJSON-1292926242-project-member] Lock "9c20d337-b8f5-4965-b707-0f87c4ab6fc4" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 868.709436] env[61806]: DEBUG oslo_concurrency.lockutils [None req-20dfd5c3-f9cd-45b7-8782-70160298a47b tempest-ServerActionsV293TestJSON-906151111 tempest-ServerActionsV293TestJSON-906151111-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.821s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 868.711731] env[61806]: DEBUG oslo_concurrency.lockutils [None req-eec47054-9fa4-495a-837f-6d4ed5d258b3 tempest-ImagesTestJSON-618852556 tempest-ImagesTestJSON-618852556-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 5.537s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 868.711966] env[61806]: DEBUG nova.objects.instance [None req-eec47054-9fa4-495a-837f-6d4ed5d258b3 tempest-ImagesTestJSON-618852556 tempest-ImagesTestJSON-618852556-project-member] Lazy-loading 'resources' on Instance uuid 28e0baab-8516-42e3-acc2-9b8eb5192f57 {{(pid=61806) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 868.803766] env[61806]: DEBUG oslo_vmware.api [None req-a7b134bb-7562-4442-8a02-daf53033761c tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] Task: {'id': task-1294757, 'name': PowerOffVM_Task, 'duration_secs': 0.263888} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 868.803766] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [None req-a7b134bb-7562-4442-8a02-daf53033761c tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] [instance: 2a3f81e3-b84f-4370-a598-2f6a607c3d47] Powered off the VM {{(pid=61806) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 868.803766] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-a7b134bb-7562-4442-8a02-daf53033761c tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] [instance: 2a3f81e3-b84f-4370-a598-2f6a607c3d47] Unregistering the VM {{(pid=61806) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 868.804070] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-839bc61d-1a6f-4f59-b9e0-cf24face00f5 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 868.848943] env[61806]: DEBUG nova.compute.manager [None req-b85317d7-e79b-4ab3-83e4-2de701507449 tempest-ServerTagsTestJSON-1292926242 tempest-ServerTagsTestJSON-1292926242-project-member] [instance: 9c20d337-b8f5-4965-b707-0f87c4ab6fc4] Starting instance... {{(pid=61806) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 868.904962] env[61806]: INFO nova.network.neutron [None req-13a0ddcb-4deb-45c0-93df-f8b645118f1e tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] [instance: 2a3f81e3-b84f-4370-a598-2f6a607c3d47] Port 64b56aa1-c5e3-4093-acee-b75f7ead331e from network info_cache is no longer associated with instance in Neutron. Removing from network info_cache. [ 868.905457] env[61806]: DEBUG nova.network.neutron [None req-13a0ddcb-4deb-45c0-93df-f8b645118f1e tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] [instance: 2a3f81e3-b84f-4370-a598-2f6a607c3d47] Updating instance_info_cache with network_info: [{"id": "a839bcaa-9990-4bad-be1f-1f06eb1c978e", "address": "fa:16:3e:0b:8a:a6", "network": {"id": "a7ffda72-e57c-41fa-a7ed-d7b12084bc9d", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-62794865-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.193", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "6e391e1267014fafbb8f5a3211299819", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8b3cfeb1-f262-4fd9-b506-8e9c0733e2d8", "external-id": "nsx-vlan-transportzone-119", "segmentation_id": 119, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa839bcaa-99", "ovs_interfaceid": "a839bcaa-9990-4bad-be1f-1f06eb1c978e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61806) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 868.908255] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-a7b134bb-7562-4442-8a02-daf53033761c tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] [instance: 2a3f81e3-b84f-4370-a598-2f6a607c3d47] Unregistered the VM {{(pid=61806) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 868.908492] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-a7b134bb-7562-4442-8a02-daf53033761c tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] [instance: 2a3f81e3-b84f-4370-a598-2f6a607c3d47] Deleting contents of the VM from datastore datastore1 {{(pid=61806) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 868.908678] env[61806]: DEBUG nova.virt.vmwareapi.ds_util [None req-a7b134bb-7562-4442-8a02-daf53033761c tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] Deleting the datastore file [datastore1] 2a3f81e3-b84f-4370-a598-2f6a607c3d47 {{(pid=61806) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 868.908966] env[61806]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-a20257ee-05c1-494f-8cba-2a95ef6b4e29 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 868.922276] env[61806]: DEBUG oslo_vmware.api [None req-a7b134bb-7562-4442-8a02-daf53033761c tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] Waiting for the task: (returnval){ [ 868.922276] env[61806]: value = "task-1294759" [ 868.922276] env[61806]: _type = "Task" [ 868.922276] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 868.930809] env[61806]: DEBUG oslo_vmware.api [None req-a7b134bb-7562-4442-8a02-daf53033761c tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] Task: {'id': task-1294759, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 869.020896] env[61806]: DEBUG oslo_concurrency.lockutils [None req-c651ea58-49e6-4c69-a5f7-88000b663e10 tempest-ServerRescueTestJSONUnderV235-749414849 tempest-ServerRescueTestJSONUnderV235-749414849-project-member] Acquiring lock "5a526856-0dcd-4b3f-8359-13a5c3b9bfc4" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 869.021147] env[61806]: DEBUG oslo_concurrency.lockutils [None req-c651ea58-49e6-4c69-a5f7-88000b663e10 tempest-ServerRescueTestJSONUnderV235-749414849 tempest-ServerRescueTestJSONUnderV235-749414849-project-member] Lock "5a526856-0dcd-4b3f-8359-13a5c3b9bfc4" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 869.159199] env[61806]: DEBUG nova.network.neutron [None req-ca966d32-e3ef-4c81-a661-22cd95d9e4ec tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] [instance: efa19aac-3ad7-42d5-a6d8-859050de6e63] Successfully updated port: fea0c8d9-43ab-4462-9427-73d9ae4a47b2 {{(pid=61806) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 869.169978] env[61806]: DEBUG nova.compute.manager [req-72cb8b39-6413-4e1d-a14d-bbe174b5efdd req-28204b7f-b430-4355-819f-32febc086130 service nova] [instance: 2a3f81e3-b84f-4370-a598-2f6a607c3d47] Received event network-vif-deleted-64b56aa1-c5e3-4093-acee-b75f7ead331e {{(pid=61806) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 869.169978] env[61806]: DEBUG nova.compute.manager [req-72cb8b39-6413-4e1d-a14d-bbe174b5efdd req-28204b7f-b430-4355-819f-32febc086130 service nova] [instance: efa19aac-3ad7-42d5-a6d8-859050de6e63] Received event network-vif-plugged-fea0c8d9-43ab-4462-9427-73d9ae4a47b2 {{(pid=61806) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 869.170204] env[61806]: DEBUG oslo_concurrency.lockutils [req-72cb8b39-6413-4e1d-a14d-bbe174b5efdd req-28204b7f-b430-4355-819f-32febc086130 service nova] Acquiring lock "efa19aac-3ad7-42d5-a6d8-859050de6e63-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 869.170412] env[61806]: DEBUG oslo_concurrency.lockutils [req-72cb8b39-6413-4e1d-a14d-bbe174b5efdd req-28204b7f-b430-4355-819f-32febc086130 service nova] Lock "efa19aac-3ad7-42d5-a6d8-859050de6e63-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 869.170591] env[61806]: DEBUG oslo_concurrency.lockutils [req-72cb8b39-6413-4e1d-a14d-bbe174b5efdd req-28204b7f-b430-4355-819f-32febc086130 service nova] Lock "efa19aac-3ad7-42d5-a6d8-859050de6e63-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 869.170763] env[61806]: DEBUG nova.compute.manager [req-72cb8b39-6413-4e1d-a14d-bbe174b5efdd req-28204b7f-b430-4355-819f-32febc086130 service nova] [instance: efa19aac-3ad7-42d5-a6d8-859050de6e63] No waiting events found dispatching network-vif-plugged-fea0c8d9-43ab-4462-9427-73d9ae4a47b2 {{(pid=61806) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 869.170970] env[61806]: WARNING nova.compute.manager [req-72cb8b39-6413-4e1d-a14d-bbe174b5efdd req-28204b7f-b430-4355-819f-32febc086130 service nova] [instance: efa19aac-3ad7-42d5-a6d8-859050de6e63] Received unexpected event network-vif-plugged-fea0c8d9-43ab-4462-9427-73d9ae4a47b2 for instance with vm_state building and task_state spawning. [ 869.228452] env[61806]: DEBUG oslo_concurrency.lockutils [None req-20dfd5c3-f9cd-45b7-8782-70160298a47b tempest-ServerActionsV293TestJSON-906151111 tempest-ServerActionsV293TestJSON-906151111-project-member] Lock "0758901a-7093-41d3-b0e2-5c519333abdd" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 9.177s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 869.369524] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-df3caded-be92-4e48-8b5d-e2793a72350b {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 869.373508] env[61806]: DEBUG oslo_concurrency.lockutils [None req-b85317d7-e79b-4ab3-83e4-2de701507449 tempest-ServerTagsTestJSON-1292926242 tempest-ServerTagsTestJSON-1292926242-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 869.379526] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-00cd34bb-7b5a-4997-8e39-dbaaa2642594 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 869.415626] env[61806]: DEBUG oslo_concurrency.lockutils [None req-13a0ddcb-4deb-45c0-93df-f8b645118f1e tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] Releasing lock "refresh_cache-2a3f81e3-b84f-4370-a598-2f6a607c3d47" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 869.418954] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f7849473-f610-4e3f-b7de-ba53a05034d0 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 869.434774] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7fa0ecb7-5907-4409-8016-043ccf43f86b {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 869.438981] env[61806]: DEBUG oslo_vmware.api [None req-a7b134bb-7562-4442-8a02-daf53033761c tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] Task: {'id': task-1294759, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.210045} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 869.441164] env[61806]: DEBUG nova.virt.vmwareapi.ds_util [None req-a7b134bb-7562-4442-8a02-daf53033761c tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] Deleted the datastore file {{(pid=61806) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 869.441164] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-a7b134bb-7562-4442-8a02-daf53033761c tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] [instance: 2a3f81e3-b84f-4370-a598-2f6a607c3d47] Deleted contents of the VM from datastore datastore1 {{(pid=61806) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 869.441164] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-a7b134bb-7562-4442-8a02-daf53033761c tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] [instance: 2a3f81e3-b84f-4370-a598-2f6a607c3d47] Instance destroyed {{(pid=61806) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 869.441164] env[61806]: INFO nova.compute.manager [None req-a7b134bb-7562-4442-8a02-daf53033761c tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] [instance: 2a3f81e3-b84f-4370-a598-2f6a607c3d47] Took 1.17 seconds to destroy the instance on the hypervisor. [ 869.441164] env[61806]: DEBUG oslo.service.loopingcall [None req-a7b134bb-7562-4442-8a02-daf53033761c tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61806) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 869.441164] env[61806]: DEBUG nova.compute.manager [-] [instance: 2a3f81e3-b84f-4370-a598-2f6a607c3d47] Deallocating network for instance {{(pid=61806) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 869.441437] env[61806]: DEBUG nova.network.neutron [-] [instance: 2a3f81e3-b84f-4370-a598-2f6a607c3d47] deallocate_for_instance() {{(pid=61806) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 869.451586] env[61806]: DEBUG nova.compute.provider_tree [None req-eec47054-9fa4-495a-837f-6d4ed5d258b3 tempest-ImagesTestJSON-618852556 tempest-ImagesTestJSON-618852556-project-member] Inventory has not changed in ProviderTree for provider: a2858be1-fd22-4e08-979e-87ad25293407 {{(pid=61806) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 869.507051] env[61806]: DEBUG neutronclient.v2_0.client [-] Error message: {"NeutronError": {"type": "PortNotFound", "message": "Port 64b56aa1-c5e3-4093-acee-b75f7ead331e could not be found.", "detail": ""}} {{(pid=61806) _handle_fault_response /opt/stack/data/venv/lib/python3.10/site-packages/neutronclient/v2_0/client.py:262}} [ 869.507299] env[61806]: DEBUG nova.network.neutron [-] Unable to show port 64b56aa1-c5e3-4093-acee-b75f7ead331e as it no longer exists. {{(pid=61806) _unbind_ports /opt/stack/nova/nova/network/neutron.py:666}} [ 869.525246] env[61806]: DEBUG nova.compute.manager [None req-c651ea58-49e6-4c69-a5f7-88000b663e10 tempest-ServerRescueTestJSONUnderV235-749414849 tempest-ServerRescueTestJSONUnderV235-749414849-project-member] [instance: 5a526856-0dcd-4b3f-8359-13a5c3b9bfc4] Starting instance... {{(pid=61806) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 869.662045] env[61806]: DEBUG oslo_concurrency.lockutils [None req-ca966d32-e3ef-4c81-a661-22cd95d9e4ec tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Acquiring lock "refresh_cache-efa19aac-3ad7-42d5-a6d8-859050de6e63" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 869.662210] env[61806]: DEBUG oslo_concurrency.lockutils [None req-ca966d32-e3ef-4c81-a661-22cd95d9e4ec tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Acquired lock "refresh_cache-efa19aac-3ad7-42d5-a6d8-859050de6e63" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 869.662364] env[61806]: DEBUG nova.network.neutron [None req-ca966d32-e3ef-4c81-a661-22cd95d9e4ec tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] [instance: efa19aac-3ad7-42d5-a6d8-859050de6e63] Building network info cache for instance {{(pid=61806) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 869.924333] env[61806]: DEBUG oslo_concurrency.lockutils [None req-13a0ddcb-4deb-45c0-93df-f8b645118f1e tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] Lock "interface-2a3f81e3-b84f-4370-a598-2f6a607c3d47-1f80e0d2-192f-4177-8212-acd6ca475851" "released" by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" :: held 10.204s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 869.957321] env[61806]: DEBUG nova.scheduler.client.report [None req-eec47054-9fa4-495a-837f-6d4ed5d258b3 tempest-ImagesTestJSON-618852556 tempest-ImagesTestJSON-618852556-project-member] Inventory has not changed for provider a2858be1-fd22-4e08-979e-87ad25293407 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 138, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61806) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 870.067258] env[61806]: DEBUG oslo_concurrency.lockutils [None req-c651ea58-49e6-4c69-a5f7-88000b663e10 tempest-ServerRescueTestJSONUnderV235-749414849 tempest-ServerRescueTestJSONUnderV235-749414849-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 870.194686] env[61806]: DEBUG nova.network.neutron [None req-ca966d32-e3ef-4c81-a661-22cd95d9e4ec tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] [instance: efa19aac-3ad7-42d5-a6d8-859050de6e63] Instance cache missing network info. {{(pid=61806) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 870.329619] env[61806]: DEBUG nova.network.neutron [None req-ca966d32-e3ef-4c81-a661-22cd95d9e4ec tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] [instance: efa19aac-3ad7-42d5-a6d8-859050de6e63] Updating instance_info_cache with network_info: [{"id": "fea0c8d9-43ab-4462-9427-73d9ae4a47b2", "address": "fa:16:3e:18:2f:85", "network": {"id": "46515a4e-610f-42fb-a424-205a03df0640", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-1133623547-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4b9d94c3d684474b99e305fb6cd9b95f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "82ca17df-257e-40e6-9ec9-310ed6f05ccb", "external-id": "nsx-vlan-transportzone-778", "segmentation_id": 778, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapfea0c8d9-43", "ovs_interfaceid": "fea0c8d9-43ab-4462-9427-73d9ae4a47b2", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61806) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 870.399644] env[61806]: DEBUG nova.network.neutron [-] [instance: 2a3f81e3-b84f-4370-a598-2f6a607c3d47] Updating instance_info_cache with network_info: [] {{(pid=61806) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 870.460066] env[61806]: DEBUG oslo_concurrency.lockutils [None req-eec47054-9fa4-495a-837f-6d4ed5d258b3 tempest-ImagesTestJSON-618852556 tempest-ImagesTestJSON-618852556-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.748s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 870.463869] env[61806]: DEBUG oslo_concurrency.lockutils [None req-b996cb64-3154-4eef-9fcd-aa24605b9020 tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 5.807s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 870.464274] env[61806]: INFO nova.compute.claims [None req-b996cb64-3154-4eef-9fcd-aa24605b9020 tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] [instance: d1802da4-a670-427d-84d7-0fcb4717e18f] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 870.485693] env[61806]: INFO nova.scheduler.client.report [None req-eec47054-9fa4-495a-837f-6d4ed5d258b3 tempest-ImagesTestJSON-618852556 tempest-ImagesTestJSON-618852556-project-member] Deleted allocations for instance 28e0baab-8516-42e3-acc2-9b8eb5192f57 [ 870.832545] env[61806]: DEBUG oslo_concurrency.lockutils [None req-ca966d32-e3ef-4c81-a661-22cd95d9e4ec tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Releasing lock "refresh_cache-efa19aac-3ad7-42d5-a6d8-859050de6e63" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 870.832870] env[61806]: DEBUG nova.compute.manager [None req-ca966d32-e3ef-4c81-a661-22cd95d9e4ec tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] [instance: efa19aac-3ad7-42d5-a6d8-859050de6e63] Instance network_info: |[{"id": "fea0c8d9-43ab-4462-9427-73d9ae4a47b2", "address": "fa:16:3e:18:2f:85", "network": {"id": "46515a4e-610f-42fb-a424-205a03df0640", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-1133623547-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4b9d94c3d684474b99e305fb6cd9b95f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "82ca17df-257e-40e6-9ec9-310ed6f05ccb", "external-id": "nsx-vlan-transportzone-778", "segmentation_id": 778, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapfea0c8d9-43", "ovs_interfaceid": "fea0c8d9-43ab-4462-9427-73d9ae4a47b2", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61806) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 870.833466] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-ca966d32-e3ef-4c81-a661-22cd95d9e4ec tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] [instance: efa19aac-3ad7-42d5-a6d8-859050de6e63] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:18:2f:85', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '82ca17df-257e-40e6-9ec9-310ed6f05ccb', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'fea0c8d9-43ab-4462-9427-73d9ae4a47b2', 'vif_model': 'vmxnet3'}] {{(pid=61806) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 870.841147] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [None req-ca966d32-e3ef-4c81-a661-22cd95d9e4ec tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Creating folder: Project (4b9d94c3d684474b99e305fb6cd9b95f). Parent ref: group-v277609. {{(pid=61806) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 870.841801] env[61806]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-1d5b68ab-b1f4-4157-801c-69a708229ec3 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 870.853587] env[61806]: INFO nova.virt.vmwareapi.vm_util [None req-ca966d32-e3ef-4c81-a661-22cd95d9e4ec tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Created folder: Project (4b9d94c3d684474b99e305fb6cd9b95f) in parent group-v277609. [ 870.853783] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [None req-ca966d32-e3ef-4c81-a661-22cd95d9e4ec tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Creating folder: Instances. Parent ref: group-v277687. {{(pid=61806) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 870.854083] env[61806]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-f41c2a68-99fc-4ff0-9259-21f69e10db47 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 870.864038] env[61806]: INFO nova.virt.vmwareapi.vm_util [None req-ca966d32-e3ef-4c81-a661-22cd95d9e4ec tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Created folder: Instances in parent group-v277687. [ 870.864302] env[61806]: DEBUG oslo.service.loopingcall [None req-ca966d32-e3ef-4c81-a661-22cd95d9e4ec tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61806) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 870.864497] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: efa19aac-3ad7-42d5-a6d8-859050de6e63] Creating VM on the ESX host {{(pid=61806) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 870.864701] env[61806]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-f1a9bee8-52e8-4640-84fc-44dd62be5e4e {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 870.884424] env[61806]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 870.884424] env[61806]: value = "task-1294763" [ 870.884424] env[61806]: _type = "Task" [ 870.884424] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 870.893569] env[61806]: DEBUG oslo_vmware.api [-] Task: {'id': task-1294763, 'name': CreateVM_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 870.902125] env[61806]: INFO nova.compute.manager [-] [instance: 2a3f81e3-b84f-4370-a598-2f6a607c3d47] Took 1.46 seconds to deallocate network for instance. [ 870.993281] env[61806]: DEBUG oslo_concurrency.lockutils [None req-eec47054-9fa4-495a-837f-6d4ed5d258b3 tempest-ImagesTestJSON-618852556 tempest-ImagesTestJSON-618852556-project-member] Lock "28e0baab-8516-42e3-acc2-9b8eb5192f57" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 10.807s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 871.216664] env[61806]: DEBUG nova.compute.manager [req-5ffb675e-f2c7-4c1c-80ad-b1930d7f32a1 req-bdcd0a67-aad8-44c2-bdfc-1c1a6d2a1bca service nova] [instance: efa19aac-3ad7-42d5-a6d8-859050de6e63] Received event network-changed-fea0c8d9-43ab-4462-9427-73d9ae4a47b2 {{(pid=61806) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 871.216878] env[61806]: DEBUG nova.compute.manager [req-5ffb675e-f2c7-4c1c-80ad-b1930d7f32a1 req-bdcd0a67-aad8-44c2-bdfc-1c1a6d2a1bca service nova] [instance: efa19aac-3ad7-42d5-a6d8-859050de6e63] Refreshing instance network info cache due to event network-changed-fea0c8d9-43ab-4462-9427-73d9ae4a47b2. {{(pid=61806) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 871.217108] env[61806]: DEBUG oslo_concurrency.lockutils [req-5ffb675e-f2c7-4c1c-80ad-b1930d7f32a1 req-bdcd0a67-aad8-44c2-bdfc-1c1a6d2a1bca service nova] Acquiring lock "refresh_cache-efa19aac-3ad7-42d5-a6d8-859050de6e63" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 871.217263] env[61806]: DEBUG oslo_concurrency.lockutils [req-5ffb675e-f2c7-4c1c-80ad-b1930d7f32a1 req-bdcd0a67-aad8-44c2-bdfc-1c1a6d2a1bca service nova] Acquired lock "refresh_cache-efa19aac-3ad7-42d5-a6d8-859050de6e63" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 871.217428] env[61806]: DEBUG nova.network.neutron [req-5ffb675e-f2c7-4c1c-80ad-b1930d7f32a1 req-bdcd0a67-aad8-44c2-bdfc-1c1a6d2a1bca service nova] [instance: efa19aac-3ad7-42d5-a6d8-859050de6e63] Refreshing network info cache for port fea0c8d9-43ab-4462-9427-73d9ae4a47b2 {{(pid=61806) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 871.395528] env[61806]: DEBUG oslo_vmware.api [-] Task: {'id': task-1294763, 'name': CreateVM_Task, 'duration_secs': 0.305608} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 871.395622] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: efa19aac-3ad7-42d5-a6d8-859050de6e63] Created VM on the ESX host {{(pid=61806) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 871.396340] env[61806]: DEBUG oslo_concurrency.lockutils [None req-ca966d32-e3ef-4c81-a661-22cd95d9e4ec tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 871.396511] env[61806]: DEBUG oslo_concurrency.lockutils [None req-ca966d32-e3ef-4c81-a661-22cd95d9e4ec tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Acquired lock "[datastore2] devstack-image-cache_base/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 871.396845] env[61806]: DEBUG oslo_concurrency.lockutils [None req-ca966d32-e3ef-4c81-a661-22cd95d9e4ec tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 871.397237] env[61806]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c095a706-d7aa-4284-ba8d-8bae581b4f31 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 871.402300] env[61806]: DEBUG oslo_vmware.api [None req-ca966d32-e3ef-4c81-a661-22cd95d9e4ec tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Waiting for the task: (returnval){ [ 871.402300] env[61806]: value = "session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]52e0998a-d1e9-0c8d-612a-988c9a4655bb" [ 871.402300] env[61806]: _type = "Task" [ 871.402300] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 871.411518] env[61806]: DEBUG oslo_concurrency.lockutils [None req-a7b134bb-7562-4442-8a02-daf53033761c tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 871.411799] env[61806]: DEBUG oslo_vmware.api [None req-ca966d32-e3ef-4c81-a661-22cd95d9e4ec tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Task: {'id': session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]52e0998a-d1e9-0c8d-612a-988c9a4655bb, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 871.588122] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-714ec453-6faa-4b8f-bb8e-314f217774ab {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 871.596850] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-88eb2110-b03d-4970-a097-71d751e1b00e {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 871.630020] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-83fbd226-584c-4497-af46-6f42d5abc38f {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 871.638781] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5011fd20-c509-400e-9519-653b27993e4e {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 871.654755] env[61806]: DEBUG nova.compute.provider_tree [None req-b996cb64-3154-4eef-9fcd-aa24605b9020 tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] Inventory has not changed in ProviderTree for provider: a2858be1-fd22-4e08-979e-87ad25293407 {{(pid=61806) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 871.813821] env[61806]: DEBUG oslo_concurrency.lockutils [None req-708d7cc8-bb4e-45d0-adad-8f77e0f096a2 tempest-ImagesTestJSON-618852556 tempest-ImagesTestJSON-618852556-project-member] Acquiring lock "e0ef0a35-82a5-495b-9d5c-5805e8306390" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 871.814112] env[61806]: DEBUG oslo_concurrency.lockutils [None req-708d7cc8-bb4e-45d0-adad-8f77e0f096a2 tempest-ImagesTestJSON-618852556 tempest-ImagesTestJSON-618852556-project-member] Lock "e0ef0a35-82a5-495b-9d5c-5805e8306390" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 871.814345] env[61806]: DEBUG oslo_concurrency.lockutils [None req-708d7cc8-bb4e-45d0-adad-8f77e0f096a2 tempest-ImagesTestJSON-618852556 tempest-ImagesTestJSON-618852556-project-member] Acquiring lock "e0ef0a35-82a5-495b-9d5c-5805e8306390-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 871.814539] env[61806]: DEBUG oslo_concurrency.lockutils [None req-708d7cc8-bb4e-45d0-adad-8f77e0f096a2 tempest-ImagesTestJSON-618852556 tempest-ImagesTestJSON-618852556-project-member] Lock "e0ef0a35-82a5-495b-9d5c-5805e8306390-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 871.814717] env[61806]: DEBUG oslo_concurrency.lockutils [None req-708d7cc8-bb4e-45d0-adad-8f77e0f096a2 tempest-ImagesTestJSON-618852556 tempest-ImagesTestJSON-618852556-project-member] Lock "e0ef0a35-82a5-495b-9d5c-5805e8306390-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 871.817217] env[61806]: INFO nova.compute.manager [None req-708d7cc8-bb4e-45d0-adad-8f77e0f096a2 tempest-ImagesTestJSON-618852556 tempest-ImagesTestJSON-618852556-project-member] [instance: e0ef0a35-82a5-495b-9d5c-5805e8306390] Terminating instance [ 871.818941] env[61806]: DEBUG nova.compute.manager [None req-708d7cc8-bb4e-45d0-adad-8f77e0f096a2 tempest-ImagesTestJSON-618852556 tempest-ImagesTestJSON-618852556-project-member] [instance: e0ef0a35-82a5-495b-9d5c-5805e8306390] Start destroying the instance on the hypervisor. {{(pid=61806) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 871.819136] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-708d7cc8-bb4e-45d0-adad-8f77e0f096a2 tempest-ImagesTestJSON-618852556 tempest-ImagesTestJSON-618852556-project-member] [instance: e0ef0a35-82a5-495b-9d5c-5805e8306390] Destroying instance {{(pid=61806) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 871.819965] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c9c1c6db-3fc1-4065-8fa2-3b949b2f771c {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 871.830768] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [None req-708d7cc8-bb4e-45d0-adad-8f77e0f096a2 tempest-ImagesTestJSON-618852556 tempest-ImagesTestJSON-618852556-project-member] [instance: e0ef0a35-82a5-495b-9d5c-5805e8306390] Powering off the VM {{(pid=61806) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 871.831132] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-deb6e1eb-2988-4abf-9ddc-b69f8f365098 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 871.837860] env[61806]: DEBUG oslo_vmware.api [None req-708d7cc8-bb4e-45d0-adad-8f77e0f096a2 tempest-ImagesTestJSON-618852556 tempest-ImagesTestJSON-618852556-project-member] Waiting for the task: (returnval){ [ 871.837860] env[61806]: value = "task-1294764" [ 871.837860] env[61806]: _type = "Task" [ 871.837860] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 871.847999] env[61806]: DEBUG oslo_vmware.api [None req-708d7cc8-bb4e-45d0-adad-8f77e0f096a2 tempest-ImagesTestJSON-618852556 tempest-ImagesTestJSON-618852556-project-member] Task: {'id': task-1294764, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 871.912767] env[61806]: DEBUG oslo_vmware.api [None req-ca966d32-e3ef-4c81-a661-22cd95d9e4ec tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Task: {'id': session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]52e0998a-d1e9-0c8d-612a-988c9a4655bb, 'name': SearchDatastore_Task, 'duration_secs': 0.010002} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 871.913136] env[61806]: DEBUG oslo_concurrency.lockutils [None req-ca966d32-e3ef-4c81-a661-22cd95d9e4ec tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Releasing lock "[datastore2] devstack-image-cache_base/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 871.913414] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-ca966d32-e3ef-4c81-a661-22cd95d9e4ec tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] [instance: efa19aac-3ad7-42d5-a6d8-859050de6e63] Processing image ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa {{(pid=61806) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 871.913744] env[61806]: DEBUG oslo_concurrency.lockutils [None req-ca966d32-e3ef-4c81-a661-22cd95d9e4ec tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa.vmdk" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 871.913936] env[61806]: DEBUG oslo_concurrency.lockutils [None req-ca966d32-e3ef-4c81-a661-22cd95d9e4ec tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Acquired lock "[datastore2] devstack-image-cache_base/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa.vmdk" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 871.914166] env[61806]: DEBUG nova.virt.vmwareapi.ds_util [None req-ca966d32-e3ef-4c81-a661-22cd95d9e4ec tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61806) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 871.914442] env[61806]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-5c9e2efa-65fc-4959-b24a-da8e4c544c57 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 871.921485] env[61806]: DEBUG nova.network.neutron [req-5ffb675e-f2c7-4c1c-80ad-b1930d7f32a1 req-bdcd0a67-aad8-44c2-bdfc-1c1a6d2a1bca service nova] [instance: efa19aac-3ad7-42d5-a6d8-859050de6e63] Updated VIF entry in instance network info cache for port fea0c8d9-43ab-4462-9427-73d9ae4a47b2. {{(pid=61806) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 871.921830] env[61806]: DEBUG nova.network.neutron [req-5ffb675e-f2c7-4c1c-80ad-b1930d7f32a1 req-bdcd0a67-aad8-44c2-bdfc-1c1a6d2a1bca service nova] [instance: efa19aac-3ad7-42d5-a6d8-859050de6e63] Updating instance_info_cache with network_info: [{"id": "fea0c8d9-43ab-4462-9427-73d9ae4a47b2", "address": "fa:16:3e:18:2f:85", "network": {"id": "46515a4e-610f-42fb-a424-205a03df0640", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-1133623547-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4b9d94c3d684474b99e305fb6cd9b95f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "82ca17df-257e-40e6-9ec9-310ed6f05ccb", "external-id": "nsx-vlan-transportzone-778", "segmentation_id": 778, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapfea0c8d9-43", "ovs_interfaceid": "fea0c8d9-43ab-4462-9427-73d9ae4a47b2", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61806) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 871.923941] env[61806]: DEBUG nova.virt.vmwareapi.ds_util [None req-ca966d32-e3ef-4c81-a661-22cd95d9e4ec tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61806) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 871.924157] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-ca966d32-e3ef-4c81-a661-22cd95d9e4ec tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61806) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 871.925308] env[61806]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ecc36d3b-5dc6-4eba-be14-6b066ef0b6f9 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 871.930871] env[61806]: DEBUG oslo_vmware.api [None req-ca966d32-e3ef-4c81-a661-22cd95d9e4ec tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Waiting for the task: (returnval){ [ 871.930871] env[61806]: value = "session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]52f2c32b-aaf8-42f2-f076-1c9b7764302b" [ 871.930871] env[61806]: _type = "Task" [ 871.930871] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 871.940594] env[61806]: DEBUG oslo_vmware.api [None req-ca966d32-e3ef-4c81-a661-22cd95d9e4ec tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Task: {'id': session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]52f2c32b-aaf8-42f2-f076-1c9b7764302b, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 872.158654] env[61806]: DEBUG nova.scheduler.client.report [None req-b996cb64-3154-4eef-9fcd-aa24605b9020 tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] Inventory has not changed for provider a2858be1-fd22-4e08-979e-87ad25293407 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 138, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61806) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 872.259684] env[61806]: DEBUG nova.virt.vmwareapi.volumeops [None req-98f42353-addf-4bdc-9891-17066a23ed33 tempest-AttachVolumeNegativeTest-1683033704 tempest-AttachVolumeNegativeTest-1683033704-project-member] [instance: 43dcce0d-4fdb-4ee1-8309-92e92d2331a9] Volume attach. Driver type: vmdk {{(pid=61806) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 872.259894] env[61806]: DEBUG nova.virt.vmwareapi.volumeops [None req-98f42353-addf-4bdc-9891-17066a23ed33 tempest-AttachVolumeNegativeTest-1683033704 tempest-AttachVolumeNegativeTest-1683033704-project-member] [instance: 43dcce0d-4fdb-4ee1-8309-92e92d2331a9] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-277686', 'volume_id': '21f9e064-7d34-44c9-9759-32f3d73dec59', 'name': 'volume-21f9e064-7d34-44c9-9759-32f3d73dec59', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '43dcce0d-4fdb-4ee1-8309-92e92d2331a9', 'attached_at': '', 'detached_at': '', 'volume_id': '21f9e064-7d34-44c9-9759-32f3d73dec59', 'serial': '21f9e064-7d34-44c9-9759-32f3d73dec59'} {{(pid=61806) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 872.260737] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-977d599c-57f3-442c-a001-14c8579bcf39 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 872.281036] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d5465545-ac02-4432-b1b1-e066cf336722 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 872.286475] env[61806]: DEBUG oslo_concurrency.lockutils [None req-7e96f1c1-fc04-4010-b728-951cd26dd38b tempest-ImagesTestJSON-618852556 tempest-ImagesTestJSON-618852556-project-member] Acquiring lock "1a73d994-5f21-4f80-8f33-5830d860a7ee" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 872.286707] env[61806]: DEBUG oslo_concurrency.lockutils [None req-7e96f1c1-fc04-4010-b728-951cd26dd38b tempest-ImagesTestJSON-618852556 tempest-ImagesTestJSON-618852556-project-member] Lock "1a73d994-5f21-4f80-8f33-5830d860a7ee" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 872.313068] env[61806]: DEBUG nova.virt.vmwareapi.volumeops [None req-98f42353-addf-4bdc-9891-17066a23ed33 tempest-AttachVolumeNegativeTest-1683033704 tempest-AttachVolumeNegativeTest-1683033704-project-member] [instance: 43dcce0d-4fdb-4ee1-8309-92e92d2331a9] Reconfiguring VM instance instance-0000003f to attach disk [datastore2] volume-21f9e064-7d34-44c9-9759-32f3d73dec59/volume-21f9e064-7d34-44c9-9759-32f3d73dec59.vmdk or device None with type thin {{(pid=61806) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 872.313608] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-8d34d9b9-2c71-4edc-9088-7a336991edf3 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 872.333234] env[61806]: DEBUG oslo_vmware.api [None req-98f42353-addf-4bdc-9891-17066a23ed33 tempest-AttachVolumeNegativeTest-1683033704 tempest-AttachVolumeNegativeTest-1683033704-project-member] Waiting for the task: (returnval){ [ 872.333234] env[61806]: value = "task-1294765" [ 872.333234] env[61806]: _type = "Task" [ 872.333234] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 872.341622] env[61806]: DEBUG oslo_vmware.api [None req-98f42353-addf-4bdc-9891-17066a23ed33 tempest-AttachVolumeNegativeTest-1683033704 tempest-AttachVolumeNegativeTest-1683033704-project-member] Task: {'id': task-1294765, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 872.349868] env[61806]: DEBUG oslo_vmware.api [None req-708d7cc8-bb4e-45d0-adad-8f77e0f096a2 tempest-ImagesTestJSON-618852556 tempest-ImagesTestJSON-618852556-project-member] Task: {'id': task-1294764, 'name': PowerOffVM_Task, 'duration_secs': 0.185019} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 872.350151] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [None req-708d7cc8-bb4e-45d0-adad-8f77e0f096a2 tempest-ImagesTestJSON-618852556 tempest-ImagesTestJSON-618852556-project-member] [instance: e0ef0a35-82a5-495b-9d5c-5805e8306390] Powered off the VM {{(pid=61806) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 872.350327] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-708d7cc8-bb4e-45d0-adad-8f77e0f096a2 tempest-ImagesTestJSON-618852556 tempest-ImagesTestJSON-618852556-project-member] [instance: e0ef0a35-82a5-495b-9d5c-5805e8306390] Unregistering the VM {{(pid=61806) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 872.350573] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-bea76dc0-8f2f-4f6d-af4b-57b9b613d443 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 872.426696] env[61806]: DEBUG oslo_concurrency.lockutils [req-5ffb675e-f2c7-4c1c-80ad-b1930d7f32a1 req-bdcd0a67-aad8-44c2-bdfc-1c1a6d2a1bca service nova] Releasing lock "refresh_cache-efa19aac-3ad7-42d5-a6d8-859050de6e63" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 872.427032] env[61806]: DEBUG nova.compute.manager [req-5ffb675e-f2c7-4c1c-80ad-b1930d7f32a1 req-bdcd0a67-aad8-44c2-bdfc-1c1a6d2a1bca service nova] [instance: 2a3f81e3-b84f-4370-a598-2f6a607c3d47] Received event network-vif-deleted-a839bcaa-9990-4bad-be1f-1f06eb1c978e {{(pid=61806) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 872.442958] env[61806]: DEBUG oslo_vmware.api [None req-ca966d32-e3ef-4c81-a661-22cd95d9e4ec tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Task: {'id': session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]52f2c32b-aaf8-42f2-f076-1c9b7764302b, 'name': SearchDatastore_Task, 'duration_secs': 0.010042} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 872.443804] env[61806]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-af94e1ca-92bb-44bd-a7e9-281bb05ce419 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 872.449555] env[61806]: DEBUG oslo_vmware.api [None req-ca966d32-e3ef-4c81-a661-22cd95d9e4ec tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Waiting for the task: (returnval){ [ 872.449555] env[61806]: value = "session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]52766bdb-6cfc-c148-44c2-d5f0579d53f4" [ 872.449555] env[61806]: _type = "Task" [ 872.449555] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 872.458766] env[61806]: DEBUG oslo_vmware.api [None req-ca966d32-e3ef-4c81-a661-22cd95d9e4ec tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Task: {'id': session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]52766bdb-6cfc-c148-44c2-d5f0579d53f4, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 872.475153] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-708d7cc8-bb4e-45d0-adad-8f77e0f096a2 tempest-ImagesTestJSON-618852556 tempest-ImagesTestJSON-618852556-project-member] [instance: e0ef0a35-82a5-495b-9d5c-5805e8306390] Unregistered the VM {{(pid=61806) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 872.475450] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-708d7cc8-bb4e-45d0-adad-8f77e0f096a2 tempest-ImagesTestJSON-618852556 tempest-ImagesTestJSON-618852556-project-member] [instance: e0ef0a35-82a5-495b-9d5c-5805e8306390] Deleting contents of the VM from datastore datastore1 {{(pid=61806) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 872.475684] env[61806]: DEBUG nova.virt.vmwareapi.ds_util [None req-708d7cc8-bb4e-45d0-adad-8f77e0f096a2 tempest-ImagesTestJSON-618852556 tempest-ImagesTestJSON-618852556-project-member] Deleting the datastore file [datastore1] e0ef0a35-82a5-495b-9d5c-5805e8306390 {{(pid=61806) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 872.475987] env[61806]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-6d64bc35-9cab-4fbb-9610-3e9eae9395ef {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 872.483255] env[61806]: DEBUG oslo_vmware.api [None req-708d7cc8-bb4e-45d0-adad-8f77e0f096a2 tempest-ImagesTestJSON-618852556 tempest-ImagesTestJSON-618852556-project-member] Waiting for the task: (returnval){ [ 872.483255] env[61806]: value = "task-1294767" [ 872.483255] env[61806]: _type = "Task" [ 872.483255] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 872.491471] env[61806]: DEBUG oslo_vmware.api [None req-708d7cc8-bb4e-45d0-adad-8f77e0f096a2 tempest-ImagesTestJSON-618852556 tempest-ImagesTestJSON-618852556-project-member] Task: {'id': task-1294767, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 872.664223] env[61806]: DEBUG oslo_concurrency.lockutils [None req-b996cb64-3154-4eef-9fcd-aa24605b9020 tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.201s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 872.664223] env[61806]: DEBUG nova.compute.manager [None req-b996cb64-3154-4eef-9fcd-aa24605b9020 tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] [instance: d1802da4-a670-427d-84d7-0fcb4717e18f] Start building networks asynchronously for instance. {{(pid=61806) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 872.666484] env[61806]: DEBUG oslo_concurrency.lockutils [None req-799f40df-a27b-4200-803a-166b5c1e29fd tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 5.922s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 872.667850] env[61806]: INFO nova.compute.claims [None req-799f40df-a27b-4200-803a-166b5c1e29fd tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] [instance: 59c22c16-3d1c-4e64-bfac-7f8886a1927e] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 872.788704] env[61806]: DEBUG nova.compute.manager [None req-7e96f1c1-fc04-4010-b728-951cd26dd38b tempest-ImagesTestJSON-618852556 tempest-ImagesTestJSON-618852556-project-member] [instance: 1a73d994-5f21-4f80-8f33-5830d860a7ee] Starting instance... {{(pid=61806) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 872.843292] env[61806]: DEBUG oslo_vmware.api [None req-98f42353-addf-4bdc-9891-17066a23ed33 tempest-AttachVolumeNegativeTest-1683033704 tempest-AttachVolumeNegativeTest-1683033704-project-member] Task: {'id': task-1294765, 'name': ReconfigVM_Task, 'duration_secs': 0.352958} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 872.843576] env[61806]: DEBUG nova.virt.vmwareapi.volumeops [None req-98f42353-addf-4bdc-9891-17066a23ed33 tempest-AttachVolumeNegativeTest-1683033704 tempest-AttachVolumeNegativeTest-1683033704-project-member] [instance: 43dcce0d-4fdb-4ee1-8309-92e92d2331a9] Reconfigured VM instance instance-0000003f to attach disk [datastore2] volume-21f9e064-7d34-44c9-9759-32f3d73dec59/volume-21f9e064-7d34-44c9-9759-32f3d73dec59.vmdk or device None with type thin {{(pid=61806) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 872.848248] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-f49baeee-6218-481e-8086-1e46ca23cfcb {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 872.866763] env[61806]: DEBUG oslo_vmware.api [None req-98f42353-addf-4bdc-9891-17066a23ed33 tempest-AttachVolumeNegativeTest-1683033704 tempest-AttachVolumeNegativeTest-1683033704-project-member] Waiting for the task: (returnval){ [ 872.866763] env[61806]: value = "task-1294768" [ 872.866763] env[61806]: _type = "Task" [ 872.866763] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 872.879374] env[61806]: DEBUG oslo_vmware.api [None req-98f42353-addf-4bdc-9891-17066a23ed33 tempest-AttachVolumeNegativeTest-1683033704 tempest-AttachVolumeNegativeTest-1683033704-project-member] Task: {'id': task-1294768, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 872.961624] env[61806]: DEBUG oslo_vmware.api [None req-ca966d32-e3ef-4c81-a661-22cd95d9e4ec tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Task: {'id': session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]52766bdb-6cfc-c148-44c2-d5f0579d53f4, 'name': SearchDatastore_Task, 'duration_secs': 0.011057} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 872.961920] env[61806]: DEBUG oslo_concurrency.lockutils [None req-ca966d32-e3ef-4c81-a661-22cd95d9e4ec tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Releasing lock "[datastore2] devstack-image-cache_base/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa.vmdk" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 872.962247] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [None req-ca966d32-e3ef-4c81-a661-22cd95d9e4ec tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa.vmdk to [datastore2] efa19aac-3ad7-42d5-a6d8-859050de6e63/efa19aac-3ad7-42d5-a6d8-859050de6e63.vmdk {{(pid=61806) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 872.962519] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-ae7834d4-4c0a-4ddc-a107-05e8837a49f6 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 872.969653] env[61806]: DEBUG oslo_vmware.api [None req-ca966d32-e3ef-4c81-a661-22cd95d9e4ec tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Waiting for the task: (returnval){ [ 872.969653] env[61806]: value = "task-1294769" [ 872.969653] env[61806]: _type = "Task" [ 872.969653] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 872.978653] env[61806]: DEBUG oslo_vmware.api [None req-ca966d32-e3ef-4c81-a661-22cd95d9e4ec tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Task: {'id': task-1294769, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 872.994039] env[61806]: DEBUG oslo_vmware.api [None req-708d7cc8-bb4e-45d0-adad-8f77e0f096a2 tempest-ImagesTestJSON-618852556 tempest-ImagesTestJSON-618852556-project-member] Task: {'id': task-1294767, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.162651} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 872.994039] env[61806]: DEBUG nova.virt.vmwareapi.ds_util [None req-708d7cc8-bb4e-45d0-adad-8f77e0f096a2 tempest-ImagesTestJSON-618852556 tempest-ImagesTestJSON-618852556-project-member] Deleted the datastore file {{(pid=61806) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 872.994039] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-708d7cc8-bb4e-45d0-adad-8f77e0f096a2 tempest-ImagesTestJSON-618852556 tempest-ImagesTestJSON-618852556-project-member] [instance: e0ef0a35-82a5-495b-9d5c-5805e8306390] Deleted contents of the VM from datastore datastore1 {{(pid=61806) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 872.994039] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-708d7cc8-bb4e-45d0-adad-8f77e0f096a2 tempest-ImagesTestJSON-618852556 tempest-ImagesTestJSON-618852556-project-member] [instance: e0ef0a35-82a5-495b-9d5c-5805e8306390] Instance destroyed {{(pid=61806) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 872.994039] env[61806]: INFO nova.compute.manager [None req-708d7cc8-bb4e-45d0-adad-8f77e0f096a2 tempest-ImagesTestJSON-618852556 tempest-ImagesTestJSON-618852556-project-member] [instance: e0ef0a35-82a5-495b-9d5c-5805e8306390] Took 1.17 seconds to destroy the instance on the hypervisor. [ 872.994410] env[61806]: DEBUG oslo.service.loopingcall [None req-708d7cc8-bb4e-45d0-adad-8f77e0f096a2 tempest-ImagesTestJSON-618852556 tempest-ImagesTestJSON-618852556-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61806) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 872.994532] env[61806]: DEBUG nova.compute.manager [-] [instance: e0ef0a35-82a5-495b-9d5c-5805e8306390] Deallocating network for instance {{(pid=61806) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 872.994646] env[61806]: DEBUG nova.network.neutron [-] [instance: e0ef0a35-82a5-495b-9d5c-5805e8306390] deallocate_for_instance() {{(pid=61806) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 873.175036] env[61806]: DEBUG nova.compute.utils [None req-b996cb64-3154-4eef-9fcd-aa24605b9020 tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] Using /dev/sd instead of None {{(pid=61806) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 873.185191] env[61806]: DEBUG nova.compute.manager [None req-b996cb64-3154-4eef-9fcd-aa24605b9020 tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] [instance: d1802da4-a670-427d-84d7-0fcb4717e18f] Allocating IP information in the background. {{(pid=61806) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 873.185191] env[61806]: DEBUG nova.network.neutron [None req-b996cb64-3154-4eef-9fcd-aa24605b9020 tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] [instance: d1802da4-a670-427d-84d7-0fcb4717e18f] allocate_for_instance() {{(pid=61806) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 873.280573] env[61806]: DEBUG nova.policy [None req-b996cb64-3154-4eef-9fcd-aa24605b9020 tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '4c5d4c6e1ad844a585c08c4644a60bbf', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '73ebcc4d79d248efb653a084a5e44302', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61806) authorize /opt/stack/nova/nova/policy.py:201}} [ 873.325802] env[61806]: DEBUG oslo_concurrency.lockutils [None req-7e96f1c1-fc04-4010-b728-951cd26dd38b tempest-ImagesTestJSON-618852556 tempest-ImagesTestJSON-618852556-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 873.379750] env[61806]: DEBUG oslo_vmware.api [None req-98f42353-addf-4bdc-9891-17066a23ed33 tempest-AttachVolumeNegativeTest-1683033704 tempest-AttachVolumeNegativeTest-1683033704-project-member] Task: {'id': task-1294768, 'name': ReconfigVM_Task, 'duration_secs': 0.157272} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 873.380113] env[61806]: DEBUG nova.virt.vmwareapi.volumeops [None req-98f42353-addf-4bdc-9891-17066a23ed33 tempest-AttachVolumeNegativeTest-1683033704 tempest-AttachVolumeNegativeTest-1683033704-project-member] [instance: 43dcce0d-4fdb-4ee1-8309-92e92d2331a9] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-277686', 'volume_id': '21f9e064-7d34-44c9-9759-32f3d73dec59', 'name': 'volume-21f9e064-7d34-44c9-9759-32f3d73dec59', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '43dcce0d-4fdb-4ee1-8309-92e92d2331a9', 'attached_at': '', 'detached_at': '', 'volume_id': '21f9e064-7d34-44c9-9759-32f3d73dec59', 'serial': '21f9e064-7d34-44c9-9759-32f3d73dec59'} {{(pid=61806) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 873.480610] env[61806]: DEBUG oslo_vmware.api [None req-ca966d32-e3ef-4c81-a661-22cd95d9e4ec tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Task: {'id': task-1294769, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.472791} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 873.483023] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [None req-ca966d32-e3ef-4c81-a661-22cd95d9e4ec tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa.vmdk to [datastore2] efa19aac-3ad7-42d5-a6d8-859050de6e63/efa19aac-3ad7-42d5-a6d8-859050de6e63.vmdk {{(pid=61806) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 873.483023] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-ca966d32-e3ef-4c81-a661-22cd95d9e4ec tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] [instance: efa19aac-3ad7-42d5-a6d8-859050de6e63] Extending root virtual disk to 1048576 {{(pid=61806) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 873.483023] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-03b183d2-7494-4ae9-96a2-feb375963a22 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 873.489083] env[61806]: DEBUG nova.compute.manager [req-7266f037-71b7-47be-95b0-a10d08e0ae11 req-0e2233d4-c260-4e5a-8bd2-58df784d942d service nova] [instance: e0ef0a35-82a5-495b-9d5c-5805e8306390] Received event network-vif-deleted-c42059da-e422-46df-be37-f9a0816b97c8 {{(pid=61806) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 873.489306] env[61806]: INFO nova.compute.manager [req-7266f037-71b7-47be-95b0-a10d08e0ae11 req-0e2233d4-c260-4e5a-8bd2-58df784d942d service nova] [instance: e0ef0a35-82a5-495b-9d5c-5805e8306390] Neutron deleted interface c42059da-e422-46df-be37-f9a0816b97c8; detaching it from the instance and deleting it from the info cache [ 873.489488] env[61806]: DEBUG nova.network.neutron [req-7266f037-71b7-47be-95b0-a10d08e0ae11 req-0e2233d4-c260-4e5a-8bd2-58df784d942d service nova] [instance: e0ef0a35-82a5-495b-9d5c-5805e8306390] Updating instance_info_cache with network_info: [] {{(pid=61806) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 873.494265] env[61806]: DEBUG oslo_vmware.api [None req-ca966d32-e3ef-4c81-a661-22cd95d9e4ec tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Waiting for the task: (returnval){ [ 873.494265] env[61806]: value = "task-1294770" [ 873.494265] env[61806]: _type = "Task" [ 873.494265] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 873.506757] env[61806]: DEBUG oslo_vmware.api [None req-ca966d32-e3ef-4c81-a661-22cd95d9e4ec tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Task: {'id': task-1294770, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 873.683242] env[61806]: DEBUG nova.compute.manager [None req-b996cb64-3154-4eef-9fcd-aa24605b9020 tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] [instance: d1802da4-a670-427d-84d7-0fcb4717e18f] Start building block device mappings for instance. {{(pid=61806) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 873.722465] env[61806]: DEBUG nova.network.neutron [None req-b996cb64-3154-4eef-9fcd-aa24605b9020 tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] [instance: d1802da4-a670-427d-84d7-0fcb4717e18f] Successfully created port: 16d93306-6614-4a1d-aa33-e18d3f79ed1e {{(pid=61806) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 873.861704] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b835b84e-04cb-4b73-92bd-a54afa226d31 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 873.870296] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fa94bcf5-02c1-4a43-b554-2858b9db6cd8 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 873.905744] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-afc5077f-9f68-4593-a0aa-047725dfb85b {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 873.914030] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b738f393-f254-4720-bc52-662b084cd912 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 873.931739] env[61806]: DEBUG nova.compute.provider_tree [None req-799f40df-a27b-4200-803a-166b5c1e29fd tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Inventory has not changed in ProviderTree for provider: a2858be1-fd22-4e08-979e-87ad25293407 {{(pid=61806) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 873.953161] env[61806]: DEBUG nova.network.neutron [-] [instance: e0ef0a35-82a5-495b-9d5c-5805e8306390] Updating instance_info_cache with network_info: [] {{(pid=61806) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 873.996328] env[61806]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-aae725e5-ad13-4cd4-ab6b-ce556401c473 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 874.008673] env[61806]: DEBUG oslo_vmware.api [None req-ca966d32-e3ef-4c81-a661-22cd95d9e4ec tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Task: {'id': task-1294770, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.072779} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 874.010048] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-ca966d32-e3ef-4c81-a661-22cd95d9e4ec tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] [instance: efa19aac-3ad7-42d5-a6d8-859050de6e63] Extended root virtual disk {{(pid=61806) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 874.010885] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-27784784-7a22-4ce1-be87-406442be5e47 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 874.016095] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-de1d0f82-4f8c-4e77-917f-956aec57b196 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 874.046857] env[61806]: DEBUG nova.virt.vmwareapi.volumeops [None req-ca966d32-e3ef-4c81-a661-22cd95d9e4ec tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] [instance: efa19aac-3ad7-42d5-a6d8-859050de6e63] Reconfiguring VM instance instance-00000049 to attach disk [datastore2] efa19aac-3ad7-42d5-a6d8-859050de6e63/efa19aac-3ad7-42d5-a6d8-859050de6e63.vmdk or device None with type sparse {{(pid=61806) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 874.052920] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-53d9d564-5810-428d-8a1a-15038ba2ea0e {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 874.067133] env[61806]: DEBUG nova.compute.manager [req-7266f037-71b7-47be-95b0-a10d08e0ae11 req-0e2233d4-c260-4e5a-8bd2-58df784d942d service nova] [instance: e0ef0a35-82a5-495b-9d5c-5805e8306390] Detach interface failed, port_id=c42059da-e422-46df-be37-f9a0816b97c8, reason: Instance e0ef0a35-82a5-495b-9d5c-5805e8306390 could not be found. {{(pid=61806) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10963}} [ 874.073878] env[61806]: DEBUG oslo_vmware.api [None req-ca966d32-e3ef-4c81-a661-22cd95d9e4ec tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Waiting for the task: (returnval){ [ 874.073878] env[61806]: value = "task-1294771" [ 874.073878] env[61806]: _type = "Task" [ 874.073878] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 874.081824] env[61806]: DEBUG oslo_vmware.api [None req-ca966d32-e3ef-4c81-a661-22cd95d9e4ec tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Task: {'id': task-1294771, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 874.303296] env[61806]: DEBUG oslo_concurrency.lockutils [None req-db5f1c5c-a7b9-4862-9010-d92ce4d061a9 tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Acquiring lock "67d1bdd8-cf8f-4fb5-bd3c-42008e4d60e3" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 874.303650] env[61806]: DEBUG oslo_concurrency.lockutils [None req-db5f1c5c-a7b9-4862-9010-d92ce4d061a9 tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Lock "67d1bdd8-cf8f-4fb5-bd3c-42008e4d60e3" acquired by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: waited 0.001s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 874.303843] env[61806]: DEBUG nova.compute.manager [None req-db5f1c5c-a7b9-4862-9010-d92ce4d061a9 tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] [instance: 67d1bdd8-cf8f-4fb5-bd3c-42008e4d60e3] Checking state {{(pid=61806) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 874.304806] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1cf4bd78-7fa4-40fb-9357-f02eeaddade4 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 874.312550] env[61806]: DEBUG nova.compute.manager [None req-db5f1c5c-a7b9-4862-9010-d92ce4d061a9 tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] [instance: 67d1bdd8-cf8f-4fb5-bd3c-42008e4d60e3] Stopping instance; current vm_state: active, current task_state: powering-off, current DB power_state: 1, current VM power_state: 1 {{(pid=61806) do_stop_instance /opt/stack/nova/nova/compute/manager.py:3368}} [ 874.313223] env[61806]: DEBUG nova.objects.instance [None req-db5f1c5c-a7b9-4862-9010-d92ce4d061a9 tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Lazy-loading 'flavor' on Instance uuid 67d1bdd8-cf8f-4fb5-bd3c-42008e4d60e3 {{(pid=61806) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 874.437206] env[61806]: DEBUG nova.scheduler.client.report [None req-799f40df-a27b-4200-803a-166b5c1e29fd tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Inventory has not changed for provider a2858be1-fd22-4e08-979e-87ad25293407 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 138, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61806) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 874.455697] env[61806]: INFO nova.compute.manager [-] [instance: e0ef0a35-82a5-495b-9d5c-5805e8306390] Took 1.46 seconds to deallocate network for instance. [ 874.463438] env[61806]: DEBUG nova.objects.instance [None req-98f42353-addf-4bdc-9891-17066a23ed33 tempest-AttachVolumeNegativeTest-1683033704 tempest-AttachVolumeNegativeTest-1683033704-project-member] Lazy-loading 'flavor' on Instance uuid 43dcce0d-4fdb-4ee1-8309-92e92d2331a9 {{(pid=61806) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 874.584337] env[61806]: DEBUG oslo_vmware.api [None req-ca966d32-e3ef-4c81-a661-22cd95d9e4ec tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Task: {'id': task-1294771, 'name': ReconfigVM_Task, 'duration_secs': 0.341871} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 874.584624] env[61806]: DEBUG nova.virt.vmwareapi.volumeops [None req-ca966d32-e3ef-4c81-a661-22cd95d9e4ec tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] [instance: efa19aac-3ad7-42d5-a6d8-859050de6e63] Reconfigured VM instance instance-00000049 to attach disk [datastore2] efa19aac-3ad7-42d5-a6d8-859050de6e63/efa19aac-3ad7-42d5-a6d8-859050de6e63.vmdk or device None with type sparse {{(pid=61806) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 874.585401] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-573efa5f-dbd2-4a05-80d1-0b7af59eee0c {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 874.594984] env[61806]: DEBUG oslo_vmware.api [None req-ca966d32-e3ef-4c81-a661-22cd95d9e4ec tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Waiting for the task: (returnval){ [ 874.594984] env[61806]: value = "task-1294772" [ 874.594984] env[61806]: _type = "Task" [ 874.594984] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 874.603888] env[61806]: DEBUG oslo_vmware.api [None req-ca966d32-e3ef-4c81-a661-22cd95d9e4ec tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Task: {'id': task-1294772, 'name': Rename_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 874.692727] env[61806]: DEBUG nova.compute.manager [None req-b996cb64-3154-4eef-9fcd-aa24605b9020 tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] [instance: d1802da4-a670-427d-84d7-0fcb4717e18f] Start spawning the instance on the hypervisor. {{(pid=61806) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 874.718483] env[61806]: DEBUG nova.virt.hardware [None req-b996cb64-3154-4eef-9fcd-aa24605b9020 tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-15T18:05:10Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-15T18:04:52Z,direct_url=,disk_format='vmdk',id=ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='84ba83fce2c349a6988173c5d6fc3b07',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-15T18:04:53Z,virtual_size=,visibility=), allow threads: False {{(pid=61806) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 874.718882] env[61806]: DEBUG nova.virt.hardware [None req-b996cb64-3154-4eef-9fcd-aa24605b9020 tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] Flavor limits 0:0:0 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 874.719175] env[61806]: DEBUG nova.virt.hardware [None req-b996cb64-3154-4eef-9fcd-aa24605b9020 tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] Image limits 0:0:0 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 874.719444] env[61806]: DEBUG nova.virt.hardware [None req-b996cb64-3154-4eef-9fcd-aa24605b9020 tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] Flavor pref 0:0:0 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 874.719626] env[61806]: DEBUG nova.virt.hardware [None req-b996cb64-3154-4eef-9fcd-aa24605b9020 tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] Image pref 0:0:0 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 874.719794] env[61806]: DEBUG nova.virt.hardware [None req-b996cb64-3154-4eef-9fcd-aa24605b9020 tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 874.720023] env[61806]: DEBUG nova.virt.hardware [None req-b996cb64-3154-4eef-9fcd-aa24605b9020 tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61806) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 874.720216] env[61806]: DEBUG nova.virt.hardware [None req-b996cb64-3154-4eef-9fcd-aa24605b9020 tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61806) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 874.720391] env[61806]: DEBUG nova.virt.hardware [None req-b996cb64-3154-4eef-9fcd-aa24605b9020 tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] Got 1 possible topologies {{(pid=61806) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 874.720561] env[61806]: DEBUG nova.virt.hardware [None req-b996cb64-3154-4eef-9fcd-aa24605b9020 tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61806) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 874.720738] env[61806]: DEBUG nova.virt.hardware [None req-b996cb64-3154-4eef-9fcd-aa24605b9020 tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61806) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 874.721852] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6fe4d05e-e42c-4e72-91f3-24f68fe35ba7 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 874.730613] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bf037d74-5e88-4113-8714-9e1297f25910 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 874.819332] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [None req-db5f1c5c-a7b9-4862-9010-d92ce4d061a9 tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] [instance: 67d1bdd8-cf8f-4fb5-bd3c-42008e4d60e3] Powering off the VM {{(pid=61806) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 874.819636] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-57840923-b31d-432d-bb1e-5d815071c26e {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 874.827506] env[61806]: DEBUG oslo_vmware.api [None req-db5f1c5c-a7b9-4862-9010-d92ce4d061a9 tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Waiting for the task: (returnval){ [ 874.827506] env[61806]: value = "task-1294773" [ 874.827506] env[61806]: _type = "Task" [ 874.827506] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 874.836674] env[61806]: DEBUG oslo_vmware.api [None req-db5f1c5c-a7b9-4862-9010-d92ce4d061a9 tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Task: {'id': task-1294773, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 874.942612] env[61806]: DEBUG oslo_concurrency.lockutils [None req-799f40df-a27b-4200-803a-166b5c1e29fd tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.276s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 874.943564] env[61806]: DEBUG nova.compute.manager [None req-799f40df-a27b-4200-803a-166b5c1e29fd tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] [instance: 59c22c16-3d1c-4e64-bfac-7f8886a1927e] Start building networks asynchronously for instance. {{(pid=61806) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 874.946659] env[61806]: DEBUG oslo_concurrency.lockutils [None req-b85317d7-e79b-4ab3-83e4-2de701507449 tempest-ServerTagsTestJSON-1292926242 tempest-ServerTagsTestJSON-1292926242-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 5.573s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 874.948184] env[61806]: INFO nova.compute.claims [None req-b85317d7-e79b-4ab3-83e4-2de701507449 tempest-ServerTagsTestJSON-1292926242 tempest-ServerTagsTestJSON-1292926242-project-member] [instance: 9c20d337-b8f5-4965-b707-0f87c4ab6fc4] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 874.962820] env[61806]: DEBUG oslo_concurrency.lockutils [None req-708d7cc8-bb4e-45d0-adad-8f77e0f096a2 tempest-ImagesTestJSON-618852556 tempest-ImagesTestJSON-618852556-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 874.968385] env[61806]: DEBUG oslo_concurrency.lockutils [None req-98f42353-addf-4bdc-9891-17066a23ed33 tempest-AttachVolumeNegativeTest-1683033704 tempest-AttachVolumeNegativeTest-1683033704-project-member] Lock "43dcce0d-4fdb-4ee1-8309-92e92d2331a9" "released" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: held 7.417s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 875.105319] env[61806]: DEBUG oslo_vmware.api [None req-ca966d32-e3ef-4c81-a661-22cd95d9e4ec tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Task: {'id': task-1294772, 'name': Rename_Task, 'duration_secs': 0.16293} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 875.105689] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [None req-ca966d32-e3ef-4c81-a661-22cd95d9e4ec tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] [instance: efa19aac-3ad7-42d5-a6d8-859050de6e63] Powering on the VM {{(pid=61806) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 875.105949] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-31ca1e36-9203-4481-a220-6aaeb6c1c89c {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 875.115131] env[61806]: DEBUG oslo_vmware.api [None req-ca966d32-e3ef-4c81-a661-22cd95d9e4ec tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Waiting for the task: (returnval){ [ 875.115131] env[61806]: value = "task-1294774" [ 875.115131] env[61806]: _type = "Task" [ 875.115131] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 875.125150] env[61806]: DEBUG oslo_vmware.api [None req-ca966d32-e3ef-4c81-a661-22cd95d9e4ec tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Task: {'id': task-1294774, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 875.297049] env[61806]: DEBUG nova.compute.manager [req-dda7ddec-39a6-438c-b997-60f96d06da8e req-fa3bd646-41a9-4de3-b8d3-e94efca82729 service nova] [instance: d1802da4-a670-427d-84d7-0fcb4717e18f] Received event network-vif-plugged-16d93306-6614-4a1d-aa33-e18d3f79ed1e {{(pid=61806) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 875.297374] env[61806]: DEBUG oslo_concurrency.lockutils [req-dda7ddec-39a6-438c-b997-60f96d06da8e req-fa3bd646-41a9-4de3-b8d3-e94efca82729 service nova] Acquiring lock "d1802da4-a670-427d-84d7-0fcb4717e18f-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 875.297676] env[61806]: DEBUG oslo_concurrency.lockutils [req-dda7ddec-39a6-438c-b997-60f96d06da8e req-fa3bd646-41a9-4de3-b8d3-e94efca82729 service nova] Lock "d1802da4-a670-427d-84d7-0fcb4717e18f-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 875.297899] env[61806]: DEBUG oslo_concurrency.lockutils [req-dda7ddec-39a6-438c-b997-60f96d06da8e req-fa3bd646-41a9-4de3-b8d3-e94efca82729 service nova] Lock "d1802da4-a670-427d-84d7-0fcb4717e18f-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 875.298080] env[61806]: DEBUG nova.compute.manager [req-dda7ddec-39a6-438c-b997-60f96d06da8e req-fa3bd646-41a9-4de3-b8d3-e94efca82729 service nova] [instance: d1802da4-a670-427d-84d7-0fcb4717e18f] No waiting events found dispatching network-vif-plugged-16d93306-6614-4a1d-aa33-e18d3f79ed1e {{(pid=61806) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 875.298315] env[61806]: WARNING nova.compute.manager [req-dda7ddec-39a6-438c-b997-60f96d06da8e req-fa3bd646-41a9-4de3-b8d3-e94efca82729 service nova] [instance: d1802da4-a670-427d-84d7-0fcb4717e18f] Received unexpected event network-vif-plugged-16d93306-6614-4a1d-aa33-e18d3f79ed1e for instance with vm_state building and task_state spawning. [ 875.337764] env[61806]: DEBUG oslo_vmware.api [None req-db5f1c5c-a7b9-4862-9010-d92ce4d061a9 tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Task: {'id': task-1294773, 'name': PowerOffVM_Task, 'duration_secs': 0.272044} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 875.338081] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [None req-db5f1c5c-a7b9-4862-9010-d92ce4d061a9 tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] [instance: 67d1bdd8-cf8f-4fb5-bd3c-42008e4d60e3] Powered off the VM {{(pid=61806) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 875.338247] env[61806]: DEBUG nova.compute.manager [None req-db5f1c5c-a7b9-4862-9010-d92ce4d061a9 tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] [instance: 67d1bdd8-cf8f-4fb5-bd3c-42008e4d60e3] Checking state {{(pid=61806) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 875.339078] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9d6295b2-70b9-4a1a-b23d-d7138e1099b3 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 875.452815] env[61806]: DEBUG nova.compute.utils [None req-799f40df-a27b-4200-803a-166b5c1e29fd tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Using /dev/sd instead of None {{(pid=61806) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 875.456222] env[61806]: DEBUG nova.compute.manager [None req-799f40df-a27b-4200-803a-166b5c1e29fd tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] [instance: 59c22c16-3d1c-4e64-bfac-7f8886a1927e] Allocating IP information in the background. {{(pid=61806) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 875.456429] env[61806]: DEBUG nova.network.neutron [None req-799f40df-a27b-4200-803a-166b5c1e29fd tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] [instance: 59c22c16-3d1c-4e64-bfac-7f8886a1927e] allocate_for_instance() {{(pid=61806) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 875.505674] env[61806]: DEBUG nova.policy [None req-799f40df-a27b-4200-803a-166b5c1e29fd tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '2b86d596e89645c2ab31332afdcbc3c2', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '7ab95fb72b5d46c3b6c7bebbccf897cd', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61806) authorize /opt/stack/nova/nova/policy.py:201}} [ 875.634910] env[61806]: DEBUG oslo_vmware.api [None req-ca966d32-e3ef-4c81-a661-22cd95d9e4ec tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Task: {'id': task-1294774, 'name': PowerOnVM_Task, 'duration_secs': 0.478673} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 875.636978] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [None req-ca966d32-e3ef-4c81-a661-22cd95d9e4ec tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] [instance: efa19aac-3ad7-42d5-a6d8-859050de6e63] Powered on the VM {{(pid=61806) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 875.636978] env[61806]: INFO nova.compute.manager [None req-ca966d32-e3ef-4c81-a661-22cd95d9e4ec tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] [instance: efa19aac-3ad7-42d5-a6d8-859050de6e63] Took 7.95 seconds to spawn the instance on the hypervisor. [ 875.636978] env[61806]: DEBUG nova.compute.manager [None req-ca966d32-e3ef-4c81-a661-22cd95d9e4ec tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] [instance: efa19aac-3ad7-42d5-a6d8-859050de6e63] Checking state {{(pid=61806) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 875.636978] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e784228b-3ace-4eec-8f18-863dbf21c604 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 875.831314] env[61806]: DEBUG nova.network.neutron [None req-b996cb64-3154-4eef-9fcd-aa24605b9020 tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] [instance: d1802da4-a670-427d-84d7-0fcb4717e18f] Successfully updated port: 16d93306-6614-4a1d-aa33-e18d3f79ed1e {{(pid=61806) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 875.836473] env[61806]: DEBUG nova.network.neutron [None req-799f40df-a27b-4200-803a-166b5c1e29fd tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] [instance: 59c22c16-3d1c-4e64-bfac-7f8886a1927e] Successfully created port: a3901838-4afa-473f-a673-faac7dee3e75 {{(pid=61806) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 875.854703] env[61806]: DEBUG oslo_concurrency.lockutils [None req-db5f1c5c-a7b9-4862-9010-d92ce4d061a9 tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Lock "67d1bdd8-cf8f-4fb5-bd3c-42008e4d60e3" "released" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: held 1.551s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 875.905165] env[61806]: DEBUG nova.compute.manager [req-9095cdfd-8912-4b94-b271-c51bdb6683ac req-384aedb3-697f-4d0d-b1ee-ab7a642db6b7 service nova] [instance: d1802da4-a670-427d-84d7-0fcb4717e18f] Received event network-changed-16d93306-6614-4a1d-aa33-e18d3f79ed1e {{(pid=61806) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 875.905372] env[61806]: DEBUG nova.compute.manager [req-9095cdfd-8912-4b94-b271-c51bdb6683ac req-384aedb3-697f-4d0d-b1ee-ab7a642db6b7 service nova] [instance: d1802da4-a670-427d-84d7-0fcb4717e18f] Refreshing instance network info cache due to event network-changed-16d93306-6614-4a1d-aa33-e18d3f79ed1e. {{(pid=61806) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 875.905582] env[61806]: DEBUG oslo_concurrency.lockutils [req-9095cdfd-8912-4b94-b271-c51bdb6683ac req-384aedb3-697f-4d0d-b1ee-ab7a642db6b7 service nova] Acquiring lock "refresh_cache-d1802da4-a670-427d-84d7-0fcb4717e18f" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 875.905721] env[61806]: DEBUG oslo_concurrency.lockutils [req-9095cdfd-8912-4b94-b271-c51bdb6683ac req-384aedb3-697f-4d0d-b1ee-ab7a642db6b7 service nova] Acquired lock "refresh_cache-d1802da4-a670-427d-84d7-0fcb4717e18f" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 875.905879] env[61806]: DEBUG nova.network.neutron [req-9095cdfd-8912-4b94-b271-c51bdb6683ac req-384aedb3-697f-4d0d-b1ee-ab7a642db6b7 service nova] [instance: d1802da4-a670-427d-84d7-0fcb4717e18f] Refreshing network info cache for port 16d93306-6614-4a1d-aa33-e18d3f79ed1e {{(pid=61806) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 875.956968] env[61806]: DEBUG nova.compute.manager [None req-799f40df-a27b-4200-803a-166b5c1e29fd tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] [instance: 59c22c16-3d1c-4e64-bfac-7f8886a1927e] Start building block device mappings for instance. {{(pid=61806) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 876.113685] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ce979738-fcb2-4b95-98f1-cec0cf3b4b2e {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 876.122630] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-97380e77-e496-4b42-a35a-f52878b34210 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 876.160729] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-00255db3-3876-48d5-88e3-2fc4d19273cf {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 876.167600] env[61806]: DEBUG oslo_concurrency.lockutils [None req-0bb2f11a-fff8-4c85-aa76-4f260c30bde9 tempest-AttachVolumeNegativeTest-1683033704 tempest-AttachVolumeNegativeTest-1683033704-project-member] Acquiring lock "daff849e-eb4f-411e-92f8-9362e3bd91e3" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 876.167830] env[61806]: DEBUG oslo_concurrency.lockutils [None req-0bb2f11a-fff8-4c85-aa76-4f260c30bde9 tempest-AttachVolumeNegativeTest-1683033704 tempest-AttachVolumeNegativeTest-1683033704-project-member] Lock "daff849e-eb4f-411e-92f8-9362e3bd91e3" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 876.168902] env[61806]: INFO nova.compute.manager [None req-ca966d32-e3ef-4c81-a661-22cd95d9e4ec tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] [instance: efa19aac-3ad7-42d5-a6d8-859050de6e63] Took 19.10 seconds to build instance. [ 876.175960] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-39f3ee72-836c-4da8-b563-dc8c32c99689 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 876.194223] env[61806]: DEBUG nova.compute.provider_tree [None req-b85317d7-e79b-4ab3-83e4-2de701507449 tempest-ServerTagsTestJSON-1292926242 tempest-ServerTagsTestJSON-1292926242-project-member] Inventory has not changed in ProviderTree for provider: a2858be1-fd22-4e08-979e-87ad25293407 {{(pid=61806) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 876.335238] env[61806]: DEBUG oslo_concurrency.lockutils [None req-b996cb64-3154-4eef-9fcd-aa24605b9020 tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] Acquiring lock "refresh_cache-d1802da4-a670-427d-84d7-0fcb4717e18f" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 876.449953] env[61806]: DEBUG nova.network.neutron [req-9095cdfd-8912-4b94-b271-c51bdb6683ac req-384aedb3-697f-4d0d-b1ee-ab7a642db6b7 service nova] [instance: d1802da4-a670-427d-84d7-0fcb4717e18f] Instance cache missing network info. {{(pid=61806) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 876.570033] env[61806]: DEBUG nova.network.neutron [req-9095cdfd-8912-4b94-b271-c51bdb6683ac req-384aedb3-697f-4d0d-b1ee-ab7a642db6b7 service nova] [instance: d1802da4-a670-427d-84d7-0fcb4717e18f] Updating instance_info_cache with network_info: [] {{(pid=61806) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 876.671541] env[61806]: DEBUG nova.compute.manager [None req-0bb2f11a-fff8-4c85-aa76-4f260c30bde9 tempest-AttachVolumeNegativeTest-1683033704 tempest-AttachVolumeNegativeTest-1683033704-project-member] [instance: daff849e-eb4f-411e-92f8-9362e3bd91e3] Starting instance... {{(pid=61806) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 876.676048] env[61806]: DEBUG oslo_concurrency.lockutils [None req-ca966d32-e3ef-4c81-a661-22cd95d9e4ec tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Lock "efa19aac-3ad7-42d5-a6d8-859050de6e63" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 20.621s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 876.696622] env[61806]: DEBUG nova.scheduler.client.report [None req-b85317d7-e79b-4ab3-83e4-2de701507449 tempest-ServerTagsTestJSON-1292926242 tempest-ServerTagsTestJSON-1292926242-project-member] Inventory has not changed for provider a2858be1-fd22-4e08-979e-87ad25293407 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 138, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61806) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 876.928531] env[61806]: DEBUG nova.objects.instance [None req-1504c565-36b9-453a-af40-d9aaf0dc82e4 tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Lazy-loading 'flavor' on Instance uuid 67d1bdd8-cf8f-4fb5-bd3c-42008e4d60e3 {{(pid=61806) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 876.969048] env[61806]: DEBUG nova.compute.manager [None req-799f40df-a27b-4200-803a-166b5c1e29fd tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] [instance: 59c22c16-3d1c-4e64-bfac-7f8886a1927e] Start spawning the instance on the hypervisor. {{(pid=61806) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 876.989417] env[61806]: DEBUG nova.virt.hardware [None req-799f40df-a27b-4200-803a-166b5c1e29fd tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-15T18:05:10Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-15T18:04:52Z,direct_url=,disk_format='vmdk',id=ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='84ba83fce2c349a6988173c5d6fc3b07',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-15T18:04:53Z,virtual_size=,visibility=), allow threads: False {{(pid=61806) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 876.991608] env[61806]: DEBUG nova.virt.hardware [None req-799f40df-a27b-4200-803a-166b5c1e29fd tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Flavor limits 0:0:0 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 876.991608] env[61806]: DEBUG nova.virt.hardware [None req-799f40df-a27b-4200-803a-166b5c1e29fd tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Image limits 0:0:0 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 876.991608] env[61806]: DEBUG nova.virt.hardware [None req-799f40df-a27b-4200-803a-166b5c1e29fd tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Flavor pref 0:0:0 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 876.991608] env[61806]: DEBUG nova.virt.hardware [None req-799f40df-a27b-4200-803a-166b5c1e29fd tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Image pref 0:0:0 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 876.991608] env[61806]: DEBUG nova.virt.hardware [None req-799f40df-a27b-4200-803a-166b5c1e29fd tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 876.991608] env[61806]: DEBUG nova.virt.hardware [None req-799f40df-a27b-4200-803a-166b5c1e29fd tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61806) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 876.991608] env[61806]: DEBUG nova.virt.hardware [None req-799f40df-a27b-4200-803a-166b5c1e29fd tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61806) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 876.991608] env[61806]: DEBUG nova.virt.hardware [None req-799f40df-a27b-4200-803a-166b5c1e29fd tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Got 1 possible topologies {{(pid=61806) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 876.992058] env[61806]: DEBUG nova.virt.hardware [None req-799f40df-a27b-4200-803a-166b5c1e29fd tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61806) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 876.992058] env[61806]: DEBUG nova.virt.hardware [None req-799f40df-a27b-4200-803a-166b5c1e29fd tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61806) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 876.992709] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0e0ec03c-99f4-415b-ad13-c450ea0fa8c1 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 877.006366] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fcb0a57d-b0c0-4c75-96a5-d723687b60c8 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 877.072825] env[61806]: DEBUG oslo_concurrency.lockutils [req-9095cdfd-8912-4b94-b271-c51bdb6683ac req-384aedb3-697f-4d0d-b1ee-ab7a642db6b7 service nova] Releasing lock "refresh_cache-d1802da4-a670-427d-84d7-0fcb4717e18f" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 877.073710] env[61806]: DEBUG oslo_concurrency.lockutils [None req-b996cb64-3154-4eef-9fcd-aa24605b9020 tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] Acquired lock "refresh_cache-d1802da4-a670-427d-84d7-0fcb4717e18f" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 877.073807] env[61806]: DEBUG nova.network.neutron [None req-b996cb64-3154-4eef-9fcd-aa24605b9020 tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] [instance: d1802da4-a670-427d-84d7-0fcb4717e18f] Building network info cache for instance {{(pid=61806) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 877.196615] env[61806]: DEBUG oslo_concurrency.lockutils [None req-0bb2f11a-fff8-4c85-aa76-4f260c30bde9 tempest-AttachVolumeNegativeTest-1683033704 tempest-AttachVolumeNegativeTest-1683033704-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 877.201528] env[61806]: DEBUG oslo_concurrency.lockutils [None req-b85317d7-e79b-4ab3-83e4-2de701507449 tempest-ServerTagsTestJSON-1292926242 tempest-ServerTagsTestJSON-1292926242-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.255s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 877.202068] env[61806]: DEBUG nova.compute.manager [None req-b85317d7-e79b-4ab3-83e4-2de701507449 tempest-ServerTagsTestJSON-1292926242 tempest-ServerTagsTestJSON-1292926242-project-member] [instance: 9c20d337-b8f5-4965-b707-0f87c4ab6fc4] Start building networks asynchronously for instance. {{(pid=61806) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 877.204518] env[61806]: DEBUG oslo_concurrency.lockutils [None req-c651ea58-49e6-4c69-a5f7-88000b663e10 tempest-ServerRescueTestJSONUnderV235-749414849 tempest-ServerRescueTestJSONUnderV235-749414849-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 7.137s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 877.205856] env[61806]: INFO nova.compute.claims [None req-c651ea58-49e6-4c69-a5f7-88000b663e10 tempest-ServerRescueTestJSONUnderV235-749414849 tempest-ServerRescueTestJSONUnderV235-749414849-project-member] [instance: 5a526856-0dcd-4b3f-8359-13a5c3b9bfc4] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 877.436651] env[61806]: DEBUG oslo_concurrency.lockutils [None req-1504c565-36b9-453a-af40-d9aaf0dc82e4 tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Acquiring lock "refresh_cache-67d1bdd8-cf8f-4fb5-bd3c-42008e4d60e3" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 877.436651] env[61806]: DEBUG oslo_concurrency.lockutils [None req-1504c565-36b9-453a-af40-d9aaf0dc82e4 tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Acquired lock "refresh_cache-67d1bdd8-cf8f-4fb5-bd3c-42008e4d60e3" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 877.436651] env[61806]: DEBUG nova.network.neutron [None req-1504c565-36b9-453a-af40-d9aaf0dc82e4 tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] [instance: 67d1bdd8-cf8f-4fb5-bd3c-42008e4d60e3] Building network info cache for instance {{(pid=61806) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 877.436910] env[61806]: DEBUG nova.objects.instance [None req-1504c565-36b9-453a-af40-d9aaf0dc82e4 tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Lazy-loading 'info_cache' on Instance uuid 67d1bdd8-cf8f-4fb5-bd3c-42008e4d60e3 {{(pid=61806) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 877.502518] env[61806]: INFO nova.compute.manager [None req-58b4b995-fdfd-465f-8927-98666a556c7c tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] [instance: efa19aac-3ad7-42d5-a6d8-859050de6e63] Rebuilding instance [ 877.541752] env[61806]: DEBUG nova.compute.manager [None req-58b4b995-fdfd-465f-8927-98666a556c7c tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] [instance: efa19aac-3ad7-42d5-a6d8-859050de6e63] Checking state {{(pid=61806) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 877.542617] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-41c25aa5-66e4-454c-b8db-44f26435be7e {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 877.585132] env[61806]: DEBUG nova.network.neutron [None req-799f40df-a27b-4200-803a-166b5c1e29fd tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] [instance: 59c22c16-3d1c-4e64-bfac-7f8886a1927e] Successfully updated port: a3901838-4afa-473f-a673-faac7dee3e75 {{(pid=61806) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 877.612889] env[61806]: DEBUG nova.network.neutron [None req-b996cb64-3154-4eef-9fcd-aa24605b9020 tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] [instance: d1802da4-a670-427d-84d7-0fcb4717e18f] Instance cache missing network info. {{(pid=61806) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 877.714315] env[61806]: DEBUG nova.compute.utils [None req-b85317d7-e79b-4ab3-83e4-2de701507449 tempest-ServerTagsTestJSON-1292926242 tempest-ServerTagsTestJSON-1292926242-project-member] Using /dev/sd instead of None {{(pid=61806) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 877.720329] env[61806]: DEBUG nova.compute.manager [None req-b85317d7-e79b-4ab3-83e4-2de701507449 tempest-ServerTagsTestJSON-1292926242 tempest-ServerTagsTestJSON-1292926242-project-member] [instance: 9c20d337-b8f5-4965-b707-0f87c4ab6fc4] Allocating IP information in the background. {{(pid=61806) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 877.720329] env[61806]: DEBUG nova.network.neutron [None req-b85317d7-e79b-4ab3-83e4-2de701507449 tempest-ServerTagsTestJSON-1292926242 tempest-ServerTagsTestJSON-1292926242-project-member] [instance: 9c20d337-b8f5-4965-b707-0f87c4ab6fc4] allocate_for_instance() {{(pid=61806) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 877.792191] env[61806]: DEBUG nova.policy [None req-b85317d7-e79b-4ab3-83e4-2de701507449 tempest-ServerTagsTestJSON-1292926242 tempest-ServerTagsTestJSON-1292926242-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'a09afbaa082f4508970f2461f716f0f8', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '6d7669e9f6a643df9a5e9fd016dcbb72', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61806) authorize /opt/stack/nova/nova/policy.py:201}} [ 877.792519] env[61806]: DEBUG nova.network.neutron [None req-b996cb64-3154-4eef-9fcd-aa24605b9020 tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] [instance: d1802da4-a670-427d-84d7-0fcb4717e18f] Updating instance_info_cache with network_info: [{"id": "16d93306-6614-4a1d-aa33-e18d3f79ed1e", "address": "fa:16:3e:41:5f:ce", "network": {"id": "f280a01b-7000-4e54-ab9a-20eab156a410", "bridge": "br-int", "label": "tempest-ServersTestJSON-2080718146-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "73ebcc4d79d248efb653a084a5e44302", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "43798f54-0c07-4417-a23f-58bb6b7e204b", "external-id": "nsx-vlan-transportzone-571", "segmentation_id": 571, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap16d93306-66", "ovs_interfaceid": "16d93306-6614-4a1d-aa33-e18d3f79ed1e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61806) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 877.937160] env[61806]: DEBUG nova.compute.manager [req-03c5e62c-6358-41b3-b38a-fe0332467673 req-c1f165e9-f27c-4cf8-949a-c65b3e4ffc3d service nova] [instance: 59c22c16-3d1c-4e64-bfac-7f8886a1927e] Received event network-vif-plugged-a3901838-4afa-473f-a673-faac7dee3e75 {{(pid=61806) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 877.937160] env[61806]: DEBUG oslo_concurrency.lockutils [req-03c5e62c-6358-41b3-b38a-fe0332467673 req-c1f165e9-f27c-4cf8-949a-c65b3e4ffc3d service nova] Acquiring lock "59c22c16-3d1c-4e64-bfac-7f8886a1927e-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 877.938702] env[61806]: DEBUG oslo_concurrency.lockutils [req-03c5e62c-6358-41b3-b38a-fe0332467673 req-c1f165e9-f27c-4cf8-949a-c65b3e4ffc3d service nova] Lock "59c22c16-3d1c-4e64-bfac-7f8886a1927e-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 877.939078] env[61806]: DEBUG oslo_concurrency.lockutils [req-03c5e62c-6358-41b3-b38a-fe0332467673 req-c1f165e9-f27c-4cf8-949a-c65b3e4ffc3d service nova] Lock "59c22c16-3d1c-4e64-bfac-7f8886a1927e-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.002s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 877.939382] env[61806]: DEBUG nova.compute.manager [req-03c5e62c-6358-41b3-b38a-fe0332467673 req-c1f165e9-f27c-4cf8-949a-c65b3e4ffc3d service nova] [instance: 59c22c16-3d1c-4e64-bfac-7f8886a1927e] No waiting events found dispatching network-vif-plugged-a3901838-4afa-473f-a673-faac7dee3e75 {{(pid=61806) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 877.939680] env[61806]: WARNING nova.compute.manager [req-03c5e62c-6358-41b3-b38a-fe0332467673 req-c1f165e9-f27c-4cf8-949a-c65b3e4ffc3d service nova] [instance: 59c22c16-3d1c-4e64-bfac-7f8886a1927e] Received unexpected event network-vif-plugged-a3901838-4afa-473f-a673-faac7dee3e75 for instance with vm_state building and task_state spawning. [ 877.939982] env[61806]: DEBUG nova.compute.manager [req-03c5e62c-6358-41b3-b38a-fe0332467673 req-c1f165e9-f27c-4cf8-949a-c65b3e4ffc3d service nova] [instance: 59c22c16-3d1c-4e64-bfac-7f8886a1927e] Received event network-changed-a3901838-4afa-473f-a673-faac7dee3e75 {{(pid=61806) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 877.940270] env[61806]: DEBUG nova.compute.manager [req-03c5e62c-6358-41b3-b38a-fe0332467673 req-c1f165e9-f27c-4cf8-949a-c65b3e4ffc3d service nova] [instance: 59c22c16-3d1c-4e64-bfac-7f8886a1927e] Refreshing instance network info cache due to event network-changed-a3901838-4afa-473f-a673-faac7dee3e75. {{(pid=61806) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 877.940597] env[61806]: DEBUG oslo_concurrency.lockutils [req-03c5e62c-6358-41b3-b38a-fe0332467673 req-c1f165e9-f27c-4cf8-949a-c65b3e4ffc3d service nova] Acquiring lock "refresh_cache-59c22c16-3d1c-4e64-bfac-7f8886a1927e" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 877.940856] env[61806]: DEBUG oslo_concurrency.lockutils [req-03c5e62c-6358-41b3-b38a-fe0332467673 req-c1f165e9-f27c-4cf8-949a-c65b3e4ffc3d service nova] Acquired lock "refresh_cache-59c22c16-3d1c-4e64-bfac-7f8886a1927e" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 877.941157] env[61806]: DEBUG nova.network.neutron [req-03c5e62c-6358-41b3-b38a-fe0332467673 req-c1f165e9-f27c-4cf8-949a-c65b3e4ffc3d service nova] [instance: 59c22c16-3d1c-4e64-bfac-7f8886a1927e] Refreshing network info cache for port a3901838-4afa-473f-a673-faac7dee3e75 {{(pid=61806) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 877.943828] env[61806]: DEBUG nova.objects.base [None req-1504c565-36b9-453a-af40-d9aaf0dc82e4 tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Object Instance<67d1bdd8-cf8f-4fb5-bd3c-42008e4d60e3> lazy-loaded attributes: flavor,info_cache {{(pid=61806) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 878.036590] env[61806]: DEBUG nova.network.neutron [None req-b85317d7-e79b-4ab3-83e4-2de701507449 tempest-ServerTagsTestJSON-1292926242 tempest-ServerTagsTestJSON-1292926242-project-member] [instance: 9c20d337-b8f5-4965-b707-0f87c4ab6fc4] Successfully created port: a7f1c67d-ce27-436f-9329-4a3933beeb5c {{(pid=61806) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 878.054226] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [None req-58b4b995-fdfd-465f-8927-98666a556c7c tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] [instance: efa19aac-3ad7-42d5-a6d8-859050de6e63] Powering off the VM {{(pid=61806) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 878.054537] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-b035ea17-5426-4845-b826-acdf807b4ad4 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 878.062699] env[61806]: DEBUG oslo_vmware.api [None req-58b4b995-fdfd-465f-8927-98666a556c7c tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Waiting for the task: (returnval){ [ 878.062699] env[61806]: value = "task-1294775" [ 878.062699] env[61806]: _type = "Task" [ 878.062699] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 878.072549] env[61806]: DEBUG oslo_vmware.api [None req-58b4b995-fdfd-465f-8927-98666a556c7c tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Task: {'id': task-1294775, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 878.087992] env[61806]: DEBUG oslo_concurrency.lockutils [None req-799f40df-a27b-4200-803a-166b5c1e29fd tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Acquiring lock "refresh_cache-59c22c16-3d1c-4e64-bfac-7f8886a1927e" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 878.222274] env[61806]: DEBUG nova.compute.manager [None req-b85317d7-e79b-4ab3-83e4-2de701507449 tempest-ServerTagsTestJSON-1292926242 tempest-ServerTagsTestJSON-1292926242-project-member] [instance: 9c20d337-b8f5-4965-b707-0f87c4ab6fc4] Start building block device mappings for instance. {{(pid=61806) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 878.295265] env[61806]: DEBUG oslo_concurrency.lockutils [None req-b996cb64-3154-4eef-9fcd-aa24605b9020 tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] Releasing lock "refresh_cache-d1802da4-a670-427d-84d7-0fcb4717e18f" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 878.295566] env[61806]: DEBUG nova.compute.manager [None req-b996cb64-3154-4eef-9fcd-aa24605b9020 tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] [instance: d1802da4-a670-427d-84d7-0fcb4717e18f] Instance network_info: |[{"id": "16d93306-6614-4a1d-aa33-e18d3f79ed1e", "address": "fa:16:3e:41:5f:ce", "network": {"id": "f280a01b-7000-4e54-ab9a-20eab156a410", "bridge": "br-int", "label": "tempest-ServersTestJSON-2080718146-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "73ebcc4d79d248efb653a084a5e44302", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "43798f54-0c07-4417-a23f-58bb6b7e204b", "external-id": "nsx-vlan-transportzone-571", "segmentation_id": 571, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap16d93306-66", "ovs_interfaceid": "16d93306-6614-4a1d-aa33-e18d3f79ed1e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61806) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 878.295959] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-b996cb64-3154-4eef-9fcd-aa24605b9020 tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] [instance: d1802da4-a670-427d-84d7-0fcb4717e18f] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:41:5f:ce', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '43798f54-0c07-4417-a23f-58bb6b7e204b', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '16d93306-6614-4a1d-aa33-e18d3f79ed1e', 'vif_model': 'vmxnet3'}] {{(pid=61806) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 878.303681] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [None req-b996cb64-3154-4eef-9fcd-aa24605b9020 tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] Creating folder: Project (73ebcc4d79d248efb653a084a5e44302). Parent ref: group-v277609. {{(pid=61806) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 878.306388] env[61806]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-a9a69848-d1af-48ac-920f-b9cc0c295912 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 878.320345] env[61806]: INFO nova.virt.vmwareapi.vm_util [None req-b996cb64-3154-4eef-9fcd-aa24605b9020 tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] Created folder: Project (73ebcc4d79d248efb653a084a5e44302) in parent group-v277609. [ 878.320576] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [None req-b996cb64-3154-4eef-9fcd-aa24605b9020 tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] Creating folder: Instances. Parent ref: group-v277690. {{(pid=61806) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 878.320817] env[61806]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-3cb6fd74-a8e7-43f4-9e08-a67e74935df5 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 878.332441] env[61806]: INFO nova.virt.vmwareapi.vm_util [None req-b996cb64-3154-4eef-9fcd-aa24605b9020 tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] Created folder: Instances in parent group-v277690. [ 878.332695] env[61806]: DEBUG oslo.service.loopingcall [None req-b996cb64-3154-4eef-9fcd-aa24605b9020 tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61806) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 878.332895] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: d1802da4-a670-427d-84d7-0fcb4717e18f] Creating VM on the ESX host {{(pid=61806) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 878.333140] env[61806]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-9b2e8482-e86f-4a5c-9b9c-b5fc7d4ba92d {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 878.354897] env[61806]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 878.354897] env[61806]: value = "task-1294778" [ 878.354897] env[61806]: _type = "Task" [ 878.354897] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 878.364306] env[61806]: DEBUG oslo_vmware.api [-] Task: {'id': task-1294778, 'name': CreateVM_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 878.401108] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ce80d781-8605-4898-ac76-178a2e93bbf3 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 878.408765] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-65f5eeeb-0426-47f6-9010-b793244e1b97 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 878.441420] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a17ee423-01ee-4cad-95e5-5ca74a2f3e73 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 878.452245] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-40bf083b-354f-4027-a9b8-bbf00aea4cae {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 878.467304] env[61806]: DEBUG nova.compute.provider_tree [None req-c651ea58-49e6-4c69-a5f7-88000b663e10 tempest-ServerRescueTestJSONUnderV235-749414849 tempest-ServerRescueTestJSONUnderV235-749414849-project-member] Inventory has not changed in ProviderTree for provider: a2858be1-fd22-4e08-979e-87ad25293407 {{(pid=61806) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 878.486871] env[61806]: DEBUG nova.network.neutron [req-03c5e62c-6358-41b3-b38a-fe0332467673 req-c1f165e9-f27c-4cf8-949a-c65b3e4ffc3d service nova] [instance: 59c22c16-3d1c-4e64-bfac-7f8886a1927e] Instance cache missing network info. {{(pid=61806) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 878.571804] env[61806]: DEBUG nova.network.neutron [req-03c5e62c-6358-41b3-b38a-fe0332467673 req-c1f165e9-f27c-4cf8-949a-c65b3e4ffc3d service nova] [instance: 59c22c16-3d1c-4e64-bfac-7f8886a1927e] Updating instance_info_cache with network_info: [] {{(pid=61806) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 878.578408] env[61806]: DEBUG oslo_vmware.api [None req-58b4b995-fdfd-465f-8927-98666a556c7c tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Task: {'id': task-1294775, 'name': PowerOffVM_Task, 'duration_secs': 0.198214} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 878.578755] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [None req-58b4b995-fdfd-465f-8927-98666a556c7c tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] [instance: efa19aac-3ad7-42d5-a6d8-859050de6e63] Powered off the VM {{(pid=61806) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 878.578990] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-58b4b995-fdfd-465f-8927-98666a556c7c tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] [instance: efa19aac-3ad7-42d5-a6d8-859050de6e63] Destroying instance {{(pid=61806) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 878.579758] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2066bfec-acf0-438f-8685-1cfb89035989 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 878.588071] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-58b4b995-fdfd-465f-8927-98666a556c7c tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] [instance: efa19aac-3ad7-42d5-a6d8-859050de6e63] Unregistering the VM {{(pid=61806) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 878.588347] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-54957e23-124e-4dd7-a0ef-1af45f36daba {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 878.686534] env[61806]: DEBUG nova.network.neutron [None req-1504c565-36b9-453a-af40-d9aaf0dc82e4 tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] [instance: 67d1bdd8-cf8f-4fb5-bd3c-42008e4d60e3] Updating instance_info_cache with network_info: [{"id": "859fb723-4a1c-4ea8-bd7c-0ed308f3a9e3", "address": "fa:16:3e:3c:40:d6", "network": {"id": "746a03b6-c1a6-4d27-98b9-0bee6376634f", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-1971382587-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.245", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c1f2d161c50f4b70987b90e6364c3779", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6a31b224-77d7-48a4-af87-312758611b8e", "external-id": "nsx-vlan-transportzone-761", "segmentation_id": 761, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap859fb723-4a", "ovs_interfaceid": "859fb723-4a1c-4ea8-bd7c-0ed308f3a9e3", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61806) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 878.715179] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-58b4b995-fdfd-465f-8927-98666a556c7c tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] [instance: efa19aac-3ad7-42d5-a6d8-859050de6e63] Unregistered the VM {{(pid=61806) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 878.715407] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-58b4b995-fdfd-465f-8927-98666a556c7c tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] [instance: efa19aac-3ad7-42d5-a6d8-859050de6e63] Deleting contents of the VM from datastore datastore2 {{(pid=61806) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 878.715592] env[61806]: DEBUG nova.virt.vmwareapi.ds_util [None req-58b4b995-fdfd-465f-8927-98666a556c7c tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Deleting the datastore file [datastore2] efa19aac-3ad7-42d5-a6d8-859050de6e63 {{(pid=61806) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 878.715855] env[61806]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-bf65fd55-8563-4c04-ba5f-06330d3d812e {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 878.722753] env[61806]: DEBUG oslo_vmware.api [None req-58b4b995-fdfd-465f-8927-98666a556c7c tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Waiting for the task: (returnval){ [ 878.722753] env[61806]: value = "task-1294780" [ 878.722753] env[61806]: _type = "Task" [ 878.722753] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 878.734929] env[61806]: DEBUG oslo_vmware.api [None req-58b4b995-fdfd-465f-8927-98666a556c7c tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Task: {'id': task-1294780, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 878.867056] env[61806]: DEBUG oslo_vmware.api [-] Task: {'id': task-1294778, 'name': CreateVM_Task, 'duration_secs': 0.318998} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 878.867235] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: d1802da4-a670-427d-84d7-0fcb4717e18f] Created VM on the ESX host {{(pid=61806) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 878.868031] env[61806]: DEBUG oslo_concurrency.lockutils [None req-b996cb64-3154-4eef-9fcd-aa24605b9020 tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 878.868214] env[61806]: DEBUG oslo_concurrency.lockutils [None req-b996cb64-3154-4eef-9fcd-aa24605b9020 tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] Acquired lock "[datastore1] devstack-image-cache_base/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 878.868546] env[61806]: DEBUG oslo_concurrency.lockutils [None req-b996cb64-3154-4eef-9fcd-aa24605b9020 tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 878.869350] env[61806]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c2b111bd-7734-41e3-ae00-57eca3b352c2 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 878.875367] env[61806]: DEBUG oslo_vmware.api [None req-b996cb64-3154-4eef-9fcd-aa24605b9020 tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] Waiting for the task: (returnval){ [ 878.875367] env[61806]: value = "session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]52644c90-5abb-9e4c-fcd3-579233d2ce1b" [ 878.875367] env[61806]: _type = "Task" [ 878.875367] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 878.883747] env[61806]: DEBUG oslo_vmware.api [None req-b996cb64-3154-4eef-9fcd-aa24605b9020 tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] Task: {'id': session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]52644c90-5abb-9e4c-fcd3-579233d2ce1b, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 878.970525] env[61806]: DEBUG nova.scheduler.client.report [None req-c651ea58-49e6-4c69-a5f7-88000b663e10 tempest-ServerRescueTestJSONUnderV235-749414849 tempest-ServerRescueTestJSONUnderV235-749414849-project-member] Inventory has not changed for provider a2858be1-fd22-4e08-979e-87ad25293407 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 138, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61806) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 879.075052] env[61806]: DEBUG oslo_concurrency.lockutils [req-03c5e62c-6358-41b3-b38a-fe0332467673 req-c1f165e9-f27c-4cf8-949a-c65b3e4ffc3d service nova] Releasing lock "refresh_cache-59c22c16-3d1c-4e64-bfac-7f8886a1927e" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 879.075448] env[61806]: DEBUG oslo_concurrency.lockutils [None req-799f40df-a27b-4200-803a-166b5c1e29fd tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Acquired lock "refresh_cache-59c22c16-3d1c-4e64-bfac-7f8886a1927e" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 879.075622] env[61806]: DEBUG nova.network.neutron [None req-799f40df-a27b-4200-803a-166b5c1e29fd tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] [instance: 59c22c16-3d1c-4e64-bfac-7f8886a1927e] Building network info cache for instance {{(pid=61806) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 879.190218] env[61806]: DEBUG oslo_concurrency.lockutils [None req-1504c565-36b9-453a-af40-d9aaf0dc82e4 tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Releasing lock "refresh_cache-67d1bdd8-cf8f-4fb5-bd3c-42008e4d60e3" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 879.234890] env[61806]: DEBUG oslo_vmware.api [None req-58b4b995-fdfd-465f-8927-98666a556c7c tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Task: {'id': task-1294780, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.135536} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 879.235187] env[61806]: DEBUG nova.virt.vmwareapi.ds_util [None req-58b4b995-fdfd-465f-8927-98666a556c7c tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Deleted the datastore file {{(pid=61806) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 879.235383] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-58b4b995-fdfd-465f-8927-98666a556c7c tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] [instance: efa19aac-3ad7-42d5-a6d8-859050de6e63] Deleted contents of the VM from datastore datastore2 {{(pid=61806) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 879.235565] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-58b4b995-fdfd-465f-8927-98666a556c7c tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] [instance: efa19aac-3ad7-42d5-a6d8-859050de6e63] Instance destroyed {{(pid=61806) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 879.238409] env[61806]: DEBUG nova.compute.manager [None req-b85317d7-e79b-4ab3-83e4-2de701507449 tempest-ServerTagsTestJSON-1292926242 tempest-ServerTagsTestJSON-1292926242-project-member] [instance: 9c20d337-b8f5-4965-b707-0f87c4ab6fc4] Start spawning the instance on the hypervisor. {{(pid=61806) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 879.266629] env[61806]: DEBUG nova.virt.hardware [None req-b85317d7-e79b-4ab3-83e4-2de701507449 tempest-ServerTagsTestJSON-1292926242 tempest-ServerTagsTestJSON-1292926242-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-15T18:05:10Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-15T18:04:52Z,direct_url=,disk_format='vmdk',id=ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='84ba83fce2c349a6988173c5d6fc3b07',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-15T18:04:53Z,virtual_size=,visibility=), allow threads: False {{(pid=61806) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 879.266882] env[61806]: DEBUG nova.virt.hardware [None req-b85317d7-e79b-4ab3-83e4-2de701507449 tempest-ServerTagsTestJSON-1292926242 tempest-ServerTagsTestJSON-1292926242-project-member] Flavor limits 0:0:0 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 879.267060] env[61806]: DEBUG nova.virt.hardware [None req-b85317d7-e79b-4ab3-83e4-2de701507449 tempest-ServerTagsTestJSON-1292926242 tempest-ServerTagsTestJSON-1292926242-project-member] Image limits 0:0:0 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 879.267254] env[61806]: DEBUG nova.virt.hardware [None req-b85317d7-e79b-4ab3-83e4-2de701507449 tempest-ServerTagsTestJSON-1292926242 tempest-ServerTagsTestJSON-1292926242-project-member] Flavor pref 0:0:0 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 879.267416] env[61806]: DEBUG nova.virt.hardware [None req-b85317d7-e79b-4ab3-83e4-2de701507449 tempest-ServerTagsTestJSON-1292926242 tempest-ServerTagsTestJSON-1292926242-project-member] Image pref 0:0:0 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 879.267559] env[61806]: DEBUG nova.virt.hardware [None req-b85317d7-e79b-4ab3-83e4-2de701507449 tempest-ServerTagsTestJSON-1292926242 tempest-ServerTagsTestJSON-1292926242-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 879.267770] env[61806]: DEBUG nova.virt.hardware [None req-b85317d7-e79b-4ab3-83e4-2de701507449 tempest-ServerTagsTestJSON-1292926242 tempest-ServerTagsTestJSON-1292926242-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61806) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 879.267933] env[61806]: DEBUG nova.virt.hardware [None req-b85317d7-e79b-4ab3-83e4-2de701507449 tempest-ServerTagsTestJSON-1292926242 tempest-ServerTagsTestJSON-1292926242-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61806) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 879.268118] env[61806]: DEBUG nova.virt.hardware [None req-b85317d7-e79b-4ab3-83e4-2de701507449 tempest-ServerTagsTestJSON-1292926242 tempest-ServerTagsTestJSON-1292926242-project-member] Got 1 possible topologies {{(pid=61806) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 879.268295] env[61806]: DEBUG nova.virt.hardware [None req-b85317d7-e79b-4ab3-83e4-2de701507449 tempest-ServerTagsTestJSON-1292926242 tempest-ServerTagsTestJSON-1292926242-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61806) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 879.269121] env[61806]: DEBUG nova.virt.hardware [None req-b85317d7-e79b-4ab3-83e4-2de701507449 tempest-ServerTagsTestJSON-1292926242 tempest-ServerTagsTestJSON-1292926242-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61806) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 879.270088] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6de229ff-2a47-4f74-977a-a8b8c45ae5f2 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 879.279536] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7efb5321-2374-4094-8838-3ebeb6bcd7dd {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 879.386944] env[61806]: DEBUG oslo_vmware.api [None req-b996cb64-3154-4eef-9fcd-aa24605b9020 tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] Task: {'id': session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]52644c90-5abb-9e4c-fcd3-579233d2ce1b, 'name': SearchDatastore_Task, 'duration_secs': 0.012128} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 879.387278] env[61806]: DEBUG oslo_concurrency.lockutils [None req-b996cb64-3154-4eef-9fcd-aa24605b9020 tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] Releasing lock "[datastore1] devstack-image-cache_base/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 879.387528] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-b996cb64-3154-4eef-9fcd-aa24605b9020 tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] [instance: d1802da4-a670-427d-84d7-0fcb4717e18f] Processing image ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa {{(pid=61806) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 879.387766] env[61806]: DEBUG oslo_concurrency.lockutils [None req-b996cb64-3154-4eef-9fcd-aa24605b9020 tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa.vmdk" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 879.387919] env[61806]: DEBUG oslo_concurrency.lockutils [None req-b996cb64-3154-4eef-9fcd-aa24605b9020 tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] Acquired lock "[datastore1] devstack-image-cache_base/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa.vmdk" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 879.388115] env[61806]: DEBUG nova.virt.vmwareapi.ds_util [None req-b996cb64-3154-4eef-9fcd-aa24605b9020 tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61806) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 879.388382] env[61806]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-34ef1090-9a19-41b1-b5d0-f8f9953c6bb3 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 879.398228] env[61806]: DEBUG nova.virt.vmwareapi.ds_util [None req-b996cb64-3154-4eef-9fcd-aa24605b9020 tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61806) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 879.398422] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-b996cb64-3154-4eef-9fcd-aa24605b9020 tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61806) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 879.399232] env[61806]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-3c4ee93b-a9ef-46fc-8e3d-3a4ea82c667e {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 879.405379] env[61806]: DEBUG oslo_vmware.api [None req-b996cb64-3154-4eef-9fcd-aa24605b9020 tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] Waiting for the task: (returnval){ [ 879.405379] env[61806]: value = "session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]5223addf-22e2-a73a-f313-63850348eadc" [ 879.405379] env[61806]: _type = "Task" [ 879.405379] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 879.413146] env[61806]: DEBUG oslo_vmware.api [None req-b996cb64-3154-4eef-9fcd-aa24605b9020 tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] Task: {'id': session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]5223addf-22e2-a73a-f313-63850348eadc, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 879.475600] env[61806]: DEBUG oslo_concurrency.lockutils [None req-c651ea58-49e6-4c69-a5f7-88000b663e10 tempest-ServerRescueTestJSONUnderV235-749414849 tempest-ServerRescueTestJSONUnderV235-749414849-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.271s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 879.476159] env[61806]: DEBUG nova.compute.manager [None req-c651ea58-49e6-4c69-a5f7-88000b663e10 tempest-ServerRescueTestJSONUnderV235-749414849 tempest-ServerRescueTestJSONUnderV235-749414849-project-member] [instance: 5a526856-0dcd-4b3f-8359-13a5c3b9bfc4] Start building networks asynchronously for instance. {{(pid=61806) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 879.478800] env[61806]: DEBUG oslo_concurrency.lockutils [None req-a7b134bb-7562-4442-8a02-daf53033761c tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 8.067s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 879.479143] env[61806]: DEBUG nova.objects.instance [None req-a7b134bb-7562-4442-8a02-daf53033761c tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] Lazy-loading 'resources' on Instance uuid 2a3f81e3-b84f-4370-a598-2f6a607c3d47 {{(pid=61806) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 879.607641] env[61806]: DEBUG nova.network.neutron [None req-799f40df-a27b-4200-803a-166b5c1e29fd tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] [instance: 59c22c16-3d1c-4e64-bfac-7f8886a1927e] Instance cache missing network info. {{(pid=61806) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 879.641541] env[61806]: DEBUG nova.network.neutron [None req-b85317d7-e79b-4ab3-83e4-2de701507449 tempest-ServerTagsTestJSON-1292926242 tempest-ServerTagsTestJSON-1292926242-project-member] [instance: 9c20d337-b8f5-4965-b707-0f87c4ab6fc4] Successfully updated port: a7f1c67d-ce27-436f-9329-4a3933beeb5c {{(pid=61806) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 879.694080] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [None req-1504c565-36b9-453a-af40-d9aaf0dc82e4 tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] [instance: 67d1bdd8-cf8f-4fb5-bd3c-42008e4d60e3] Powering on the VM {{(pid=61806) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 879.694466] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-af77a72b-31fa-47c4-a30c-181b6e5faff8 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 879.702993] env[61806]: DEBUG oslo_vmware.api [None req-1504c565-36b9-453a-af40-d9aaf0dc82e4 tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Waiting for the task: (returnval){ [ 879.702993] env[61806]: value = "task-1294781" [ 879.702993] env[61806]: _type = "Task" [ 879.702993] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 879.713713] env[61806]: DEBUG oslo_vmware.api [None req-1504c565-36b9-453a-af40-d9aaf0dc82e4 tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Task: {'id': task-1294781, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 879.746583] env[61806]: DEBUG nova.network.neutron [None req-799f40df-a27b-4200-803a-166b5c1e29fd tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] [instance: 59c22c16-3d1c-4e64-bfac-7f8886a1927e] Updating instance_info_cache with network_info: [{"id": "a3901838-4afa-473f-a673-faac7dee3e75", "address": "fa:16:3e:57:5a:0d", "network": {"id": "b7e78559-d885-4b59-9c7f-9d966b424d04", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-1456721354-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7ab95fb72b5d46c3b6c7bebbccf897cd", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d891a662-7da0-4be5-8d0d-01b4ff70552a", "external-id": "nsx-vlan-transportzone-815", "segmentation_id": 815, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa3901838-4a", "ovs_interfaceid": "a3901838-4afa-473f-a673-faac7dee3e75", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61806) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 879.916524] env[61806]: DEBUG oslo_vmware.api [None req-b996cb64-3154-4eef-9fcd-aa24605b9020 tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] Task: {'id': session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]5223addf-22e2-a73a-f313-63850348eadc, 'name': SearchDatastore_Task, 'duration_secs': 0.018922} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 879.917522] env[61806]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-2f516ace-03e9-4173-a302-31b5673005cf {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 879.924373] env[61806]: DEBUG oslo_vmware.api [None req-b996cb64-3154-4eef-9fcd-aa24605b9020 tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] Waiting for the task: (returnval){ [ 879.924373] env[61806]: value = "session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]526ee85d-c4ff-955c-e6ac-67836a284199" [ 879.924373] env[61806]: _type = "Task" [ 879.924373] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 879.934566] env[61806]: DEBUG oslo_vmware.api [None req-b996cb64-3154-4eef-9fcd-aa24605b9020 tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] Task: {'id': session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]526ee85d-c4ff-955c-e6ac-67836a284199, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 879.962901] env[61806]: DEBUG nova.compute.manager [req-c9e1b6f9-7795-4013-b214-a80af10f899d req-73e536d4-e93d-45d4-8ed4-5bbe97c39b3f service nova] [instance: 9c20d337-b8f5-4965-b707-0f87c4ab6fc4] Received event network-vif-plugged-a7f1c67d-ce27-436f-9329-4a3933beeb5c {{(pid=61806) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 879.963165] env[61806]: DEBUG oslo_concurrency.lockutils [req-c9e1b6f9-7795-4013-b214-a80af10f899d req-73e536d4-e93d-45d4-8ed4-5bbe97c39b3f service nova] Acquiring lock "9c20d337-b8f5-4965-b707-0f87c4ab6fc4-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 879.963376] env[61806]: DEBUG oslo_concurrency.lockutils [req-c9e1b6f9-7795-4013-b214-a80af10f899d req-73e536d4-e93d-45d4-8ed4-5bbe97c39b3f service nova] Lock "9c20d337-b8f5-4965-b707-0f87c4ab6fc4-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 879.963509] env[61806]: DEBUG oslo_concurrency.lockutils [req-c9e1b6f9-7795-4013-b214-a80af10f899d req-73e536d4-e93d-45d4-8ed4-5bbe97c39b3f service nova] Lock "9c20d337-b8f5-4965-b707-0f87c4ab6fc4-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 879.963677] env[61806]: DEBUG nova.compute.manager [req-c9e1b6f9-7795-4013-b214-a80af10f899d req-73e536d4-e93d-45d4-8ed4-5bbe97c39b3f service nova] [instance: 9c20d337-b8f5-4965-b707-0f87c4ab6fc4] No waiting events found dispatching network-vif-plugged-a7f1c67d-ce27-436f-9329-4a3933beeb5c {{(pid=61806) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 879.963860] env[61806]: WARNING nova.compute.manager [req-c9e1b6f9-7795-4013-b214-a80af10f899d req-73e536d4-e93d-45d4-8ed4-5bbe97c39b3f service nova] [instance: 9c20d337-b8f5-4965-b707-0f87c4ab6fc4] Received unexpected event network-vif-plugged-a7f1c67d-ce27-436f-9329-4a3933beeb5c for instance with vm_state building and task_state spawning. [ 879.964033] env[61806]: DEBUG nova.compute.manager [req-c9e1b6f9-7795-4013-b214-a80af10f899d req-73e536d4-e93d-45d4-8ed4-5bbe97c39b3f service nova] [instance: 9c20d337-b8f5-4965-b707-0f87c4ab6fc4] Received event network-changed-a7f1c67d-ce27-436f-9329-4a3933beeb5c {{(pid=61806) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 879.964227] env[61806]: DEBUG nova.compute.manager [req-c9e1b6f9-7795-4013-b214-a80af10f899d req-73e536d4-e93d-45d4-8ed4-5bbe97c39b3f service nova] [instance: 9c20d337-b8f5-4965-b707-0f87c4ab6fc4] Refreshing instance network info cache due to event network-changed-a7f1c67d-ce27-436f-9329-4a3933beeb5c. {{(pid=61806) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 879.964419] env[61806]: DEBUG oslo_concurrency.lockutils [req-c9e1b6f9-7795-4013-b214-a80af10f899d req-73e536d4-e93d-45d4-8ed4-5bbe97c39b3f service nova] Acquiring lock "refresh_cache-9c20d337-b8f5-4965-b707-0f87c4ab6fc4" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 879.964561] env[61806]: DEBUG oslo_concurrency.lockutils [req-c9e1b6f9-7795-4013-b214-a80af10f899d req-73e536d4-e93d-45d4-8ed4-5bbe97c39b3f service nova] Acquired lock "refresh_cache-9c20d337-b8f5-4965-b707-0f87c4ab6fc4" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 879.964719] env[61806]: DEBUG nova.network.neutron [req-c9e1b6f9-7795-4013-b214-a80af10f899d req-73e536d4-e93d-45d4-8ed4-5bbe97c39b3f service nova] [instance: 9c20d337-b8f5-4965-b707-0f87c4ab6fc4] Refreshing network info cache for port a7f1c67d-ce27-436f-9329-4a3933beeb5c {{(pid=61806) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 879.982294] env[61806]: DEBUG nova.compute.utils [None req-c651ea58-49e6-4c69-a5f7-88000b663e10 tempest-ServerRescueTestJSONUnderV235-749414849 tempest-ServerRescueTestJSONUnderV235-749414849-project-member] Using /dev/sd instead of None {{(pid=61806) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 879.986192] env[61806]: DEBUG nova.compute.manager [None req-c651ea58-49e6-4c69-a5f7-88000b663e10 tempest-ServerRescueTestJSONUnderV235-749414849 tempest-ServerRescueTestJSONUnderV235-749414849-project-member] [instance: 5a526856-0dcd-4b3f-8359-13a5c3b9bfc4] Allocating IP information in the background. {{(pid=61806) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 879.986363] env[61806]: DEBUG nova.network.neutron [None req-c651ea58-49e6-4c69-a5f7-88000b663e10 tempest-ServerRescueTestJSONUnderV235-749414849 tempest-ServerRescueTestJSONUnderV235-749414849-project-member] [instance: 5a526856-0dcd-4b3f-8359-13a5c3b9bfc4] allocate_for_instance() {{(pid=61806) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 880.023913] env[61806]: DEBUG nova.policy [None req-c651ea58-49e6-4c69-a5f7-88000b663e10 tempest-ServerRescueTestJSONUnderV235-749414849 tempest-ServerRescueTestJSONUnderV235-749414849-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'cfbde9681e234ad08b176673120c4ea1', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '2ba59612579b47a4bde7f82d951ac3ff', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61806) authorize /opt/stack/nova/nova/policy.py:201}} [ 880.129669] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ed02bdf9-9ea4-4637-bc90-b52e5eda8053 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 880.142315] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c6b1050a-95a2-43a5-a227-bbae1a2e5515 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 880.146024] env[61806]: DEBUG oslo_concurrency.lockutils [None req-b85317d7-e79b-4ab3-83e4-2de701507449 tempest-ServerTagsTestJSON-1292926242 tempest-ServerTagsTestJSON-1292926242-project-member] Acquiring lock "refresh_cache-9c20d337-b8f5-4965-b707-0f87c4ab6fc4" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 880.173726] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-90b03648-8087-4522-ae2c-de5cc4a5dfd8 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 880.183657] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c646da17-4192-43b0-bab0-e45bdba94a67 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 880.197953] env[61806]: DEBUG nova.compute.provider_tree [None req-a7b134bb-7562-4442-8a02-daf53033761c tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] Inventory has not changed in ProviderTree for provider: a2858be1-fd22-4e08-979e-87ad25293407 {{(pid=61806) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 880.212372] env[61806]: DEBUG oslo_vmware.api [None req-1504c565-36b9-453a-af40-d9aaf0dc82e4 tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Task: {'id': task-1294781, 'name': PowerOnVM_Task, 'duration_secs': 0.463064} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 880.212627] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [None req-1504c565-36b9-453a-af40-d9aaf0dc82e4 tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] [instance: 67d1bdd8-cf8f-4fb5-bd3c-42008e4d60e3] Powered on the VM {{(pid=61806) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 880.212814] env[61806]: DEBUG nova.compute.manager [None req-1504c565-36b9-453a-af40-d9aaf0dc82e4 tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] [instance: 67d1bdd8-cf8f-4fb5-bd3c-42008e4d60e3] Checking state {{(pid=61806) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 880.213594] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b611fb26-b27d-44d4-9c2a-7741480c0fd0 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 880.251560] env[61806]: DEBUG oslo_concurrency.lockutils [None req-799f40df-a27b-4200-803a-166b5c1e29fd tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Releasing lock "refresh_cache-59c22c16-3d1c-4e64-bfac-7f8886a1927e" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 880.252175] env[61806]: DEBUG nova.compute.manager [None req-799f40df-a27b-4200-803a-166b5c1e29fd tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] [instance: 59c22c16-3d1c-4e64-bfac-7f8886a1927e] Instance network_info: |[{"id": "a3901838-4afa-473f-a673-faac7dee3e75", "address": "fa:16:3e:57:5a:0d", "network": {"id": "b7e78559-d885-4b59-9c7f-9d966b424d04", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-1456721354-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7ab95fb72b5d46c3b6c7bebbccf897cd", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d891a662-7da0-4be5-8d0d-01b4ff70552a", "external-id": "nsx-vlan-transportzone-815", "segmentation_id": 815, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa3901838-4a", "ovs_interfaceid": "a3901838-4afa-473f-a673-faac7dee3e75", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61806) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 880.252334] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-799f40df-a27b-4200-803a-166b5c1e29fd tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] [instance: 59c22c16-3d1c-4e64-bfac-7f8886a1927e] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:57:5a:0d', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'd891a662-7da0-4be5-8d0d-01b4ff70552a', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'a3901838-4afa-473f-a673-faac7dee3e75', 'vif_model': 'vmxnet3'}] {{(pid=61806) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 880.260451] env[61806]: DEBUG oslo.service.loopingcall [None req-799f40df-a27b-4200-803a-166b5c1e29fd tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61806) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 880.261342] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 59c22c16-3d1c-4e64-bfac-7f8886a1927e] Creating VM on the ESX host {{(pid=61806) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 880.261594] env[61806]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-e49b1b1c-6b48-4d0a-be86-54092046a9ba {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 880.287097] env[61806]: DEBUG nova.virt.hardware [None req-58b4b995-fdfd-465f-8927-98666a556c7c tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-15T18:05:10Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-15T18:04:52Z,direct_url=,disk_format='vmdk',id=ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='84ba83fce2c349a6988173c5d6fc3b07',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-15T18:04:53Z,virtual_size=,visibility=), allow threads: False {{(pid=61806) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 880.287386] env[61806]: DEBUG nova.virt.hardware [None req-58b4b995-fdfd-465f-8927-98666a556c7c tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Flavor limits 0:0:0 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 880.287553] env[61806]: DEBUG nova.virt.hardware [None req-58b4b995-fdfd-465f-8927-98666a556c7c tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Image limits 0:0:0 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 880.287736] env[61806]: DEBUG nova.virt.hardware [None req-58b4b995-fdfd-465f-8927-98666a556c7c tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Flavor pref 0:0:0 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 880.287886] env[61806]: DEBUG nova.virt.hardware [None req-58b4b995-fdfd-465f-8927-98666a556c7c tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Image pref 0:0:0 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 880.288048] env[61806]: DEBUG nova.virt.hardware [None req-58b4b995-fdfd-465f-8927-98666a556c7c tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 880.288290] env[61806]: DEBUG nova.virt.hardware [None req-58b4b995-fdfd-465f-8927-98666a556c7c tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61806) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 880.288598] env[61806]: DEBUG nova.virt.hardware [None req-58b4b995-fdfd-465f-8927-98666a556c7c tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61806) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 880.288655] env[61806]: DEBUG nova.virt.hardware [None req-58b4b995-fdfd-465f-8927-98666a556c7c tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Got 1 possible topologies {{(pid=61806) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 880.288795] env[61806]: DEBUG nova.virt.hardware [None req-58b4b995-fdfd-465f-8927-98666a556c7c tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61806) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 880.288974] env[61806]: DEBUG nova.virt.hardware [None req-58b4b995-fdfd-465f-8927-98666a556c7c tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61806) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 880.290194] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0be13ce4-ff74-4754-b2dc-b349ce04cfef {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 880.294325] env[61806]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 880.294325] env[61806]: value = "task-1294782" [ 880.294325] env[61806]: _type = "Task" [ 880.294325] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 880.301454] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-705c41b1-0281-4750-8ff9-8da6a1fd1c9b {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 880.306566] env[61806]: DEBUG nova.network.neutron [None req-c651ea58-49e6-4c69-a5f7-88000b663e10 tempest-ServerRescueTestJSONUnderV235-749414849 tempest-ServerRescueTestJSONUnderV235-749414849-project-member] [instance: 5a526856-0dcd-4b3f-8359-13a5c3b9bfc4] Successfully created port: 58782499-bc1a-4be1-9d48-655c6b2f11d7 {{(pid=61806) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 880.311746] env[61806]: DEBUG oslo_vmware.api [-] Task: {'id': task-1294782, 'name': CreateVM_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 880.322928] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-58b4b995-fdfd-465f-8927-98666a556c7c tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] [instance: efa19aac-3ad7-42d5-a6d8-859050de6e63] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:18:2f:85', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '82ca17df-257e-40e6-9ec9-310ed6f05ccb', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'fea0c8d9-43ab-4462-9427-73d9ae4a47b2', 'vif_model': 'vmxnet3'}] {{(pid=61806) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 880.330536] env[61806]: DEBUG oslo.service.loopingcall [None req-58b4b995-fdfd-465f-8927-98666a556c7c tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61806) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 880.330784] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: efa19aac-3ad7-42d5-a6d8-859050de6e63] Creating VM on the ESX host {{(pid=61806) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 880.331036] env[61806]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-9c02cb00-7867-4256-bfdd-8834c267dae1 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 880.351169] env[61806]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 880.351169] env[61806]: value = "task-1294783" [ 880.351169] env[61806]: _type = "Task" [ 880.351169] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 880.359503] env[61806]: DEBUG oslo_vmware.api [-] Task: {'id': task-1294783, 'name': CreateVM_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 880.436502] env[61806]: DEBUG oslo_vmware.api [None req-b996cb64-3154-4eef-9fcd-aa24605b9020 tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] Task: {'id': session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]526ee85d-c4ff-955c-e6ac-67836a284199, 'name': SearchDatastore_Task, 'duration_secs': 0.019781} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 880.436803] env[61806]: DEBUG oslo_concurrency.lockutils [None req-b996cb64-3154-4eef-9fcd-aa24605b9020 tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] Releasing lock "[datastore1] devstack-image-cache_base/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa.vmdk" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 880.437136] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [None req-b996cb64-3154-4eef-9fcd-aa24605b9020 tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa.vmdk to [datastore1] d1802da4-a670-427d-84d7-0fcb4717e18f/d1802da4-a670-427d-84d7-0fcb4717e18f.vmdk {{(pid=61806) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 880.437499] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-e3de939b-e799-4c75-a95f-26a70b2db070 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 880.445616] env[61806]: DEBUG oslo_vmware.api [None req-b996cb64-3154-4eef-9fcd-aa24605b9020 tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] Waiting for the task: (returnval){ [ 880.445616] env[61806]: value = "task-1294784" [ 880.445616] env[61806]: _type = "Task" [ 880.445616] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 880.453548] env[61806]: DEBUG oslo_vmware.api [None req-b996cb64-3154-4eef-9fcd-aa24605b9020 tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] Task: {'id': task-1294784, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 880.487084] env[61806]: DEBUG nova.compute.manager [None req-c651ea58-49e6-4c69-a5f7-88000b663e10 tempest-ServerRescueTestJSONUnderV235-749414849 tempest-ServerRescueTestJSONUnderV235-749414849-project-member] [instance: 5a526856-0dcd-4b3f-8359-13a5c3b9bfc4] Start building block device mappings for instance. {{(pid=61806) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 880.497949] env[61806]: DEBUG nova.network.neutron [req-c9e1b6f9-7795-4013-b214-a80af10f899d req-73e536d4-e93d-45d4-8ed4-5bbe97c39b3f service nova] [instance: 9c20d337-b8f5-4965-b707-0f87c4ab6fc4] Instance cache missing network info. {{(pid=61806) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 880.579032] env[61806]: DEBUG nova.network.neutron [req-c9e1b6f9-7795-4013-b214-a80af10f899d req-73e536d4-e93d-45d4-8ed4-5bbe97c39b3f service nova] [instance: 9c20d337-b8f5-4965-b707-0f87c4ab6fc4] Updating instance_info_cache with network_info: [] {{(pid=61806) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 880.701593] env[61806]: DEBUG nova.scheduler.client.report [None req-a7b134bb-7562-4442-8a02-daf53033761c tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] Inventory has not changed for provider a2858be1-fd22-4e08-979e-87ad25293407 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 138, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61806) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 880.805683] env[61806]: DEBUG oslo_vmware.api [-] Task: {'id': task-1294782, 'name': CreateVM_Task} progress is 25%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 880.860937] env[61806]: DEBUG oslo_vmware.api [-] Task: {'id': task-1294783, 'name': CreateVM_Task, 'duration_secs': 0.395972} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 880.861146] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: efa19aac-3ad7-42d5-a6d8-859050de6e63] Created VM on the ESX host {{(pid=61806) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 880.861848] env[61806]: DEBUG oslo_concurrency.lockutils [None req-58b4b995-fdfd-465f-8927-98666a556c7c tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 880.862034] env[61806]: DEBUG oslo_concurrency.lockutils [None req-58b4b995-fdfd-465f-8927-98666a556c7c tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Acquired lock "[datastore2] devstack-image-cache_base/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 880.862384] env[61806]: DEBUG oslo_concurrency.lockutils [None req-58b4b995-fdfd-465f-8927-98666a556c7c tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 880.862660] env[61806]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e1f1776b-3ed6-4546-812c-ed357aab0110 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 880.868602] env[61806]: DEBUG oslo_vmware.api [None req-58b4b995-fdfd-465f-8927-98666a556c7c tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Waiting for the task: (returnval){ [ 880.868602] env[61806]: value = "session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]52384abd-8c14-c468-1469-d2acfda74a59" [ 880.868602] env[61806]: _type = "Task" [ 880.868602] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 880.878244] env[61806]: DEBUG oslo_vmware.api [None req-58b4b995-fdfd-465f-8927-98666a556c7c tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Task: {'id': session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]52384abd-8c14-c468-1469-d2acfda74a59, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 880.956621] env[61806]: DEBUG oslo_vmware.api [None req-b996cb64-3154-4eef-9fcd-aa24605b9020 tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] Task: {'id': task-1294784, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 881.082126] env[61806]: DEBUG oslo_concurrency.lockutils [req-c9e1b6f9-7795-4013-b214-a80af10f899d req-73e536d4-e93d-45d4-8ed4-5bbe97c39b3f service nova] Releasing lock "refresh_cache-9c20d337-b8f5-4965-b707-0f87c4ab6fc4" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 881.082126] env[61806]: DEBUG oslo_concurrency.lockutils [None req-b85317d7-e79b-4ab3-83e4-2de701507449 tempest-ServerTagsTestJSON-1292926242 tempest-ServerTagsTestJSON-1292926242-project-member] Acquired lock "refresh_cache-9c20d337-b8f5-4965-b707-0f87c4ab6fc4" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 881.082126] env[61806]: DEBUG nova.network.neutron [None req-b85317d7-e79b-4ab3-83e4-2de701507449 tempest-ServerTagsTestJSON-1292926242 tempest-ServerTagsTestJSON-1292926242-project-member] [instance: 9c20d337-b8f5-4965-b707-0f87c4ab6fc4] Building network info cache for instance {{(pid=61806) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 881.209019] env[61806]: DEBUG oslo_concurrency.lockutils [None req-a7b134bb-7562-4442-8a02-daf53033761c tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.728s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 881.209180] env[61806]: DEBUG oslo_concurrency.lockutils [None req-7e96f1c1-fc04-4010-b728-951cd26dd38b tempest-ImagesTestJSON-618852556 tempest-ImagesTestJSON-618852556-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 7.883s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 881.211492] env[61806]: INFO nova.compute.claims [None req-7e96f1c1-fc04-4010-b728-951cd26dd38b tempest-ImagesTestJSON-618852556 tempest-ImagesTestJSON-618852556-project-member] [instance: 1a73d994-5f21-4f80-8f33-5830d860a7ee] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 881.235208] env[61806]: INFO nova.scheduler.client.report [None req-a7b134bb-7562-4442-8a02-daf53033761c tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] Deleted allocations for instance 2a3f81e3-b84f-4370-a598-2f6a607c3d47 [ 881.305112] env[61806]: DEBUG oslo_vmware.api [-] Task: {'id': task-1294782, 'name': CreateVM_Task} progress is 99%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 881.380323] env[61806]: DEBUG oslo_vmware.api [None req-58b4b995-fdfd-465f-8927-98666a556c7c tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Task: {'id': session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]52384abd-8c14-c468-1469-d2acfda74a59, 'name': SearchDatastore_Task, 'duration_secs': 0.014637} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 881.380847] env[61806]: DEBUG oslo_concurrency.lockutils [None req-58b4b995-fdfd-465f-8927-98666a556c7c tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Releasing lock "[datastore2] devstack-image-cache_base/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 881.381301] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-58b4b995-fdfd-465f-8927-98666a556c7c tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] [instance: efa19aac-3ad7-42d5-a6d8-859050de6e63] Processing image ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa {{(pid=61806) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 881.381664] env[61806]: DEBUG oslo_concurrency.lockutils [None req-58b4b995-fdfd-465f-8927-98666a556c7c tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa.vmdk" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 881.381960] env[61806]: DEBUG oslo_concurrency.lockutils [None req-58b4b995-fdfd-465f-8927-98666a556c7c tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Acquired lock "[datastore2] devstack-image-cache_base/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa.vmdk" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 881.382285] env[61806]: DEBUG nova.virt.vmwareapi.ds_util [None req-58b4b995-fdfd-465f-8927-98666a556c7c tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61806) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 881.382878] env[61806]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-be013aaa-c977-48e1-b02b-66151dffe417 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 881.395022] env[61806]: DEBUG nova.virt.vmwareapi.ds_util [None req-58b4b995-fdfd-465f-8927-98666a556c7c tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61806) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 881.395243] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-58b4b995-fdfd-465f-8927-98666a556c7c tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61806) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 881.396313] env[61806]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-deb49313-4978-4ff5-8212-465ab68d455f {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 881.402714] env[61806]: DEBUG oslo_vmware.api [None req-58b4b995-fdfd-465f-8927-98666a556c7c tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Waiting for the task: (returnval){ [ 881.402714] env[61806]: value = "session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]5289ca80-53df-ac7c-994e-43f43361cdd4" [ 881.402714] env[61806]: _type = "Task" [ 881.402714] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 881.411630] env[61806]: DEBUG oslo_vmware.api [None req-58b4b995-fdfd-465f-8927-98666a556c7c tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Task: {'id': session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]5289ca80-53df-ac7c-994e-43f43361cdd4, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 881.457114] env[61806]: DEBUG oslo_vmware.api [None req-b996cb64-3154-4eef-9fcd-aa24605b9020 tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] Task: {'id': task-1294784, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.645696} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 881.457432] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [None req-b996cb64-3154-4eef-9fcd-aa24605b9020 tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa.vmdk to [datastore1] d1802da4-a670-427d-84d7-0fcb4717e18f/d1802da4-a670-427d-84d7-0fcb4717e18f.vmdk {{(pid=61806) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 881.457675] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-b996cb64-3154-4eef-9fcd-aa24605b9020 tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] [instance: d1802da4-a670-427d-84d7-0fcb4717e18f] Extending root virtual disk to 1048576 {{(pid=61806) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 881.457940] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-98d9e667-0bad-48e2-a33d-89e0a3d17ac6 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 881.465283] env[61806]: DEBUG oslo_vmware.api [None req-b996cb64-3154-4eef-9fcd-aa24605b9020 tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] Waiting for the task: (returnval){ [ 881.465283] env[61806]: value = "task-1294785" [ 881.465283] env[61806]: _type = "Task" [ 881.465283] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 881.475254] env[61806]: DEBUG oslo_vmware.api [None req-b996cb64-3154-4eef-9fcd-aa24605b9020 tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] Task: {'id': task-1294785, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 881.497690] env[61806]: DEBUG nova.compute.manager [None req-c651ea58-49e6-4c69-a5f7-88000b663e10 tempest-ServerRescueTestJSONUnderV235-749414849 tempest-ServerRescueTestJSONUnderV235-749414849-project-member] [instance: 5a526856-0dcd-4b3f-8359-13a5c3b9bfc4] Start spawning the instance on the hypervisor. {{(pid=61806) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 881.524878] env[61806]: DEBUG nova.virt.hardware [None req-c651ea58-49e6-4c69-a5f7-88000b663e10 tempest-ServerRescueTestJSONUnderV235-749414849 tempest-ServerRescueTestJSONUnderV235-749414849-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-15T18:05:10Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-15T18:04:52Z,direct_url=,disk_format='vmdk',id=ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='84ba83fce2c349a6988173c5d6fc3b07',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-15T18:04:53Z,virtual_size=,visibility=), allow threads: False {{(pid=61806) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 881.525152] env[61806]: DEBUG nova.virt.hardware [None req-c651ea58-49e6-4c69-a5f7-88000b663e10 tempest-ServerRescueTestJSONUnderV235-749414849 tempest-ServerRescueTestJSONUnderV235-749414849-project-member] Flavor limits 0:0:0 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 881.525319] env[61806]: DEBUG nova.virt.hardware [None req-c651ea58-49e6-4c69-a5f7-88000b663e10 tempest-ServerRescueTestJSONUnderV235-749414849 tempest-ServerRescueTestJSONUnderV235-749414849-project-member] Image limits 0:0:0 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 881.525511] env[61806]: DEBUG nova.virt.hardware [None req-c651ea58-49e6-4c69-a5f7-88000b663e10 tempest-ServerRescueTestJSONUnderV235-749414849 tempest-ServerRescueTestJSONUnderV235-749414849-project-member] Flavor pref 0:0:0 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 881.525667] env[61806]: DEBUG nova.virt.hardware [None req-c651ea58-49e6-4c69-a5f7-88000b663e10 tempest-ServerRescueTestJSONUnderV235-749414849 tempest-ServerRescueTestJSONUnderV235-749414849-project-member] Image pref 0:0:0 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 881.525819] env[61806]: DEBUG nova.virt.hardware [None req-c651ea58-49e6-4c69-a5f7-88000b663e10 tempest-ServerRescueTestJSONUnderV235-749414849 tempest-ServerRescueTestJSONUnderV235-749414849-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 881.526102] env[61806]: DEBUG nova.virt.hardware [None req-c651ea58-49e6-4c69-a5f7-88000b663e10 tempest-ServerRescueTestJSONUnderV235-749414849 tempest-ServerRescueTestJSONUnderV235-749414849-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61806) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 881.526346] env[61806]: DEBUG nova.virt.hardware [None req-c651ea58-49e6-4c69-a5f7-88000b663e10 tempest-ServerRescueTestJSONUnderV235-749414849 tempest-ServerRescueTestJSONUnderV235-749414849-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61806) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 881.526572] env[61806]: DEBUG nova.virt.hardware [None req-c651ea58-49e6-4c69-a5f7-88000b663e10 tempest-ServerRescueTestJSONUnderV235-749414849 tempest-ServerRescueTestJSONUnderV235-749414849-project-member] Got 1 possible topologies {{(pid=61806) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 881.526781] env[61806]: DEBUG nova.virt.hardware [None req-c651ea58-49e6-4c69-a5f7-88000b663e10 tempest-ServerRescueTestJSONUnderV235-749414849 tempest-ServerRescueTestJSONUnderV235-749414849-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61806) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 881.527009] env[61806]: DEBUG nova.virt.hardware [None req-c651ea58-49e6-4c69-a5f7-88000b663e10 tempest-ServerRescueTestJSONUnderV235-749414849 tempest-ServerRescueTestJSONUnderV235-749414849-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61806) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 881.527906] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-beb3adb3-46cc-4346-b186-6280d1446a5f {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 881.536665] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3f980b9a-ff68-474f-969b-196309e9ccf4 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 881.618866] env[61806]: DEBUG nova.network.neutron [None req-b85317d7-e79b-4ab3-83e4-2de701507449 tempest-ServerTagsTestJSON-1292926242 tempest-ServerTagsTestJSON-1292926242-project-member] [instance: 9c20d337-b8f5-4965-b707-0f87c4ab6fc4] Instance cache missing network info. {{(pid=61806) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 881.743547] env[61806]: DEBUG oslo_concurrency.lockutils [None req-a7b134bb-7562-4442-8a02-daf53033761c tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] Lock "2a3f81e3-b84f-4370-a598-2f6a607c3d47" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 13.479s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 881.785769] env[61806]: DEBUG nova.network.neutron [None req-b85317d7-e79b-4ab3-83e4-2de701507449 tempest-ServerTagsTestJSON-1292926242 tempest-ServerTagsTestJSON-1292926242-project-member] [instance: 9c20d337-b8f5-4965-b707-0f87c4ab6fc4] Updating instance_info_cache with network_info: [{"id": "a7f1c67d-ce27-436f-9329-4a3933beeb5c", "address": "fa:16:3e:a5:59:92", "network": {"id": "761dc50a-7ca4-4ff1-8d14-699d8459e809", "bridge": "br-int", "label": "tempest-ServerTagsTestJSON-1083705816-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "6d7669e9f6a643df9a5e9fd016dcbb72", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "29945b49-d2b8-449d-9531-437917f49839", "external-id": "nsx-vlan-transportzone-787", "segmentation_id": 787, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa7f1c67d-ce", "ovs_interfaceid": "a7f1c67d-ce27-436f-9329-4a3933beeb5c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61806) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 881.805853] env[61806]: DEBUG oslo_vmware.api [-] Task: {'id': task-1294782, 'name': CreateVM_Task, 'duration_secs': 1.120759} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 881.806038] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 59c22c16-3d1c-4e64-bfac-7f8886a1927e] Created VM on the ESX host {{(pid=61806) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 881.806727] env[61806]: DEBUG oslo_concurrency.lockutils [None req-799f40df-a27b-4200-803a-166b5c1e29fd tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 881.806915] env[61806]: DEBUG oslo_concurrency.lockutils [None req-799f40df-a27b-4200-803a-166b5c1e29fd tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Acquired lock "[datastore1] devstack-image-cache_base/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 881.807325] env[61806]: DEBUG oslo_concurrency.lockutils [None req-799f40df-a27b-4200-803a-166b5c1e29fd tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 881.810766] env[61806]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-1ad822bb-3617-47b9-a277-74e0c4ec1a0c {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 881.813390] env[61806]: DEBUG oslo_vmware.api [None req-799f40df-a27b-4200-803a-166b5c1e29fd tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Waiting for the task: (returnval){ [ 881.813390] env[61806]: value = "session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]525a403a-fb48-070e-56e7-b5712415f24f" [ 881.813390] env[61806]: _type = "Task" [ 881.813390] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 881.822420] env[61806]: DEBUG oslo_vmware.api [None req-799f40df-a27b-4200-803a-166b5c1e29fd tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Task: {'id': session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]525a403a-fb48-070e-56e7-b5712415f24f, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 881.888719] env[61806]: DEBUG nova.compute.manager [req-21f7c4cc-0266-4ecc-9753-66573a766dcd req-726ae928-5761-45e5-9992-e7cdc11dd595 service nova] [instance: 5a526856-0dcd-4b3f-8359-13a5c3b9bfc4] Received event network-vif-plugged-58782499-bc1a-4be1-9d48-655c6b2f11d7 {{(pid=61806) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 881.888982] env[61806]: DEBUG oslo_concurrency.lockutils [req-21f7c4cc-0266-4ecc-9753-66573a766dcd req-726ae928-5761-45e5-9992-e7cdc11dd595 service nova] Acquiring lock "5a526856-0dcd-4b3f-8359-13a5c3b9bfc4-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 881.889181] env[61806]: DEBUG oslo_concurrency.lockutils [req-21f7c4cc-0266-4ecc-9753-66573a766dcd req-726ae928-5761-45e5-9992-e7cdc11dd595 service nova] Lock "5a526856-0dcd-4b3f-8359-13a5c3b9bfc4-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 881.889354] env[61806]: DEBUG oslo_concurrency.lockutils [req-21f7c4cc-0266-4ecc-9753-66573a766dcd req-726ae928-5761-45e5-9992-e7cdc11dd595 service nova] Lock "5a526856-0dcd-4b3f-8359-13a5c3b9bfc4-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 881.889524] env[61806]: DEBUG nova.compute.manager [req-21f7c4cc-0266-4ecc-9753-66573a766dcd req-726ae928-5761-45e5-9992-e7cdc11dd595 service nova] [instance: 5a526856-0dcd-4b3f-8359-13a5c3b9bfc4] No waiting events found dispatching network-vif-plugged-58782499-bc1a-4be1-9d48-655c6b2f11d7 {{(pid=61806) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 881.889695] env[61806]: WARNING nova.compute.manager [req-21f7c4cc-0266-4ecc-9753-66573a766dcd req-726ae928-5761-45e5-9992-e7cdc11dd595 service nova] [instance: 5a526856-0dcd-4b3f-8359-13a5c3b9bfc4] Received unexpected event network-vif-plugged-58782499-bc1a-4be1-9d48-655c6b2f11d7 for instance with vm_state building and task_state spawning. [ 881.915378] env[61806]: DEBUG oslo_vmware.api [None req-58b4b995-fdfd-465f-8927-98666a556c7c tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Task: {'id': session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]5289ca80-53df-ac7c-994e-43f43361cdd4, 'name': SearchDatastore_Task, 'duration_secs': 0.013362} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 881.916133] env[61806]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-655bb21d-92c1-404b-9e36-173dc7e42639 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 881.921438] env[61806]: DEBUG oslo_vmware.api [None req-58b4b995-fdfd-465f-8927-98666a556c7c tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Waiting for the task: (returnval){ [ 881.921438] env[61806]: value = "session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]52c09167-8d82-3c7f-9e63-09d6aee462c3" [ 881.921438] env[61806]: _type = "Task" [ 881.921438] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 881.929242] env[61806]: DEBUG oslo_vmware.api [None req-58b4b995-fdfd-465f-8927-98666a556c7c tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Task: {'id': session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]52c09167-8d82-3c7f-9e63-09d6aee462c3, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 881.952866] env[61806]: DEBUG nova.network.neutron [None req-c651ea58-49e6-4c69-a5f7-88000b663e10 tempest-ServerRescueTestJSONUnderV235-749414849 tempest-ServerRescueTestJSONUnderV235-749414849-project-member] [instance: 5a526856-0dcd-4b3f-8359-13a5c3b9bfc4] Successfully updated port: 58782499-bc1a-4be1-9d48-655c6b2f11d7 {{(pid=61806) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 881.976199] env[61806]: DEBUG oslo_vmware.api [None req-b996cb64-3154-4eef-9fcd-aa24605b9020 tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] Task: {'id': task-1294785, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.071078} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 881.976486] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-b996cb64-3154-4eef-9fcd-aa24605b9020 tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] [instance: d1802da4-a670-427d-84d7-0fcb4717e18f] Extended root virtual disk {{(pid=61806) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 881.977654] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-44d779c8-e423-49df-a3e4-0208701f913f {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 882.001810] env[61806]: DEBUG nova.virt.vmwareapi.volumeops [None req-b996cb64-3154-4eef-9fcd-aa24605b9020 tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] [instance: d1802da4-a670-427d-84d7-0fcb4717e18f] Reconfiguring VM instance instance-0000004a to attach disk [datastore1] d1802da4-a670-427d-84d7-0fcb4717e18f/d1802da4-a670-427d-84d7-0fcb4717e18f.vmdk or device None with type sparse {{(pid=61806) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 882.002193] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-4cb9f470-43d9-4a43-9045-39a1771575a8 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 882.023783] env[61806]: DEBUG oslo_vmware.api [None req-b996cb64-3154-4eef-9fcd-aa24605b9020 tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] Waiting for the task: (returnval){ [ 882.023783] env[61806]: value = "task-1294786" [ 882.023783] env[61806]: _type = "Task" [ 882.023783] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 882.031784] env[61806]: DEBUG oslo_vmware.api [None req-b996cb64-3154-4eef-9fcd-aa24605b9020 tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] Task: {'id': task-1294786, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 882.289585] env[61806]: DEBUG oslo_concurrency.lockutils [None req-b85317d7-e79b-4ab3-83e4-2de701507449 tempest-ServerTagsTestJSON-1292926242 tempest-ServerTagsTestJSON-1292926242-project-member] Releasing lock "refresh_cache-9c20d337-b8f5-4965-b707-0f87c4ab6fc4" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 882.289951] env[61806]: DEBUG nova.compute.manager [None req-b85317d7-e79b-4ab3-83e4-2de701507449 tempest-ServerTagsTestJSON-1292926242 tempest-ServerTagsTestJSON-1292926242-project-member] [instance: 9c20d337-b8f5-4965-b707-0f87c4ab6fc4] Instance network_info: |[{"id": "a7f1c67d-ce27-436f-9329-4a3933beeb5c", "address": "fa:16:3e:a5:59:92", "network": {"id": "761dc50a-7ca4-4ff1-8d14-699d8459e809", "bridge": "br-int", "label": "tempest-ServerTagsTestJSON-1083705816-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "6d7669e9f6a643df9a5e9fd016dcbb72", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "29945b49-d2b8-449d-9531-437917f49839", "external-id": "nsx-vlan-transportzone-787", "segmentation_id": 787, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa7f1c67d-ce", "ovs_interfaceid": "a7f1c67d-ce27-436f-9329-4a3933beeb5c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61806) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 882.293065] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-b85317d7-e79b-4ab3-83e4-2de701507449 tempest-ServerTagsTestJSON-1292926242 tempest-ServerTagsTestJSON-1292926242-project-member] [instance: 9c20d337-b8f5-4965-b707-0f87c4ab6fc4] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:a5:59:92', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '29945b49-d2b8-449d-9531-437917f49839', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'a7f1c67d-ce27-436f-9329-4a3933beeb5c', 'vif_model': 'vmxnet3'}] {{(pid=61806) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 882.300680] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [None req-b85317d7-e79b-4ab3-83e4-2de701507449 tempest-ServerTagsTestJSON-1292926242 tempest-ServerTagsTestJSON-1292926242-project-member] Creating folder: Project (6d7669e9f6a643df9a5e9fd016dcbb72). Parent ref: group-v277609. {{(pid=61806) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 882.301265] env[61806]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-8d8bfc15-e848-49d6-beeb-5bf9b3907632 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 882.315267] env[61806]: INFO nova.virt.vmwareapi.vm_util [None req-b85317d7-e79b-4ab3-83e4-2de701507449 tempest-ServerTagsTestJSON-1292926242 tempest-ServerTagsTestJSON-1292926242-project-member] Created folder: Project (6d7669e9f6a643df9a5e9fd016dcbb72) in parent group-v277609. [ 882.315492] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [None req-b85317d7-e79b-4ab3-83e4-2de701507449 tempest-ServerTagsTestJSON-1292926242 tempest-ServerTagsTestJSON-1292926242-project-member] Creating folder: Instances. Parent ref: group-v277695. {{(pid=61806) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 882.319119] env[61806]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-98258f1b-c0ef-4616-8c78-7793781b33db {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 882.329854] env[61806]: DEBUG oslo_vmware.api [None req-799f40df-a27b-4200-803a-166b5c1e29fd tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Task: {'id': session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]525a403a-fb48-070e-56e7-b5712415f24f, 'name': SearchDatastore_Task, 'duration_secs': 0.011849} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 882.332238] env[61806]: DEBUG oslo_concurrency.lockutils [None req-799f40df-a27b-4200-803a-166b5c1e29fd tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Releasing lock "[datastore1] devstack-image-cache_base/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 882.332238] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-799f40df-a27b-4200-803a-166b5c1e29fd tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] [instance: 59c22c16-3d1c-4e64-bfac-7f8886a1927e] Processing image ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa {{(pid=61806) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 882.332238] env[61806]: DEBUG oslo_concurrency.lockutils [None req-799f40df-a27b-4200-803a-166b5c1e29fd tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa.vmdk" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 882.332238] env[61806]: DEBUG oslo_concurrency.lockutils [None req-799f40df-a27b-4200-803a-166b5c1e29fd tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Acquired lock "[datastore1] devstack-image-cache_base/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa.vmdk" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 882.332238] env[61806]: DEBUG nova.virt.vmwareapi.ds_util [None req-799f40df-a27b-4200-803a-166b5c1e29fd tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61806) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 882.332238] env[61806]: INFO nova.virt.vmwareapi.vm_util [None req-b85317d7-e79b-4ab3-83e4-2de701507449 tempest-ServerTagsTestJSON-1292926242 tempest-ServerTagsTestJSON-1292926242-project-member] Created folder: Instances in parent group-v277695. [ 882.332238] env[61806]: DEBUG oslo.service.loopingcall [None req-b85317d7-e79b-4ab3-83e4-2de701507449 tempest-ServerTagsTestJSON-1292926242 tempest-ServerTagsTestJSON-1292926242-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61806) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 882.332571] env[61806]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-7331c46d-2a12-4209-9ecf-fdf3dddd91df {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 882.334137] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 9c20d337-b8f5-4965-b707-0f87c4ab6fc4] Creating VM on the ESX host {{(pid=61806) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 882.335126] env[61806]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-48e16d8a-0af4-4466-81ae-523852563b18 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 882.357259] env[61806]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 882.357259] env[61806]: value = "task-1294789" [ 882.357259] env[61806]: _type = "Task" [ 882.357259] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 882.361466] env[61806]: DEBUG nova.virt.vmwareapi.ds_util [None req-799f40df-a27b-4200-803a-166b5c1e29fd tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61806) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 882.361644] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-799f40df-a27b-4200-803a-166b5c1e29fd tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61806) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 882.365018] env[61806]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-52a98004-469e-48e4-ae17-ac2fbf7b2717 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 882.374400] env[61806]: DEBUG oslo_vmware.api [-] Task: {'id': task-1294789, 'name': CreateVM_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 882.374713] env[61806]: DEBUG oslo_vmware.api [None req-799f40df-a27b-4200-803a-166b5c1e29fd tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Waiting for the task: (returnval){ [ 882.374713] env[61806]: value = "session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]520ba2cf-7693-b305-db38-306f80f00f3c" [ 882.374713] env[61806]: _type = "Task" [ 882.374713] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 882.385775] env[61806]: DEBUG oslo_vmware.api [None req-799f40df-a27b-4200-803a-166b5c1e29fd tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Task: {'id': session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]520ba2cf-7693-b305-db38-306f80f00f3c, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 882.396381] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ec7be512-7492-4114-98d4-c275bc4e774c {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 882.410157] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-67ee867b-776b-42f2-9e85-5be00f3bb99d {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 882.449509] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e74fb0eb-17b8-4bf3-be32-81715604fa3b {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 882.455159] env[61806]: DEBUG oslo_concurrency.lockutils [None req-c651ea58-49e6-4c69-a5f7-88000b663e10 tempest-ServerRescueTestJSONUnderV235-749414849 tempest-ServerRescueTestJSONUnderV235-749414849-project-member] Acquiring lock "refresh_cache-5a526856-0dcd-4b3f-8359-13a5c3b9bfc4" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 882.455338] env[61806]: DEBUG oslo_concurrency.lockutils [None req-c651ea58-49e6-4c69-a5f7-88000b663e10 tempest-ServerRescueTestJSONUnderV235-749414849 tempest-ServerRescueTestJSONUnderV235-749414849-project-member] Acquired lock "refresh_cache-5a526856-0dcd-4b3f-8359-13a5c3b9bfc4" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 882.455455] env[61806]: DEBUG nova.network.neutron [None req-c651ea58-49e6-4c69-a5f7-88000b663e10 tempest-ServerRescueTestJSONUnderV235-749414849 tempest-ServerRescueTestJSONUnderV235-749414849-project-member] [instance: 5a526856-0dcd-4b3f-8359-13a5c3b9bfc4] Building network info cache for instance {{(pid=61806) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 882.462512] env[61806]: DEBUG oslo_vmware.api [None req-58b4b995-fdfd-465f-8927-98666a556c7c tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Task: {'id': session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]52c09167-8d82-3c7f-9e63-09d6aee462c3, 'name': SearchDatastore_Task, 'duration_secs': 0.011729} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 882.463834] env[61806]: DEBUG oslo_concurrency.lockutils [None req-58b4b995-fdfd-465f-8927-98666a556c7c tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Releasing lock "[datastore2] devstack-image-cache_base/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa.vmdk" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 882.463834] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [None req-58b4b995-fdfd-465f-8927-98666a556c7c tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa.vmdk to [datastore2] efa19aac-3ad7-42d5-a6d8-859050de6e63/efa19aac-3ad7-42d5-a6d8-859050de6e63.vmdk {{(pid=61806) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 882.464947] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e3f548c9-bb01-41fb-8433-9a98ad7aba89 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 882.469449] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-c8c2056f-7a9e-477b-b73e-bdf5b13b20e0 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 882.485317] env[61806]: DEBUG nova.compute.provider_tree [None req-7e96f1c1-fc04-4010-b728-951cd26dd38b tempest-ImagesTestJSON-618852556 tempest-ImagesTestJSON-618852556-project-member] Inventory has not changed in ProviderTree for provider: a2858be1-fd22-4e08-979e-87ad25293407 {{(pid=61806) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 882.487859] env[61806]: DEBUG oslo_vmware.api [None req-58b4b995-fdfd-465f-8927-98666a556c7c tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Waiting for the task: (returnval){ [ 882.487859] env[61806]: value = "task-1294790" [ 882.487859] env[61806]: _type = "Task" [ 882.487859] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 882.501352] env[61806]: DEBUG oslo_vmware.api [None req-58b4b995-fdfd-465f-8927-98666a556c7c tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Task: {'id': task-1294790, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 882.535237] env[61806]: DEBUG oslo_vmware.api [None req-b996cb64-3154-4eef-9fcd-aa24605b9020 tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] Task: {'id': task-1294786, 'name': ReconfigVM_Task, 'duration_secs': 0.29495} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 882.535595] env[61806]: DEBUG nova.virt.vmwareapi.volumeops [None req-b996cb64-3154-4eef-9fcd-aa24605b9020 tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] [instance: d1802da4-a670-427d-84d7-0fcb4717e18f] Reconfigured VM instance instance-0000004a to attach disk [datastore1] d1802da4-a670-427d-84d7-0fcb4717e18f/d1802da4-a670-427d-84d7-0fcb4717e18f.vmdk or device None with type sparse {{(pid=61806) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 882.536246] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-afe8e55c-9df5-4345-85ef-490f80d895d4 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 882.544261] env[61806]: DEBUG oslo_vmware.api [None req-b996cb64-3154-4eef-9fcd-aa24605b9020 tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] Waiting for the task: (returnval){ [ 882.544261] env[61806]: value = "task-1294791" [ 882.544261] env[61806]: _type = "Task" [ 882.544261] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 882.554759] env[61806]: DEBUG oslo_vmware.api [None req-b996cb64-3154-4eef-9fcd-aa24605b9020 tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] Task: {'id': task-1294791, 'name': Rename_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 882.872032] env[61806]: DEBUG oslo_vmware.api [-] Task: {'id': task-1294789, 'name': CreateVM_Task, 'duration_secs': 0.388982} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 882.872032] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 9c20d337-b8f5-4965-b707-0f87c4ab6fc4] Created VM on the ESX host {{(pid=61806) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 882.872032] env[61806]: DEBUG oslo_concurrency.lockutils [None req-b85317d7-e79b-4ab3-83e4-2de701507449 tempest-ServerTagsTestJSON-1292926242 tempest-ServerTagsTestJSON-1292926242-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 882.872032] env[61806]: DEBUG oslo_concurrency.lockutils [None req-b85317d7-e79b-4ab3-83e4-2de701507449 tempest-ServerTagsTestJSON-1292926242 tempest-ServerTagsTestJSON-1292926242-project-member] Acquired lock "[datastore2] devstack-image-cache_base/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 882.872032] env[61806]: DEBUG oslo_concurrency.lockutils [None req-b85317d7-e79b-4ab3-83e4-2de701507449 tempest-ServerTagsTestJSON-1292926242 tempest-ServerTagsTestJSON-1292926242-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 882.872032] env[61806]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-9193a28d-ecba-4cd8-9fda-9d7fa54b0e15 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 882.880759] env[61806]: DEBUG oslo_vmware.api [None req-b85317d7-e79b-4ab3-83e4-2de701507449 tempest-ServerTagsTestJSON-1292926242 tempest-ServerTagsTestJSON-1292926242-project-member] Waiting for the task: (returnval){ [ 882.880759] env[61806]: value = "session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]52685583-04f6-be5a-feae-ddbee5712345" [ 882.880759] env[61806]: _type = "Task" [ 882.880759] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 882.891315] env[61806]: DEBUG oslo_vmware.api [None req-799f40df-a27b-4200-803a-166b5c1e29fd tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Task: {'id': session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]520ba2cf-7693-b305-db38-306f80f00f3c, 'name': SearchDatastore_Task, 'duration_secs': 0.012778} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 882.897135] env[61806]: DEBUG oslo_vmware.api [None req-b85317d7-e79b-4ab3-83e4-2de701507449 tempest-ServerTagsTestJSON-1292926242 tempest-ServerTagsTestJSON-1292926242-project-member] Task: {'id': session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]52685583-04f6-be5a-feae-ddbee5712345, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 882.897135] env[61806]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-8d183c0e-6298-4bbd-a42f-b2ca72b294b0 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 882.904170] env[61806]: DEBUG oslo_vmware.api [None req-799f40df-a27b-4200-803a-166b5c1e29fd tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Waiting for the task: (returnval){ [ 882.904170] env[61806]: value = "session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]523ec540-fe90-8888-2ae2-6eef992b091b" [ 882.904170] env[61806]: _type = "Task" [ 882.904170] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 882.914663] env[61806]: DEBUG oslo_vmware.api [None req-799f40df-a27b-4200-803a-166b5c1e29fd tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Task: {'id': session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]523ec540-fe90-8888-2ae2-6eef992b091b, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 882.992084] env[61806]: DEBUG nova.scheduler.client.report [None req-7e96f1c1-fc04-4010-b728-951cd26dd38b tempest-ImagesTestJSON-618852556 tempest-ImagesTestJSON-618852556-project-member] Inventory has not changed for provider a2858be1-fd22-4e08-979e-87ad25293407 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 138, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61806) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 883.017396] env[61806]: DEBUG oslo_vmware.api [None req-58b4b995-fdfd-465f-8927-98666a556c7c tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Task: {'id': task-1294790, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 883.018442] env[61806]: DEBUG nova.network.neutron [None req-c651ea58-49e6-4c69-a5f7-88000b663e10 tempest-ServerRescueTestJSONUnderV235-749414849 tempest-ServerRescueTestJSONUnderV235-749414849-project-member] [instance: 5a526856-0dcd-4b3f-8359-13a5c3b9bfc4] Instance cache missing network info. {{(pid=61806) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 883.054977] env[61806]: DEBUG oslo_vmware.api [None req-b996cb64-3154-4eef-9fcd-aa24605b9020 tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] Task: {'id': task-1294791, 'name': Rename_Task, 'duration_secs': 0.207725} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 883.055292] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [None req-b996cb64-3154-4eef-9fcd-aa24605b9020 tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] [instance: d1802da4-a670-427d-84d7-0fcb4717e18f] Powering on the VM {{(pid=61806) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 883.055671] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-776a0bd1-b99a-44f2-bb20-b93cf9de7dfd {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 883.062741] env[61806]: DEBUG oslo_vmware.api [None req-b996cb64-3154-4eef-9fcd-aa24605b9020 tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] Waiting for the task: (returnval){ [ 883.062741] env[61806]: value = "task-1294792" [ 883.062741] env[61806]: _type = "Task" [ 883.062741] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 883.076919] env[61806]: DEBUG oslo_vmware.api [None req-b996cb64-3154-4eef-9fcd-aa24605b9020 tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] Task: {'id': task-1294792, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 883.149427] env[61806]: DEBUG nova.network.neutron [None req-c651ea58-49e6-4c69-a5f7-88000b663e10 tempest-ServerRescueTestJSONUnderV235-749414849 tempest-ServerRescueTestJSONUnderV235-749414849-project-member] [instance: 5a526856-0dcd-4b3f-8359-13a5c3b9bfc4] Updating instance_info_cache with network_info: [{"id": "58782499-bc1a-4be1-9d48-655c6b2f11d7", "address": "fa:16:3e:52:7a:2b", "network": {"id": "b4fa0830-265e-47bf-8029-5944534312e7", "bridge": "br-int", "label": "tempest-ServerRescueTestJSONUnderV235-1460638664-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "2ba59612579b47a4bde7f82d951ac3ff", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8a34aa30-95be-4b18-98ca-1f2d81f7e9e6", "external-id": "nsx-vlan-transportzone-234", "segmentation_id": 234, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap58782499-bc", "ovs_interfaceid": "58782499-bc1a-4be1-9d48-655c6b2f11d7", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61806) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 883.395117] env[61806]: DEBUG oslo_vmware.api [None req-b85317d7-e79b-4ab3-83e4-2de701507449 tempest-ServerTagsTestJSON-1292926242 tempest-ServerTagsTestJSON-1292926242-project-member] Task: {'id': session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]52685583-04f6-be5a-feae-ddbee5712345, 'name': SearchDatastore_Task, 'duration_secs': 0.049236} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 883.395547] env[61806]: DEBUG oslo_concurrency.lockutils [None req-b85317d7-e79b-4ab3-83e4-2de701507449 tempest-ServerTagsTestJSON-1292926242 tempest-ServerTagsTestJSON-1292926242-project-member] Releasing lock "[datastore2] devstack-image-cache_base/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 883.395790] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-b85317d7-e79b-4ab3-83e4-2de701507449 tempest-ServerTagsTestJSON-1292926242 tempest-ServerTagsTestJSON-1292926242-project-member] [instance: 9c20d337-b8f5-4965-b707-0f87c4ab6fc4] Processing image ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa {{(pid=61806) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 883.396239] env[61806]: DEBUG oslo_concurrency.lockutils [None req-b85317d7-e79b-4ab3-83e4-2de701507449 tempest-ServerTagsTestJSON-1292926242 tempest-ServerTagsTestJSON-1292926242-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa.vmdk" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 883.396630] env[61806]: DEBUG oslo_concurrency.lockutils [None req-b85317d7-e79b-4ab3-83e4-2de701507449 tempest-ServerTagsTestJSON-1292926242 tempest-ServerTagsTestJSON-1292926242-project-member] Acquired lock "[datastore2] devstack-image-cache_base/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa.vmdk" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 883.396939] env[61806]: DEBUG nova.virt.vmwareapi.ds_util [None req-b85317d7-e79b-4ab3-83e4-2de701507449 tempest-ServerTagsTestJSON-1292926242 tempest-ServerTagsTestJSON-1292926242-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61806) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 883.397379] env[61806]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-acd168eb-43b2-476d-b6f3-0a34af4cc95f {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 883.410297] env[61806]: DEBUG nova.virt.vmwareapi.ds_util [None req-b85317d7-e79b-4ab3-83e4-2de701507449 tempest-ServerTagsTestJSON-1292926242 tempest-ServerTagsTestJSON-1292926242-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61806) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 883.410496] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-b85317d7-e79b-4ab3-83e4-2de701507449 tempest-ServerTagsTestJSON-1292926242 tempest-ServerTagsTestJSON-1292926242-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61806) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 883.411730] env[61806]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b2544dc6-379f-49b4-8a0f-230fbc74a7f5 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 883.417712] env[61806]: DEBUG oslo_vmware.api [None req-799f40df-a27b-4200-803a-166b5c1e29fd tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Task: {'id': session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]523ec540-fe90-8888-2ae2-6eef992b091b, 'name': SearchDatastore_Task, 'duration_secs': 0.022318} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 883.418627] env[61806]: DEBUG oslo_concurrency.lockutils [None req-799f40df-a27b-4200-803a-166b5c1e29fd tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Releasing lock "[datastore1] devstack-image-cache_base/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa.vmdk" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 883.419126] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [None req-799f40df-a27b-4200-803a-166b5c1e29fd tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa.vmdk to [datastore1] 59c22c16-3d1c-4e64-bfac-7f8886a1927e/59c22c16-3d1c-4e64-bfac-7f8886a1927e.vmdk {{(pid=61806) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 883.419486] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-2b10e5f4-53cc-4a50-8382-56a29a9e6a11 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 883.422625] env[61806]: DEBUG oslo_vmware.api [None req-b85317d7-e79b-4ab3-83e4-2de701507449 tempest-ServerTagsTestJSON-1292926242 tempest-ServerTagsTestJSON-1292926242-project-member] Waiting for the task: (returnval){ [ 883.422625] env[61806]: value = "session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]529e710b-d118-1c01-9923-28eb1a0349cd" [ 883.422625] env[61806]: _type = "Task" [ 883.422625] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 883.428975] env[61806]: DEBUG oslo_vmware.api [None req-799f40df-a27b-4200-803a-166b5c1e29fd tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Waiting for the task: (returnval){ [ 883.428975] env[61806]: value = "task-1294793" [ 883.428975] env[61806]: _type = "Task" [ 883.428975] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 883.436880] env[61806]: DEBUG oslo_vmware.api [None req-b85317d7-e79b-4ab3-83e4-2de701507449 tempest-ServerTagsTestJSON-1292926242 tempest-ServerTagsTestJSON-1292926242-project-member] Task: {'id': session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]529e710b-d118-1c01-9923-28eb1a0349cd, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 883.441774] env[61806]: DEBUG oslo_vmware.api [None req-799f40df-a27b-4200-803a-166b5c1e29fd tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Task: {'id': task-1294793, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 883.505688] env[61806]: DEBUG oslo_concurrency.lockutils [None req-7e96f1c1-fc04-4010-b728-951cd26dd38b tempest-ImagesTestJSON-618852556 tempest-ImagesTestJSON-618852556-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.297s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 883.506201] env[61806]: DEBUG nova.compute.manager [None req-7e96f1c1-fc04-4010-b728-951cd26dd38b tempest-ImagesTestJSON-618852556 tempest-ImagesTestJSON-618852556-project-member] [instance: 1a73d994-5f21-4f80-8f33-5830d860a7ee] Start building networks asynchronously for instance. {{(pid=61806) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 883.509199] env[61806]: DEBUG oslo_vmware.api [None req-58b4b995-fdfd-465f-8927-98666a556c7c tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Task: {'id': task-1294790, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.601912} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 883.509460] env[61806]: DEBUG oslo_concurrency.lockutils [None req-708d7cc8-bb4e-45d0-adad-8f77e0f096a2 tempest-ImagesTestJSON-618852556 tempest-ImagesTestJSON-618852556-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 8.547s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 883.509831] env[61806]: DEBUG nova.objects.instance [None req-708d7cc8-bb4e-45d0-adad-8f77e0f096a2 tempest-ImagesTestJSON-618852556 tempest-ImagesTestJSON-618852556-project-member] Lazy-loading 'resources' on Instance uuid e0ef0a35-82a5-495b-9d5c-5805e8306390 {{(pid=61806) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 883.511120] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [None req-58b4b995-fdfd-465f-8927-98666a556c7c tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa.vmdk to [datastore2] efa19aac-3ad7-42d5-a6d8-859050de6e63/efa19aac-3ad7-42d5-a6d8-859050de6e63.vmdk {{(pid=61806) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 883.511463] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-58b4b995-fdfd-465f-8927-98666a556c7c tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] [instance: efa19aac-3ad7-42d5-a6d8-859050de6e63] Extending root virtual disk to 1048576 {{(pid=61806) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 883.511825] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-f2e168e4-6f97-4102-9b9a-8f76d35eee8b {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 883.520753] env[61806]: DEBUG oslo_vmware.api [None req-58b4b995-fdfd-465f-8927-98666a556c7c tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Waiting for the task: (returnval){ [ 883.520753] env[61806]: value = "task-1294794" [ 883.520753] env[61806]: _type = "Task" [ 883.520753] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 883.533270] env[61806]: DEBUG oslo_vmware.api [None req-58b4b995-fdfd-465f-8927-98666a556c7c tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Task: {'id': task-1294794, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 883.572937] env[61806]: DEBUG oslo_vmware.api [None req-b996cb64-3154-4eef-9fcd-aa24605b9020 tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] Task: {'id': task-1294792, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 883.651821] env[61806]: DEBUG oslo_concurrency.lockutils [None req-c651ea58-49e6-4c69-a5f7-88000b663e10 tempest-ServerRescueTestJSONUnderV235-749414849 tempest-ServerRescueTestJSONUnderV235-749414849-project-member] Releasing lock "refresh_cache-5a526856-0dcd-4b3f-8359-13a5c3b9bfc4" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 883.652170] env[61806]: DEBUG nova.compute.manager [None req-c651ea58-49e6-4c69-a5f7-88000b663e10 tempest-ServerRescueTestJSONUnderV235-749414849 tempest-ServerRescueTestJSONUnderV235-749414849-project-member] [instance: 5a526856-0dcd-4b3f-8359-13a5c3b9bfc4] Instance network_info: |[{"id": "58782499-bc1a-4be1-9d48-655c6b2f11d7", "address": "fa:16:3e:52:7a:2b", "network": {"id": "b4fa0830-265e-47bf-8029-5944534312e7", "bridge": "br-int", "label": "tempest-ServerRescueTestJSONUnderV235-1460638664-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "2ba59612579b47a4bde7f82d951ac3ff", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8a34aa30-95be-4b18-98ca-1f2d81f7e9e6", "external-id": "nsx-vlan-transportzone-234", "segmentation_id": 234, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap58782499-bc", "ovs_interfaceid": "58782499-bc1a-4be1-9d48-655c6b2f11d7", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61806) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 883.652635] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-c651ea58-49e6-4c69-a5f7-88000b663e10 tempest-ServerRescueTestJSONUnderV235-749414849 tempest-ServerRescueTestJSONUnderV235-749414849-project-member] [instance: 5a526856-0dcd-4b3f-8359-13a5c3b9bfc4] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:52:7a:2b', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '8a34aa30-95be-4b18-98ca-1f2d81f7e9e6', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '58782499-bc1a-4be1-9d48-655c6b2f11d7', 'vif_model': 'vmxnet3'}] {{(pid=61806) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 883.660188] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [None req-c651ea58-49e6-4c69-a5f7-88000b663e10 tempest-ServerRescueTestJSONUnderV235-749414849 tempest-ServerRescueTestJSONUnderV235-749414849-project-member] Creating folder: Project (2ba59612579b47a4bde7f82d951ac3ff). Parent ref: group-v277609. {{(pid=61806) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 883.660482] env[61806]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-2ffcdb85-3166-4ff5-8635-969be6406ead {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 883.672525] env[61806]: INFO nova.virt.vmwareapi.vm_util [None req-c651ea58-49e6-4c69-a5f7-88000b663e10 tempest-ServerRescueTestJSONUnderV235-749414849 tempest-ServerRescueTestJSONUnderV235-749414849-project-member] Created folder: Project (2ba59612579b47a4bde7f82d951ac3ff) in parent group-v277609. [ 883.672851] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [None req-c651ea58-49e6-4c69-a5f7-88000b663e10 tempest-ServerRescueTestJSONUnderV235-749414849 tempest-ServerRescueTestJSONUnderV235-749414849-project-member] Creating folder: Instances. Parent ref: group-v277698. {{(pid=61806) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 883.673166] env[61806]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-4280e1d6-df14-4d2f-b19d-cb3554a17df7 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 883.683392] env[61806]: INFO nova.virt.vmwareapi.vm_util [None req-c651ea58-49e6-4c69-a5f7-88000b663e10 tempest-ServerRescueTestJSONUnderV235-749414849 tempest-ServerRescueTestJSONUnderV235-749414849-project-member] Created folder: Instances in parent group-v277698. [ 883.683661] env[61806]: DEBUG oslo.service.loopingcall [None req-c651ea58-49e6-4c69-a5f7-88000b663e10 tempest-ServerRescueTestJSONUnderV235-749414849 tempest-ServerRescueTestJSONUnderV235-749414849-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61806) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 883.683839] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 5a526856-0dcd-4b3f-8359-13a5c3b9bfc4] Creating VM on the ESX host {{(pid=61806) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 883.684077] env[61806]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-e0904598-8496-47e6-961f-3341a2e79c3a {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 883.703816] env[61806]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 883.703816] env[61806]: value = "task-1294797" [ 883.703816] env[61806]: _type = "Task" [ 883.703816] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 883.712296] env[61806]: DEBUG oslo_vmware.api [-] Task: {'id': task-1294797, 'name': CreateVM_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 883.913303] env[61806]: DEBUG nova.compute.manager [req-b2bbef49-94eb-4b26-ab49-02410cc3a1d4 req-e8927c34-d897-4896-81a3-ec939d0ae624 service nova] [instance: 5a526856-0dcd-4b3f-8359-13a5c3b9bfc4] Received event network-changed-58782499-bc1a-4be1-9d48-655c6b2f11d7 {{(pid=61806) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 883.913575] env[61806]: DEBUG nova.compute.manager [req-b2bbef49-94eb-4b26-ab49-02410cc3a1d4 req-e8927c34-d897-4896-81a3-ec939d0ae624 service nova] [instance: 5a526856-0dcd-4b3f-8359-13a5c3b9bfc4] Refreshing instance network info cache due to event network-changed-58782499-bc1a-4be1-9d48-655c6b2f11d7. {{(pid=61806) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 883.913760] env[61806]: DEBUG oslo_concurrency.lockutils [req-b2bbef49-94eb-4b26-ab49-02410cc3a1d4 req-e8927c34-d897-4896-81a3-ec939d0ae624 service nova] Acquiring lock "refresh_cache-5a526856-0dcd-4b3f-8359-13a5c3b9bfc4" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 883.913919] env[61806]: DEBUG oslo_concurrency.lockutils [req-b2bbef49-94eb-4b26-ab49-02410cc3a1d4 req-e8927c34-d897-4896-81a3-ec939d0ae624 service nova] Acquired lock "refresh_cache-5a526856-0dcd-4b3f-8359-13a5c3b9bfc4" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 883.914098] env[61806]: DEBUG nova.network.neutron [req-b2bbef49-94eb-4b26-ab49-02410cc3a1d4 req-e8927c34-d897-4896-81a3-ec939d0ae624 service nova] [instance: 5a526856-0dcd-4b3f-8359-13a5c3b9bfc4] Refreshing network info cache for port 58782499-bc1a-4be1-9d48-655c6b2f11d7 {{(pid=61806) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 883.937813] env[61806]: DEBUG oslo_vmware.api [None req-b85317d7-e79b-4ab3-83e4-2de701507449 tempest-ServerTagsTestJSON-1292926242 tempest-ServerTagsTestJSON-1292926242-project-member] Task: {'id': session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]529e710b-d118-1c01-9923-28eb1a0349cd, 'name': SearchDatastore_Task, 'duration_secs': 0.013007} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 883.939013] env[61806]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-72df2464-492e-49fd-8a42-0addf0ffc1c4 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 883.944423] env[61806]: DEBUG oslo_vmware.api [None req-799f40df-a27b-4200-803a-166b5c1e29fd tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Task: {'id': task-1294793, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 883.948098] env[61806]: DEBUG oslo_vmware.api [None req-b85317d7-e79b-4ab3-83e4-2de701507449 tempest-ServerTagsTestJSON-1292926242 tempest-ServerTagsTestJSON-1292926242-project-member] Waiting for the task: (returnval){ [ 883.948098] env[61806]: value = "session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]52710191-e6a4-bb3a-a6a2-15076e6ddc64" [ 883.948098] env[61806]: _type = "Task" [ 883.948098] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 883.956582] env[61806]: DEBUG oslo_vmware.api [None req-b85317d7-e79b-4ab3-83e4-2de701507449 tempest-ServerTagsTestJSON-1292926242 tempest-ServerTagsTestJSON-1292926242-project-member] Task: {'id': session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]52710191-e6a4-bb3a-a6a2-15076e6ddc64, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 884.011222] env[61806]: DEBUG nova.compute.utils [None req-7e96f1c1-fc04-4010-b728-951cd26dd38b tempest-ImagesTestJSON-618852556 tempest-ImagesTestJSON-618852556-project-member] Using /dev/sd instead of None {{(pid=61806) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 884.013289] env[61806]: DEBUG nova.compute.manager [None req-7e96f1c1-fc04-4010-b728-951cd26dd38b tempest-ImagesTestJSON-618852556 tempest-ImagesTestJSON-618852556-project-member] [instance: 1a73d994-5f21-4f80-8f33-5830d860a7ee] Allocating IP information in the background. {{(pid=61806) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 884.013520] env[61806]: DEBUG nova.network.neutron [None req-7e96f1c1-fc04-4010-b728-951cd26dd38b tempest-ImagesTestJSON-618852556 tempest-ImagesTestJSON-618852556-project-member] [instance: 1a73d994-5f21-4f80-8f33-5830d860a7ee] allocate_for_instance() {{(pid=61806) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 884.032619] env[61806]: DEBUG oslo_vmware.api [None req-58b4b995-fdfd-465f-8927-98666a556c7c tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Task: {'id': task-1294794, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.305863} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 884.035504] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-58b4b995-fdfd-465f-8927-98666a556c7c tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] [instance: efa19aac-3ad7-42d5-a6d8-859050de6e63] Extended root virtual disk {{(pid=61806) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 884.036537] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3b634ae9-a446-48ba-beca-9f49df7815cf {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 884.059490] env[61806]: DEBUG nova.virt.vmwareapi.volumeops [None req-58b4b995-fdfd-465f-8927-98666a556c7c tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] [instance: efa19aac-3ad7-42d5-a6d8-859050de6e63] Reconfiguring VM instance instance-00000049 to attach disk [datastore2] efa19aac-3ad7-42d5-a6d8-859050de6e63/efa19aac-3ad7-42d5-a6d8-859050de6e63.vmdk or device None with type sparse {{(pid=61806) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 884.063305] env[61806]: DEBUG nova.policy [None req-7e96f1c1-fc04-4010-b728-951cd26dd38b tempest-ImagesTestJSON-618852556 tempest-ImagesTestJSON-618852556-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'e6fa11a9b63f4cd6b04baf3115431167', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '4b9aa91aff1d4008ac5096902b77f852', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61806) authorize /opt/stack/nova/nova/policy.py:201}} [ 884.064848] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-434e2e6f-dfc9-4810-b873-dd82666e2a09 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 884.087990] env[61806]: DEBUG oslo_vmware.api [None req-b996cb64-3154-4eef-9fcd-aa24605b9020 tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] Task: {'id': task-1294792, 'name': PowerOnVM_Task, 'duration_secs': 0.58566} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 884.091645] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [None req-b996cb64-3154-4eef-9fcd-aa24605b9020 tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] [instance: d1802da4-a670-427d-84d7-0fcb4717e18f] Powered on the VM {{(pid=61806) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 884.092012] env[61806]: INFO nova.compute.manager [None req-b996cb64-3154-4eef-9fcd-aa24605b9020 tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] [instance: d1802da4-a670-427d-84d7-0fcb4717e18f] Took 9.40 seconds to spawn the instance on the hypervisor. [ 884.092076] env[61806]: DEBUG nova.compute.manager [None req-b996cb64-3154-4eef-9fcd-aa24605b9020 tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] [instance: d1802da4-a670-427d-84d7-0fcb4717e18f] Checking state {{(pid=61806) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 884.092390] env[61806]: DEBUG oslo_vmware.api [None req-58b4b995-fdfd-465f-8927-98666a556c7c tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Waiting for the task: (returnval){ [ 884.092390] env[61806]: value = "task-1294798" [ 884.092390] env[61806]: _type = "Task" [ 884.092390] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 884.093383] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b4bda559-f756-46ba-ba47-df4ed848ca56 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 884.104725] env[61806]: DEBUG oslo_vmware.api [None req-58b4b995-fdfd-465f-8927-98666a556c7c tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Task: {'id': task-1294798, 'name': ReconfigVM_Task} progress is 10%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 884.219378] env[61806]: DEBUG oslo_vmware.api [-] Task: {'id': task-1294797, 'name': CreateVM_Task} progress is 25%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 884.223340] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3e35e379-7dee-4057-8947-957c74867c97 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 884.232960] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ae254d8f-e14d-4264-bb79-2d0581d005eb {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 884.270394] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c03ed0fa-1741-44e5-bd03-02f6cb582aa6 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 884.279557] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b45f9abe-7923-44ae-b7bd-8cfc99fe56a0 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 884.296803] env[61806]: DEBUG nova.compute.provider_tree [None req-708d7cc8-bb4e-45d0-adad-8f77e0f096a2 tempest-ImagesTestJSON-618852556 tempest-ImagesTestJSON-618852556-project-member] Inventory has not changed in ProviderTree for provider: a2858be1-fd22-4e08-979e-87ad25293407 {{(pid=61806) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 884.356826] env[61806]: DEBUG oslo_concurrency.lockutils [None req-095bcbe6-2af7-4761-a8ad-c73adfe5ab9d tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] Acquiring lock "6f6ba57a-e2d6-4749-a53a-e263861cb1c0" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 884.356948] env[61806]: DEBUG oslo_concurrency.lockutils [None req-095bcbe6-2af7-4761-a8ad-c73adfe5ab9d tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] Lock "6f6ba57a-e2d6-4749-a53a-e263861cb1c0" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 884.399084] env[61806]: DEBUG nova.network.neutron [None req-7e96f1c1-fc04-4010-b728-951cd26dd38b tempest-ImagesTestJSON-618852556 tempest-ImagesTestJSON-618852556-project-member] [instance: 1a73d994-5f21-4f80-8f33-5830d860a7ee] Successfully created port: 4017a509-57e9-4526-a107-3d54b7ed6684 {{(pid=61806) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 884.444203] env[61806]: DEBUG oslo_vmware.api [None req-799f40df-a27b-4200-803a-166b5c1e29fd tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Task: {'id': task-1294793, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 884.461067] env[61806]: DEBUG oslo_vmware.api [None req-b85317d7-e79b-4ab3-83e4-2de701507449 tempest-ServerTagsTestJSON-1292926242 tempest-ServerTagsTestJSON-1292926242-project-member] Task: {'id': session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]52710191-e6a4-bb3a-a6a2-15076e6ddc64, 'name': SearchDatastore_Task, 'duration_secs': 0.029815} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 884.461439] env[61806]: DEBUG oslo_concurrency.lockutils [None req-b85317d7-e79b-4ab3-83e4-2de701507449 tempest-ServerTagsTestJSON-1292926242 tempest-ServerTagsTestJSON-1292926242-project-member] Releasing lock "[datastore2] devstack-image-cache_base/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa.vmdk" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 884.461709] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [None req-b85317d7-e79b-4ab3-83e4-2de701507449 tempest-ServerTagsTestJSON-1292926242 tempest-ServerTagsTestJSON-1292926242-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa.vmdk to [datastore2] 9c20d337-b8f5-4965-b707-0f87c4ab6fc4/9c20d337-b8f5-4965-b707-0f87c4ab6fc4.vmdk {{(pid=61806) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 884.461990] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-970eb5d4-525c-4d38-8035-80e1fbe62a24 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 884.473428] env[61806]: DEBUG oslo_vmware.api [None req-b85317d7-e79b-4ab3-83e4-2de701507449 tempest-ServerTagsTestJSON-1292926242 tempest-ServerTagsTestJSON-1292926242-project-member] Waiting for the task: (returnval){ [ 884.473428] env[61806]: value = "task-1294799" [ 884.473428] env[61806]: _type = "Task" [ 884.473428] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 884.484017] env[61806]: DEBUG oslo_vmware.api [None req-b85317d7-e79b-4ab3-83e4-2de701507449 tempest-ServerTagsTestJSON-1292926242 tempest-ServerTagsTestJSON-1292926242-project-member] Task: {'id': task-1294799, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 884.518921] env[61806]: DEBUG nova.compute.manager [None req-7e96f1c1-fc04-4010-b728-951cd26dd38b tempest-ImagesTestJSON-618852556 tempest-ImagesTestJSON-618852556-project-member] [instance: 1a73d994-5f21-4f80-8f33-5830d860a7ee] Start building block device mappings for instance. {{(pid=61806) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 884.607166] env[61806]: DEBUG oslo_vmware.api [None req-58b4b995-fdfd-465f-8927-98666a556c7c tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Task: {'id': task-1294798, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 884.621730] env[61806]: INFO nova.compute.manager [None req-b996cb64-3154-4eef-9fcd-aa24605b9020 tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] [instance: d1802da4-a670-427d-84d7-0fcb4717e18f] Took 19.98 seconds to build instance. [ 884.690081] env[61806]: DEBUG nova.network.neutron [req-b2bbef49-94eb-4b26-ab49-02410cc3a1d4 req-e8927c34-d897-4896-81a3-ec939d0ae624 service nova] [instance: 5a526856-0dcd-4b3f-8359-13a5c3b9bfc4] Updated VIF entry in instance network info cache for port 58782499-bc1a-4be1-9d48-655c6b2f11d7. {{(pid=61806) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 884.690081] env[61806]: DEBUG nova.network.neutron [req-b2bbef49-94eb-4b26-ab49-02410cc3a1d4 req-e8927c34-d897-4896-81a3-ec939d0ae624 service nova] [instance: 5a526856-0dcd-4b3f-8359-13a5c3b9bfc4] Updating instance_info_cache with network_info: [{"id": "58782499-bc1a-4be1-9d48-655c6b2f11d7", "address": "fa:16:3e:52:7a:2b", "network": {"id": "b4fa0830-265e-47bf-8029-5944534312e7", "bridge": "br-int", "label": "tempest-ServerRescueTestJSONUnderV235-1460638664-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "2ba59612579b47a4bde7f82d951ac3ff", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8a34aa30-95be-4b18-98ca-1f2d81f7e9e6", "external-id": "nsx-vlan-transportzone-234", "segmentation_id": 234, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap58782499-bc", "ovs_interfaceid": "58782499-bc1a-4be1-9d48-655c6b2f11d7", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61806) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 884.718417] env[61806]: DEBUG oslo_vmware.api [-] Task: {'id': task-1294797, 'name': CreateVM_Task, 'duration_secs': 0.654372} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 884.718417] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 5a526856-0dcd-4b3f-8359-13a5c3b9bfc4] Created VM on the ESX host {{(pid=61806) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 884.719368] env[61806]: DEBUG oslo_concurrency.lockutils [None req-c651ea58-49e6-4c69-a5f7-88000b663e10 tempest-ServerRescueTestJSONUnderV235-749414849 tempest-ServerRescueTestJSONUnderV235-749414849-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 884.720142] env[61806]: DEBUG oslo_concurrency.lockutils [None req-c651ea58-49e6-4c69-a5f7-88000b663e10 tempest-ServerRescueTestJSONUnderV235-749414849 tempest-ServerRescueTestJSONUnderV235-749414849-project-member] Acquired lock "[datastore2] devstack-image-cache_base/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 884.720267] env[61806]: DEBUG oslo_concurrency.lockutils [None req-c651ea58-49e6-4c69-a5f7-88000b663e10 tempest-ServerRescueTestJSONUnderV235-749414849 tempest-ServerRescueTestJSONUnderV235-749414849-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 884.720628] env[61806]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-69d92522-1880-44f6-a592-4ead34996e23 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 884.726167] env[61806]: DEBUG oslo_vmware.api [None req-c651ea58-49e6-4c69-a5f7-88000b663e10 tempest-ServerRescueTestJSONUnderV235-749414849 tempest-ServerRescueTestJSONUnderV235-749414849-project-member] Waiting for the task: (returnval){ [ 884.726167] env[61806]: value = "session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]5280b012-a84c-8921-2221-15ef0d9c7fa4" [ 884.726167] env[61806]: _type = "Task" [ 884.726167] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 884.735742] env[61806]: DEBUG oslo_vmware.api [None req-c651ea58-49e6-4c69-a5f7-88000b663e10 tempest-ServerRescueTestJSONUnderV235-749414849 tempest-ServerRescueTestJSONUnderV235-749414849-project-member] Task: {'id': session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]5280b012-a84c-8921-2221-15ef0d9c7fa4, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 884.799909] env[61806]: DEBUG nova.scheduler.client.report [None req-708d7cc8-bb4e-45d0-adad-8f77e0f096a2 tempest-ImagesTestJSON-618852556 tempest-ImagesTestJSON-618852556-project-member] Inventory has not changed for provider a2858be1-fd22-4e08-979e-87ad25293407 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 138, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61806) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 884.860107] env[61806]: DEBUG nova.compute.manager [None req-095bcbe6-2af7-4761-a8ad-c73adfe5ab9d tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] [instance: 6f6ba57a-e2d6-4749-a53a-e263861cb1c0] Starting instance... {{(pid=61806) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 884.942011] env[61806]: DEBUG oslo_vmware.api [None req-799f40df-a27b-4200-803a-166b5c1e29fd tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Task: {'id': task-1294793, 'name': CopyVirtualDisk_Task, 'duration_secs': 1.090309} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 884.942375] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [None req-799f40df-a27b-4200-803a-166b5c1e29fd tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa.vmdk to [datastore1] 59c22c16-3d1c-4e64-bfac-7f8886a1927e/59c22c16-3d1c-4e64-bfac-7f8886a1927e.vmdk {{(pid=61806) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 884.942561] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-799f40df-a27b-4200-803a-166b5c1e29fd tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] [instance: 59c22c16-3d1c-4e64-bfac-7f8886a1927e] Extending root virtual disk to 1048576 {{(pid=61806) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 884.942846] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-18bc6a44-f47d-4b27-91e1-cecec8c9fc51 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 884.950476] env[61806]: DEBUG oslo_vmware.api [None req-799f40df-a27b-4200-803a-166b5c1e29fd tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Waiting for the task: (returnval){ [ 884.950476] env[61806]: value = "task-1294800" [ 884.950476] env[61806]: _type = "Task" [ 884.950476] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 884.962472] env[61806]: DEBUG oslo_vmware.api [None req-799f40df-a27b-4200-803a-166b5c1e29fd tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Task: {'id': task-1294800, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 884.984091] env[61806]: DEBUG oslo_vmware.api [None req-b85317d7-e79b-4ab3-83e4-2de701507449 tempest-ServerTagsTestJSON-1292926242 tempest-ServerTagsTestJSON-1292926242-project-member] Task: {'id': task-1294799, 'name': CopyVirtualDisk_Task} progress is 4%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 885.110134] env[61806]: DEBUG oslo_vmware.api [None req-58b4b995-fdfd-465f-8927-98666a556c7c tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Task: {'id': task-1294798, 'name': ReconfigVM_Task, 'duration_secs': 0.579429} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 885.110464] env[61806]: DEBUG nova.virt.vmwareapi.volumeops [None req-58b4b995-fdfd-465f-8927-98666a556c7c tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] [instance: efa19aac-3ad7-42d5-a6d8-859050de6e63] Reconfigured VM instance instance-00000049 to attach disk [datastore2] efa19aac-3ad7-42d5-a6d8-859050de6e63/efa19aac-3ad7-42d5-a6d8-859050de6e63.vmdk or device None with type sparse {{(pid=61806) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 885.111157] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-6fccd2c1-2da4-43b7-9cd7-b9b8fb4f4c20 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 885.119465] env[61806]: DEBUG oslo_vmware.api [None req-58b4b995-fdfd-465f-8927-98666a556c7c tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Waiting for the task: (returnval){ [ 885.119465] env[61806]: value = "task-1294801" [ 885.119465] env[61806]: _type = "Task" [ 885.119465] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 885.130404] env[61806]: DEBUG oslo_concurrency.lockutils [None req-b996cb64-3154-4eef-9fcd-aa24605b9020 tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] Lock "d1802da4-a670-427d-84d7-0fcb4717e18f" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 21.499s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 885.130713] env[61806]: DEBUG oslo_vmware.api [None req-58b4b995-fdfd-465f-8927-98666a556c7c tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Task: {'id': task-1294801, 'name': Rename_Task} progress is 5%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 885.192490] env[61806]: DEBUG oslo_concurrency.lockutils [req-b2bbef49-94eb-4b26-ab49-02410cc3a1d4 req-e8927c34-d897-4896-81a3-ec939d0ae624 service nova] Releasing lock "refresh_cache-5a526856-0dcd-4b3f-8359-13a5c3b9bfc4" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 885.238610] env[61806]: DEBUG oslo_vmware.api [None req-c651ea58-49e6-4c69-a5f7-88000b663e10 tempest-ServerRescueTestJSONUnderV235-749414849 tempest-ServerRescueTestJSONUnderV235-749414849-project-member] Task: {'id': session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]5280b012-a84c-8921-2221-15ef0d9c7fa4, 'name': SearchDatastore_Task, 'duration_secs': 0.016326} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 885.238956] env[61806]: DEBUG oslo_concurrency.lockutils [None req-c651ea58-49e6-4c69-a5f7-88000b663e10 tempest-ServerRescueTestJSONUnderV235-749414849 tempest-ServerRescueTestJSONUnderV235-749414849-project-member] Releasing lock "[datastore2] devstack-image-cache_base/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 885.239221] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-c651ea58-49e6-4c69-a5f7-88000b663e10 tempest-ServerRescueTestJSONUnderV235-749414849 tempest-ServerRescueTestJSONUnderV235-749414849-project-member] [instance: 5a526856-0dcd-4b3f-8359-13a5c3b9bfc4] Processing image ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa {{(pid=61806) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 885.239533] env[61806]: DEBUG oslo_concurrency.lockutils [None req-c651ea58-49e6-4c69-a5f7-88000b663e10 tempest-ServerRescueTestJSONUnderV235-749414849 tempest-ServerRescueTestJSONUnderV235-749414849-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa.vmdk" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 885.239729] env[61806]: DEBUG oslo_concurrency.lockutils [None req-c651ea58-49e6-4c69-a5f7-88000b663e10 tempest-ServerRescueTestJSONUnderV235-749414849 tempest-ServerRescueTestJSONUnderV235-749414849-project-member] Acquired lock "[datastore2] devstack-image-cache_base/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa.vmdk" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 885.239936] env[61806]: DEBUG nova.virt.vmwareapi.ds_util [None req-c651ea58-49e6-4c69-a5f7-88000b663e10 tempest-ServerRescueTestJSONUnderV235-749414849 tempest-ServerRescueTestJSONUnderV235-749414849-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61806) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 885.240255] env[61806]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-16873d85-be31-4157-bda4-8eadfc0efb0e {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 885.257552] env[61806]: DEBUG nova.virt.vmwareapi.ds_util [None req-c651ea58-49e6-4c69-a5f7-88000b663e10 tempest-ServerRescueTestJSONUnderV235-749414849 tempest-ServerRescueTestJSONUnderV235-749414849-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61806) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 885.257784] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-c651ea58-49e6-4c69-a5f7-88000b663e10 tempest-ServerRescueTestJSONUnderV235-749414849 tempest-ServerRescueTestJSONUnderV235-749414849-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61806) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 885.258942] env[61806]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-977e99bb-c19e-4868-8840-897061fe1c30 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 885.268655] env[61806]: DEBUG oslo_vmware.api [None req-c651ea58-49e6-4c69-a5f7-88000b663e10 tempest-ServerRescueTestJSONUnderV235-749414849 tempest-ServerRescueTestJSONUnderV235-749414849-project-member] Waiting for the task: (returnval){ [ 885.268655] env[61806]: value = "session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]521c8865-5a3c-2354-3bce-293cb8ee9b60" [ 885.268655] env[61806]: _type = "Task" [ 885.268655] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 885.279324] env[61806]: DEBUG oslo_vmware.api [None req-c651ea58-49e6-4c69-a5f7-88000b663e10 tempest-ServerRescueTestJSONUnderV235-749414849 tempest-ServerRescueTestJSONUnderV235-749414849-project-member] Task: {'id': session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]521c8865-5a3c-2354-3bce-293cb8ee9b60, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 885.305359] env[61806]: DEBUG oslo_concurrency.lockutils [None req-708d7cc8-bb4e-45d0-adad-8f77e0f096a2 tempest-ImagesTestJSON-618852556 tempest-ImagesTestJSON-618852556-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.796s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 885.307836] env[61806]: DEBUG oslo_concurrency.lockutils [None req-0bb2f11a-fff8-4c85-aa76-4f260c30bde9 tempest-AttachVolumeNegativeTest-1683033704 tempest-AttachVolumeNegativeTest-1683033704-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 8.111s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 885.309522] env[61806]: INFO nova.compute.claims [None req-0bb2f11a-fff8-4c85-aa76-4f260c30bde9 tempest-AttachVolumeNegativeTest-1683033704 tempest-AttachVolumeNegativeTest-1683033704-project-member] [instance: daff849e-eb4f-411e-92f8-9362e3bd91e3] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 885.332590] env[61806]: INFO nova.scheduler.client.report [None req-708d7cc8-bb4e-45d0-adad-8f77e0f096a2 tempest-ImagesTestJSON-618852556 tempest-ImagesTestJSON-618852556-project-member] Deleted allocations for instance e0ef0a35-82a5-495b-9d5c-5805e8306390 [ 885.380124] env[61806]: DEBUG oslo_concurrency.lockutils [None req-095bcbe6-2af7-4761-a8ad-c73adfe5ab9d tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 885.461285] env[61806]: DEBUG oslo_vmware.api [None req-799f40df-a27b-4200-803a-166b5c1e29fd tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Task: {'id': task-1294800, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.102775} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 885.461551] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-799f40df-a27b-4200-803a-166b5c1e29fd tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] [instance: 59c22c16-3d1c-4e64-bfac-7f8886a1927e] Extended root virtual disk {{(pid=61806) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 885.462359] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1e59cf87-07ad-4800-9ce9-8f50be8733fc {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 885.486314] env[61806]: DEBUG nova.virt.vmwareapi.volumeops [None req-799f40df-a27b-4200-803a-166b5c1e29fd tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] [instance: 59c22c16-3d1c-4e64-bfac-7f8886a1927e] Reconfiguring VM instance instance-0000004b to attach disk [datastore1] 59c22c16-3d1c-4e64-bfac-7f8886a1927e/59c22c16-3d1c-4e64-bfac-7f8886a1927e.vmdk or device None with type sparse {{(pid=61806) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 885.489728] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-2f51acd2-1618-4191-aa61-b297e0a73675 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 885.510215] env[61806]: DEBUG oslo_vmware.api [None req-b85317d7-e79b-4ab3-83e4-2de701507449 tempest-ServerTagsTestJSON-1292926242 tempest-ServerTagsTestJSON-1292926242-project-member] Task: {'id': task-1294799, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.867152} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 885.511540] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [None req-b85317d7-e79b-4ab3-83e4-2de701507449 tempest-ServerTagsTestJSON-1292926242 tempest-ServerTagsTestJSON-1292926242-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa.vmdk to [datastore2] 9c20d337-b8f5-4965-b707-0f87c4ab6fc4/9c20d337-b8f5-4965-b707-0f87c4ab6fc4.vmdk {{(pid=61806) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 885.511774] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-b85317d7-e79b-4ab3-83e4-2de701507449 tempest-ServerTagsTestJSON-1292926242 tempest-ServerTagsTestJSON-1292926242-project-member] [instance: 9c20d337-b8f5-4965-b707-0f87c4ab6fc4] Extending root virtual disk to 1048576 {{(pid=61806) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 885.512118] env[61806]: DEBUG oslo_vmware.api [None req-799f40df-a27b-4200-803a-166b5c1e29fd tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Waiting for the task: (returnval){ [ 885.512118] env[61806]: value = "task-1294802" [ 885.512118] env[61806]: _type = "Task" [ 885.512118] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 885.512325] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-3863380b-f912-4495-926c-4d6427b0f147 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 885.522850] env[61806]: DEBUG oslo_vmware.api [None req-799f40df-a27b-4200-803a-166b5c1e29fd tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Task: {'id': task-1294802, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 885.524258] env[61806]: DEBUG oslo_vmware.api [None req-b85317d7-e79b-4ab3-83e4-2de701507449 tempest-ServerTagsTestJSON-1292926242 tempest-ServerTagsTestJSON-1292926242-project-member] Waiting for the task: (returnval){ [ 885.524258] env[61806]: value = "task-1294803" [ 885.524258] env[61806]: _type = "Task" [ 885.524258] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 885.528533] env[61806]: DEBUG nova.compute.manager [None req-7e96f1c1-fc04-4010-b728-951cd26dd38b tempest-ImagesTestJSON-618852556 tempest-ImagesTestJSON-618852556-project-member] [instance: 1a73d994-5f21-4f80-8f33-5830d860a7ee] Start spawning the instance on the hypervisor. {{(pid=61806) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 885.537370] env[61806]: DEBUG oslo_vmware.api [None req-b85317d7-e79b-4ab3-83e4-2de701507449 tempest-ServerTagsTestJSON-1292926242 tempest-ServerTagsTestJSON-1292926242-project-member] Task: {'id': task-1294803, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 885.554825] env[61806]: DEBUG nova.virt.hardware [None req-7e96f1c1-fc04-4010-b728-951cd26dd38b tempest-ImagesTestJSON-618852556 tempest-ImagesTestJSON-618852556-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-15T18:05:10Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-15T18:04:52Z,direct_url=,disk_format='vmdk',id=ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='84ba83fce2c349a6988173c5d6fc3b07',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-15T18:04:53Z,virtual_size=,visibility=), allow threads: False {{(pid=61806) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 885.555091] env[61806]: DEBUG nova.virt.hardware [None req-7e96f1c1-fc04-4010-b728-951cd26dd38b tempest-ImagesTestJSON-618852556 tempest-ImagesTestJSON-618852556-project-member] Flavor limits 0:0:0 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 885.555275] env[61806]: DEBUG nova.virt.hardware [None req-7e96f1c1-fc04-4010-b728-951cd26dd38b tempest-ImagesTestJSON-618852556 tempest-ImagesTestJSON-618852556-project-member] Image limits 0:0:0 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 885.555518] env[61806]: DEBUG nova.virt.hardware [None req-7e96f1c1-fc04-4010-b728-951cd26dd38b tempest-ImagesTestJSON-618852556 tempest-ImagesTestJSON-618852556-project-member] Flavor pref 0:0:0 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 885.555646] env[61806]: DEBUG nova.virt.hardware [None req-7e96f1c1-fc04-4010-b728-951cd26dd38b tempest-ImagesTestJSON-618852556 tempest-ImagesTestJSON-618852556-project-member] Image pref 0:0:0 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 885.555801] env[61806]: DEBUG nova.virt.hardware [None req-7e96f1c1-fc04-4010-b728-951cd26dd38b tempest-ImagesTestJSON-618852556 tempest-ImagesTestJSON-618852556-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 885.556030] env[61806]: DEBUG nova.virt.hardware [None req-7e96f1c1-fc04-4010-b728-951cd26dd38b tempest-ImagesTestJSON-618852556 tempest-ImagesTestJSON-618852556-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61806) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 885.556197] env[61806]: DEBUG nova.virt.hardware [None req-7e96f1c1-fc04-4010-b728-951cd26dd38b tempest-ImagesTestJSON-618852556 tempest-ImagesTestJSON-618852556-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61806) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 885.556375] env[61806]: DEBUG nova.virt.hardware [None req-7e96f1c1-fc04-4010-b728-951cd26dd38b tempest-ImagesTestJSON-618852556 tempest-ImagesTestJSON-618852556-project-member] Got 1 possible topologies {{(pid=61806) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 885.556546] env[61806]: DEBUG nova.virt.hardware [None req-7e96f1c1-fc04-4010-b728-951cd26dd38b tempest-ImagesTestJSON-618852556 tempest-ImagesTestJSON-618852556-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61806) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 885.556750] env[61806]: DEBUG nova.virt.hardware [None req-7e96f1c1-fc04-4010-b728-951cd26dd38b tempest-ImagesTestJSON-618852556 tempest-ImagesTestJSON-618852556-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61806) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 885.557647] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cb385828-096d-4431-9ed0-59434ad739b3 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 885.566215] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bd3ea91c-da5d-4f2b-9d93-d0ccb9b54009 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 885.630142] env[61806]: DEBUG oslo_vmware.api [None req-58b4b995-fdfd-465f-8927-98666a556c7c tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Task: {'id': task-1294801, 'name': Rename_Task, 'duration_secs': 0.403997} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 885.630448] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [None req-58b4b995-fdfd-465f-8927-98666a556c7c tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] [instance: efa19aac-3ad7-42d5-a6d8-859050de6e63] Powering on the VM {{(pid=61806) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 885.630730] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-10e885a4-1b6b-4aa9-bce2-d6a25bc38328 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 885.638292] env[61806]: DEBUG oslo_vmware.api [None req-58b4b995-fdfd-465f-8927-98666a556c7c tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Waiting for the task: (returnval){ [ 885.638292] env[61806]: value = "task-1294804" [ 885.638292] env[61806]: _type = "Task" [ 885.638292] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 885.649037] env[61806]: DEBUG oslo_vmware.api [None req-58b4b995-fdfd-465f-8927-98666a556c7c tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Task: {'id': task-1294804, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 885.780567] env[61806]: DEBUG oslo_vmware.api [None req-c651ea58-49e6-4c69-a5f7-88000b663e10 tempest-ServerRescueTestJSONUnderV235-749414849 tempest-ServerRescueTestJSONUnderV235-749414849-project-member] Task: {'id': session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]521c8865-5a3c-2354-3bce-293cb8ee9b60, 'name': SearchDatastore_Task, 'duration_secs': 0.042362} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 885.781516] env[61806]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-9f45db71-74f9-401a-9e2f-4b1d4c59968d {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 885.787680] env[61806]: DEBUG oslo_vmware.api [None req-c651ea58-49e6-4c69-a5f7-88000b663e10 tempest-ServerRescueTestJSONUnderV235-749414849 tempest-ServerRescueTestJSONUnderV235-749414849-project-member] Waiting for the task: (returnval){ [ 885.787680] env[61806]: value = "session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]5242b270-544a-d1e8-d40a-18570a2c5a96" [ 885.787680] env[61806]: _type = "Task" [ 885.787680] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 885.796951] env[61806]: DEBUG oslo_vmware.api [None req-c651ea58-49e6-4c69-a5f7-88000b663e10 tempest-ServerRescueTestJSONUnderV235-749414849 tempest-ServerRescueTestJSONUnderV235-749414849-project-member] Task: {'id': session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]5242b270-544a-d1e8-d40a-18570a2c5a96, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 885.839971] env[61806]: DEBUG oslo_concurrency.lockutils [None req-708d7cc8-bb4e-45d0-adad-8f77e0f096a2 tempest-ImagesTestJSON-618852556 tempest-ImagesTestJSON-618852556-project-member] Lock "e0ef0a35-82a5-495b-9d5c-5805e8306390" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 14.026s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 886.038997] env[61806]: DEBUG oslo_vmware.api [None req-799f40df-a27b-4200-803a-166b5c1e29fd tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Task: {'id': task-1294802, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 886.044822] env[61806]: DEBUG oslo_vmware.api [None req-b85317d7-e79b-4ab3-83e4-2de701507449 tempest-ServerTagsTestJSON-1292926242 tempest-ServerTagsTestJSON-1292926242-project-member] Task: {'id': task-1294803, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.088561} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 886.045109] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-b85317d7-e79b-4ab3-83e4-2de701507449 tempest-ServerTagsTestJSON-1292926242 tempest-ServerTagsTestJSON-1292926242-project-member] [instance: 9c20d337-b8f5-4965-b707-0f87c4ab6fc4] Extended root virtual disk {{(pid=61806) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 886.045953] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3a4550f8-885c-499a-b3a9-f5bb7ee9e37c {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 886.050698] env[61806]: DEBUG nova.compute.manager [req-1f3fa0fc-edf5-4d6d-9458-a6f4026a5145 req-99d6fc98-de90-443c-844d-f8d4e0aa24a9 service nova] [instance: 1a73d994-5f21-4f80-8f33-5830d860a7ee] Received event network-vif-plugged-4017a509-57e9-4526-a107-3d54b7ed6684 {{(pid=61806) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 886.050907] env[61806]: DEBUG oslo_concurrency.lockutils [req-1f3fa0fc-edf5-4d6d-9458-a6f4026a5145 req-99d6fc98-de90-443c-844d-f8d4e0aa24a9 service nova] Acquiring lock "1a73d994-5f21-4f80-8f33-5830d860a7ee-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 886.051127] env[61806]: DEBUG oslo_concurrency.lockutils [req-1f3fa0fc-edf5-4d6d-9458-a6f4026a5145 req-99d6fc98-de90-443c-844d-f8d4e0aa24a9 service nova] Lock "1a73d994-5f21-4f80-8f33-5830d860a7ee-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 886.051334] env[61806]: DEBUG oslo_concurrency.lockutils [req-1f3fa0fc-edf5-4d6d-9458-a6f4026a5145 req-99d6fc98-de90-443c-844d-f8d4e0aa24a9 service nova] Lock "1a73d994-5f21-4f80-8f33-5830d860a7ee-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 886.051511] env[61806]: DEBUG nova.compute.manager [req-1f3fa0fc-edf5-4d6d-9458-a6f4026a5145 req-99d6fc98-de90-443c-844d-f8d4e0aa24a9 service nova] [instance: 1a73d994-5f21-4f80-8f33-5830d860a7ee] No waiting events found dispatching network-vif-plugged-4017a509-57e9-4526-a107-3d54b7ed6684 {{(pid=61806) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 886.051675] env[61806]: WARNING nova.compute.manager [req-1f3fa0fc-edf5-4d6d-9458-a6f4026a5145 req-99d6fc98-de90-443c-844d-f8d4e0aa24a9 service nova] [instance: 1a73d994-5f21-4f80-8f33-5830d860a7ee] Received unexpected event network-vif-plugged-4017a509-57e9-4526-a107-3d54b7ed6684 for instance with vm_state building and task_state spawning. [ 886.082810] env[61806]: DEBUG nova.virt.vmwareapi.volumeops [None req-b85317d7-e79b-4ab3-83e4-2de701507449 tempest-ServerTagsTestJSON-1292926242 tempest-ServerTagsTestJSON-1292926242-project-member] [instance: 9c20d337-b8f5-4965-b707-0f87c4ab6fc4] Reconfiguring VM instance instance-0000004c to attach disk [datastore2] 9c20d337-b8f5-4965-b707-0f87c4ab6fc4/9c20d337-b8f5-4965-b707-0f87c4ab6fc4.vmdk or device None with type sparse {{(pid=61806) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 886.083892] env[61806]: DEBUG nova.network.neutron [None req-7e96f1c1-fc04-4010-b728-951cd26dd38b tempest-ImagesTestJSON-618852556 tempest-ImagesTestJSON-618852556-project-member] [instance: 1a73d994-5f21-4f80-8f33-5830d860a7ee] Successfully updated port: 4017a509-57e9-4526-a107-3d54b7ed6684 {{(pid=61806) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 886.085370] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-3024e54a-a9ad-467f-9311-867edbde891e {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 886.108475] env[61806]: DEBUG oslo_vmware.api [None req-b85317d7-e79b-4ab3-83e4-2de701507449 tempest-ServerTagsTestJSON-1292926242 tempest-ServerTagsTestJSON-1292926242-project-member] Waiting for the task: (returnval){ [ 886.108475] env[61806]: value = "task-1294805" [ 886.108475] env[61806]: _type = "Task" [ 886.108475] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 886.118905] env[61806]: DEBUG oslo_vmware.api [None req-b85317d7-e79b-4ab3-83e4-2de701507449 tempest-ServerTagsTestJSON-1292926242 tempest-ServerTagsTestJSON-1292926242-project-member] Task: {'id': task-1294805, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 886.149546] env[61806]: DEBUG oslo_vmware.api [None req-58b4b995-fdfd-465f-8927-98666a556c7c tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Task: {'id': task-1294804, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 886.300416] env[61806]: DEBUG oslo_vmware.api [None req-c651ea58-49e6-4c69-a5f7-88000b663e10 tempest-ServerRescueTestJSONUnderV235-749414849 tempest-ServerRescueTestJSONUnderV235-749414849-project-member] Task: {'id': session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]5242b270-544a-d1e8-d40a-18570a2c5a96, 'name': SearchDatastore_Task, 'duration_secs': 0.033237} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 886.300850] env[61806]: DEBUG oslo_concurrency.lockutils [None req-c651ea58-49e6-4c69-a5f7-88000b663e10 tempest-ServerRescueTestJSONUnderV235-749414849 tempest-ServerRescueTestJSONUnderV235-749414849-project-member] Releasing lock "[datastore2] devstack-image-cache_base/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa.vmdk" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 886.301114] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [None req-c651ea58-49e6-4c69-a5f7-88000b663e10 tempest-ServerRescueTestJSONUnderV235-749414849 tempest-ServerRescueTestJSONUnderV235-749414849-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa.vmdk to [datastore2] 5a526856-0dcd-4b3f-8359-13a5c3b9bfc4/5a526856-0dcd-4b3f-8359-13a5c3b9bfc4.vmdk {{(pid=61806) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 886.301411] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-058ea9de-d628-47ac-9895-a802054d7cd6 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 886.310376] env[61806]: DEBUG oslo_vmware.api [None req-c651ea58-49e6-4c69-a5f7-88000b663e10 tempest-ServerRescueTestJSONUnderV235-749414849 tempest-ServerRescueTestJSONUnderV235-749414849-project-member] Waiting for the task: (returnval){ [ 886.310376] env[61806]: value = "task-1294806" [ 886.310376] env[61806]: _type = "Task" [ 886.310376] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 886.322972] env[61806]: DEBUG oslo_vmware.api [None req-c651ea58-49e6-4c69-a5f7-88000b663e10 tempest-ServerRescueTestJSONUnderV235-749414849 tempest-ServerRescueTestJSONUnderV235-749414849-project-member] Task: {'id': task-1294806, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 886.470467] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4199b2de-f51b-4eb0-9046-e7b25ca8c23d {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 886.480018] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a9cd3a37-7c49-4265-b071-b494b9f1d7be {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 886.514836] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4a6a089e-12dd-4870-a662-3fd702344725 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 886.527332] env[61806]: DEBUG oslo_vmware.api [None req-799f40df-a27b-4200-803a-166b5c1e29fd tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Task: {'id': task-1294802, 'name': ReconfigVM_Task, 'duration_secs': 0.92021} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 886.529810] env[61806]: DEBUG nova.virt.vmwareapi.volumeops [None req-799f40df-a27b-4200-803a-166b5c1e29fd tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] [instance: 59c22c16-3d1c-4e64-bfac-7f8886a1927e] Reconfigured VM instance instance-0000004b to attach disk [datastore1] 59c22c16-3d1c-4e64-bfac-7f8886a1927e/59c22c16-3d1c-4e64-bfac-7f8886a1927e.vmdk or device None with type sparse {{(pid=61806) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 886.530602] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-dbf00cab-f46e-432c-a8b5-d697555beaba {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 886.533522] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fc16eccd-acd9-4110-b618-9209bd94e9ad {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 886.551675] env[61806]: DEBUG nova.compute.provider_tree [None req-0bb2f11a-fff8-4c85-aa76-4f260c30bde9 tempest-AttachVolumeNegativeTest-1683033704 tempest-AttachVolumeNegativeTest-1683033704-project-member] Inventory has not changed in ProviderTree for provider: a2858be1-fd22-4e08-979e-87ad25293407 {{(pid=61806) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 886.554677] env[61806]: DEBUG oslo_vmware.api [None req-799f40df-a27b-4200-803a-166b5c1e29fd tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Waiting for the task: (returnval){ [ 886.554677] env[61806]: value = "task-1294807" [ 886.554677] env[61806]: _type = "Task" [ 886.554677] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 886.567049] env[61806]: DEBUG oslo_vmware.api [None req-799f40df-a27b-4200-803a-166b5c1e29fd tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Task: {'id': task-1294807, 'name': Rename_Task} progress is 10%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 886.601752] env[61806]: DEBUG oslo_concurrency.lockutils [None req-7e96f1c1-fc04-4010-b728-951cd26dd38b tempest-ImagesTestJSON-618852556 tempest-ImagesTestJSON-618852556-project-member] Acquiring lock "refresh_cache-1a73d994-5f21-4f80-8f33-5830d860a7ee" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 886.601967] env[61806]: DEBUG oslo_concurrency.lockutils [None req-7e96f1c1-fc04-4010-b728-951cd26dd38b tempest-ImagesTestJSON-618852556 tempest-ImagesTestJSON-618852556-project-member] Acquired lock "refresh_cache-1a73d994-5f21-4f80-8f33-5830d860a7ee" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 886.602105] env[61806]: DEBUG nova.network.neutron [None req-7e96f1c1-fc04-4010-b728-951cd26dd38b tempest-ImagesTestJSON-618852556 tempest-ImagesTestJSON-618852556-project-member] [instance: 1a73d994-5f21-4f80-8f33-5830d860a7ee] Building network info cache for instance {{(pid=61806) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 886.622749] env[61806]: DEBUG oslo_vmware.api [None req-b85317d7-e79b-4ab3-83e4-2de701507449 tempest-ServerTagsTestJSON-1292926242 tempest-ServerTagsTestJSON-1292926242-project-member] Task: {'id': task-1294805, 'name': ReconfigVM_Task, 'duration_secs': 0.332732} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 886.623455] env[61806]: DEBUG nova.virt.vmwareapi.volumeops [None req-b85317d7-e79b-4ab3-83e4-2de701507449 tempest-ServerTagsTestJSON-1292926242 tempest-ServerTagsTestJSON-1292926242-project-member] [instance: 9c20d337-b8f5-4965-b707-0f87c4ab6fc4] Reconfigured VM instance instance-0000004c to attach disk [datastore2] 9c20d337-b8f5-4965-b707-0f87c4ab6fc4/9c20d337-b8f5-4965-b707-0f87c4ab6fc4.vmdk or device None with type sparse {{(pid=61806) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 886.624189] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-a9f2a16d-e836-4d31-bfa6-b68cfea6c958 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 886.633349] env[61806]: DEBUG oslo_vmware.api [None req-b85317d7-e79b-4ab3-83e4-2de701507449 tempest-ServerTagsTestJSON-1292926242 tempest-ServerTagsTestJSON-1292926242-project-member] Waiting for the task: (returnval){ [ 886.633349] env[61806]: value = "task-1294808" [ 886.633349] env[61806]: _type = "Task" [ 886.633349] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 886.647255] env[61806]: DEBUG oslo_vmware.api [None req-b85317d7-e79b-4ab3-83e4-2de701507449 tempest-ServerTagsTestJSON-1292926242 tempest-ServerTagsTestJSON-1292926242-project-member] Task: {'id': task-1294808, 'name': Rename_Task} progress is 5%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 886.653717] env[61806]: DEBUG oslo_vmware.api [None req-58b4b995-fdfd-465f-8927-98666a556c7c tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Task: {'id': task-1294804, 'name': PowerOnVM_Task, 'duration_secs': 0.891649} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 886.654123] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [None req-58b4b995-fdfd-465f-8927-98666a556c7c tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] [instance: efa19aac-3ad7-42d5-a6d8-859050de6e63] Powered on the VM {{(pid=61806) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 886.654442] env[61806]: DEBUG nova.compute.manager [None req-58b4b995-fdfd-465f-8927-98666a556c7c tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] [instance: efa19aac-3ad7-42d5-a6d8-859050de6e63] Checking state {{(pid=61806) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 886.655545] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c86632a7-25cb-4a58-a6ed-5e951f6c22cc {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 886.821625] env[61806]: DEBUG oslo_vmware.api [None req-c651ea58-49e6-4c69-a5f7-88000b663e10 tempest-ServerRescueTestJSONUnderV235-749414849 tempest-ServerRescueTestJSONUnderV235-749414849-project-member] Task: {'id': task-1294806, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 886.867647] env[61806]: DEBUG oslo_concurrency.lockutils [None req-80f76c83-748f-43fe-b146-5517b19ec0ae tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] Acquiring lock "ee002e7c-8965-494f-a1af-0b6f3b74eceb" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 886.867897] env[61806]: DEBUG oslo_concurrency.lockutils [None req-80f76c83-748f-43fe-b146-5517b19ec0ae tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] Lock "ee002e7c-8965-494f-a1af-0b6f3b74eceb" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 887.057036] env[61806]: DEBUG nova.scheduler.client.report [None req-0bb2f11a-fff8-4c85-aa76-4f260c30bde9 tempest-AttachVolumeNegativeTest-1683033704 tempest-AttachVolumeNegativeTest-1683033704-project-member] Inventory has not changed for provider a2858be1-fd22-4e08-979e-87ad25293407 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 138, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61806) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 887.069941] env[61806]: DEBUG oslo_vmware.api [None req-799f40df-a27b-4200-803a-166b5c1e29fd tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Task: {'id': task-1294807, 'name': Rename_Task, 'duration_secs': 0.350904} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 887.070296] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [None req-799f40df-a27b-4200-803a-166b5c1e29fd tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] [instance: 59c22c16-3d1c-4e64-bfac-7f8886a1927e] Powering on the VM {{(pid=61806) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 887.070556] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-d8018f41-07fc-49c6-84b3-18c8ed994e2f {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 887.078166] env[61806]: DEBUG oslo_vmware.api [None req-799f40df-a27b-4200-803a-166b5c1e29fd tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Waiting for the task: (returnval){ [ 887.078166] env[61806]: value = "task-1294809" [ 887.078166] env[61806]: _type = "Task" [ 887.078166] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 887.087386] env[61806]: DEBUG oslo_vmware.api [None req-799f40df-a27b-4200-803a-166b5c1e29fd tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Task: {'id': task-1294809, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 887.133469] env[61806]: DEBUG nova.network.neutron [None req-7e96f1c1-fc04-4010-b728-951cd26dd38b tempest-ImagesTestJSON-618852556 tempest-ImagesTestJSON-618852556-project-member] [instance: 1a73d994-5f21-4f80-8f33-5830d860a7ee] Instance cache missing network info. {{(pid=61806) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 887.145513] env[61806]: DEBUG oslo_vmware.api [None req-b85317d7-e79b-4ab3-83e4-2de701507449 tempest-ServerTagsTestJSON-1292926242 tempest-ServerTagsTestJSON-1292926242-project-member] Task: {'id': task-1294808, 'name': Rename_Task, 'duration_secs': 0.258373} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 887.146029] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [None req-b85317d7-e79b-4ab3-83e4-2de701507449 tempest-ServerTagsTestJSON-1292926242 tempest-ServerTagsTestJSON-1292926242-project-member] [instance: 9c20d337-b8f5-4965-b707-0f87c4ab6fc4] Powering on the VM {{(pid=61806) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 887.146129] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-3d51bf55-ecf9-4acc-b88e-d203b8261ef3 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 887.155793] env[61806]: DEBUG oslo_vmware.api [None req-b85317d7-e79b-4ab3-83e4-2de701507449 tempest-ServerTagsTestJSON-1292926242 tempest-ServerTagsTestJSON-1292926242-project-member] Waiting for the task: (returnval){ [ 887.155793] env[61806]: value = "task-1294810" [ 887.155793] env[61806]: _type = "Task" [ 887.155793] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 887.164663] env[61806]: DEBUG oslo_vmware.api [None req-b85317d7-e79b-4ab3-83e4-2de701507449 tempest-ServerTagsTestJSON-1292926242 tempest-ServerTagsTestJSON-1292926242-project-member] Task: {'id': task-1294810, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 887.174946] env[61806]: DEBUG oslo_concurrency.lockutils [None req-58b4b995-fdfd-465f-8927-98666a556c7c tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 887.324212] env[61806]: DEBUG oslo_vmware.api [None req-c651ea58-49e6-4c69-a5f7-88000b663e10 tempest-ServerRescueTestJSONUnderV235-749414849 tempest-ServerRescueTestJSONUnderV235-749414849-project-member] Task: {'id': task-1294806, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.545814} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 887.324445] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [None req-c651ea58-49e6-4c69-a5f7-88000b663e10 tempest-ServerRescueTestJSONUnderV235-749414849 tempest-ServerRescueTestJSONUnderV235-749414849-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa.vmdk to [datastore2] 5a526856-0dcd-4b3f-8359-13a5c3b9bfc4/5a526856-0dcd-4b3f-8359-13a5c3b9bfc4.vmdk {{(pid=61806) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 887.324640] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-c651ea58-49e6-4c69-a5f7-88000b663e10 tempest-ServerRescueTestJSONUnderV235-749414849 tempest-ServerRescueTestJSONUnderV235-749414849-project-member] [instance: 5a526856-0dcd-4b3f-8359-13a5c3b9bfc4] Extending root virtual disk to 1048576 {{(pid=61806) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 887.324950] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-59e440da-3ff0-4720-abc9-e24c680bf76f {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 887.334031] env[61806]: DEBUG oslo_vmware.api [None req-c651ea58-49e6-4c69-a5f7-88000b663e10 tempest-ServerRescueTestJSONUnderV235-749414849 tempest-ServerRescueTestJSONUnderV235-749414849-project-member] Waiting for the task: (returnval){ [ 887.334031] env[61806]: value = "task-1294811" [ 887.334031] env[61806]: _type = "Task" [ 887.334031] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 887.334270] env[61806]: DEBUG nova.network.neutron [None req-7e96f1c1-fc04-4010-b728-951cd26dd38b tempest-ImagesTestJSON-618852556 tempest-ImagesTestJSON-618852556-project-member] [instance: 1a73d994-5f21-4f80-8f33-5830d860a7ee] Updating instance_info_cache with network_info: [{"id": "4017a509-57e9-4526-a107-3d54b7ed6684", "address": "fa:16:3e:40:38:fb", "network": {"id": "288ebc87-a30f-439b-9885-9cdda68a9591", "bridge": "br-int", "label": "tempest-ImagesTestJSON-914411285-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4b9aa91aff1d4008ac5096902b77f852", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8cbc9b8f-ce19-4262-bf4d-88cd4f259a1c", "external-id": "nsx-vlan-transportzone-630", "segmentation_id": 630, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4017a509-57", "ovs_interfaceid": "4017a509-57e9-4526-a107-3d54b7ed6684", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61806) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 887.346147] env[61806]: DEBUG oslo_vmware.api [None req-c651ea58-49e6-4c69-a5f7-88000b663e10 tempest-ServerRescueTestJSONUnderV235-749414849 tempest-ServerRescueTestJSONUnderV235-749414849-project-member] Task: {'id': task-1294811, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 887.370760] env[61806]: DEBUG nova.compute.manager [None req-80f76c83-748f-43fe-b146-5517b19ec0ae tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] [instance: ee002e7c-8965-494f-a1af-0b6f3b74eceb] Starting instance... {{(pid=61806) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 887.564971] env[61806]: DEBUG oslo_concurrency.lockutils [None req-0bb2f11a-fff8-4c85-aa76-4f260c30bde9 tempest-AttachVolumeNegativeTest-1683033704 tempest-AttachVolumeNegativeTest-1683033704-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.257s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 887.565767] env[61806]: DEBUG nova.compute.manager [None req-0bb2f11a-fff8-4c85-aa76-4f260c30bde9 tempest-AttachVolumeNegativeTest-1683033704 tempest-AttachVolumeNegativeTest-1683033704-project-member] [instance: daff849e-eb4f-411e-92f8-9362e3bd91e3] Start building networks asynchronously for instance. {{(pid=61806) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 887.569097] env[61806]: DEBUG oslo_concurrency.lockutils [None req-095bcbe6-2af7-4761-a8ad-c73adfe5ab9d tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 2.189s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 887.571715] env[61806]: INFO nova.compute.claims [None req-095bcbe6-2af7-4761-a8ad-c73adfe5ab9d tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] [instance: 6f6ba57a-e2d6-4749-a53a-e263861cb1c0] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 887.589062] env[61806]: DEBUG oslo_vmware.api [None req-799f40df-a27b-4200-803a-166b5c1e29fd tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Task: {'id': task-1294809, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 887.666721] env[61806]: DEBUG oslo_vmware.api [None req-b85317d7-e79b-4ab3-83e4-2de701507449 tempest-ServerTagsTestJSON-1292926242 tempest-ServerTagsTestJSON-1292926242-project-member] Task: {'id': task-1294810, 'name': PowerOnVM_Task, 'duration_secs': 0.497817} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 887.667958] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [None req-b85317d7-e79b-4ab3-83e4-2de701507449 tempest-ServerTagsTestJSON-1292926242 tempest-ServerTagsTestJSON-1292926242-project-member] [instance: 9c20d337-b8f5-4965-b707-0f87c4ab6fc4] Powered on the VM {{(pid=61806) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 887.667958] env[61806]: INFO nova.compute.manager [None req-b85317d7-e79b-4ab3-83e4-2de701507449 tempest-ServerTagsTestJSON-1292926242 tempest-ServerTagsTestJSON-1292926242-project-member] [instance: 9c20d337-b8f5-4965-b707-0f87c4ab6fc4] Took 8.43 seconds to spawn the instance on the hypervisor. [ 887.667958] env[61806]: DEBUG nova.compute.manager [None req-b85317d7-e79b-4ab3-83e4-2de701507449 tempest-ServerTagsTestJSON-1292926242 tempest-ServerTagsTestJSON-1292926242-project-member] [instance: 9c20d337-b8f5-4965-b707-0f87c4ab6fc4] Checking state {{(pid=61806) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 887.668232] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f9ff8817-c2b3-400c-88d0-5bc713db96f6 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 887.841039] env[61806]: DEBUG oslo_concurrency.lockutils [None req-7e96f1c1-fc04-4010-b728-951cd26dd38b tempest-ImagesTestJSON-618852556 tempest-ImagesTestJSON-618852556-project-member] Releasing lock "refresh_cache-1a73d994-5f21-4f80-8f33-5830d860a7ee" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 887.841352] env[61806]: DEBUG nova.compute.manager [None req-7e96f1c1-fc04-4010-b728-951cd26dd38b tempest-ImagesTestJSON-618852556 tempest-ImagesTestJSON-618852556-project-member] [instance: 1a73d994-5f21-4f80-8f33-5830d860a7ee] Instance network_info: |[{"id": "4017a509-57e9-4526-a107-3d54b7ed6684", "address": "fa:16:3e:40:38:fb", "network": {"id": "288ebc87-a30f-439b-9885-9cdda68a9591", "bridge": "br-int", "label": "tempest-ImagesTestJSON-914411285-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4b9aa91aff1d4008ac5096902b77f852", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8cbc9b8f-ce19-4262-bf4d-88cd4f259a1c", "external-id": "nsx-vlan-transportzone-630", "segmentation_id": 630, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4017a509-57", "ovs_interfaceid": "4017a509-57e9-4526-a107-3d54b7ed6684", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61806) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 887.842126] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-7e96f1c1-fc04-4010-b728-951cd26dd38b tempest-ImagesTestJSON-618852556 tempest-ImagesTestJSON-618852556-project-member] [instance: 1a73d994-5f21-4f80-8f33-5830d860a7ee] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:40:38:fb', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '8cbc9b8f-ce19-4262-bf4d-88cd4f259a1c', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '4017a509-57e9-4526-a107-3d54b7ed6684', 'vif_model': 'vmxnet3'}] {{(pid=61806) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 887.850063] env[61806]: DEBUG oslo.service.loopingcall [None req-7e96f1c1-fc04-4010-b728-951cd26dd38b tempest-ImagesTestJSON-618852556 tempest-ImagesTestJSON-618852556-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61806) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 887.850705] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 1a73d994-5f21-4f80-8f33-5830d860a7ee] Creating VM on the ESX host {{(pid=61806) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 887.850958] env[61806]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-5be900d5-0639-427b-92ed-5abe04eff73e {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 887.869090] env[61806]: DEBUG oslo_vmware.api [None req-c651ea58-49e6-4c69-a5f7-88000b663e10 tempest-ServerRescueTestJSONUnderV235-749414849 tempest-ServerRescueTestJSONUnderV235-749414849-project-member] Task: {'id': task-1294811, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.072118} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 887.869735] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-c651ea58-49e6-4c69-a5f7-88000b663e10 tempest-ServerRescueTestJSONUnderV235-749414849 tempest-ServerRescueTestJSONUnderV235-749414849-project-member] [instance: 5a526856-0dcd-4b3f-8359-13a5c3b9bfc4] Extended root virtual disk {{(pid=61806) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 887.870590] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2d488749-fada-4a27-8fed-0c52e30dbc04 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 887.891119] env[61806]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 887.891119] env[61806]: value = "task-1294812" [ 887.891119] env[61806]: _type = "Task" [ 887.891119] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 887.900229] env[61806]: DEBUG nova.virt.vmwareapi.volumeops [None req-c651ea58-49e6-4c69-a5f7-88000b663e10 tempest-ServerRescueTestJSONUnderV235-749414849 tempest-ServerRescueTestJSONUnderV235-749414849-project-member] [instance: 5a526856-0dcd-4b3f-8359-13a5c3b9bfc4] Reconfiguring VM instance instance-0000004d to attach disk [datastore2] 5a526856-0dcd-4b3f-8359-13a5c3b9bfc4/5a526856-0dcd-4b3f-8359-13a5c3b9bfc4.vmdk or device None with type sparse {{(pid=61806) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 887.902044] env[61806]: DEBUG oslo_concurrency.lockutils [None req-80f76c83-748f-43fe-b146-5517b19ec0ae tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 887.902044] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-459daaf9-d692-44cd-9445-e22a0eefe41b {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 887.927212] env[61806]: DEBUG oslo_vmware.api [-] Task: {'id': task-1294812, 'name': CreateVM_Task} progress is 15%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 887.928533] env[61806]: DEBUG oslo_vmware.api [None req-c651ea58-49e6-4c69-a5f7-88000b663e10 tempest-ServerRescueTestJSONUnderV235-749414849 tempest-ServerRescueTestJSONUnderV235-749414849-project-member] Waiting for the task: (returnval){ [ 887.928533] env[61806]: value = "task-1294813" [ 887.928533] env[61806]: _type = "Task" [ 887.928533] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 887.938911] env[61806]: DEBUG oslo_vmware.api [None req-c651ea58-49e6-4c69-a5f7-88000b663e10 tempest-ServerRescueTestJSONUnderV235-749414849 tempest-ServerRescueTestJSONUnderV235-749414849-project-member] Task: {'id': task-1294813, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 888.014640] env[61806]: DEBUG oslo_concurrency.lockutils [None req-f488cd5c-2a2f-4173-ba8b-c601bc6a140a tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Acquiring lock "efa19aac-3ad7-42d5-a6d8-859050de6e63" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 888.014965] env[61806]: DEBUG oslo_concurrency.lockutils [None req-f488cd5c-2a2f-4173-ba8b-c601bc6a140a tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Lock "efa19aac-3ad7-42d5-a6d8-859050de6e63" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 888.015207] env[61806]: DEBUG oslo_concurrency.lockutils [None req-f488cd5c-2a2f-4173-ba8b-c601bc6a140a tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Acquiring lock "efa19aac-3ad7-42d5-a6d8-859050de6e63-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 888.015416] env[61806]: DEBUG oslo_concurrency.lockutils [None req-f488cd5c-2a2f-4173-ba8b-c601bc6a140a tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Lock "efa19aac-3ad7-42d5-a6d8-859050de6e63-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 888.015783] env[61806]: DEBUG oslo_concurrency.lockutils [None req-f488cd5c-2a2f-4173-ba8b-c601bc6a140a tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Lock "efa19aac-3ad7-42d5-a6d8-859050de6e63-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 888.017759] env[61806]: INFO nova.compute.manager [None req-f488cd5c-2a2f-4173-ba8b-c601bc6a140a tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] [instance: efa19aac-3ad7-42d5-a6d8-859050de6e63] Terminating instance [ 888.019706] env[61806]: DEBUG nova.compute.manager [None req-f488cd5c-2a2f-4173-ba8b-c601bc6a140a tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] [instance: efa19aac-3ad7-42d5-a6d8-859050de6e63] Start destroying the instance on the hypervisor. {{(pid=61806) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 888.020080] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-f488cd5c-2a2f-4173-ba8b-c601bc6a140a tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] [instance: efa19aac-3ad7-42d5-a6d8-859050de6e63] Destroying instance {{(pid=61806) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 888.020743] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2f279f0c-52fd-4ca3-b33c-b1c025234c1a {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 888.760602] env[61806]: DEBUG nova.compute.utils [None req-0bb2f11a-fff8-4c85-aa76-4f260c30bde9 tempest-AttachVolumeNegativeTest-1683033704 tempest-AttachVolumeNegativeTest-1683033704-project-member] Using /dev/sd instead of None {{(pid=61806) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 888.767667] env[61806]: DEBUG nova.compute.manager [req-8d0aae47-5aba-452d-a53d-d2cb80d31716 req-69a8bca3-617b-4b55-b7c4-c25c80ad22e0 service nova] [instance: 1a73d994-5f21-4f80-8f33-5830d860a7ee] Received event network-changed-4017a509-57e9-4526-a107-3d54b7ed6684 {{(pid=61806) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 888.767868] env[61806]: DEBUG nova.compute.manager [req-8d0aae47-5aba-452d-a53d-d2cb80d31716 req-69a8bca3-617b-4b55-b7c4-c25c80ad22e0 service nova] [instance: 1a73d994-5f21-4f80-8f33-5830d860a7ee] Refreshing instance network info cache due to event network-changed-4017a509-57e9-4526-a107-3d54b7ed6684. {{(pid=61806) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 888.768095] env[61806]: DEBUG oslo_concurrency.lockutils [req-8d0aae47-5aba-452d-a53d-d2cb80d31716 req-69a8bca3-617b-4b55-b7c4-c25c80ad22e0 service nova] Acquiring lock "refresh_cache-1a73d994-5f21-4f80-8f33-5830d860a7ee" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 888.768252] env[61806]: DEBUG oslo_concurrency.lockutils [req-8d0aae47-5aba-452d-a53d-d2cb80d31716 req-69a8bca3-617b-4b55-b7c4-c25c80ad22e0 service nova] Acquired lock "refresh_cache-1a73d994-5f21-4f80-8f33-5830d860a7ee" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 888.768422] env[61806]: DEBUG nova.network.neutron [req-8d0aae47-5aba-452d-a53d-d2cb80d31716 req-69a8bca3-617b-4b55-b7c4-c25c80ad22e0 service nova] [instance: 1a73d994-5f21-4f80-8f33-5830d860a7ee] Refreshing network info cache for port 4017a509-57e9-4526-a107-3d54b7ed6684 {{(pid=61806) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 888.772809] env[61806]: DEBUG nova.compute.manager [None req-0bb2f11a-fff8-4c85-aa76-4f260c30bde9 tempest-AttachVolumeNegativeTest-1683033704 tempest-AttachVolumeNegativeTest-1683033704-project-member] [instance: daff849e-eb4f-411e-92f8-9362e3bd91e3] Allocating IP information in the background. {{(pid=61806) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 888.772910] env[61806]: DEBUG nova.network.neutron [None req-0bb2f11a-fff8-4c85-aa76-4f260c30bde9 tempest-AttachVolumeNegativeTest-1683033704 tempest-AttachVolumeNegativeTest-1683033704-project-member] [instance: daff849e-eb4f-411e-92f8-9362e3bd91e3] allocate_for_instance() {{(pid=61806) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 888.783220] env[61806]: DEBUG oslo_concurrency.lockutils [None req-6d4e5237-b28c-45e9-8ae0-8ed79347c566 tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Acquiring lock "75f1aded-d1da-4ddc-be29-f1fb93799364" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 888.783474] env[61806]: DEBUG oslo_concurrency.lockutils [None req-6d4e5237-b28c-45e9-8ae0-8ed79347c566 tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Lock "75f1aded-d1da-4ddc-be29-f1fb93799364" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 888.789386] env[61806]: INFO nova.compute.manager [None req-b85317d7-e79b-4ab3-83e4-2de701507449 tempest-ServerTagsTestJSON-1292926242 tempest-ServerTagsTestJSON-1292926242-project-member] [instance: 9c20d337-b8f5-4965-b707-0f87c4ab6fc4] Took 19.43 seconds to build instance. [ 888.796020] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [None req-f488cd5c-2a2f-4173-ba8b-c601bc6a140a tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] [instance: efa19aac-3ad7-42d5-a6d8-859050de6e63] Powering off the VM {{(pid=61806) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 888.796020] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-18f7f250-6069-4121-825e-6c9e4e571f21 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 888.803426] env[61806]: DEBUG oslo_vmware.api [-] Task: {'id': task-1294812, 'name': CreateVM_Task, 'duration_secs': 0.588106} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 888.803675] env[61806]: DEBUG oslo_vmware.api [None req-799f40df-a27b-4200-803a-166b5c1e29fd tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Task: {'id': task-1294809, 'name': PowerOnVM_Task, 'duration_secs': 1.334601} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 888.804165] env[61806]: DEBUG oslo_vmware.api [None req-c651ea58-49e6-4c69-a5f7-88000b663e10 tempest-ServerRescueTestJSONUnderV235-749414849 tempest-ServerRescueTestJSONUnderV235-749414849-project-member] Task: {'id': task-1294813, 'name': ReconfigVM_Task, 'duration_secs': 0.634648} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 888.804784] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 1a73d994-5f21-4f80-8f33-5830d860a7ee] Created VM on the ESX host {{(pid=61806) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 888.805047] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [None req-799f40df-a27b-4200-803a-166b5c1e29fd tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] [instance: 59c22c16-3d1c-4e64-bfac-7f8886a1927e] Powered on the VM {{(pid=61806) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 888.805270] env[61806]: INFO nova.compute.manager [None req-799f40df-a27b-4200-803a-166b5c1e29fd tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] [instance: 59c22c16-3d1c-4e64-bfac-7f8886a1927e] Took 11.84 seconds to spawn the instance on the hypervisor. [ 888.805457] env[61806]: DEBUG nova.compute.manager [None req-799f40df-a27b-4200-803a-166b5c1e29fd tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] [instance: 59c22c16-3d1c-4e64-bfac-7f8886a1927e] Checking state {{(pid=61806) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 888.805767] env[61806]: DEBUG nova.virt.vmwareapi.volumeops [None req-c651ea58-49e6-4c69-a5f7-88000b663e10 tempest-ServerRescueTestJSONUnderV235-749414849 tempest-ServerRescueTestJSONUnderV235-749414849-project-member] [instance: 5a526856-0dcd-4b3f-8359-13a5c3b9bfc4] Reconfigured VM instance instance-0000004d to attach disk [datastore2] 5a526856-0dcd-4b3f-8359-13a5c3b9bfc4/5a526856-0dcd-4b3f-8359-13a5c3b9bfc4.vmdk or device None with type sparse {{(pid=61806) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 888.806895] env[61806]: DEBUG oslo_concurrency.lockutils [None req-7e96f1c1-fc04-4010-b728-951cd26dd38b tempest-ImagesTestJSON-618852556 tempest-ImagesTestJSON-618852556-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 888.807065] env[61806]: DEBUG oslo_concurrency.lockutils [None req-7e96f1c1-fc04-4010-b728-951cd26dd38b tempest-ImagesTestJSON-618852556 tempest-ImagesTestJSON-618852556-project-member] Acquired lock "[datastore2] devstack-image-cache_base/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 888.807392] env[61806]: DEBUG oslo_concurrency.lockutils [None req-7e96f1c1-fc04-4010-b728-951cd26dd38b tempest-ImagesTestJSON-618852556 tempest-ImagesTestJSON-618852556-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 888.808223] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a2f2c88e-1ea3-4c57-957c-a6d995560919 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 888.810718] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-37e3b30b-1246-4120-97e7-a1e33ce9c95b {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 888.814298] env[61806]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d7533e1d-5662-4177-a2f1-bbd33867c753 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 888.816515] env[61806]: DEBUG oslo_vmware.api [None req-f488cd5c-2a2f-4173-ba8b-c601bc6a140a tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Waiting for the task: (returnval){ [ 888.816515] env[61806]: value = "task-1294814" [ 888.816515] env[61806]: _type = "Task" [ 888.816515] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 888.827317] env[61806]: DEBUG oslo_vmware.api [None req-7e96f1c1-fc04-4010-b728-951cd26dd38b tempest-ImagesTestJSON-618852556 tempest-ImagesTestJSON-618852556-project-member] Waiting for the task: (returnval){ [ 888.827317] env[61806]: value = "session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]525996f3-aea8-67cf-7beb-c5d62f454f0f" [ 888.827317] env[61806]: _type = "Task" [ 888.827317] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 888.830561] env[61806]: DEBUG oslo_vmware.api [None req-c651ea58-49e6-4c69-a5f7-88000b663e10 tempest-ServerRescueTestJSONUnderV235-749414849 tempest-ServerRescueTestJSONUnderV235-749414849-project-member] Waiting for the task: (returnval){ [ 888.830561] env[61806]: value = "task-1294815" [ 888.830561] env[61806]: _type = "Task" [ 888.830561] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 888.838997] env[61806]: DEBUG oslo_vmware.api [None req-f488cd5c-2a2f-4173-ba8b-c601bc6a140a tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Task: {'id': task-1294814, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 888.840873] env[61806]: DEBUG nova.policy [None req-0bb2f11a-fff8-4c85-aa76-4f260c30bde9 tempest-AttachVolumeNegativeTest-1683033704 tempest-AttachVolumeNegativeTest-1683033704-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '62b5e014cfeb4403a1edf89108c55e11', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '22e47882c2c44ce492dff6b1c5d782e2', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61806) authorize /opt/stack/nova/nova/policy.py:201}} [ 888.852897] env[61806]: DEBUG oslo_vmware.api [None req-c651ea58-49e6-4c69-a5f7-88000b663e10 tempest-ServerRescueTestJSONUnderV235-749414849 tempest-ServerRescueTestJSONUnderV235-749414849-project-member] Task: {'id': task-1294815, 'name': Rename_Task} progress is 14%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 888.853233] env[61806]: DEBUG oslo_vmware.api [None req-7e96f1c1-fc04-4010-b728-951cd26dd38b tempest-ImagesTestJSON-618852556 tempest-ImagesTestJSON-618852556-project-member] Task: {'id': session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]525996f3-aea8-67cf-7beb-c5d62f454f0f, 'name': SearchDatastore_Task, 'duration_secs': 0.011263} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 888.853496] env[61806]: DEBUG oslo_concurrency.lockutils [None req-7e96f1c1-fc04-4010-b728-951cd26dd38b tempest-ImagesTestJSON-618852556 tempest-ImagesTestJSON-618852556-project-member] Releasing lock "[datastore2] devstack-image-cache_base/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 888.853759] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-7e96f1c1-fc04-4010-b728-951cd26dd38b tempest-ImagesTestJSON-618852556 tempest-ImagesTestJSON-618852556-project-member] [instance: 1a73d994-5f21-4f80-8f33-5830d860a7ee] Processing image ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa {{(pid=61806) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 888.853999] env[61806]: DEBUG oslo_concurrency.lockutils [None req-7e96f1c1-fc04-4010-b728-951cd26dd38b tempest-ImagesTestJSON-618852556 tempest-ImagesTestJSON-618852556-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa.vmdk" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 888.855558] env[61806]: DEBUG oslo_concurrency.lockutils [None req-7e96f1c1-fc04-4010-b728-951cd26dd38b tempest-ImagesTestJSON-618852556 tempest-ImagesTestJSON-618852556-project-member] Acquired lock "[datastore2] devstack-image-cache_base/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa.vmdk" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 888.855558] env[61806]: DEBUG nova.virt.vmwareapi.ds_util [None req-7e96f1c1-fc04-4010-b728-951cd26dd38b tempest-ImagesTestJSON-618852556 tempest-ImagesTestJSON-618852556-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61806) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 888.855558] env[61806]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-f530ec02-42ec-4235-8ff7-bea1fc0ea8d7 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 888.863877] env[61806]: DEBUG nova.virt.vmwareapi.ds_util [None req-7e96f1c1-fc04-4010-b728-951cd26dd38b tempest-ImagesTestJSON-618852556 tempest-ImagesTestJSON-618852556-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61806) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 888.864070] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-7e96f1c1-fc04-4010-b728-951cd26dd38b tempest-ImagesTestJSON-618852556 tempest-ImagesTestJSON-618852556-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61806) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 888.866563] env[61806]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-41a31ea9-fe95-427f-9c20-b8ff823cf253 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 888.871405] env[61806]: DEBUG oslo_vmware.api [None req-7e96f1c1-fc04-4010-b728-951cd26dd38b tempest-ImagesTestJSON-618852556 tempest-ImagesTestJSON-618852556-project-member] Waiting for the task: (returnval){ [ 888.871405] env[61806]: value = "session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]5214fac7-50bc-a170-3c3b-d42cb9c36f1c" [ 888.871405] env[61806]: _type = "Task" [ 888.871405] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 888.880151] env[61806]: DEBUG oslo_vmware.api [None req-7e96f1c1-fc04-4010-b728-951cd26dd38b tempest-ImagesTestJSON-618852556 tempest-ImagesTestJSON-618852556-project-member] Task: {'id': session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]5214fac7-50bc-a170-3c3b-d42cb9c36f1c, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 889.138259] env[61806]: DEBUG nova.network.neutron [None req-0bb2f11a-fff8-4c85-aa76-4f260c30bde9 tempest-AttachVolumeNegativeTest-1683033704 tempest-AttachVolumeNegativeTest-1683033704-project-member] [instance: daff849e-eb4f-411e-92f8-9362e3bd91e3] Successfully created port: 9fdf6139-082b-49e0-8ce8-7870749b2c7a {{(pid=61806) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 889.276685] env[61806]: DEBUG nova.compute.manager [None req-0bb2f11a-fff8-4c85-aa76-4f260c30bde9 tempest-AttachVolumeNegativeTest-1683033704 tempest-AttachVolumeNegativeTest-1683033704-project-member] [instance: daff849e-eb4f-411e-92f8-9362e3bd91e3] Start building block device mappings for instance. {{(pid=61806) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 889.291318] env[61806]: DEBUG nova.compute.manager [None req-6d4e5237-b28c-45e9-8ae0-8ed79347c566 tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] [instance: 75f1aded-d1da-4ddc-be29-f1fb93799364] Starting instance... {{(pid=61806) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 889.294539] env[61806]: DEBUG oslo_concurrency.lockutils [None req-b85317d7-e79b-4ab3-83e4-2de701507449 tempest-ServerTagsTestJSON-1292926242 tempest-ServerTagsTestJSON-1292926242-project-member] Lock "9c20d337-b8f5-4965-b707-0f87c4ab6fc4" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 20.949s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 889.336715] env[61806]: DEBUG oslo_vmware.api [None req-f488cd5c-2a2f-4173-ba8b-c601bc6a140a tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Task: {'id': task-1294814, 'name': PowerOffVM_Task, 'duration_secs': 0.199555} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 889.341020] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [None req-f488cd5c-2a2f-4173-ba8b-c601bc6a140a tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] [instance: efa19aac-3ad7-42d5-a6d8-859050de6e63] Powered off the VM {{(pid=61806) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 889.341020] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-f488cd5c-2a2f-4173-ba8b-c601bc6a140a tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] [instance: efa19aac-3ad7-42d5-a6d8-859050de6e63] Unregistering the VM {{(pid=61806) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 889.345347] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-5dfd6986-7ad1-4add-9895-51685038b387 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 889.349020] env[61806]: INFO nova.compute.manager [None req-799f40df-a27b-4200-803a-166b5c1e29fd tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] [instance: 59c22c16-3d1c-4e64-bfac-7f8886a1927e] Took 22.62 seconds to build instance. [ 889.355059] env[61806]: DEBUG oslo_vmware.api [None req-c651ea58-49e6-4c69-a5f7-88000b663e10 tempest-ServerRescueTestJSONUnderV235-749414849 tempest-ServerRescueTestJSONUnderV235-749414849-project-member] Task: {'id': task-1294815, 'name': Rename_Task, 'duration_secs': 0.204056} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 889.355328] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [None req-c651ea58-49e6-4c69-a5f7-88000b663e10 tempest-ServerRescueTestJSONUnderV235-749414849 tempest-ServerRescueTestJSONUnderV235-749414849-project-member] [instance: 5a526856-0dcd-4b3f-8359-13a5c3b9bfc4] Powering on the VM {{(pid=61806) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 889.355565] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-cfd947ec-b010-4c57-8659-9bfb560bf246 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 889.367264] env[61806]: DEBUG oslo_vmware.api [None req-c651ea58-49e6-4c69-a5f7-88000b663e10 tempest-ServerRescueTestJSONUnderV235-749414849 tempest-ServerRescueTestJSONUnderV235-749414849-project-member] Waiting for the task: (returnval){ [ 889.367264] env[61806]: value = "task-1294817" [ 889.367264] env[61806]: _type = "Task" [ 889.367264] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 889.381811] env[61806]: DEBUG oslo_vmware.api [None req-c651ea58-49e6-4c69-a5f7-88000b663e10 tempest-ServerRescueTestJSONUnderV235-749414849 tempest-ServerRescueTestJSONUnderV235-749414849-project-member] Task: {'id': task-1294817, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 889.387607] env[61806]: DEBUG oslo_vmware.api [None req-7e96f1c1-fc04-4010-b728-951cd26dd38b tempest-ImagesTestJSON-618852556 tempest-ImagesTestJSON-618852556-project-member] Task: {'id': session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]5214fac7-50bc-a170-3c3b-d42cb9c36f1c, 'name': SearchDatastore_Task, 'duration_secs': 0.009423} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 889.388600] env[61806]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-8c67ed89-f15e-434f-a22a-56da7d960f29 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 889.397366] env[61806]: DEBUG oslo_vmware.api [None req-7e96f1c1-fc04-4010-b728-951cd26dd38b tempest-ImagesTestJSON-618852556 tempest-ImagesTestJSON-618852556-project-member] Waiting for the task: (returnval){ [ 889.397366] env[61806]: value = "session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]528c9bff-dbce-0aea-edb1-e496a4969366" [ 889.397366] env[61806]: _type = "Task" [ 889.397366] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 889.410718] env[61806]: DEBUG oslo_vmware.api [None req-7e96f1c1-fc04-4010-b728-951cd26dd38b tempest-ImagesTestJSON-618852556 tempest-ImagesTestJSON-618852556-project-member] Task: {'id': session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]528c9bff-dbce-0aea-edb1-e496a4969366, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 889.430709] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-f488cd5c-2a2f-4173-ba8b-c601bc6a140a tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] [instance: efa19aac-3ad7-42d5-a6d8-859050de6e63] Unregistered the VM {{(pid=61806) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 889.430951] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-f488cd5c-2a2f-4173-ba8b-c601bc6a140a tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] [instance: efa19aac-3ad7-42d5-a6d8-859050de6e63] Deleting contents of the VM from datastore datastore2 {{(pid=61806) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 889.431203] env[61806]: DEBUG nova.virt.vmwareapi.ds_util [None req-f488cd5c-2a2f-4173-ba8b-c601bc6a140a tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Deleting the datastore file [datastore2] efa19aac-3ad7-42d5-a6d8-859050de6e63 {{(pid=61806) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 889.431550] env[61806]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-f02ed629-f2dd-409d-b205-4c9fec5250b6 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 889.440187] env[61806]: DEBUG oslo_vmware.api [None req-f488cd5c-2a2f-4173-ba8b-c601bc6a140a tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Waiting for the task: (returnval){ [ 889.440187] env[61806]: value = "task-1294818" [ 889.440187] env[61806]: _type = "Task" [ 889.440187] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 889.451873] env[61806]: DEBUG oslo_vmware.api [None req-f488cd5c-2a2f-4173-ba8b-c601bc6a140a tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Task: {'id': task-1294818, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 889.507732] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5d5d710f-dec7-4335-a297-87f43cc7557d {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 889.516165] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6c0b57a4-aa27-4d8f-ab0f-b4d6f7f31acc {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 889.552096] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fea06263-51fd-4909-b4d2-af8e3ec60a8f {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 889.561217] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-644e5c1e-46a5-43f7-b02d-071ca5ef5bca {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 889.575887] env[61806]: DEBUG nova.compute.provider_tree [None req-095bcbe6-2af7-4761-a8ad-c73adfe5ab9d tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] Inventory has not changed in ProviderTree for provider: a2858be1-fd22-4e08-979e-87ad25293407 {{(pid=61806) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 889.682042] env[61806]: DEBUG nova.network.neutron [req-8d0aae47-5aba-452d-a53d-d2cb80d31716 req-69a8bca3-617b-4b55-b7c4-c25c80ad22e0 service nova] [instance: 1a73d994-5f21-4f80-8f33-5830d860a7ee] Updated VIF entry in instance network info cache for port 4017a509-57e9-4526-a107-3d54b7ed6684. {{(pid=61806) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 889.682624] env[61806]: DEBUG nova.network.neutron [req-8d0aae47-5aba-452d-a53d-d2cb80d31716 req-69a8bca3-617b-4b55-b7c4-c25c80ad22e0 service nova] [instance: 1a73d994-5f21-4f80-8f33-5830d860a7ee] Updating instance_info_cache with network_info: [{"id": "4017a509-57e9-4526-a107-3d54b7ed6684", "address": "fa:16:3e:40:38:fb", "network": {"id": "288ebc87-a30f-439b-9885-9cdda68a9591", "bridge": "br-int", "label": "tempest-ImagesTestJSON-914411285-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4b9aa91aff1d4008ac5096902b77f852", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8cbc9b8f-ce19-4262-bf4d-88cd4f259a1c", "external-id": "nsx-vlan-transportzone-630", "segmentation_id": 630, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4017a509-57", "ovs_interfaceid": "4017a509-57e9-4526-a107-3d54b7ed6684", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61806) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 889.814082] env[61806]: DEBUG oslo_concurrency.lockutils [None req-6d4e5237-b28c-45e9-8ae0-8ed79347c566 tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 889.849765] env[61806]: DEBUG oslo_concurrency.lockutils [None req-799f40df-a27b-4200-803a-166b5c1e29fd tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Lock "59c22c16-3d1c-4e64-bfac-7f8886a1927e" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 24.137s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 889.878633] env[61806]: DEBUG oslo_vmware.api [None req-c651ea58-49e6-4c69-a5f7-88000b663e10 tempest-ServerRescueTestJSONUnderV235-749414849 tempest-ServerRescueTestJSONUnderV235-749414849-project-member] Task: {'id': task-1294817, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 889.907233] env[61806]: DEBUG oslo_vmware.api [None req-7e96f1c1-fc04-4010-b728-951cd26dd38b tempest-ImagesTestJSON-618852556 tempest-ImagesTestJSON-618852556-project-member] Task: {'id': session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]528c9bff-dbce-0aea-edb1-e496a4969366, 'name': SearchDatastore_Task, 'duration_secs': 0.014455} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 889.907519] env[61806]: DEBUG oslo_concurrency.lockutils [None req-7e96f1c1-fc04-4010-b728-951cd26dd38b tempest-ImagesTestJSON-618852556 tempest-ImagesTestJSON-618852556-project-member] Releasing lock "[datastore2] devstack-image-cache_base/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa.vmdk" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 889.907777] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [None req-7e96f1c1-fc04-4010-b728-951cd26dd38b tempest-ImagesTestJSON-618852556 tempest-ImagesTestJSON-618852556-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa.vmdk to [datastore2] 1a73d994-5f21-4f80-8f33-5830d860a7ee/1a73d994-5f21-4f80-8f33-5830d860a7ee.vmdk {{(pid=61806) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 889.908052] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-a1804a31-4161-47f9-8abf-70a821729b91 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 889.915377] env[61806]: DEBUG oslo_vmware.api [None req-7e96f1c1-fc04-4010-b728-951cd26dd38b tempest-ImagesTestJSON-618852556 tempest-ImagesTestJSON-618852556-project-member] Waiting for the task: (returnval){ [ 889.915377] env[61806]: value = "task-1294819" [ 889.915377] env[61806]: _type = "Task" [ 889.915377] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 889.923953] env[61806]: DEBUG oslo_vmware.api [None req-7e96f1c1-fc04-4010-b728-951cd26dd38b tempest-ImagesTestJSON-618852556 tempest-ImagesTestJSON-618852556-project-member] Task: {'id': task-1294819, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 889.953462] env[61806]: DEBUG oslo_vmware.api [None req-f488cd5c-2a2f-4173-ba8b-c601bc6a140a tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Task: {'id': task-1294818, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.227495} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 889.953692] env[61806]: DEBUG nova.virt.vmwareapi.ds_util [None req-f488cd5c-2a2f-4173-ba8b-c601bc6a140a tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Deleted the datastore file {{(pid=61806) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 889.953892] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-f488cd5c-2a2f-4173-ba8b-c601bc6a140a tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] [instance: efa19aac-3ad7-42d5-a6d8-859050de6e63] Deleted contents of the VM from datastore datastore2 {{(pid=61806) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 889.954087] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-f488cd5c-2a2f-4173-ba8b-c601bc6a140a tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] [instance: efa19aac-3ad7-42d5-a6d8-859050de6e63] Instance destroyed {{(pid=61806) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 889.954281] env[61806]: INFO nova.compute.manager [None req-f488cd5c-2a2f-4173-ba8b-c601bc6a140a tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] [instance: efa19aac-3ad7-42d5-a6d8-859050de6e63] Took 1.93 seconds to destroy the instance on the hypervisor. [ 889.954534] env[61806]: DEBUG oslo.service.loopingcall [None req-f488cd5c-2a2f-4173-ba8b-c601bc6a140a tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61806) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 889.954758] env[61806]: DEBUG nova.compute.manager [-] [instance: efa19aac-3ad7-42d5-a6d8-859050de6e63] Deallocating network for instance {{(pid=61806) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 889.954851] env[61806]: DEBUG nova.network.neutron [-] [instance: efa19aac-3ad7-42d5-a6d8-859050de6e63] deallocate_for_instance() {{(pid=61806) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 890.078676] env[61806]: DEBUG nova.scheduler.client.report [None req-095bcbe6-2af7-4761-a8ad-c73adfe5ab9d tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] Inventory has not changed for provider a2858be1-fd22-4e08-979e-87ad25293407 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 138, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61806) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 890.187163] env[61806]: DEBUG oslo_concurrency.lockutils [req-8d0aae47-5aba-452d-a53d-d2cb80d31716 req-69a8bca3-617b-4b55-b7c4-c25c80ad22e0 service nova] Releasing lock "refresh_cache-1a73d994-5f21-4f80-8f33-5830d860a7ee" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 890.248650] env[61806]: DEBUG oslo_concurrency.lockutils [None req-18f7d4e8-522e-4e97-b6f4-8662acf8d0f8 tempest-ServerTagsTestJSON-1292926242 tempest-ServerTagsTestJSON-1292926242-project-member] Acquiring lock "9c20d337-b8f5-4965-b707-0f87c4ab6fc4" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 890.249022] env[61806]: DEBUG oslo_concurrency.lockutils [None req-18f7d4e8-522e-4e97-b6f4-8662acf8d0f8 tempest-ServerTagsTestJSON-1292926242 tempest-ServerTagsTestJSON-1292926242-project-member] Lock "9c20d337-b8f5-4965-b707-0f87c4ab6fc4" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 890.249265] env[61806]: DEBUG oslo_concurrency.lockutils [None req-18f7d4e8-522e-4e97-b6f4-8662acf8d0f8 tempest-ServerTagsTestJSON-1292926242 tempest-ServerTagsTestJSON-1292926242-project-member] Acquiring lock "9c20d337-b8f5-4965-b707-0f87c4ab6fc4-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 890.249468] env[61806]: DEBUG oslo_concurrency.lockutils [None req-18f7d4e8-522e-4e97-b6f4-8662acf8d0f8 tempest-ServerTagsTestJSON-1292926242 tempest-ServerTagsTestJSON-1292926242-project-member] Lock "9c20d337-b8f5-4965-b707-0f87c4ab6fc4-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 890.249671] env[61806]: DEBUG oslo_concurrency.lockutils [None req-18f7d4e8-522e-4e97-b6f4-8662acf8d0f8 tempest-ServerTagsTestJSON-1292926242 tempest-ServerTagsTestJSON-1292926242-project-member] Lock "9c20d337-b8f5-4965-b707-0f87c4ab6fc4-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 890.254919] env[61806]: INFO nova.compute.manager [None req-18f7d4e8-522e-4e97-b6f4-8662acf8d0f8 tempest-ServerTagsTestJSON-1292926242 tempest-ServerTagsTestJSON-1292926242-project-member] [instance: 9c20d337-b8f5-4965-b707-0f87c4ab6fc4] Terminating instance [ 890.258200] env[61806]: DEBUG nova.compute.manager [None req-18f7d4e8-522e-4e97-b6f4-8662acf8d0f8 tempest-ServerTagsTestJSON-1292926242 tempest-ServerTagsTestJSON-1292926242-project-member] [instance: 9c20d337-b8f5-4965-b707-0f87c4ab6fc4] Start destroying the instance on the hypervisor. {{(pid=61806) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 890.258835] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-18f7d4e8-522e-4e97-b6f4-8662acf8d0f8 tempest-ServerTagsTestJSON-1292926242 tempest-ServerTagsTestJSON-1292926242-project-member] [instance: 9c20d337-b8f5-4965-b707-0f87c4ab6fc4] Destroying instance {{(pid=61806) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 890.259720] env[61806]: DEBUG nova.compute.manager [req-9f71b756-9421-45ee-9ba2-44ed3c0ed532 req-eeab4892-964b-4bf3-9a68-19412c2765f9 service nova] [instance: efa19aac-3ad7-42d5-a6d8-859050de6e63] Received event network-vif-deleted-fea0c8d9-43ab-4462-9427-73d9ae4a47b2 {{(pid=61806) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 890.259943] env[61806]: INFO nova.compute.manager [req-9f71b756-9421-45ee-9ba2-44ed3c0ed532 req-eeab4892-964b-4bf3-9a68-19412c2765f9 service nova] [instance: efa19aac-3ad7-42d5-a6d8-859050de6e63] Neutron deleted interface fea0c8d9-43ab-4462-9427-73d9ae4a47b2; detaching it from the instance and deleting it from the info cache [ 890.260165] env[61806]: DEBUG nova.network.neutron [req-9f71b756-9421-45ee-9ba2-44ed3c0ed532 req-eeab4892-964b-4bf3-9a68-19412c2765f9 service nova] [instance: efa19aac-3ad7-42d5-a6d8-859050de6e63] Updating instance_info_cache with network_info: [] {{(pid=61806) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 890.262041] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-943ed673-0258-4a0d-b33a-376d45f2dd8e {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 890.272117] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [None req-18f7d4e8-522e-4e97-b6f4-8662acf8d0f8 tempest-ServerTagsTestJSON-1292926242 tempest-ServerTagsTestJSON-1292926242-project-member] [instance: 9c20d337-b8f5-4965-b707-0f87c4ab6fc4] Powering off the VM {{(pid=61806) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 890.272992] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-668329f0-63b7-4238-9363-6b07f4765b77 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 890.281825] env[61806]: DEBUG oslo_vmware.api [None req-18f7d4e8-522e-4e97-b6f4-8662acf8d0f8 tempest-ServerTagsTestJSON-1292926242 tempest-ServerTagsTestJSON-1292926242-project-member] Waiting for the task: (returnval){ [ 890.281825] env[61806]: value = "task-1294820" [ 890.281825] env[61806]: _type = "Task" [ 890.281825] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 890.287838] env[61806]: DEBUG nova.compute.manager [None req-0bb2f11a-fff8-4c85-aa76-4f260c30bde9 tempest-AttachVolumeNegativeTest-1683033704 tempest-AttachVolumeNegativeTest-1683033704-project-member] [instance: daff849e-eb4f-411e-92f8-9362e3bd91e3] Start spawning the instance on the hypervisor. {{(pid=61806) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 890.297202] env[61806]: DEBUG oslo_vmware.api [None req-18f7d4e8-522e-4e97-b6f4-8662acf8d0f8 tempest-ServerTagsTestJSON-1292926242 tempest-ServerTagsTestJSON-1292926242-project-member] Task: {'id': task-1294820, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 890.317604] env[61806]: DEBUG nova.virt.hardware [None req-0bb2f11a-fff8-4c85-aa76-4f260c30bde9 tempest-AttachVolumeNegativeTest-1683033704 tempest-AttachVolumeNegativeTest-1683033704-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-15T18:05:10Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-15T18:04:52Z,direct_url=,disk_format='vmdk',id=ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='84ba83fce2c349a6988173c5d6fc3b07',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-15T18:04:53Z,virtual_size=,visibility=), allow threads: False {{(pid=61806) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 890.318110] env[61806]: DEBUG nova.virt.hardware [None req-0bb2f11a-fff8-4c85-aa76-4f260c30bde9 tempest-AttachVolumeNegativeTest-1683033704 tempest-AttachVolumeNegativeTest-1683033704-project-member] Flavor limits 0:0:0 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 890.318339] env[61806]: DEBUG nova.virt.hardware [None req-0bb2f11a-fff8-4c85-aa76-4f260c30bde9 tempest-AttachVolumeNegativeTest-1683033704 tempest-AttachVolumeNegativeTest-1683033704-project-member] Image limits 0:0:0 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 890.318594] env[61806]: DEBUG nova.virt.hardware [None req-0bb2f11a-fff8-4c85-aa76-4f260c30bde9 tempest-AttachVolumeNegativeTest-1683033704 tempest-AttachVolumeNegativeTest-1683033704-project-member] Flavor pref 0:0:0 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 890.318886] env[61806]: DEBUG nova.virt.hardware [None req-0bb2f11a-fff8-4c85-aa76-4f260c30bde9 tempest-AttachVolumeNegativeTest-1683033704 tempest-AttachVolumeNegativeTest-1683033704-project-member] Image pref 0:0:0 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 890.319205] env[61806]: DEBUG nova.virt.hardware [None req-0bb2f11a-fff8-4c85-aa76-4f260c30bde9 tempest-AttachVolumeNegativeTest-1683033704 tempest-AttachVolumeNegativeTest-1683033704-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 890.319583] env[61806]: DEBUG nova.virt.hardware [None req-0bb2f11a-fff8-4c85-aa76-4f260c30bde9 tempest-AttachVolumeNegativeTest-1683033704 tempest-AttachVolumeNegativeTest-1683033704-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61806) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 890.319913] env[61806]: DEBUG nova.virt.hardware [None req-0bb2f11a-fff8-4c85-aa76-4f260c30bde9 tempest-AttachVolumeNegativeTest-1683033704 tempest-AttachVolumeNegativeTest-1683033704-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61806) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 890.320206] env[61806]: DEBUG nova.virt.hardware [None req-0bb2f11a-fff8-4c85-aa76-4f260c30bde9 tempest-AttachVolumeNegativeTest-1683033704 tempest-AttachVolumeNegativeTest-1683033704-project-member] Got 1 possible topologies {{(pid=61806) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 890.320485] env[61806]: DEBUG nova.virt.hardware [None req-0bb2f11a-fff8-4c85-aa76-4f260c30bde9 tempest-AttachVolumeNegativeTest-1683033704 tempest-AttachVolumeNegativeTest-1683033704-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61806) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 890.320795] env[61806]: DEBUG nova.virt.hardware [None req-0bb2f11a-fff8-4c85-aa76-4f260c30bde9 tempest-AttachVolumeNegativeTest-1683033704 tempest-AttachVolumeNegativeTest-1683033704-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61806) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 890.322215] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-36793dab-1c16-4a1a-a1aa-1463fe8f7400 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 890.334345] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7e1c4719-d9da-41ce-9f9a-436fccf37372 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 890.380467] env[61806]: DEBUG oslo_vmware.api [None req-c651ea58-49e6-4c69-a5f7-88000b663e10 tempest-ServerRescueTestJSONUnderV235-749414849 tempest-ServerRescueTestJSONUnderV235-749414849-project-member] Task: {'id': task-1294817, 'name': PowerOnVM_Task, 'duration_secs': 0.725925} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 890.380847] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [None req-c651ea58-49e6-4c69-a5f7-88000b663e10 tempest-ServerRescueTestJSONUnderV235-749414849 tempest-ServerRescueTestJSONUnderV235-749414849-project-member] [instance: 5a526856-0dcd-4b3f-8359-13a5c3b9bfc4] Powered on the VM {{(pid=61806) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 890.381159] env[61806]: INFO nova.compute.manager [None req-c651ea58-49e6-4c69-a5f7-88000b663e10 tempest-ServerRescueTestJSONUnderV235-749414849 tempest-ServerRescueTestJSONUnderV235-749414849-project-member] [instance: 5a526856-0dcd-4b3f-8359-13a5c3b9bfc4] Took 8.88 seconds to spawn the instance on the hypervisor. [ 890.381353] env[61806]: DEBUG nova.compute.manager [None req-c651ea58-49e6-4c69-a5f7-88000b663e10 tempest-ServerRescueTestJSONUnderV235-749414849 tempest-ServerRescueTestJSONUnderV235-749414849-project-member] [instance: 5a526856-0dcd-4b3f-8359-13a5c3b9bfc4] Checking state {{(pid=61806) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 890.382289] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b0130831-8daa-4ac2-a844-3a1b4011d8de {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 890.427847] env[61806]: DEBUG oslo_vmware.api [None req-7e96f1c1-fc04-4010-b728-951cd26dd38b tempest-ImagesTestJSON-618852556 tempest-ImagesTestJSON-618852556-project-member] Task: {'id': task-1294819, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 890.586123] env[61806]: DEBUG oslo_concurrency.lockutils [None req-095bcbe6-2af7-4761-a8ad-c73adfe5ab9d tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 3.015s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 890.586123] env[61806]: DEBUG nova.compute.manager [None req-095bcbe6-2af7-4761-a8ad-c73adfe5ab9d tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] [instance: 6f6ba57a-e2d6-4749-a53a-e263861cb1c0] Start building networks asynchronously for instance. {{(pid=61806) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 890.589547] env[61806]: DEBUG oslo_concurrency.lockutils [None req-58b4b995-fdfd-465f-8927-98666a556c7c tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: waited 3.415s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 890.589906] env[61806]: DEBUG nova.objects.instance [None req-58b4b995-fdfd-465f-8927-98666a556c7c tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] [instance: efa19aac-3ad7-42d5-a6d8-859050de6e63] Trying to apply a migration context that does not seem to be set for this instance {{(pid=61806) apply_migration_context /opt/stack/nova/nova/objects/instance.py:1067}} [ 890.593690] env[61806]: DEBUG nova.compute.manager [req-3443fe33-7b41-42ed-9179-435bd9b4de5f req-12d692ef-c473-49b6-bab5-267fa6a72f32 service nova] [instance: daff849e-eb4f-411e-92f8-9362e3bd91e3] Received event network-vif-plugged-9fdf6139-082b-49e0-8ce8-7870749b2c7a {{(pid=61806) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 890.594064] env[61806]: DEBUG oslo_concurrency.lockutils [req-3443fe33-7b41-42ed-9179-435bd9b4de5f req-12d692ef-c473-49b6-bab5-267fa6a72f32 service nova] Acquiring lock "daff849e-eb4f-411e-92f8-9362e3bd91e3-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 890.594418] env[61806]: DEBUG oslo_concurrency.lockutils [req-3443fe33-7b41-42ed-9179-435bd9b4de5f req-12d692ef-c473-49b6-bab5-267fa6a72f32 service nova] Lock "daff849e-eb4f-411e-92f8-9362e3bd91e3-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 890.594712] env[61806]: DEBUG oslo_concurrency.lockutils [req-3443fe33-7b41-42ed-9179-435bd9b4de5f req-12d692ef-c473-49b6-bab5-267fa6a72f32 service nova] Lock "daff849e-eb4f-411e-92f8-9362e3bd91e3-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 890.594988] env[61806]: DEBUG nova.compute.manager [req-3443fe33-7b41-42ed-9179-435bd9b4de5f req-12d692ef-c473-49b6-bab5-267fa6a72f32 service nova] [instance: daff849e-eb4f-411e-92f8-9362e3bd91e3] No waiting events found dispatching network-vif-plugged-9fdf6139-082b-49e0-8ce8-7870749b2c7a {{(pid=61806) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 890.595294] env[61806]: WARNING nova.compute.manager [req-3443fe33-7b41-42ed-9179-435bd9b4de5f req-12d692ef-c473-49b6-bab5-267fa6a72f32 service nova] [instance: daff849e-eb4f-411e-92f8-9362e3bd91e3] Received unexpected event network-vif-plugged-9fdf6139-082b-49e0-8ce8-7870749b2c7a for instance with vm_state building and task_state spawning. [ 890.728866] env[61806]: DEBUG nova.network.neutron [-] [instance: efa19aac-3ad7-42d5-a6d8-859050de6e63] Updating instance_info_cache with network_info: [] {{(pid=61806) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 890.767578] env[61806]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-786774d4-2170-436a-8ea9-66a8a214c374 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 890.778664] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-583ba01b-0da3-40c5-9cd7-e3b0eb7c8500 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 890.797939] env[61806]: DEBUG oslo_vmware.api [None req-18f7d4e8-522e-4e97-b6f4-8662acf8d0f8 tempest-ServerTagsTestJSON-1292926242 tempest-ServerTagsTestJSON-1292926242-project-member] Task: {'id': task-1294820, 'name': PowerOffVM_Task, 'duration_secs': 0.321249} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 890.798244] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [None req-18f7d4e8-522e-4e97-b6f4-8662acf8d0f8 tempest-ServerTagsTestJSON-1292926242 tempest-ServerTagsTestJSON-1292926242-project-member] [instance: 9c20d337-b8f5-4965-b707-0f87c4ab6fc4] Powered off the VM {{(pid=61806) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 890.798710] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-18f7d4e8-522e-4e97-b6f4-8662acf8d0f8 tempest-ServerTagsTestJSON-1292926242 tempest-ServerTagsTestJSON-1292926242-project-member] [instance: 9c20d337-b8f5-4965-b707-0f87c4ab6fc4] Unregistering the VM {{(pid=61806) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 890.799617] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-f5bd917f-0a81-4849-aa2b-e500ef4a0d00 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 890.803050] env[61806]: DEBUG oslo_concurrency.lockutils [None req-76aa021b-a2c9-4c97-bbb8-a4fc0140ddae tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Acquiring lock "59c22c16-3d1c-4e64-bfac-7f8886a1927e" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 890.803341] env[61806]: DEBUG oslo_concurrency.lockutils [None req-76aa021b-a2c9-4c97-bbb8-a4fc0140ddae tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Lock "59c22c16-3d1c-4e64-bfac-7f8886a1927e" acquired by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: waited 0.000s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 890.812061] env[61806]: DEBUG nova.compute.manager [req-9f71b756-9421-45ee-9ba2-44ed3c0ed532 req-eeab4892-964b-4bf3-9a68-19412c2765f9 service nova] [instance: efa19aac-3ad7-42d5-a6d8-859050de6e63] Detach interface failed, port_id=fea0c8d9-43ab-4462-9427-73d9ae4a47b2, reason: Instance efa19aac-3ad7-42d5-a6d8-859050de6e63 could not be found. {{(pid=61806) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10963}} [ 890.885941] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-18f7d4e8-522e-4e97-b6f4-8662acf8d0f8 tempest-ServerTagsTestJSON-1292926242 tempest-ServerTagsTestJSON-1292926242-project-member] [instance: 9c20d337-b8f5-4965-b707-0f87c4ab6fc4] Unregistered the VM {{(pid=61806) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 890.886237] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-18f7d4e8-522e-4e97-b6f4-8662acf8d0f8 tempest-ServerTagsTestJSON-1292926242 tempest-ServerTagsTestJSON-1292926242-project-member] [instance: 9c20d337-b8f5-4965-b707-0f87c4ab6fc4] Deleting contents of the VM from datastore datastore2 {{(pid=61806) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 890.886442] env[61806]: DEBUG nova.virt.vmwareapi.ds_util [None req-18f7d4e8-522e-4e97-b6f4-8662acf8d0f8 tempest-ServerTagsTestJSON-1292926242 tempest-ServerTagsTestJSON-1292926242-project-member] Deleting the datastore file [datastore2] 9c20d337-b8f5-4965-b707-0f87c4ab6fc4 {{(pid=61806) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 890.886727] env[61806]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-9dc23c37-c625-40f2-8a85-f7488ae3a354 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 890.893781] env[61806]: DEBUG oslo_vmware.api [None req-18f7d4e8-522e-4e97-b6f4-8662acf8d0f8 tempest-ServerTagsTestJSON-1292926242 tempest-ServerTagsTestJSON-1292926242-project-member] Waiting for the task: (returnval){ [ 890.893781] env[61806]: value = "task-1294822" [ 890.893781] env[61806]: _type = "Task" [ 890.893781] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 890.906579] env[61806]: DEBUG oslo_vmware.api [None req-18f7d4e8-522e-4e97-b6f4-8662acf8d0f8 tempest-ServerTagsTestJSON-1292926242 tempest-ServerTagsTestJSON-1292926242-project-member] Task: {'id': task-1294822, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 890.907102] env[61806]: INFO nova.compute.manager [None req-c651ea58-49e6-4c69-a5f7-88000b663e10 tempest-ServerRescueTestJSONUnderV235-749414849 tempest-ServerRescueTestJSONUnderV235-749414849-project-member] [instance: 5a526856-0dcd-4b3f-8359-13a5c3b9bfc4] Took 20.88 seconds to build instance. [ 890.927514] env[61806]: DEBUG oslo_vmware.api [None req-7e96f1c1-fc04-4010-b728-951cd26dd38b tempest-ImagesTestJSON-618852556 tempest-ImagesTestJSON-618852556-project-member] Task: {'id': task-1294819, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.645188} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 890.927791] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [None req-7e96f1c1-fc04-4010-b728-951cd26dd38b tempest-ImagesTestJSON-618852556 tempest-ImagesTestJSON-618852556-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa.vmdk to [datastore2] 1a73d994-5f21-4f80-8f33-5830d860a7ee/1a73d994-5f21-4f80-8f33-5830d860a7ee.vmdk {{(pid=61806) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 890.928011] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-7e96f1c1-fc04-4010-b728-951cd26dd38b tempest-ImagesTestJSON-618852556 tempest-ImagesTestJSON-618852556-project-member] [instance: 1a73d994-5f21-4f80-8f33-5830d860a7ee] Extending root virtual disk to 1048576 {{(pid=61806) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 890.928279] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-f64822f9-5d81-4cb4-8cca-0381eb091939 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 890.936799] env[61806]: DEBUG oslo_vmware.api [None req-7e96f1c1-fc04-4010-b728-951cd26dd38b tempest-ImagesTestJSON-618852556 tempest-ImagesTestJSON-618852556-project-member] Waiting for the task: (returnval){ [ 890.936799] env[61806]: value = "task-1294823" [ 890.936799] env[61806]: _type = "Task" [ 890.936799] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 890.947968] env[61806]: DEBUG oslo_vmware.api [None req-7e96f1c1-fc04-4010-b728-951cd26dd38b tempest-ImagesTestJSON-618852556 tempest-ImagesTestJSON-618852556-project-member] Task: {'id': task-1294823, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 891.097945] env[61806]: DEBUG nova.compute.utils [None req-095bcbe6-2af7-4761-a8ad-c73adfe5ab9d tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] Using /dev/sd instead of None {{(pid=61806) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 891.102142] env[61806]: DEBUG nova.compute.manager [None req-095bcbe6-2af7-4761-a8ad-c73adfe5ab9d tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] [instance: 6f6ba57a-e2d6-4749-a53a-e263861cb1c0] Allocating IP information in the background. {{(pid=61806) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 891.102322] env[61806]: DEBUG nova.network.neutron [None req-095bcbe6-2af7-4761-a8ad-c73adfe5ab9d tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] [instance: 6f6ba57a-e2d6-4749-a53a-e263861cb1c0] allocate_for_instance() {{(pid=61806) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 891.142494] env[61806]: DEBUG nova.policy [None req-095bcbe6-2af7-4761-a8ad-c73adfe5ab9d tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'b1eac9072ccb4b3bbd23f240e941a76e', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '6e391e1267014fafbb8f5a3211299819', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61806) authorize /opt/stack/nova/nova/policy.py:201}} [ 891.153923] env[61806]: DEBUG nova.network.neutron [None req-0bb2f11a-fff8-4c85-aa76-4f260c30bde9 tempest-AttachVolumeNegativeTest-1683033704 tempest-AttachVolumeNegativeTest-1683033704-project-member] [instance: daff849e-eb4f-411e-92f8-9362e3bd91e3] Successfully updated port: 9fdf6139-082b-49e0-8ce8-7870749b2c7a {{(pid=61806) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 891.219644] env[61806]: DEBUG nova.compute.manager [req-95491071-9279-4c74-b0a5-8aa9ddd28556 req-66e4c594-eca2-4898-969e-94c3efc75329 service nova] [instance: daff849e-eb4f-411e-92f8-9362e3bd91e3] Received event network-changed-9fdf6139-082b-49e0-8ce8-7870749b2c7a {{(pid=61806) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 891.219696] env[61806]: DEBUG nova.compute.manager [req-95491071-9279-4c74-b0a5-8aa9ddd28556 req-66e4c594-eca2-4898-969e-94c3efc75329 service nova] [instance: daff849e-eb4f-411e-92f8-9362e3bd91e3] Refreshing instance network info cache due to event network-changed-9fdf6139-082b-49e0-8ce8-7870749b2c7a. {{(pid=61806) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 891.219916] env[61806]: DEBUG oslo_concurrency.lockutils [req-95491071-9279-4c74-b0a5-8aa9ddd28556 req-66e4c594-eca2-4898-969e-94c3efc75329 service nova] Acquiring lock "refresh_cache-daff849e-eb4f-411e-92f8-9362e3bd91e3" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 891.220078] env[61806]: DEBUG oslo_concurrency.lockutils [req-95491071-9279-4c74-b0a5-8aa9ddd28556 req-66e4c594-eca2-4898-969e-94c3efc75329 service nova] Acquired lock "refresh_cache-daff849e-eb4f-411e-92f8-9362e3bd91e3" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 891.220264] env[61806]: DEBUG nova.network.neutron [req-95491071-9279-4c74-b0a5-8aa9ddd28556 req-66e4c594-eca2-4898-969e-94c3efc75329 service nova] [instance: daff849e-eb4f-411e-92f8-9362e3bd91e3] Refreshing network info cache for port 9fdf6139-082b-49e0-8ce8-7870749b2c7a {{(pid=61806) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 891.232432] env[61806]: INFO nova.compute.manager [-] [instance: efa19aac-3ad7-42d5-a6d8-859050de6e63] Took 1.28 seconds to deallocate network for instance. [ 891.314694] env[61806]: DEBUG nova.compute.utils [None req-76aa021b-a2c9-4c97-bbb8-a4fc0140ddae tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Using /dev/sd instead of None {{(pid=61806) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 891.401026] env[61806]: DEBUG nova.network.neutron [None req-095bcbe6-2af7-4761-a8ad-c73adfe5ab9d tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] [instance: 6f6ba57a-e2d6-4749-a53a-e263861cb1c0] Successfully created port: 013c9f0d-65b1-4279-a823-a7f1c14dfb91 {{(pid=61806) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 891.406320] env[61806]: DEBUG oslo_vmware.api [None req-18f7d4e8-522e-4e97-b6f4-8662acf8d0f8 tempest-ServerTagsTestJSON-1292926242 tempest-ServerTagsTestJSON-1292926242-project-member] Task: {'id': task-1294822, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.249576} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 891.406595] env[61806]: DEBUG nova.virt.vmwareapi.ds_util [None req-18f7d4e8-522e-4e97-b6f4-8662acf8d0f8 tempest-ServerTagsTestJSON-1292926242 tempest-ServerTagsTestJSON-1292926242-project-member] Deleted the datastore file {{(pid=61806) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 891.406821] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-18f7d4e8-522e-4e97-b6f4-8662acf8d0f8 tempest-ServerTagsTestJSON-1292926242 tempest-ServerTagsTestJSON-1292926242-project-member] [instance: 9c20d337-b8f5-4965-b707-0f87c4ab6fc4] Deleted contents of the VM from datastore datastore2 {{(pid=61806) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 891.407017] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-18f7d4e8-522e-4e97-b6f4-8662acf8d0f8 tempest-ServerTagsTestJSON-1292926242 tempest-ServerTagsTestJSON-1292926242-project-member] [instance: 9c20d337-b8f5-4965-b707-0f87c4ab6fc4] Instance destroyed {{(pid=61806) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 891.407358] env[61806]: INFO nova.compute.manager [None req-18f7d4e8-522e-4e97-b6f4-8662acf8d0f8 tempest-ServerTagsTestJSON-1292926242 tempest-ServerTagsTestJSON-1292926242-project-member] [instance: 9c20d337-b8f5-4965-b707-0f87c4ab6fc4] Took 1.15 seconds to destroy the instance on the hypervisor. [ 891.407440] env[61806]: DEBUG oslo.service.loopingcall [None req-18f7d4e8-522e-4e97-b6f4-8662acf8d0f8 tempest-ServerTagsTestJSON-1292926242 tempest-ServerTagsTestJSON-1292926242-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61806) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 891.407637] env[61806]: DEBUG nova.compute.manager [-] [instance: 9c20d337-b8f5-4965-b707-0f87c4ab6fc4] Deallocating network for instance {{(pid=61806) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 891.407732] env[61806]: DEBUG nova.network.neutron [-] [instance: 9c20d337-b8f5-4965-b707-0f87c4ab6fc4] deallocate_for_instance() {{(pid=61806) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 891.409433] env[61806]: DEBUG oslo_concurrency.lockutils [None req-c651ea58-49e6-4c69-a5f7-88000b663e10 tempest-ServerRescueTestJSONUnderV235-749414849 tempest-ServerRescueTestJSONUnderV235-749414849-project-member] Lock "5a526856-0dcd-4b3f-8359-13a5c3b9bfc4" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 22.388s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 891.447925] env[61806]: DEBUG oslo_vmware.api [None req-7e96f1c1-fc04-4010-b728-951cd26dd38b tempest-ImagesTestJSON-618852556 tempest-ImagesTestJSON-618852556-project-member] Task: {'id': task-1294823, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.163676} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 891.448209] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-7e96f1c1-fc04-4010-b728-951cd26dd38b tempest-ImagesTestJSON-618852556 tempest-ImagesTestJSON-618852556-project-member] [instance: 1a73d994-5f21-4f80-8f33-5830d860a7ee] Extended root virtual disk {{(pid=61806) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 891.448992] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-adfedb6e-7edb-4ef0-a5cc-57559a09d605 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 891.472050] env[61806]: DEBUG nova.virt.vmwareapi.volumeops [None req-7e96f1c1-fc04-4010-b728-951cd26dd38b tempest-ImagesTestJSON-618852556 tempest-ImagesTestJSON-618852556-project-member] [instance: 1a73d994-5f21-4f80-8f33-5830d860a7ee] Reconfiguring VM instance instance-0000004e to attach disk [datastore2] 1a73d994-5f21-4f80-8f33-5830d860a7ee/1a73d994-5f21-4f80-8f33-5830d860a7ee.vmdk or device None with type sparse {{(pid=61806) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 891.472384] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-712f9374-fe07-48c3-a86c-9b59fabbae45 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 891.495380] env[61806]: DEBUG oslo_vmware.api [None req-7e96f1c1-fc04-4010-b728-951cd26dd38b tempest-ImagesTestJSON-618852556 tempest-ImagesTestJSON-618852556-project-member] Waiting for the task: (returnval){ [ 891.495380] env[61806]: value = "task-1294824" [ 891.495380] env[61806]: _type = "Task" [ 891.495380] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 891.507764] env[61806]: DEBUG oslo_vmware.api [None req-7e96f1c1-fc04-4010-b728-951cd26dd38b tempest-ImagesTestJSON-618852556 tempest-ImagesTestJSON-618852556-project-member] Task: {'id': task-1294824, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 891.603417] env[61806]: DEBUG nova.compute.manager [None req-095bcbe6-2af7-4761-a8ad-c73adfe5ab9d tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] [instance: 6f6ba57a-e2d6-4749-a53a-e263861cb1c0] Start building block device mappings for instance. {{(pid=61806) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 891.607035] env[61806]: DEBUG oslo_concurrency.lockutils [None req-58b4b995-fdfd-465f-8927-98666a556c7c tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: held 1.017s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 891.607975] env[61806]: DEBUG oslo_concurrency.lockutils [None req-80f76c83-748f-43fe-b146-5517b19ec0ae tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 3.707s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 891.609422] env[61806]: INFO nova.compute.claims [None req-80f76c83-748f-43fe-b146-5517b19ec0ae tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] [instance: ee002e7c-8965-494f-a1af-0b6f3b74eceb] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 891.659607] env[61806]: DEBUG oslo_concurrency.lockutils [None req-0bb2f11a-fff8-4c85-aa76-4f260c30bde9 tempest-AttachVolumeNegativeTest-1683033704 tempest-AttachVolumeNegativeTest-1683033704-project-member] Acquiring lock "refresh_cache-daff849e-eb4f-411e-92f8-9362e3bd91e3" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 891.739704] env[61806]: DEBUG oslo_concurrency.lockutils [None req-f488cd5c-2a2f-4173-ba8b-c601bc6a140a tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 891.755830] env[61806]: DEBUG nova.network.neutron [req-95491071-9279-4c74-b0a5-8aa9ddd28556 req-66e4c594-eca2-4898-969e-94c3efc75329 service nova] [instance: daff849e-eb4f-411e-92f8-9362e3bd91e3] Instance cache missing network info. {{(pid=61806) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 891.817495] env[61806]: DEBUG oslo_concurrency.lockutils [None req-76aa021b-a2c9-4c97-bbb8-a4fc0140ddae tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Lock "59c22c16-3d1c-4e64-bfac-7f8886a1927e" "released" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: held 1.014s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 891.842401] env[61806]: DEBUG nova.network.neutron [req-95491071-9279-4c74-b0a5-8aa9ddd28556 req-66e4c594-eca2-4898-969e-94c3efc75329 service nova] [instance: daff849e-eb4f-411e-92f8-9362e3bd91e3] Updating instance_info_cache with network_info: [] {{(pid=61806) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 892.007010] env[61806]: DEBUG oslo_vmware.api [None req-7e96f1c1-fc04-4010-b728-951cd26dd38b tempest-ImagesTestJSON-618852556 tempest-ImagesTestJSON-618852556-project-member] Task: {'id': task-1294824, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 892.344982] env[61806]: DEBUG oslo_concurrency.lockutils [req-95491071-9279-4c74-b0a5-8aa9ddd28556 req-66e4c594-eca2-4898-969e-94c3efc75329 service nova] Releasing lock "refresh_cache-daff849e-eb4f-411e-92f8-9362e3bd91e3" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 892.345575] env[61806]: DEBUG oslo_concurrency.lockutils [None req-0bb2f11a-fff8-4c85-aa76-4f260c30bde9 tempest-AttachVolumeNegativeTest-1683033704 tempest-AttachVolumeNegativeTest-1683033704-project-member] Acquired lock "refresh_cache-daff849e-eb4f-411e-92f8-9362e3bd91e3" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 892.345867] env[61806]: DEBUG nova.network.neutron [None req-0bb2f11a-fff8-4c85-aa76-4f260c30bde9 tempest-AttachVolumeNegativeTest-1683033704 tempest-AttachVolumeNegativeTest-1683033704-project-member] [instance: daff849e-eb4f-411e-92f8-9362e3bd91e3] Building network info cache for instance {{(pid=61806) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 892.419210] env[61806]: INFO nova.compute.manager [None req-f9554f19-c70d-44f1-bec2-80b2a52a5572 tempest-ServerRescueTestJSONUnderV235-749414849 tempest-ServerRescueTestJSONUnderV235-749414849-project-member] [instance: 5a526856-0dcd-4b3f-8359-13a5c3b9bfc4] Rescuing [ 892.419569] env[61806]: DEBUG oslo_concurrency.lockutils [None req-f9554f19-c70d-44f1-bec2-80b2a52a5572 tempest-ServerRescueTestJSONUnderV235-749414849 tempest-ServerRescueTestJSONUnderV235-749414849-project-member] Acquiring lock "refresh_cache-5a526856-0dcd-4b3f-8359-13a5c3b9bfc4" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 892.419792] env[61806]: DEBUG oslo_concurrency.lockutils [None req-f9554f19-c70d-44f1-bec2-80b2a52a5572 tempest-ServerRescueTestJSONUnderV235-749414849 tempest-ServerRescueTestJSONUnderV235-749414849-project-member] Acquired lock "refresh_cache-5a526856-0dcd-4b3f-8359-13a5c3b9bfc4" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 892.420071] env[61806]: DEBUG nova.network.neutron [None req-f9554f19-c70d-44f1-bec2-80b2a52a5572 tempest-ServerRescueTestJSONUnderV235-749414849 tempest-ServerRescueTestJSONUnderV235-749414849-project-member] [instance: 5a526856-0dcd-4b3f-8359-13a5c3b9bfc4] Building network info cache for instance {{(pid=61806) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 892.454913] env[61806]: DEBUG nova.network.neutron [-] [instance: 9c20d337-b8f5-4965-b707-0f87c4ab6fc4] Updating instance_info_cache with network_info: [] {{(pid=61806) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 892.506725] env[61806]: DEBUG oslo_vmware.api [None req-7e96f1c1-fc04-4010-b728-951cd26dd38b tempest-ImagesTestJSON-618852556 tempest-ImagesTestJSON-618852556-project-member] Task: {'id': task-1294824, 'name': ReconfigVM_Task, 'duration_secs': 0.919557} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 892.507044] env[61806]: DEBUG nova.virt.vmwareapi.volumeops [None req-7e96f1c1-fc04-4010-b728-951cd26dd38b tempest-ImagesTestJSON-618852556 tempest-ImagesTestJSON-618852556-project-member] [instance: 1a73d994-5f21-4f80-8f33-5830d860a7ee] Reconfigured VM instance instance-0000004e to attach disk [datastore2] 1a73d994-5f21-4f80-8f33-5830d860a7ee/1a73d994-5f21-4f80-8f33-5830d860a7ee.vmdk or device None with type sparse {{(pid=61806) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 892.507661] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-1d402140-5381-40c3-8571-6f0694301ffc {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 892.514624] env[61806]: DEBUG oslo_vmware.api [None req-7e96f1c1-fc04-4010-b728-951cd26dd38b tempest-ImagesTestJSON-618852556 tempest-ImagesTestJSON-618852556-project-member] Waiting for the task: (returnval){ [ 892.514624] env[61806]: value = "task-1294825" [ 892.514624] env[61806]: _type = "Task" [ 892.514624] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 892.522591] env[61806]: DEBUG oslo_vmware.api [None req-7e96f1c1-fc04-4010-b728-951cd26dd38b tempest-ImagesTestJSON-618852556 tempest-ImagesTestJSON-618852556-project-member] Task: {'id': task-1294825, 'name': Rename_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 892.627015] env[61806]: DEBUG nova.compute.manager [None req-095bcbe6-2af7-4761-a8ad-c73adfe5ab9d tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] [instance: 6f6ba57a-e2d6-4749-a53a-e263861cb1c0] Start spawning the instance on the hypervisor. {{(pid=61806) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 892.653741] env[61806]: DEBUG nova.virt.hardware [None req-095bcbe6-2af7-4761-a8ad-c73adfe5ab9d tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-15T18:05:10Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-15T18:04:52Z,direct_url=,disk_format='vmdk',id=ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='84ba83fce2c349a6988173c5d6fc3b07',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-15T18:04:53Z,virtual_size=,visibility=), allow threads: False {{(pid=61806) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 892.654013] env[61806]: DEBUG nova.virt.hardware [None req-095bcbe6-2af7-4761-a8ad-c73adfe5ab9d tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] Flavor limits 0:0:0 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 892.654698] env[61806]: DEBUG nova.virt.hardware [None req-095bcbe6-2af7-4761-a8ad-c73adfe5ab9d tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] Image limits 0:0:0 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 892.654698] env[61806]: DEBUG nova.virt.hardware [None req-095bcbe6-2af7-4761-a8ad-c73adfe5ab9d tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] Flavor pref 0:0:0 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 892.654698] env[61806]: DEBUG nova.virt.hardware [None req-095bcbe6-2af7-4761-a8ad-c73adfe5ab9d tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] Image pref 0:0:0 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 892.654834] env[61806]: DEBUG nova.virt.hardware [None req-095bcbe6-2af7-4761-a8ad-c73adfe5ab9d tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 892.654981] env[61806]: DEBUG nova.virt.hardware [None req-095bcbe6-2af7-4761-a8ad-c73adfe5ab9d tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61806) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 892.655160] env[61806]: DEBUG nova.virt.hardware [None req-095bcbe6-2af7-4761-a8ad-c73adfe5ab9d tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61806) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 892.655335] env[61806]: DEBUG nova.virt.hardware [None req-095bcbe6-2af7-4761-a8ad-c73adfe5ab9d tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] Got 1 possible topologies {{(pid=61806) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 892.655504] env[61806]: DEBUG nova.virt.hardware [None req-095bcbe6-2af7-4761-a8ad-c73adfe5ab9d tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61806) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 892.655680] env[61806]: DEBUG nova.virt.hardware [None req-095bcbe6-2af7-4761-a8ad-c73adfe5ab9d tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61806) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 892.656602] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-acf99735-209f-42a7-a1b7-d300fa539628 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 892.668185] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ada1e5be-b9ef-474a-ae72-7852af6c2419 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 892.796204] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2638d865-9760-490a-9c27-8fa7eed36388 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 892.804619] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5b0bd932-baff-44b8-b164-1b83f56aaea9 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 892.836944] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ebd94fe2-9030-476b-9b91-f35711fc7ddd {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 892.841385] env[61806]: DEBUG nova.compute.manager [req-f46742cd-0aac-4925-afc8-27b63933306a req-b2f8fd56-c742-4c2f-9e58-cf39458d9434 service nova] [instance: 6f6ba57a-e2d6-4749-a53a-e263861cb1c0] Received event network-vif-plugged-013c9f0d-65b1-4279-a823-a7f1c14dfb91 {{(pid=61806) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 892.841548] env[61806]: DEBUG oslo_concurrency.lockutils [req-f46742cd-0aac-4925-afc8-27b63933306a req-b2f8fd56-c742-4c2f-9e58-cf39458d9434 service nova] Acquiring lock "6f6ba57a-e2d6-4749-a53a-e263861cb1c0-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 892.841759] env[61806]: DEBUG oslo_concurrency.lockutils [req-f46742cd-0aac-4925-afc8-27b63933306a req-b2f8fd56-c742-4c2f-9e58-cf39458d9434 service nova] Lock "6f6ba57a-e2d6-4749-a53a-e263861cb1c0-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 892.841930] env[61806]: DEBUG oslo_concurrency.lockutils [req-f46742cd-0aac-4925-afc8-27b63933306a req-b2f8fd56-c742-4c2f-9e58-cf39458d9434 service nova] Lock "6f6ba57a-e2d6-4749-a53a-e263861cb1c0-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 892.842112] env[61806]: DEBUG nova.compute.manager [req-f46742cd-0aac-4925-afc8-27b63933306a req-b2f8fd56-c742-4c2f-9e58-cf39458d9434 service nova] [instance: 6f6ba57a-e2d6-4749-a53a-e263861cb1c0] No waiting events found dispatching network-vif-plugged-013c9f0d-65b1-4279-a823-a7f1c14dfb91 {{(pid=61806) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 892.842284] env[61806]: WARNING nova.compute.manager [req-f46742cd-0aac-4925-afc8-27b63933306a req-b2f8fd56-c742-4c2f-9e58-cf39458d9434 service nova] [instance: 6f6ba57a-e2d6-4749-a53a-e263861cb1c0] Received unexpected event network-vif-plugged-013c9f0d-65b1-4279-a823-a7f1c14dfb91 for instance with vm_state building and task_state spawning. [ 892.848912] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2aad089b-c974-458b-9ccb-f3aa69849948 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 892.866221] env[61806]: DEBUG nova.compute.provider_tree [None req-80f76c83-748f-43fe-b146-5517b19ec0ae tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] Inventory has not changed in ProviderTree for provider: a2858be1-fd22-4e08-979e-87ad25293407 {{(pid=61806) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 892.884837] env[61806]: DEBUG nova.network.neutron [None req-0bb2f11a-fff8-4c85-aa76-4f260c30bde9 tempest-AttachVolumeNegativeTest-1683033704 tempest-AttachVolumeNegativeTest-1683033704-project-member] [instance: daff849e-eb4f-411e-92f8-9362e3bd91e3] Instance cache missing network info. {{(pid=61806) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 892.887399] env[61806]: DEBUG oslo_concurrency.lockutils [None req-76aa021b-a2c9-4c97-bbb8-a4fc0140ddae tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Acquiring lock "59c22c16-3d1c-4e64-bfac-7f8886a1927e" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 892.887673] env[61806]: DEBUG oslo_concurrency.lockutils [None req-76aa021b-a2c9-4c97-bbb8-a4fc0140ddae tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Lock "59c22c16-3d1c-4e64-bfac-7f8886a1927e" acquired by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: waited 0.000s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 892.887907] env[61806]: INFO nova.compute.manager [None req-76aa021b-a2c9-4c97-bbb8-a4fc0140ddae tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] [instance: 59c22c16-3d1c-4e64-bfac-7f8886a1927e] Attaching volume 564a9f8e-1bec-410b-9d7d-ab9bf7b6d83b to /dev/sdb [ 892.934060] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6da0824e-de1c-4746-ab43-388f383840ce {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 892.941403] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e587f7de-8fb2-4b33-b049-f9667f253647 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 892.957745] env[61806]: INFO nova.compute.manager [-] [instance: 9c20d337-b8f5-4965-b707-0f87c4ab6fc4] Took 1.55 seconds to deallocate network for instance. [ 892.958214] env[61806]: DEBUG nova.virt.block_device [None req-76aa021b-a2c9-4c97-bbb8-a4fc0140ddae tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] [instance: 59c22c16-3d1c-4e64-bfac-7f8886a1927e] Updating existing volume attachment record: db318029-21c8-48ed-a57d-cd66da81408a {{(pid=61806) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 892.992908] env[61806]: DEBUG nova.network.neutron [None req-095bcbe6-2af7-4761-a8ad-c73adfe5ab9d tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] [instance: 6f6ba57a-e2d6-4749-a53a-e263861cb1c0] Successfully updated port: 013c9f0d-65b1-4279-a823-a7f1c14dfb91 {{(pid=61806) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 893.025908] env[61806]: DEBUG oslo_vmware.api [None req-7e96f1c1-fc04-4010-b728-951cd26dd38b tempest-ImagesTestJSON-618852556 tempest-ImagesTestJSON-618852556-project-member] Task: {'id': task-1294825, 'name': Rename_Task, 'duration_secs': 0.204135} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 893.026731] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [None req-7e96f1c1-fc04-4010-b728-951cd26dd38b tempest-ImagesTestJSON-618852556 tempest-ImagesTestJSON-618852556-project-member] [instance: 1a73d994-5f21-4f80-8f33-5830d860a7ee] Powering on the VM {{(pid=61806) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 893.026731] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-aecc25c3-520a-4eb7-88eb-c7b6b20ba35a {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 893.034767] env[61806]: DEBUG oslo_vmware.api [None req-7e96f1c1-fc04-4010-b728-951cd26dd38b tempest-ImagesTestJSON-618852556 tempest-ImagesTestJSON-618852556-project-member] Waiting for the task: (returnval){ [ 893.034767] env[61806]: value = "task-1294826" [ 893.034767] env[61806]: _type = "Task" [ 893.034767] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 893.045466] env[61806]: DEBUG oslo_vmware.api [None req-7e96f1c1-fc04-4010-b728-951cd26dd38b tempest-ImagesTestJSON-618852556 tempest-ImagesTestJSON-618852556-project-member] Task: {'id': task-1294826, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 893.052918] env[61806]: DEBUG nova.network.neutron [None req-0bb2f11a-fff8-4c85-aa76-4f260c30bde9 tempest-AttachVolumeNegativeTest-1683033704 tempest-AttachVolumeNegativeTest-1683033704-project-member] [instance: daff849e-eb4f-411e-92f8-9362e3bd91e3] Updating instance_info_cache with network_info: [{"id": "9fdf6139-082b-49e0-8ce8-7870749b2c7a", "address": "fa:16:3e:a5:41:88", "network": {"id": "d636ca55-0737-48e4-8d9d-7fe0f797f49e", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-2090334349-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "22e47882c2c44ce492dff6b1c5d782e2", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6f493cd8-1cb4-42a1-8d56-bfa6ac7cf563", "external-id": "nsx-vlan-transportzone-931", "segmentation_id": 931, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9fdf6139-08", "ovs_interfaceid": "9fdf6139-082b-49e0-8ce8-7870749b2c7a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61806) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 893.201927] env[61806]: DEBUG nova.network.neutron [None req-f9554f19-c70d-44f1-bec2-80b2a52a5572 tempest-ServerRescueTestJSONUnderV235-749414849 tempest-ServerRescueTestJSONUnderV235-749414849-project-member] [instance: 5a526856-0dcd-4b3f-8359-13a5c3b9bfc4] Updating instance_info_cache with network_info: [{"id": "58782499-bc1a-4be1-9d48-655c6b2f11d7", "address": "fa:16:3e:52:7a:2b", "network": {"id": "b4fa0830-265e-47bf-8029-5944534312e7", "bridge": "br-int", "label": "tempest-ServerRescueTestJSONUnderV235-1460638664-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "2ba59612579b47a4bde7f82d951ac3ff", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8a34aa30-95be-4b18-98ca-1f2d81f7e9e6", "external-id": "nsx-vlan-transportzone-234", "segmentation_id": 234, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap58782499-bc", "ovs_interfaceid": "58782499-bc1a-4be1-9d48-655c6b2f11d7", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61806) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 893.296291] env[61806]: DEBUG nova.compute.manager [req-500f4bbb-e94e-4ab2-96eb-81b9cd077e31 req-da36abd8-f9a1-4681-a178-9b3d91d9dcc9 service nova] [instance: 9c20d337-b8f5-4965-b707-0f87c4ab6fc4] Received event network-vif-deleted-a7f1c67d-ce27-436f-9329-4a3933beeb5c {{(pid=61806) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 893.370141] env[61806]: DEBUG nova.scheduler.client.report [None req-80f76c83-748f-43fe-b146-5517b19ec0ae tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] Inventory has not changed for provider a2858be1-fd22-4e08-979e-87ad25293407 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 138, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61806) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 893.469561] env[61806]: DEBUG oslo_concurrency.lockutils [None req-18f7d4e8-522e-4e97-b6f4-8662acf8d0f8 tempest-ServerTagsTestJSON-1292926242 tempest-ServerTagsTestJSON-1292926242-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 893.496583] env[61806]: DEBUG oslo_concurrency.lockutils [None req-095bcbe6-2af7-4761-a8ad-c73adfe5ab9d tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] Acquiring lock "refresh_cache-6f6ba57a-e2d6-4749-a53a-e263861cb1c0" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 893.496760] env[61806]: DEBUG oslo_concurrency.lockutils [None req-095bcbe6-2af7-4761-a8ad-c73adfe5ab9d tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] Acquired lock "refresh_cache-6f6ba57a-e2d6-4749-a53a-e263861cb1c0" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 893.496916] env[61806]: DEBUG nova.network.neutron [None req-095bcbe6-2af7-4761-a8ad-c73adfe5ab9d tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] [instance: 6f6ba57a-e2d6-4749-a53a-e263861cb1c0] Building network info cache for instance {{(pid=61806) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 893.544974] env[61806]: DEBUG oslo_vmware.api [None req-7e96f1c1-fc04-4010-b728-951cd26dd38b tempest-ImagesTestJSON-618852556 tempest-ImagesTestJSON-618852556-project-member] Task: {'id': task-1294826, 'name': PowerOnVM_Task} progress is 88%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 893.557078] env[61806]: DEBUG oslo_concurrency.lockutils [None req-0bb2f11a-fff8-4c85-aa76-4f260c30bde9 tempest-AttachVolumeNegativeTest-1683033704 tempest-AttachVolumeNegativeTest-1683033704-project-member] Releasing lock "refresh_cache-daff849e-eb4f-411e-92f8-9362e3bd91e3" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 893.557540] env[61806]: DEBUG nova.compute.manager [None req-0bb2f11a-fff8-4c85-aa76-4f260c30bde9 tempest-AttachVolumeNegativeTest-1683033704 tempest-AttachVolumeNegativeTest-1683033704-project-member] [instance: daff849e-eb4f-411e-92f8-9362e3bd91e3] Instance network_info: |[{"id": "9fdf6139-082b-49e0-8ce8-7870749b2c7a", "address": "fa:16:3e:a5:41:88", "network": {"id": "d636ca55-0737-48e4-8d9d-7fe0f797f49e", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-2090334349-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "22e47882c2c44ce492dff6b1c5d782e2", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6f493cd8-1cb4-42a1-8d56-bfa6ac7cf563", "external-id": "nsx-vlan-transportzone-931", "segmentation_id": 931, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9fdf6139-08", "ovs_interfaceid": "9fdf6139-082b-49e0-8ce8-7870749b2c7a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61806) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 893.558137] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-0bb2f11a-fff8-4c85-aa76-4f260c30bde9 tempest-AttachVolumeNegativeTest-1683033704 tempest-AttachVolumeNegativeTest-1683033704-project-member] [instance: daff849e-eb4f-411e-92f8-9362e3bd91e3] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:a5:41:88', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '6f493cd8-1cb4-42a1-8d56-bfa6ac7cf563', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '9fdf6139-082b-49e0-8ce8-7870749b2c7a', 'vif_model': 'vmxnet3'}] {{(pid=61806) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 893.566717] env[61806]: DEBUG oslo.service.loopingcall [None req-0bb2f11a-fff8-4c85-aa76-4f260c30bde9 tempest-AttachVolumeNegativeTest-1683033704 tempest-AttachVolumeNegativeTest-1683033704-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61806) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 893.566984] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: daff849e-eb4f-411e-92f8-9362e3bd91e3] Creating VM on the ESX host {{(pid=61806) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 893.567256] env[61806]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-998add65-3d9f-4fbd-97eb-862dc2f9e27e {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 893.589632] env[61806]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 893.589632] env[61806]: value = "task-1294830" [ 893.589632] env[61806]: _type = "Task" [ 893.589632] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 893.597974] env[61806]: DEBUG oslo_vmware.api [-] Task: {'id': task-1294830, 'name': CreateVM_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 893.704717] env[61806]: DEBUG oslo_concurrency.lockutils [None req-f9554f19-c70d-44f1-bec2-80b2a52a5572 tempest-ServerRescueTestJSONUnderV235-749414849 tempest-ServerRescueTestJSONUnderV235-749414849-project-member] Releasing lock "refresh_cache-5a526856-0dcd-4b3f-8359-13a5c3b9bfc4" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 893.875660] env[61806]: DEBUG oslo_concurrency.lockutils [None req-80f76c83-748f-43fe-b146-5517b19ec0ae tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.267s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 893.876180] env[61806]: DEBUG nova.compute.manager [None req-80f76c83-748f-43fe-b146-5517b19ec0ae tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] [instance: ee002e7c-8965-494f-a1af-0b6f3b74eceb] Start building networks asynchronously for instance. {{(pid=61806) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 893.879319] env[61806]: DEBUG oslo_concurrency.lockutils [None req-6d4e5237-b28c-45e9-8ae0-8ed79347c566 tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 4.065s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 893.880830] env[61806]: INFO nova.compute.claims [None req-6d4e5237-b28c-45e9-8ae0-8ed79347c566 tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] [instance: 75f1aded-d1da-4ddc-be29-f1fb93799364] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 894.034952] env[61806]: DEBUG nova.network.neutron [None req-095bcbe6-2af7-4761-a8ad-c73adfe5ab9d tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] [instance: 6f6ba57a-e2d6-4749-a53a-e263861cb1c0] Instance cache missing network info. {{(pid=61806) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 894.046357] env[61806]: DEBUG oslo_vmware.api [None req-7e96f1c1-fc04-4010-b728-951cd26dd38b tempest-ImagesTestJSON-618852556 tempest-ImagesTestJSON-618852556-project-member] Task: {'id': task-1294826, 'name': PowerOnVM_Task} progress is 88%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 894.102063] env[61806]: DEBUG oslo_vmware.api [-] Task: {'id': task-1294830, 'name': CreateVM_Task, 'duration_secs': 0.373214} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 894.104720] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: daff849e-eb4f-411e-92f8-9362e3bd91e3] Created VM on the ESX host {{(pid=61806) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 894.105350] env[61806]: DEBUG oslo_concurrency.lockutils [None req-0bb2f11a-fff8-4c85-aa76-4f260c30bde9 tempest-AttachVolumeNegativeTest-1683033704 tempest-AttachVolumeNegativeTest-1683033704-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 894.105551] env[61806]: DEBUG oslo_concurrency.lockutils [None req-0bb2f11a-fff8-4c85-aa76-4f260c30bde9 tempest-AttachVolumeNegativeTest-1683033704 tempest-AttachVolumeNegativeTest-1683033704-project-member] Acquired lock "[datastore2] devstack-image-cache_base/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 894.105887] env[61806]: DEBUG oslo_concurrency.lockutils [None req-0bb2f11a-fff8-4c85-aa76-4f260c30bde9 tempest-AttachVolumeNegativeTest-1683033704 tempest-AttachVolumeNegativeTest-1683033704-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 894.106872] env[61806]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-6c6aec3d-e491-480c-a9db-aaa109bd1770 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 894.111422] env[61806]: DEBUG oslo_vmware.api [None req-0bb2f11a-fff8-4c85-aa76-4f260c30bde9 tempest-AttachVolumeNegativeTest-1683033704 tempest-AttachVolumeNegativeTest-1683033704-project-member] Waiting for the task: (returnval){ [ 894.111422] env[61806]: value = "session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]52810f1a-0c6c-68b5-76ac-aa9f9e82a778" [ 894.111422] env[61806]: _type = "Task" [ 894.111422] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 894.119657] env[61806]: DEBUG oslo_vmware.api [None req-0bb2f11a-fff8-4c85-aa76-4f260c30bde9 tempest-AttachVolumeNegativeTest-1683033704 tempest-AttachVolumeNegativeTest-1683033704-project-member] Task: {'id': session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]52810f1a-0c6c-68b5-76ac-aa9f9e82a778, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 894.205456] env[61806]: DEBUG nova.network.neutron [None req-095bcbe6-2af7-4761-a8ad-c73adfe5ab9d tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] [instance: 6f6ba57a-e2d6-4749-a53a-e263861cb1c0] Updating instance_info_cache with network_info: [{"id": "013c9f0d-65b1-4279-a823-a7f1c14dfb91", "address": "fa:16:3e:15:56:4e", "network": {"id": "a7ffda72-e57c-41fa-a7ed-d7b12084bc9d", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-62794865-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "6e391e1267014fafbb8f5a3211299819", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8b3cfeb1-f262-4fd9-b506-8e9c0733e2d8", "external-id": "nsx-vlan-transportzone-119", "segmentation_id": 119, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap013c9f0d-65", "ovs_interfaceid": "013c9f0d-65b1-4279-a823-a7f1c14dfb91", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61806) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 894.238377] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [None req-f9554f19-c70d-44f1-bec2-80b2a52a5572 tempest-ServerRescueTestJSONUnderV235-749414849 tempest-ServerRescueTestJSONUnderV235-749414849-project-member] [instance: 5a526856-0dcd-4b3f-8359-13a5c3b9bfc4] Powering off the VM {{(pid=61806) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 894.238656] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-8e8e6ac3-a5c3-4d89-bf51-b1eae53d84a2 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 894.246788] env[61806]: DEBUG oslo_vmware.api [None req-f9554f19-c70d-44f1-bec2-80b2a52a5572 tempest-ServerRescueTestJSONUnderV235-749414849 tempest-ServerRescueTestJSONUnderV235-749414849-project-member] Waiting for the task: (returnval){ [ 894.246788] env[61806]: value = "task-1294831" [ 894.246788] env[61806]: _type = "Task" [ 894.246788] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 894.255516] env[61806]: DEBUG oslo_vmware.api [None req-f9554f19-c70d-44f1-bec2-80b2a52a5572 tempest-ServerRescueTestJSONUnderV235-749414849 tempest-ServerRescueTestJSONUnderV235-749414849-project-member] Task: {'id': task-1294831, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 894.385086] env[61806]: DEBUG nova.compute.utils [None req-80f76c83-748f-43fe-b146-5517b19ec0ae tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] Using /dev/sd instead of None {{(pid=61806) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 894.388633] env[61806]: DEBUG nova.compute.manager [None req-80f76c83-748f-43fe-b146-5517b19ec0ae tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] [instance: ee002e7c-8965-494f-a1af-0b6f3b74eceb] Allocating IP information in the background. {{(pid=61806) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 894.388775] env[61806]: DEBUG nova.network.neutron [None req-80f76c83-748f-43fe-b146-5517b19ec0ae tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] [instance: ee002e7c-8965-494f-a1af-0b6f3b74eceb] allocate_for_instance() {{(pid=61806) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 894.429529] env[61806]: DEBUG nova.policy [None req-80f76c83-748f-43fe-b146-5517b19ec0ae tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '4c5d4c6e1ad844a585c08c4644a60bbf', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '73ebcc4d79d248efb653a084a5e44302', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61806) authorize /opt/stack/nova/nova/policy.py:201}} [ 894.546480] env[61806]: DEBUG oslo_vmware.api [None req-7e96f1c1-fc04-4010-b728-951cd26dd38b tempest-ImagesTestJSON-618852556 tempest-ImagesTestJSON-618852556-project-member] Task: {'id': task-1294826, 'name': PowerOnVM_Task, 'duration_secs': 1.336672} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 894.546792] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [None req-7e96f1c1-fc04-4010-b728-951cd26dd38b tempest-ImagesTestJSON-618852556 tempest-ImagesTestJSON-618852556-project-member] [instance: 1a73d994-5f21-4f80-8f33-5830d860a7ee] Powered on the VM {{(pid=61806) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 894.547071] env[61806]: INFO nova.compute.manager [None req-7e96f1c1-fc04-4010-b728-951cd26dd38b tempest-ImagesTestJSON-618852556 tempest-ImagesTestJSON-618852556-project-member] [instance: 1a73d994-5f21-4f80-8f33-5830d860a7ee] Took 9.02 seconds to spawn the instance on the hypervisor. [ 894.547370] env[61806]: DEBUG nova.compute.manager [None req-7e96f1c1-fc04-4010-b728-951cd26dd38b tempest-ImagesTestJSON-618852556 tempest-ImagesTestJSON-618852556-project-member] [instance: 1a73d994-5f21-4f80-8f33-5830d860a7ee] Checking state {{(pid=61806) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 894.548162] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6f65d6e7-ef82-4dab-b6d4-b9a5e4e91799 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 894.623804] env[61806]: DEBUG oslo_vmware.api [None req-0bb2f11a-fff8-4c85-aa76-4f260c30bde9 tempest-AttachVolumeNegativeTest-1683033704 tempest-AttachVolumeNegativeTest-1683033704-project-member] Task: {'id': session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]52810f1a-0c6c-68b5-76ac-aa9f9e82a778, 'name': SearchDatastore_Task, 'duration_secs': 0.010631} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 894.624031] env[61806]: DEBUG oslo_concurrency.lockutils [None req-0bb2f11a-fff8-4c85-aa76-4f260c30bde9 tempest-AttachVolumeNegativeTest-1683033704 tempest-AttachVolumeNegativeTest-1683033704-project-member] Releasing lock "[datastore2] devstack-image-cache_base/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 894.624245] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-0bb2f11a-fff8-4c85-aa76-4f260c30bde9 tempest-AttachVolumeNegativeTest-1683033704 tempest-AttachVolumeNegativeTest-1683033704-project-member] [instance: daff849e-eb4f-411e-92f8-9362e3bd91e3] Processing image ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa {{(pid=61806) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 894.624514] env[61806]: DEBUG oslo_concurrency.lockutils [None req-0bb2f11a-fff8-4c85-aa76-4f260c30bde9 tempest-AttachVolumeNegativeTest-1683033704 tempest-AttachVolumeNegativeTest-1683033704-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa.vmdk" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 894.624674] env[61806]: DEBUG oslo_concurrency.lockutils [None req-0bb2f11a-fff8-4c85-aa76-4f260c30bde9 tempest-AttachVolumeNegativeTest-1683033704 tempest-AttachVolumeNegativeTest-1683033704-project-member] Acquired lock "[datastore2] devstack-image-cache_base/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa.vmdk" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 894.624862] env[61806]: DEBUG nova.virt.vmwareapi.ds_util [None req-0bb2f11a-fff8-4c85-aa76-4f260c30bde9 tempest-AttachVolumeNegativeTest-1683033704 tempest-AttachVolumeNegativeTest-1683033704-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61806) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 894.625149] env[61806]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-7c54fdcc-3e33-4fae-af84-0c9fc436aaf3 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 894.634631] env[61806]: DEBUG nova.virt.vmwareapi.ds_util [None req-0bb2f11a-fff8-4c85-aa76-4f260c30bde9 tempest-AttachVolumeNegativeTest-1683033704 tempest-AttachVolumeNegativeTest-1683033704-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61806) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 894.634820] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-0bb2f11a-fff8-4c85-aa76-4f260c30bde9 tempest-AttachVolumeNegativeTest-1683033704 tempest-AttachVolumeNegativeTest-1683033704-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61806) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 894.635606] env[61806]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d85e798f-c5b0-4d36-9453-250f49f3883f {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 894.641869] env[61806]: DEBUG oslo_vmware.api [None req-0bb2f11a-fff8-4c85-aa76-4f260c30bde9 tempest-AttachVolumeNegativeTest-1683033704 tempest-AttachVolumeNegativeTest-1683033704-project-member] Waiting for the task: (returnval){ [ 894.641869] env[61806]: value = "session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]52c0d649-ba63-0c42-7913-4ab00a4c4901" [ 894.641869] env[61806]: _type = "Task" [ 894.641869] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 894.649670] env[61806]: DEBUG oslo_vmware.api [None req-0bb2f11a-fff8-4c85-aa76-4f260c30bde9 tempest-AttachVolumeNegativeTest-1683033704 tempest-AttachVolumeNegativeTest-1683033704-project-member] Task: {'id': session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]52c0d649-ba63-0c42-7913-4ab00a4c4901, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 894.707912] env[61806]: DEBUG oslo_concurrency.lockutils [None req-095bcbe6-2af7-4761-a8ad-c73adfe5ab9d tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] Releasing lock "refresh_cache-6f6ba57a-e2d6-4749-a53a-e263861cb1c0" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 894.708269] env[61806]: DEBUG nova.compute.manager [None req-095bcbe6-2af7-4761-a8ad-c73adfe5ab9d tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] [instance: 6f6ba57a-e2d6-4749-a53a-e263861cb1c0] Instance network_info: |[{"id": "013c9f0d-65b1-4279-a823-a7f1c14dfb91", "address": "fa:16:3e:15:56:4e", "network": {"id": "a7ffda72-e57c-41fa-a7ed-d7b12084bc9d", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-62794865-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "6e391e1267014fafbb8f5a3211299819", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8b3cfeb1-f262-4fd9-b506-8e9c0733e2d8", "external-id": "nsx-vlan-transportzone-119", "segmentation_id": 119, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap013c9f0d-65", "ovs_interfaceid": "013c9f0d-65b1-4279-a823-a7f1c14dfb91", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61806) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 894.708718] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-095bcbe6-2af7-4761-a8ad-c73adfe5ab9d tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] [instance: 6f6ba57a-e2d6-4749-a53a-e263861cb1c0] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:15:56:4e', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '8b3cfeb1-f262-4fd9-b506-8e9c0733e2d8', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '013c9f0d-65b1-4279-a823-a7f1c14dfb91', 'vif_model': 'vmxnet3'}] {{(pid=61806) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 894.716930] env[61806]: DEBUG oslo.service.loopingcall [None req-095bcbe6-2af7-4761-a8ad-c73adfe5ab9d tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61806) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 894.716930] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 6f6ba57a-e2d6-4749-a53a-e263861cb1c0] Creating VM on the ESX host {{(pid=61806) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 894.716930] env[61806]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-7301a4b5-2b3f-4322-a71b-8aee14e8a567 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 894.737246] env[61806]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 894.737246] env[61806]: value = "task-1294832" [ 894.737246] env[61806]: _type = "Task" [ 894.737246] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 894.741614] env[61806]: DEBUG nova.network.neutron [None req-80f76c83-748f-43fe-b146-5517b19ec0ae tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] [instance: ee002e7c-8965-494f-a1af-0b6f3b74eceb] Successfully created port: 5c3f2277-5769-4c2e-8eb9-df0a99f35008 {{(pid=61806) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 894.746800] env[61806]: DEBUG oslo_vmware.api [-] Task: {'id': task-1294832, 'name': CreateVM_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 894.759021] env[61806]: DEBUG oslo_vmware.api [None req-f9554f19-c70d-44f1-bec2-80b2a52a5572 tempest-ServerRescueTestJSONUnderV235-749414849 tempest-ServerRescueTestJSONUnderV235-749414849-project-member] Task: {'id': task-1294831, 'name': PowerOffVM_Task, 'duration_secs': 0.198507} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 894.759218] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [None req-f9554f19-c70d-44f1-bec2-80b2a52a5572 tempest-ServerRescueTestJSONUnderV235-749414849 tempest-ServerRescueTestJSONUnderV235-749414849-project-member] [instance: 5a526856-0dcd-4b3f-8359-13a5c3b9bfc4] Powered off the VM {{(pid=61806) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 894.760014] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cde45324-835d-49fe-961f-6f91cf30a0be {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 894.784955] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1d0b1a71-78ea-405f-b21a-5156e358537f {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 894.814810] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [None req-f9554f19-c70d-44f1-bec2-80b2a52a5572 tempest-ServerRescueTestJSONUnderV235-749414849 tempest-ServerRescueTestJSONUnderV235-749414849-project-member] [instance: 5a526856-0dcd-4b3f-8359-13a5c3b9bfc4] Powering off the VM {{(pid=61806) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 894.815143] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-b7566b6e-e1ce-473b-a709-673b73c971b7 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 894.822356] env[61806]: DEBUG oslo_vmware.api [None req-f9554f19-c70d-44f1-bec2-80b2a52a5572 tempest-ServerRescueTestJSONUnderV235-749414849 tempest-ServerRescueTestJSONUnderV235-749414849-project-member] Waiting for the task: (returnval){ [ 894.822356] env[61806]: value = "task-1294833" [ 894.822356] env[61806]: _type = "Task" [ 894.822356] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 894.830906] env[61806]: DEBUG oslo_vmware.api [None req-f9554f19-c70d-44f1-bec2-80b2a52a5572 tempest-ServerRescueTestJSONUnderV235-749414849 tempest-ServerRescueTestJSONUnderV235-749414849-project-member] Task: {'id': task-1294833, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 894.868642] env[61806]: DEBUG nova.compute.manager [req-e6438e33-b3ee-4c86-8570-393a24aa8915 req-49fbdcab-5351-4a11-bafb-6a9a9b338d7d service nova] [instance: 6f6ba57a-e2d6-4749-a53a-e263861cb1c0] Received event network-changed-013c9f0d-65b1-4279-a823-a7f1c14dfb91 {{(pid=61806) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 894.868907] env[61806]: DEBUG nova.compute.manager [req-e6438e33-b3ee-4c86-8570-393a24aa8915 req-49fbdcab-5351-4a11-bafb-6a9a9b338d7d service nova] [instance: 6f6ba57a-e2d6-4749-a53a-e263861cb1c0] Refreshing instance network info cache due to event network-changed-013c9f0d-65b1-4279-a823-a7f1c14dfb91. {{(pid=61806) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 894.869541] env[61806]: DEBUG oslo_concurrency.lockutils [req-e6438e33-b3ee-4c86-8570-393a24aa8915 req-49fbdcab-5351-4a11-bafb-6a9a9b338d7d service nova] Acquiring lock "refresh_cache-6f6ba57a-e2d6-4749-a53a-e263861cb1c0" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 894.869778] env[61806]: DEBUG oslo_concurrency.lockutils [req-e6438e33-b3ee-4c86-8570-393a24aa8915 req-49fbdcab-5351-4a11-bafb-6a9a9b338d7d service nova] Acquired lock "refresh_cache-6f6ba57a-e2d6-4749-a53a-e263861cb1c0" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 894.872122] env[61806]: DEBUG nova.network.neutron [req-e6438e33-b3ee-4c86-8570-393a24aa8915 req-49fbdcab-5351-4a11-bafb-6a9a9b338d7d service nova] [instance: 6f6ba57a-e2d6-4749-a53a-e263861cb1c0] Refreshing network info cache for port 013c9f0d-65b1-4279-a823-a7f1c14dfb91 {{(pid=61806) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 894.889531] env[61806]: DEBUG nova.compute.manager [None req-80f76c83-748f-43fe-b146-5517b19ec0ae tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] [instance: ee002e7c-8965-494f-a1af-0b6f3b74eceb] Start building block device mappings for instance. {{(pid=61806) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 895.064456] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-da7de45c-3999-4689-a123-ed7c3a434c84 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 895.069668] env[61806]: INFO nova.compute.manager [None req-7e96f1c1-fc04-4010-b728-951cd26dd38b tempest-ImagesTestJSON-618852556 tempest-ImagesTestJSON-618852556-project-member] [instance: 1a73d994-5f21-4f80-8f33-5830d860a7ee] Took 21.78 seconds to build instance. [ 895.074028] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3c01ec8f-e7cf-4956-b0f9-7407b5c033ea {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 895.104760] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e7495c21-6abc-4856-bc37-e39086a81dcf {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 895.113017] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8d1d059d-74d7-4c35-a531-eb6eb04fd6a4 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 895.126963] env[61806]: DEBUG nova.compute.provider_tree [None req-6d4e5237-b28c-45e9-8ae0-8ed79347c566 tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Inventory has not changed in ProviderTree for provider: a2858be1-fd22-4e08-979e-87ad25293407 {{(pid=61806) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 895.153517] env[61806]: DEBUG oslo_vmware.api [None req-0bb2f11a-fff8-4c85-aa76-4f260c30bde9 tempest-AttachVolumeNegativeTest-1683033704 tempest-AttachVolumeNegativeTest-1683033704-project-member] Task: {'id': session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]52c0d649-ba63-0c42-7913-4ab00a4c4901, 'name': SearchDatastore_Task, 'duration_secs': 0.010896} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 895.154322] env[61806]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-2bf4da3f-7670-4a52-816d-b57fe031e185 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 895.160276] env[61806]: DEBUG oslo_vmware.api [None req-0bb2f11a-fff8-4c85-aa76-4f260c30bde9 tempest-AttachVolumeNegativeTest-1683033704 tempest-AttachVolumeNegativeTest-1683033704-project-member] Waiting for the task: (returnval){ [ 895.160276] env[61806]: value = "session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]525fdf50-c2c4-5ab9-8d12-267983287e70" [ 895.160276] env[61806]: _type = "Task" [ 895.160276] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 895.169890] env[61806]: DEBUG oslo_vmware.api [None req-0bb2f11a-fff8-4c85-aa76-4f260c30bde9 tempest-AttachVolumeNegativeTest-1683033704 tempest-AttachVolumeNegativeTest-1683033704-project-member] Task: {'id': session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]525fdf50-c2c4-5ab9-8d12-267983287e70, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 895.247195] env[61806]: DEBUG oslo_vmware.api [-] Task: {'id': task-1294832, 'name': CreateVM_Task, 'duration_secs': 0.355908} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 895.247363] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 6f6ba57a-e2d6-4749-a53a-e263861cb1c0] Created VM on the ESX host {{(pid=61806) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 895.248054] env[61806]: DEBUG oslo_concurrency.lockutils [None req-095bcbe6-2af7-4761-a8ad-c73adfe5ab9d tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 895.248229] env[61806]: DEBUG oslo_concurrency.lockutils [None req-095bcbe6-2af7-4761-a8ad-c73adfe5ab9d tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] Acquired lock "[datastore2] devstack-image-cache_base/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 895.248565] env[61806]: DEBUG oslo_concurrency.lockutils [None req-095bcbe6-2af7-4761-a8ad-c73adfe5ab9d tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 895.248817] env[61806]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-8bc3483d-0260-4042-9d4c-74c75b52cbe2 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 895.253930] env[61806]: DEBUG oslo_vmware.api [None req-095bcbe6-2af7-4761-a8ad-c73adfe5ab9d tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] Waiting for the task: (returnval){ [ 895.253930] env[61806]: value = "session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]520e7896-9cb5-3b04-6510-298574ad2974" [ 895.253930] env[61806]: _type = "Task" [ 895.253930] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 895.262368] env[61806]: DEBUG oslo_vmware.api [None req-095bcbe6-2af7-4761-a8ad-c73adfe5ab9d tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] Task: {'id': session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]520e7896-9cb5-3b04-6510-298574ad2974, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 895.333713] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [None req-f9554f19-c70d-44f1-bec2-80b2a52a5572 tempest-ServerRescueTestJSONUnderV235-749414849 tempest-ServerRescueTestJSONUnderV235-749414849-project-member] [instance: 5a526856-0dcd-4b3f-8359-13a5c3b9bfc4] VM already powered off {{(pid=61806) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1509}} [ 895.334019] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-f9554f19-c70d-44f1-bec2-80b2a52a5572 tempest-ServerRescueTestJSONUnderV235-749414849 tempest-ServerRescueTestJSONUnderV235-749414849-project-member] [instance: 5a526856-0dcd-4b3f-8359-13a5c3b9bfc4] Processing image ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa {{(pid=61806) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 895.334249] env[61806]: DEBUG oslo_concurrency.lockutils [None req-f9554f19-c70d-44f1-bec2-80b2a52a5572 tempest-ServerRescueTestJSONUnderV235-749414849 tempest-ServerRescueTestJSONUnderV235-749414849-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa.vmdk" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 895.571704] env[61806]: DEBUG oslo_concurrency.lockutils [None req-7e96f1c1-fc04-4010-b728-951cd26dd38b tempest-ImagesTestJSON-618852556 tempest-ImagesTestJSON-618852556-project-member] Lock "1a73d994-5f21-4f80-8f33-5830d860a7ee" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 23.285s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 895.590932] env[61806]: DEBUG nova.network.neutron [req-e6438e33-b3ee-4c86-8570-393a24aa8915 req-49fbdcab-5351-4a11-bafb-6a9a9b338d7d service nova] [instance: 6f6ba57a-e2d6-4749-a53a-e263861cb1c0] Updated VIF entry in instance network info cache for port 013c9f0d-65b1-4279-a823-a7f1c14dfb91. {{(pid=61806) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 895.591311] env[61806]: DEBUG nova.network.neutron [req-e6438e33-b3ee-4c86-8570-393a24aa8915 req-49fbdcab-5351-4a11-bafb-6a9a9b338d7d service nova] [instance: 6f6ba57a-e2d6-4749-a53a-e263861cb1c0] Updating instance_info_cache with network_info: [{"id": "013c9f0d-65b1-4279-a823-a7f1c14dfb91", "address": "fa:16:3e:15:56:4e", "network": {"id": "a7ffda72-e57c-41fa-a7ed-d7b12084bc9d", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-62794865-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "6e391e1267014fafbb8f5a3211299819", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8b3cfeb1-f262-4fd9-b506-8e9c0733e2d8", "external-id": "nsx-vlan-transportzone-119", "segmentation_id": 119, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap013c9f0d-65", "ovs_interfaceid": "013c9f0d-65b1-4279-a823-a7f1c14dfb91", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61806) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 895.630117] env[61806]: DEBUG nova.scheduler.client.report [None req-6d4e5237-b28c-45e9-8ae0-8ed79347c566 tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Inventory has not changed for provider a2858be1-fd22-4e08-979e-87ad25293407 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 138, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61806) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 895.672486] env[61806]: DEBUG oslo_vmware.api [None req-0bb2f11a-fff8-4c85-aa76-4f260c30bde9 tempest-AttachVolumeNegativeTest-1683033704 tempest-AttachVolumeNegativeTest-1683033704-project-member] Task: {'id': session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]525fdf50-c2c4-5ab9-8d12-267983287e70, 'name': SearchDatastore_Task, 'duration_secs': 0.02503} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 895.672767] env[61806]: DEBUG oslo_concurrency.lockutils [None req-0bb2f11a-fff8-4c85-aa76-4f260c30bde9 tempest-AttachVolumeNegativeTest-1683033704 tempest-AttachVolumeNegativeTest-1683033704-project-member] Releasing lock "[datastore2] devstack-image-cache_base/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa.vmdk" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 895.673164] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [None req-0bb2f11a-fff8-4c85-aa76-4f260c30bde9 tempest-AttachVolumeNegativeTest-1683033704 tempest-AttachVolumeNegativeTest-1683033704-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa.vmdk to [datastore2] daff849e-eb4f-411e-92f8-9362e3bd91e3/daff849e-eb4f-411e-92f8-9362e3bd91e3.vmdk {{(pid=61806) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 895.673615] env[61806]: DEBUG oslo_concurrency.lockutils [None req-f9554f19-c70d-44f1-bec2-80b2a52a5572 tempest-ServerRescueTestJSONUnderV235-749414849 tempest-ServerRescueTestJSONUnderV235-749414849-project-member] Acquired lock "[datastore2] devstack-image-cache_base/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa.vmdk" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 895.673903] env[61806]: DEBUG nova.virt.vmwareapi.ds_util [None req-f9554f19-c70d-44f1-bec2-80b2a52a5572 tempest-ServerRescueTestJSONUnderV235-749414849 tempest-ServerRescueTestJSONUnderV235-749414849-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61806) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 895.674197] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-9e427e28-53e2-4b9f-b704-55a97aca1544 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 895.676403] env[61806]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-f0f02302-6678-47d5-ba68-8978c219a87d {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 895.683831] env[61806]: DEBUG oslo_vmware.api [None req-0bb2f11a-fff8-4c85-aa76-4f260c30bde9 tempest-AttachVolumeNegativeTest-1683033704 tempest-AttachVolumeNegativeTest-1683033704-project-member] Waiting for the task: (returnval){ [ 895.683831] env[61806]: value = "task-1294835" [ 895.683831] env[61806]: _type = "Task" [ 895.683831] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 895.688519] env[61806]: DEBUG nova.virt.vmwareapi.ds_util [None req-f9554f19-c70d-44f1-bec2-80b2a52a5572 tempest-ServerRescueTestJSONUnderV235-749414849 tempest-ServerRescueTestJSONUnderV235-749414849-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61806) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 895.688718] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-f9554f19-c70d-44f1-bec2-80b2a52a5572 tempest-ServerRescueTestJSONUnderV235-749414849 tempest-ServerRescueTestJSONUnderV235-749414849-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61806) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 895.689848] env[61806]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-53dbc107-cd93-49f2-bf41-c0da4e954128 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 895.695427] env[61806]: DEBUG oslo_vmware.api [None req-0bb2f11a-fff8-4c85-aa76-4f260c30bde9 tempest-AttachVolumeNegativeTest-1683033704 tempest-AttachVolumeNegativeTest-1683033704-project-member] Task: {'id': task-1294835, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 895.699029] env[61806]: DEBUG oslo_vmware.api [None req-f9554f19-c70d-44f1-bec2-80b2a52a5572 tempest-ServerRescueTestJSONUnderV235-749414849 tempest-ServerRescueTestJSONUnderV235-749414849-project-member] Waiting for the task: (returnval){ [ 895.699029] env[61806]: value = "session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]52ebe804-14b9-4e4a-4c85-e0c9fcba9e14" [ 895.699029] env[61806]: _type = "Task" [ 895.699029] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 895.707409] env[61806]: DEBUG oslo_vmware.api [None req-f9554f19-c70d-44f1-bec2-80b2a52a5572 tempest-ServerRescueTestJSONUnderV235-749414849 tempest-ServerRescueTestJSONUnderV235-749414849-project-member] Task: {'id': session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]52ebe804-14b9-4e4a-4c85-e0c9fcba9e14, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 895.766164] env[61806]: DEBUG oslo_vmware.api [None req-095bcbe6-2af7-4761-a8ad-c73adfe5ab9d tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] Task: {'id': session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]520e7896-9cb5-3b04-6510-298574ad2974, 'name': SearchDatastore_Task, 'duration_secs': 0.009903} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 895.766547] env[61806]: DEBUG oslo_concurrency.lockutils [None req-095bcbe6-2af7-4761-a8ad-c73adfe5ab9d tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] Releasing lock "[datastore2] devstack-image-cache_base/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 895.766812] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-095bcbe6-2af7-4761-a8ad-c73adfe5ab9d tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] [instance: 6f6ba57a-e2d6-4749-a53a-e263861cb1c0] Processing image ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa {{(pid=61806) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 895.767077] env[61806]: DEBUG oslo_concurrency.lockutils [None req-095bcbe6-2af7-4761-a8ad-c73adfe5ab9d tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa.vmdk" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 895.909521] env[61806]: DEBUG nova.compute.manager [None req-80f76c83-748f-43fe-b146-5517b19ec0ae tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] [instance: ee002e7c-8965-494f-a1af-0b6f3b74eceb] Start spawning the instance on the hypervisor. {{(pid=61806) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 895.933071] env[61806]: DEBUG nova.virt.hardware [None req-80f76c83-748f-43fe-b146-5517b19ec0ae tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-15T18:05:10Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-15T18:04:52Z,direct_url=,disk_format='vmdk',id=ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='84ba83fce2c349a6988173c5d6fc3b07',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-15T18:04:53Z,virtual_size=,visibility=), allow threads: False {{(pid=61806) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 895.933483] env[61806]: DEBUG nova.virt.hardware [None req-80f76c83-748f-43fe-b146-5517b19ec0ae tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] Flavor limits 0:0:0 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 895.933669] env[61806]: DEBUG nova.virt.hardware [None req-80f76c83-748f-43fe-b146-5517b19ec0ae tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] Image limits 0:0:0 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 895.933864] env[61806]: DEBUG nova.virt.hardware [None req-80f76c83-748f-43fe-b146-5517b19ec0ae tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] Flavor pref 0:0:0 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 895.934032] env[61806]: DEBUG nova.virt.hardware [None req-80f76c83-748f-43fe-b146-5517b19ec0ae tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] Image pref 0:0:0 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 895.934194] env[61806]: DEBUG nova.virt.hardware [None req-80f76c83-748f-43fe-b146-5517b19ec0ae tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 895.934471] env[61806]: DEBUG nova.virt.hardware [None req-80f76c83-748f-43fe-b146-5517b19ec0ae tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61806) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 895.934648] env[61806]: DEBUG nova.virt.hardware [None req-80f76c83-748f-43fe-b146-5517b19ec0ae tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61806) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 895.934829] env[61806]: DEBUG nova.virt.hardware [None req-80f76c83-748f-43fe-b146-5517b19ec0ae tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] Got 1 possible topologies {{(pid=61806) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 895.935012] env[61806]: DEBUG nova.virt.hardware [None req-80f76c83-748f-43fe-b146-5517b19ec0ae tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61806) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 895.935207] env[61806]: DEBUG nova.virt.hardware [None req-80f76c83-748f-43fe-b146-5517b19ec0ae tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61806) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 895.936160] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d3c36b1b-486b-4e5d-b132-74d40fb5deac {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 895.946019] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5d97c0e9-80f6-4fad-86f8-7c412c5ae23f {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 896.094486] env[61806]: DEBUG oslo_concurrency.lockutils [req-e6438e33-b3ee-4c86-8570-393a24aa8915 req-49fbdcab-5351-4a11-bafb-6a9a9b338d7d service nova] Releasing lock "refresh_cache-6f6ba57a-e2d6-4749-a53a-e263861cb1c0" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 896.135056] env[61806]: DEBUG oslo_concurrency.lockutils [None req-6d4e5237-b28c-45e9-8ae0-8ed79347c566 tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.256s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 896.135772] env[61806]: DEBUG nova.compute.manager [None req-6d4e5237-b28c-45e9-8ae0-8ed79347c566 tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] [instance: 75f1aded-d1da-4ddc-be29-f1fb93799364] Start building networks asynchronously for instance. {{(pid=61806) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 896.138824] env[61806]: DEBUG oslo_concurrency.lockutils [None req-f488cd5c-2a2f-4173-ba8b-c601bc6a140a tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 4.400s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 896.139147] env[61806]: DEBUG nova.objects.instance [None req-f488cd5c-2a2f-4173-ba8b-c601bc6a140a tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Lazy-loading 'resources' on Instance uuid efa19aac-3ad7-42d5-a6d8-859050de6e63 {{(pid=61806) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 896.195299] env[61806]: DEBUG oslo_vmware.api [None req-0bb2f11a-fff8-4c85-aa76-4f260c30bde9 tempest-AttachVolumeNegativeTest-1683033704 tempest-AttachVolumeNegativeTest-1683033704-project-member] Task: {'id': task-1294835, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.46302} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 896.195580] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [None req-0bb2f11a-fff8-4c85-aa76-4f260c30bde9 tempest-AttachVolumeNegativeTest-1683033704 tempest-AttachVolumeNegativeTest-1683033704-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa.vmdk to [datastore2] daff849e-eb4f-411e-92f8-9362e3bd91e3/daff849e-eb4f-411e-92f8-9362e3bd91e3.vmdk {{(pid=61806) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 896.195840] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-0bb2f11a-fff8-4c85-aa76-4f260c30bde9 tempest-AttachVolumeNegativeTest-1683033704 tempest-AttachVolumeNegativeTest-1683033704-project-member] [instance: daff849e-eb4f-411e-92f8-9362e3bd91e3] Extending root virtual disk to 1048576 {{(pid=61806) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 896.196121] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-c19ac9a9-e4f0-4547-b934-cc5bf18d7d66 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 896.203970] env[61806]: DEBUG oslo_vmware.api [None req-0bb2f11a-fff8-4c85-aa76-4f260c30bde9 tempest-AttachVolumeNegativeTest-1683033704 tempest-AttachVolumeNegativeTest-1683033704-project-member] Waiting for the task: (returnval){ [ 896.203970] env[61806]: value = "task-1294836" [ 896.203970] env[61806]: _type = "Task" [ 896.203970] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 896.210960] env[61806]: DEBUG oslo_vmware.api [None req-f9554f19-c70d-44f1-bec2-80b2a52a5572 tempest-ServerRescueTestJSONUnderV235-749414849 tempest-ServerRescueTestJSONUnderV235-749414849-project-member] Task: {'id': session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]52ebe804-14b9-4e4a-4c85-e0c9fcba9e14, 'name': SearchDatastore_Task, 'duration_secs': 0.010401} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 896.214361] env[61806]: DEBUG oslo_vmware.api [None req-0bb2f11a-fff8-4c85-aa76-4f260c30bde9 tempest-AttachVolumeNegativeTest-1683033704 tempest-AttachVolumeNegativeTest-1683033704-project-member] Task: {'id': task-1294836, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 896.214589] env[61806]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-82751e00-bc44-427c-b488-6a3f78359614 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 896.220414] env[61806]: DEBUG oslo_vmware.api [None req-f9554f19-c70d-44f1-bec2-80b2a52a5572 tempest-ServerRescueTestJSONUnderV235-749414849 tempest-ServerRescueTestJSONUnderV235-749414849-project-member] Waiting for the task: (returnval){ [ 896.220414] env[61806]: value = "session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]528abac8-5a84-407c-0a13-be6c94643111" [ 896.220414] env[61806]: _type = "Task" [ 896.220414] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 896.228966] env[61806]: DEBUG oslo_vmware.api [None req-f9554f19-c70d-44f1-bec2-80b2a52a5572 tempest-ServerRescueTestJSONUnderV235-749414849 tempest-ServerRescueTestJSONUnderV235-749414849-project-member] Task: {'id': session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]528abac8-5a84-407c-0a13-be6c94643111, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 896.267512] env[61806]: DEBUG nova.network.neutron [None req-80f76c83-748f-43fe-b146-5517b19ec0ae tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] [instance: ee002e7c-8965-494f-a1af-0b6f3b74eceb] Successfully updated port: 5c3f2277-5769-4c2e-8eb9-df0a99f35008 {{(pid=61806) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 896.642147] env[61806]: DEBUG nova.compute.utils [None req-6d4e5237-b28c-45e9-8ae0-8ed79347c566 tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Using /dev/sd instead of None {{(pid=61806) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 896.643696] env[61806]: DEBUG nova.compute.manager [None req-6d4e5237-b28c-45e9-8ae0-8ed79347c566 tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] [instance: 75f1aded-d1da-4ddc-be29-f1fb93799364] Allocating IP information in the background. {{(pid=61806) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 896.643858] env[61806]: DEBUG nova.network.neutron [None req-6d4e5237-b28c-45e9-8ae0-8ed79347c566 tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] [instance: 75f1aded-d1da-4ddc-be29-f1fb93799364] allocate_for_instance() {{(pid=61806) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 896.695895] env[61806]: DEBUG nova.policy [None req-6d4e5237-b28c-45e9-8ae0-8ed79347c566 tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'b2631ce1f6dd40a4bfc5da7f90cde260', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '4b9d94c3d684474b99e305fb6cd9b95f', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61806) authorize /opt/stack/nova/nova/policy.py:201}} [ 896.714457] env[61806]: DEBUG oslo_vmware.api [None req-0bb2f11a-fff8-4c85-aa76-4f260c30bde9 tempest-AttachVolumeNegativeTest-1683033704 tempest-AttachVolumeNegativeTest-1683033704-project-member] Task: {'id': task-1294836, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.082668} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 896.715162] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-0bb2f11a-fff8-4c85-aa76-4f260c30bde9 tempest-AttachVolumeNegativeTest-1683033704 tempest-AttachVolumeNegativeTest-1683033704-project-member] [instance: daff849e-eb4f-411e-92f8-9362e3bd91e3] Extended root virtual disk {{(pid=61806) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 896.715606] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3d4bf712-c847-4d7e-9057-3ba970ee1c60 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 896.737764] env[61806]: DEBUG nova.virt.vmwareapi.volumeops [None req-0bb2f11a-fff8-4c85-aa76-4f260c30bde9 tempest-AttachVolumeNegativeTest-1683033704 tempest-AttachVolumeNegativeTest-1683033704-project-member] [instance: daff849e-eb4f-411e-92f8-9362e3bd91e3] Reconfiguring VM instance instance-0000004f to attach disk [datastore2] daff849e-eb4f-411e-92f8-9362e3bd91e3/daff849e-eb4f-411e-92f8-9362e3bd91e3.vmdk or device None with type sparse {{(pid=61806) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 896.743210] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-4899d7dd-4472-4c7b-bafb-392535d426d8 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 896.757544] env[61806]: DEBUG nova.compute.manager [None req-464e7d68-de9c-4f8c-9477-932cd9673dfd tempest-ImagesTestJSON-618852556 tempest-ImagesTestJSON-618852556-project-member] [instance: 1a73d994-5f21-4f80-8f33-5830d860a7ee] Checking state {{(pid=61806) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 896.758345] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-83491986-c870-4b44-a65c-fa52a4c96c29 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 896.767898] env[61806]: DEBUG oslo_vmware.api [None req-f9554f19-c70d-44f1-bec2-80b2a52a5572 tempest-ServerRescueTestJSONUnderV235-749414849 tempest-ServerRescueTestJSONUnderV235-749414849-project-member] Task: {'id': session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]528abac8-5a84-407c-0a13-be6c94643111, 'name': SearchDatastore_Task, 'duration_secs': 0.020888} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 896.772595] env[61806]: DEBUG oslo_concurrency.lockutils [None req-f9554f19-c70d-44f1-bec2-80b2a52a5572 tempest-ServerRescueTestJSONUnderV235-749414849 tempest-ServerRescueTestJSONUnderV235-749414849-project-member] Releasing lock "[datastore2] devstack-image-cache_base/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa.vmdk" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 896.773341] env[61806]: DEBUG nova.virt.vmwareapi.ds_util [None req-f9554f19-c70d-44f1-bec2-80b2a52a5572 tempest-ServerRescueTestJSONUnderV235-749414849 tempest-ServerRescueTestJSONUnderV235-749414849-project-member] Copying virtual disk from [datastore2] devstack-image-cache_base/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa.vmdk to [datastore2] 5a526856-0dcd-4b3f-8359-13a5c3b9bfc4/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa-rescue.vmdk. {{(pid=61806) disk_copy /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:281}} [ 896.775537] env[61806]: DEBUG oslo_concurrency.lockutils [None req-80f76c83-748f-43fe-b146-5517b19ec0ae tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] Acquiring lock "refresh_cache-ee002e7c-8965-494f-a1af-0b6f3b74eceb" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 896.775729] env[61806]: DEBUG oslo_concurrency.lockutils [None req-80f76c83-748f-43fe-b146-5517b19ec0ae tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] Acquired lock "refresh_cache-ee002e7c-8965-494f-a1af-0b6f3b74eceb" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 896.775846] env[61806]: DEBUG nova.network.neutron [None req-80f76c83-748f-43fe-b146-5517b19ec0ae tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] [instance: ee002e7c-8965-494f-a1af-0b6f3b74eceb] Building network info cache for instance {{(pid=61806) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 896.776857] env[61806]: DEBUG oslo_vmware.api [None req-0bb2f11a-fff8-4c85-aa76-4f260c30bde9 tempest-AttachVolumeNegativeTest-1683033704 tempest-AttachVolumeNegativeTest-1683033704-project-member] Waiting for the task: (returnval){ [ 896.776857] env[61806]: value = "task-1294837" [ 896.776857] env[61806]: _type = "Task" [ 896.776857] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 896.777351] env[61806]: DEBUG oslo_concurrency.lockutils [None req-095bcbe6-2af7-4761-a8ad-c73adfe5ab9d tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] Acquired lock "[datastore2] devstack-image-cache_base/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa.vmdk" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 896.777630] env[61806]: DEBUG nova.virt.vmwareapi.ds_util [None req-095bcbe6-2af7-4761-a8ad-c73adfe5ab9d tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61806) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 896.777924] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-dc42c454-0a77-4169-8e5e-0679c5e5f9ee {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 896.782398] env[61806]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-907065a5-82a6-4856-acb5-7fb68ae06ab3 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 896.798137] env[61806]: DEBUG oslo_vmware.api [None req-0bb2f11a-fff8-4c85-aa76-4f260c30bde9 tempest-AttachVolumeNegativeTest-1683033704 tempest-AttachVolumeNegativeTest-1683033704-project-member] Task: {'id': task-1294837, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 896.799150] env[61806]: DEBUG oslo_vmware.api [None req-f9554f19-c70d-44f1-bec2-80b2a52a5572 tempest-ServerRescueTestJSONUnderV235-749414849 tempest-ServerRescueTestJSONUnderV235-749414849-project-member] Waiting for the task: (returnval){ [ 896.799150] env[61806]: value = "task-1294838" [ 896.799150] env[61806]: _type = "Task" [ 896.799150] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 896.800459] env[61806]: DEBUG nova.virt.vmwareapi.ds_util [None req-095bcbe6-2af7-4761-a8ad-c73adfe5ab9d tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61806) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 896.800708] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-095bcbe6-2af7-4761-a8ad-c73adfe5ab9d tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61806) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 896.805207] env[61806]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-edeb7f8e-0ef1-43e5-8286-d734ecc47f5e {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 896.814265] env[61806]: DEBUG oslo_vmware.api [None req-f9554f19-c70d-44f1-bec2-80b2a52a5572 tempest-ServerRescueTestJSONUnderV235-749414849 tempest-ServerRescueTestJSONUnderV235-749414849-project-member] Task: {'id': task-1294838, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 896.817576] env[61806]: DEBUG oslo_vmware.api [None req-095bcbe6-2af7-4761-a8ad-c73adfe5ab9d tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] Waiting for the task: (returnval){ [ 896.817576] env[61806]: value = "session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]5202fe10-fe76-2985-f20a-637703486958" [ 896.817576] env[61806]: _type = "Task" [ 896.817576] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 896.828082] env[61806]: DEBUG oslo_vmware.api [None req-095bcbe6-2af7-4761-a8ad-c73adfe5ab9d tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] Task: {'id': session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]5202fe10-fe76-2985-f20a-637703486958, 'name': SearchDatastore_Task, 'duration_secs': 0.013188} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 896.831481] env[61806]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-7dac3286-74e1-4620-89d6-1d30afa7b21a {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 896.837113] env[61806]: DEBUG oslo_vmware.api [None req-095bcbe6-2af7-4761-a8ad-c73adfe5ab9d tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] Waiting for the task: (returnval){ [ 896.837113] env[61806]: value = "session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]52ddb997-500a-ab5c-1577-0464bf0fca91" [ 896.837113] env[61806]: _type = "Task" [ 896.837113] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 896.846794] env[61806]: DEBUG oslo_vmware.api [None req-095bcbe6-2af7-4761-a8ad-c73adfe5ab9d tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] Task: {'id': session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]52ddb997-500a-ab5c-1577-0464bf0fca91, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 896.880622] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-13505ac8-df84-4d98-8cc2-222d4848eee2 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 896.890027] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ea0cf6cb-b679-401d-ab7d-ccc4f8eec1fe {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 896.926677] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-87e667ff-a72a-4f2d-9925-88894c05a57f {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 896.934121] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bdd36ac1-83f5-48e8-ad87-c5fa4d83b52b {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 896.949916] env[61806]: DEBUG nova.compute.provider_tree [None req-f488cd5c-2a2f-4173-ba8b-c601bc6a140a tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Inventory has not changed in ProviderTree for provider: a2858be1-fd22-4e08-979e-87ad25293407 {{(pid=61806) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 896.953081] env[61806]: DEBUG nova.compute.manager [req-b664b826-71c0-46bc-897e-681e8fff2d30 req-7e1d6e26-a5e0-4074-915b-e1de3ce3e431 service nova] [instance: ee002e7c-8965-494f-a1af-0b6f3b74eceb] Received event network-vif-plugged-5c3f2277-5769-4c2e-8eb9-df0a99f35008 {{(pid=61806) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 896.953312] env[61806]: DEBUG oslo_concurrency.lockutils [req-b664b826-71c0-46bc-897e-681e8fff2d30 req-7e1d6e26-a5e0-4074-915b-e1de3ce3e431 service nova] Acquiring lock "ee002e7c-8965-494f-a1af-0b6f3b74eceb-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 896.953542] env[61806]: DEBUG oslo_concurrency.lockutils [req-b664b826-71c0-46bc-897e-681e8fff2d30 req-7e1d6e26-a5e0-4074-915b-e1de3ce3e431 service nova] Lock "ee002e7c-8965-494f-a1af-0b6f3b74eceb-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 896.953720] env[61806]: DEBUG oslo_concurrency.lockutils [req-b664b826-71c0-46bc-897e-681e8fff2d30 req-7e1d6e26-a5e0-4074-915b-e1de3ce3e431 service nova] Lock "ee002e7c-8965-494f-a1af-0b6f3b74eceb-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 896.953914] env[61806]: DEBUG nova.compute.manager [req-b664b826-71c0-46bc-897e-681e8fff2d30 req-7e1d6e26-a5e0-4074-915b-e1de3ce3e431 service nova] [instance: ee002e7c-8965-494f-a1af-0b6f3b74eceb] No waiting events found dispatching network-vif-plugged-5c3f2277-5769-4c2e-8eb9-df0a99f35008 {{(pid=61806) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 896.954106] env[61806]: WARNING nova.compute.manager [req-b664b826-71c0-46bc-897e-681e8fff2d30 req-7e1d6e26-a5e0-4074-915b-e1de3ce3e431 service nova] [instance: ee002e7c-8965-494f-a1af-0b6f3b74eceb] Received unexpected event network-vif-plugged-5c3f2277-5769-4c2e-8eb9-df0a99f35008 for instance with vm_state building and task_state spawning. [ 896.954290] env[61806]: DEBUG nova.compute.manager [req-b664b826-71c0-46bc-897e-681e8fff2d30 req-7e1d6e26-a5e0-4074-915b-e1de3ce3e431 service nova] [instance: ee002e7c-8965-494f-a1af-0b6f3b74eceb] Received event network-changed-5c3f2277-5769-4c2e-8eb9-df0a99f35008 {{(pid=61806) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 896.954462] env[61806]: DEBUG nova.compute.manager [req-b664b826-71c0-46bc-897e-681e8fff2d30 req-7e1d6e26-a5e0-4074-915b-e1de3ce3e431 service nova] [instance: ee002e7c-8965-494f-a1af-0b6f3b74eceb] Refreshing instance network info cache due to event network-changed-5c3f2277-5769-4c2e-8eb9-df0a99f35008. {{(pid=61806) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 896.954653] env[61806]: DEBUG oslo_concurrency.lockutils [req-b664b826-71c0-46bc-897e-681e8fff2d30 req-7e1d6e26-a5e0-4074-915b-e1de3ce3e431 service nova] Acquiring lock "refresh_cache-ee002e7c-8965-494f-a1af-0b6f3b74eceb" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 897.013468] env[61806]: DEBUG nova.network.neutron [None req-6d4e5237-b28c-45e9-8ae0-8ed79347c566 tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] [instance: 75f1aded-d1da-4ddc-be29-f1fb93799364] Successfully created port: a06a82f0-c5d3-48b7-bd4b-d2725bd7eaec {{(pid=61806) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 897.149361] env[61806]: DEBUG nova.compute.manager [None req-6d4e5237-b28c-45e9-8ae0-8ed79347c566 tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] [instance: 75f1aded-d1da-4ddc-be29-f1fb93799364] Start building block device mappings for instance. {{(pid=61806) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 897.284574] env[61806]: INFO nova.compute.manager [None req-464e7d68-de9c-4f8c-9477-932cd9673dfd tempest-ImagesTestJSON-618852556 tempest-ImagesTestJSON-618852556-project-member] [instance: 1a73d994-5f21-4f80-8f33-5830d860a7ee] instance snapshotting [ 897.294125] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b13c980c-dbdb-449c-b614-f78c54601c30 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 897.303255] env[61806]: DEBUG oslo_vmware.api [None req-0bb2f11a-fff8-4c85-aa76-4f260c30bde9 tempest-AttachVolumeNegativeTest-1683033704 tempest-AttachVolumeNegativeTest-1683033704-project-member] Task: {'id': task-1294837, 'name': ReconfigVM_Task, 'duration_secs': 0.305689} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 897.320097] env[61806]: DEBUG nova.virt.vmwareapi.volumeops [None req-0bb2f11a-fff8-4c85-aa76-4f260c30bde9 tempest-AttachVolumeNegativeTest-1683033704 tempest-AttachVolumeNegativeTest-1683033704-project-member] [instance: daff849e-eb4f-411e-92f8-9362e3bd91e3] Reconfigured VM instance instance-0000004f to attach disk [datastore2] daff849e-eb4f-411e-92f8-9362e3bd91e3/daff849e-eb4f-411e-92f8-9362e3bd91e3.vmdk or device None with type sparse {{(pid=61806) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 897.321451] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-965fcc80-ee5a-4dd8-89bf-921cef7412ca {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 897.323728] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3d54d211-4a6d-4dcc-b628-623a1e8ff52a {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 897.326889] env[61806]: DEBUG nova.network.neutron [None req-80f76c83-748f-43fe-b146-5517b19ec0ae tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] [instance: ee002e7c-8965-494f-a1af-0b6f3b74eceb] Instance cache missing network info. {{(pid=61806) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 897.335050] env[61806]: DEBUG oslo_vmware.api [None req-f9554f19-c70d-44f1-bec2-80b2a52a5572 tempest-ServerRescueTestJSONUnderV235-749414849 tempest-ServerRescueTestJSONUnderV235-749414849-project-member] Task: {'id': task-1294838, 'name': CopyVirtualDisk_Task} progress is 25%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 897.341751] env[61806]: DEBUG oslo_vmware.api [None req-0bb2f11a-fff8-4c85-aa76-4f260c30bde9 tempest-AttachVolumeNegativeTest-1683033704 tempest-AttachVolumeNegativeTest-1683033704-project-member] Waiting for the task: (returnval){ [ 897.341751] env[61806]: value = "task-1294839" [ 897.341751] env[61806]: _type = "Task" [ 897.341751] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 897.367028] env[61806]: DEBUG oslo_vmware.api [None req-0bb2f11a-fff8-4c85-aa76-4f260c30bde9 tempest-AttachVolumeNegativeTest-1683033704 tempest-AttachVolumeNegativeTest-1683033704-project-member] Task: {'id': task-1294839, 'name': Rename_Task} progress is 14%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 897.367252] env[61806]: DEBUG oslo_vmware.api [None req-095bcbe6-2af7-4761-a8ad-c73adfe5ab9d tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] Task: {'id': session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]52ddb997-500a-ab5c-1577-0464bf0fca91, 'name': SearchDatastore_Task, 'duration_secs': 0.01368} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 897.367466] env[61806]: DEBUG oslo_concurrency.lockutils [None req-095bcbe6-2af7-4761-a8ad-c73adfe5ab9d tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] Releasing lock "[datastore2] devstack-image-cache_base/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa.vmdk" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 897.368118] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [None req-095bcbe6-2af7-4761-a8ad-c73adfe5ab9d tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa.vmdk to [datastore2] 6f6ba57a-e2d6-4749-a53a-e263861cb1c0/6f6ba57a-e2d6-4749-a53a-e263861cb1c0.vmdk {{(pid=61806) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 897.368274] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-cac037b8-9d94-48cc-b1c2-5c8842f74968 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 897.377033] env[61806]: DEBUG oslo_vmware.api [None req-095bcbe6-2af7-4761-a8ad-c73adfe5ab9d tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] Waiting for the task: (returnval){ [ 897.377033] env[61806]: value = "task-1294840" [ 897.377033] env[61806]: _type = "Task" [ 897.377033] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 897.391842] env[61806]: DEBUG oslo_vmware.api [None req-095bcbe6-2af7-4761-a8ad-c73adfe5ab9d tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] Task: {'id': task-1294840, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 897.456168] env[61806]: DEBUG nova.scheduler.client.report [None req-f488cd5c-2a2f-4173-ba8b-c601bc6a140a tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Inventory has not changed for provider a2858be1-fd22-4e08-979e-87ad25293407 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 138, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61806) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 897.536610] env[61806]: DEBUG nova.virt.vmwareapi.volumeops [None req-76aa021b-a2c9-4c97-bbb8-a4fc0140ddae tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] [instance: 59c22c16-3d1c-4e64-bfac-7f8886a1927e] Volume attach. Driver type: vmdk {{(pid=61806) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 897.536745] env[61806]: DEBUG nova.virt.vmwareapi.volumeops [None req-76aa021b-a2c9-4c97-bbb8-a4fc0140ddae tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] [instance: 59c22c16-3d1c-4e64-bfac-7f8886a1927e] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-277703', 'volume_id': '564a9f8e-1bec-410b-9d7d-ab9bf7b6d83b', 'name': 'volume-564a9f8e-1bec-410b-9d7d-ab9bf7b6d83b', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '59c22c16-3d1c-4e64-bfac-7f8886a1927e', 'attached_at': '', 'detached_at': '', 'volume_id': '564a9f8e-1bec-410b-9d7d-ab9bf7b6d83b', 'serial': '564a9f8e-1bec-410b-9d7d-ab9bf7b6d83b'} {{(pid=61806) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 897.537693] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-08939129-b9c9-4697-b8e3-75b2478876dc {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 897.558330] env[61806]: DEBUG nova.network.neutron [None req-80f76c83-748f-43fe-b146-5517b19ec0ae tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] [instance: ee002e7c-8965-494f-a1af-0b6f3b74eceb] Updating instance_info_cache with network_info: [{"id": "5c3f2277-5769-4c2e-8eb9-df0a99f35008", "address": "fa:16:3e:94:7f:ef", "network": {"id": "f280a01b-7000-4e54-ab9a-20eab156a410", "bridge": "br-int", "label": "tempest-ServersTestJSON-2080718146-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "73ebcc4d79d248efb653a084a5e44302", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "43798f54-0c07-4417-a23f-58bb6b7e204b", "external-id": "nsx-vlan-transportzone-571", "segmentation_id": 571, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap5c3f2277-57", "ovs_interfaceid": "5c3f2277-5769-4c2e-8eb9-df0a99f35008", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61806) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 897.560725] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-31ec7e73-2097-4fb9-9dc8-873c1cac9746 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 897.589610] env[61806]: DEBUG nova.virt.vmwareapi.volumeops [None req-76aa021b-a2c9-4c97-bbb8-a4fc0140ddae tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] [instance: 59c22c16-3d1c-4e64-bfac-7f8886a1927e] Reconfiguring VM instance instance-0000004b to attach disk [datastore2] volume-564a9f8e-1bec-410b-9d7d-ab9bf7b6d83b/volume-564a9f8e-1bec-410b-9d7d-ab9bf7b6d83b.vmdk or device None with type thin {{(pid=61806) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 897.590682] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-2e4b1b48-35af-4c7f-a69d-36d43398f749 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 897.612236] env[61806]: DEBUG oslo_vmware.api [None req-76aa021b-a2c9-4c97-bbb8-a4fc0140ddae tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Waiting for the task: (returnval){ [ 897.612236] env[61806]: value = "task-1294841" [ 897.612236] env[61806]: _type = "Task" [ 897.612236] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 897.622660] env[61806]: DEBUG oslo_vmware.api [None req-76aa021b-a2c9-4c97-bbb8-a4fc0140ddae tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Task: {'id': task-1294841, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 897.814068] env[61806]: DEBUG oslo_vmware.api [None req-f9554f19-c70d-44f1-bec2-80b2a52a5572 tempest-ServerRescueTestJSONUnderV235-749414849 tempest-ServerRescueTestJSONUnderV235-749414849-project-member] Task: {'id': task-1294838, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.743408} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 897.814489] env[61806]: INFO nova.virt.vmwareapi.ds_util [None req-f9554f19-c70d-44f1-bec2-80b2a52a5572 tempest-ServerRescueTestJSONUnderV235-749414849 tempest-ServerRescueTestJSONUnderV235-749414849-project-member] Copied virtual disk from [datastore2] devstack-image-cache_base/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa.vmdk to [datastore2] 5a526856-0dcd-4b3f-8359-13a5c3b9bfc4/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa-rescue.vmdk. [ 897.815398] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a8f93ca9-6f17-42ac-955f-7bd98246cf63 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 897.842364] env[61806]: DEBUG nova.virt.vmwareapi.volumeops [None req-f9554f19-c70d-44f1-bec2-80b2a52a5572 tempest-ServerRescueTestJSONUnderV235-749414849 tempest-ServerRescueTestJSONUnderV235-749414849-project-member] [instance: 5a526856-0dcd-4b3f-8359-13a5c3b9bfc4] Reconfiguring VM instance instance-0000004d to attach disk [datastore2] 5a526856-0dcd-4b3f-8359-13a5c3b9bfc4/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa-rescue.vmdk or device None with type thin {{(pid=61806) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 897.843613] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-464e7d68-de9c-4f8c-9477-932cd9673dfd tempest-ImagesTestJSON-618852556 tempest-ImagesTestJSON-618852556-project-member] [instance: 1a73d994-5f21-4f80-8f33-5830d860a7ee] Creating Snapshot of the VM instance {{(pid=61806) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 897.843918] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-698addbe-cde4-40c6-b039-3960cd16a078 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 897.857238] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-af08f63c-df78-4277-ae58-dac42f1aa030 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 897.870800] env[61806]: DEBUG oslo_vmware.api [None req-0bb2f11a-fff8-4c85-aa76-4f260c30bde9 tempest-AttachVolumeNegativeTest-1683033704 tempest-AttachVolumeNegativeTest-1683033704-project-member] Task: {'id': task-1294839, 'name': Rename_Task, 'duration_secs': 0.287155} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 897.873710] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [None req-0bb2f11a-fff8-4c85-aa76-4f260c30bde9 tempest-AttachVolumeNegativeTest-1683033704 tempest-AttachVolumeNegativeTest-1683033704-project-member] [instance: daff849e-eb4f-411e-92f8-9362e3bd91e3] Powering on the VM {{(pid=61806) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 897.874090] env[61806]: DEBUG oslo_vmware.api [None req-f9554f19-c70d-44f1-bec2-80b2a52a5572 tempest-ServerRescueTestJSONUnderV235-749414849 tempest-ServerRescueTestJSONUnderV235-749414849-project-member] Waiting for the task: (returnval){ [ 897.874090] env[61806]: value = "task-1294842" [ 897.874090] env[61806]: _type = "Task" [ 897.874090] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 897.874351] env[61806]: DEBUG oslo_vmware.api [None req-464e7d68-de9c-4f8c-9477-932cd9673dfd tempest-ImagesTestJSON-618852556 tempest-ImagesTestJSON-618852556-project-member] Waiting for the task: (returnval){ [ 897.874351] env[61806]: value = "task-1294843" [ 897.874351] env[61806]: _type = "Task" [ 897.874351] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 897.874580] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-465381f4-5841-40af-8b7f-4ba347fba24c {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 897.888957] env[61806]: DEBUG oslo_vmware.api [None req-0bb2f11a-fff8-4c85-aa76-4f260c30bde9 tempest-AttachVolumeNegativeTest-1683033704 tempest-AttachVolumeNegativeTest-1683033704-project-member] Waiting for the task: (returnval){ [ 897.888957] env[61806]: value = "task-1294844" [ 897.888957] env[61806]: _type = "Task" [ 897.888957] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 897.896070] env[61806]: DEBUG oslo_vmware.api [None req-f9554f19-c70d-44f1-bec2-80b2a52a5572 tempest-ServerRescueTestJSONUnderV235-749414849 tempest-ServerRescueTestJSONUnderV235-749414849-project-member] Task: {'id': task-1294842, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 897.896352] env[61806]: DEBUG oslo_vmware.api [None req-464e7d68-de9c-4f8c-9477-932cd9673dfd tempest-ImagesTestJSON-618852556 tempest-ImagesTestJSON-618852556-project-member] Task: {'id': task-1294843, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 897.903870] env[61806]: DEBUG oslo_vmware.api [None req-095bcbe6-2af7-4761-a8ad-c73adfe5ab9d tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] Task: {'id': task-1294840, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 897.907106] env[61806]: DEBUG oslo_vmware.api [None req-0bb2f11a-fff8-4c85-aa76-4f260c30bde9 tempest-AttachVolumeNegativeTest-1683033704 tempest-AttachVolumeNegativeTest-1683033704-project-member] Task: {'id': task-1294844, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 897.961775] env[61806]: DEBUG oslo_concurrency.lockutils [None req-f488cd5c-2a2f-4173-ba8b-c601bc6a140a tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.823s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 897.964430] env[61806]: DEBUG oslo_concurrency.lockutils [None req-18f7d4e8-522e-4e97-b6f4-8662acf8d0f8 tempest-ServerTagsTestJSON-1292926242 tempest-ServerTagsTestJSON-1292926242-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 4.495s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 897.964706] env[61806]: DEBUG nova.objects.instance [None req-18f7d4e8-522e-4e97-b6f4-8662acf8d0f8 tempest-ServerTagsTestJSON-1292926242 tempest-ServerTagsTestJSON-1292926242-project-member] Lazy-loading 'resources' on Instance uuid 9c20d337-b8f5-4965-b707-0f87c4ab6fc4 {{(pid=61806) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 897.987198] env[61806]: INFO nova.scheduler.client.report [None req-f488cd5c-2a2f-4173-ba8b-c601bc6a140a tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Deleted allocations for instance efa19aac-3ad7-42d5-a6d8-859050de6e63 [ 898.064013] env[61806]: DEBUG oslo_concurrency.lockutils [None req-80f76c83-748f-43fe-b146-5517b19ec0ae tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] Releasing lock "refresh_cache-ee002e7c-8965-494f-a1af-0b6f3b74eceb" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 898.064450] env[61806]: DEBUG nova.compute.manager [None req-80f76c83-748f-43fe-b146-5517b19ec0ae tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] [instance: ee002e7c-8965-494f-a1af-0b6f3b74eceb] Instance network_info: |[{"id": "5c3f2277-5769-4c2e-8eb9-df0a99f35008", "address": "fa:16:3e:94:7f:ef", "network": {"id": "f280a01b-7000-4e54-ab9a-20eab156a410", "bridge": "br-int", "label": "tempest-ServersTestJSON-2080718146-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "73ebcc4d79d248efb653a084a5e44302", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "43798f54-0c07-4417-a23f-58bb6b7e204b", "external-id": "nsx-vlan-transportzone-571", "segmentation_id": 571, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap5c3f2277-57", "ovs_interfaceid": "5c3f2277-5769-4c2e-8eb9-df0a99f35008", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61806) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 898.064865] env[61806]: DEBUG oslo_concurrency.lockutils [req-b664b826-71c0-46bc-897e-681e8fff2d30 req-7e1d6e26-a5e0-4074-915b-e1de3ce3e431 service nova] Acquired lock "refresh_cache-ee002e7c-8965-494f-a1af-0b6f3b74eceb" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 898.065129] env[61806]: DEBUG nova.network.neutron [req-b664b826-71c0-46bc-897e-681e8fff2d30 req-7e1d6e26-a5e0-4074-915b-e1de3ce3e431 service nova] [instance: ee002e7c-8965-494f-a1af-0b6f3b74eceb] Refreshing network info cache for port 5c3f2277-5769-4c2e-8eb9-df0a99f35008 {{(pid=61806) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 898.066379] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-80f76c83-748f-43fe-b146-5517b19ec0ae tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] [instance: ee002e7c-8965-494f-a1af-0b6f3b74eceb] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:94:7f:ef', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '43798f54-0c07-4417-a23f-58bb6b7e204b', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '5c3f2277-5769-4c2e-8eb9-df0a99f35008', 'vif_model': 'vmxnet3'}] {{(pid=61806) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 898.075669] env[61806]: DEBUG oslo.service.loopingcall [None req-80f76c83-748f-43fe-b146-5517b19ec0ae tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61806) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 898.079176] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: ee002e7c-8965-494f-a1af-0b6f3b74eceb] Creating VM on the ESX host {{(pid=61806) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 898.079755] env[61806]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-e095d587-442c-42a7-9fdf-01be51794834 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 898.104238] env[61806]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 898.104238] env[61806]: value = "task-1294845" [ 898.104238] env[61806]: _type = "Task" [ 898.104238] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 898.113577] env[61806]: DEBUG oslo_vmware.api [-] Task: {'id': task-1294845, 'name': CreateVM_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 898.122553] env[61806]: DEBUG oslo_vmware.api [None req-76aa021b-a2c9-4c97-bbb8-a4fc0140ddae tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Task: {'id': task-1294841, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 898.159986] env[61806]: DEBUG nova.compute.manager [None req-6d4e5237-b28c-45e9-8ae0-8ed79347c566 tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] [instance: 75f1aded-d1da-4ddc-be29-f1fb93799364] Start spawning the instance on the hypervisor. {{(pid=61806) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 898.190586] env[61806]: DEBUG nova.virt.hardware [None req-6d4e5237-b28c-45e9-8ae0-8ed79347c566 tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-15T18:05:10Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-15T18:04:52Z,direct_url=,disk_format='vmdk',id=ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='84ba83fce2c349a6988173c5d6fc3b07',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-15T18:04:53Z,virtual_size=,visibility=), allow threads: False {{(pid=61806) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 898.190897] env[61806]: DEBUG nova.virt.hardware [None req-6d4e5237-b28c-45e9-8ae0-8ed79347c566 tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Flavor limits 0:0:0 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 898.191083] env[61806]: DEBUG nova.virt.hardware [None req-6d4e5237-b28c-45e9-8ae0-8ed79347c566 tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Image limits 0:0:0 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 898.191281] env[61806]: DEBUG nova.virt.hardware [None req-6d4e5237-b28c-45e9-8ae0-8ed79347c566 tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Flavor pref 0:0:0 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 898.191434] env[61806]: DEBUG nova.virt.hardware [None req-6d4e5237-b28c-45e9-8ae0-8ed79347c566 tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Image pref 0:0:0 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 898.191589] env[61806]: DEBUG nova.virt.hardware [None req-6d4e5237-b28c-45e9-8ae0-8ed79347c566 tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 898.191812] env[61806]: DEBUG nova.virt.hardware [None req-6d4e5237-b28c-45e9-8ae0-8ed79347c566 tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61806) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 898.191978] env[61806]: DEBUG nova.virt.hardware [None req-6d4e5237-b28c-45e9-8ae0-8ed79347c566 tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61806) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 898.192167] env[61806]: DEBUG nova.virt.hardware [None req-6d4e5237-b28c-45e9-8ae0-8ed79347c566 tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Got 1 possible topologies {{(pid=61806) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 898.192336] env[61806]: DEBUG nova.virt.hardware [None req-6d4e5237-b28c-45e9-8ae0-8ed79347c566 tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61806) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 898.192529] env[61806]: DEBUG nova.virt.hardware [None req-6d4e5237-b28c-45e9-8ae0-8ed79347c566 tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61806) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 898.193444] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d7da18c4-18fe-4801-8dae-17be66406062 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 898.202623] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6c176d60-ff5b-4ce8-937f-88e4f22c7051 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 898.369313] env[61806]: DEBUG nova.network.neutron [req-b664b826-71c0-46bc-897e-681e8fff2d30 req-7e1d6e26-a5e0-4074-915b-e1de3ce3e431 service nova] [instance: ee002e7c-8965-494f-a1af-0b6f3b74eceb] Updated VIF entry in instance network info cache for port 5c3f2277-5769-4c2e-8eb9-df0a99f35008. {{(pid=61806) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 898.370045] env[61806]: DEBUG nova.network.neutron [req-b664b826-71c0-46bc-897e-681e8fff2d30 req-7e1d6e26-a5e0-4074-915b-e1de3ce3e431 service nova] [instance: ee002e7c-8965-494f-a1af-0b6f3b74eceb] Updating instance_info_cache with network_info: [{"id": "5c3f2277-5769-4c2e-8eb9-df0a99f35008", "address": "fa:16:3e:94:7f:ef", "network": {"id": "f280a01b-7000-4e54-ab9a-20eab156a410", "bridge": "br-int", "label": "tempest-ServersTestJSON-2080718146-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "73ebcc4d79d248efb653a084a5e44302", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "43798f54-0c07-4417-a23f-58bb6b7e204b", "external-id": "nsx-vlan-transportzone-571", "segmentation_id": 571, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap5c3f2277-57", "ovs_interfaceid": "5c3f2277-5769-4c2e-8eb9-df0a99f35008", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61806) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 898.392765] env[61806]: DEBUG oslo_vmware.api [None req-f9554f19-c70d-44f1-bec2-80b2a52a5572 tempest-ServerRescueTestJSONUnderV235-749414849 tempest-ServerRescueTestJSONUnderV235-749414849-project-member] Task: {'id': task-1294842, 'name': ReconfigVM_Task, 'duration_secs': 0.515659} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 898.396423] env[61806]: DEBUG nova.virt.vmwareapi.volumeops [None req-f9554f19-c70d-44f1-bec2-80b2a52a5572 tempest-ServerRescueTestJSONUnderV235-749414849 tempest-ServerRescueTestJSONUnderV235-749414849-project-member] [instance: 5a526856-0dcd-4b3f-8359-13a5c3b9bfc4] Reconfigured VM instance instance-0000004d to attach disk [datastore2] 5a526856-0dcd-4b3f-8359-13a5c3b9bfc4/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa-rescue.vmdk or device None with type thin {{(pid=61806) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 898.397086] env[61806]: DEBUG oslo_vmware.api [None req-464e7d68-de9c-4f8c-9477-932cd9673dfd tempest-ImagesTestJSON-618852556 tempest-ImagesTestJSON-618852556-project-member] Task: {'id': task-1294843, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 898.398097] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-51177e70-ab86-47b0-95fd-8d063df0eb8a {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 898.409755] env[61806]: DEBUG oslo_vmware.api [None req-095bcbe6-2af7-4761-a8ad-c73adfe5ab9d tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] Task: {'id': task-1294840, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.63411} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 898.410947] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [None req-095bcbe6-2af7-4761-a8ad-c73adfe5ab9d tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa.vmdk to [datastore2] 6f6ba57a-e2d6-4749-a53a-e263861cb1c0/6f6ba57a-e2d6-4749-a53a-e263861cb1c0.vmdk {{(pid=61806) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 898.411243] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-095bcbe6-2af7-4761-a8ad-c73adfe5ab9d tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] [instance: 6f6ba57a-e2d6-4749-a53a-e263861cb1c0] Extending root virtual disk to 1048576 {{(pid=61806) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 898.411541] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-f2fc925a-4b33-4836-aaa0-3ad187436eb3 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 898.436993] env[61806]: DEBUG oslo_vmware.api [None req-0bb2f11a-fff8-4c85-aa76-4f260c30bde9 tempest-AttachVolumeNegativeTest-1683033704 tempest-AttachVolumeNegativeTest-1683033704-project-member] Task: {'id': task-1294844, 'name': PowerOnVM_Task} progress is 88%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 898.442907] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-b004252c-d64f-4485-9299-ba185df4ae0d {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 898.455626] env[61806]: DEBUG oslo_vmware.api [None req-095bcbe6-2af7-4761-a8ad-c73adfe5ab9d tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] Waiting for the task: (returnval){ [ 898.455626] env[61806]: value = "task-1294846" [ 898.455626] env[61806]: _type = "Task" [ 898.455626] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 898.461544] env[61806]: DEBUG oslo_vmware.api [None req-f9554f19-c70d-44f1-bec2-80b2a52a5572 tempest-ServerRescueTestJSONUnderV235-749414849 tempest-ServerRescueTestJSONUnderV235-749414849-project-member] Waiting for the task: (returnval){ [ 898.461544] env[61806]: value = "task-1294847" [ 898.461544] env[61806]: _type = "Task" [ 898.461544] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 898.465401] env[61806]: DEBUG oslo_vmware.api [None req-095bcbe6-2af7-4761-a8ad-c73adfe5ab9d tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] Task: {'id': task-1294846, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 898.477377] env[61806]: DEBUG oslo_vmware.api [None req-f9554f19-c70d-44f1-bec2-80b2a52a5572 tempest-ServerRescueTestJSONUnderV235-749414849 tempest-ServerRescueTestJSONUnderV235-749414849-project-member] Task: {'id': task-1294847, 'name': ReconfigVM_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 898.495061] env[61806]: DEBUG oslo_concurrency.lockutils [None req-f488cd5c-2a2f-4173-ba8b-c601bc6a140a tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Lock "efa19aac-3ad7-42d5-a6d8-859050de6e63" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 10.480s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 898.619939] env[61806]: DEBUG oslo_vmware.api [-] Task: {'id': task-1294845, 'name': CreateVM_Task, 'duration_secs': 0.456866} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 898.624821] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: ee002e7c-8965-494f-a1af-0b6f3b74eceb] Created VM on the ESX host {{(pid=61806) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 898.626446] env[61806]: DEBUG oslo_concurrency.lockutils [None req-80f76c83-748f-43fe-b146-5517b19ec0ae tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 898.626446] env[61806]: DEBUG oslo_concurrency.lockutils [None req-80f76c83-748f-43fe-b146-5517b19ec0ae tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] Acquired lock "[datastore2] devstack-image-cache_base/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 898.626729] env[61806]: DEBUG oslo_concurrency.lockutils [None req-80f76c83-748f-43fe-b146-5517b19ec0ae tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 898.630177] env[61806]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-9b5cc01e-a313-4f61-af5d-1ac171350729 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 898.632340] env[61806]: DEBUG oslo_vmware.api [None req-76aa021b-a2c9-4c97-bbb8-a4fc0140ddae tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Task: {'id': task-1294841, 'name': ReconfigVM_Task, 'duration_secs': 0.906516} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 898.633084] env[61806]: DEBUG nova.virt.vmwareapi.volumeops [None req-76aa021b-a2c9-4c97-bbb8-a4fc0140ddae tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] [instance: 59c22c16-3d1c-4e64-bfac-7f8886a1927e] Reconfigured VM instance instance-0000004b to attach disk [datastore2] volume-564a9f8e-1bec-410b-9d7d-ab9bf7b6d83b/volume-564a9f8e-1bec-410b-9d7d-ab9bf7b6d83b.vmdk or device None with type thin {{(pid=61806) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 898.643078] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-fec54da3-8e99-49d2-b386-d304463aba3f {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 898.657267] env[61806]: DEBUG oslo_vmware.api [None req-80f76c83-748f-43fe-b146-5517b19ec0ae tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] Waiting for the task: (returnval){ [ 898.657267] env[61806]: value = "session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]52fb275e-1e2b-e7fd-709a-748b9fa81e66" [ 898.657267] env[61806]: _type = "Task" [ 898.657267] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 898.672698] env[61806]: DEBUG oslo_vmware.api [None req-76aa021b-a2c9-4c97-bbb8-a4fc0140ddae tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Waiting for the task: (returnval){ [ 898.672698] env[61806]: value = "task-1294848" [ 898.672698] env[61806]: _type = "Task" [ 898.672698] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 898.678047] env[61806]: DEBUG nova.network.neutron [None req-6d4e5237-b28c-45e9-8ae0-8ed79347c566 tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] [instance: 75f1aded-d1da-4ddc-be29-f1fb93799364] Successfully updated port: a06a82f0-c5d3-48b7-bd4b-d2725bd7eaec {{(pid=61806) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 898.681644] env[61806]: DEBUG oslo_vmware.api [None req-80f76c83-748f-43fe-b146-5517b19ec0ae tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] Task: {'id': session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]52fb275e-1e2b-e7fd-709a-748b9fa81e66, 'name': SearchDatastore_Task, 'duration_secs': 0.037399} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 898.681644] env[61806]: DEBUG oslo_concurrency.lockutils [None req-80f76c83-748f-43fe-b146-5517b19ec0ae tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] Releasing lock "[datastore2] devstack-image-cache_base/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 898.681644] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-80f76c83-748f-43fe-b146-5517b19ec0ae tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] [instance: ee002e7c-8965-494f-a1af-0b6f3b74eceb] Processing image ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa {{(pid=61806) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 898.681644] env[61806]: DEBUG oslo_concurrency.lockutils [None req-80f76c83-748f-43fe-b146-5517b19ec0ae tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa.vmdk" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 898.681644] env[61806]: DEBUG oslo_concurrency.lockutils [None req-80f76c83-748f-43fe-b146-5517b19ec0ae tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] Acquired lock "[datastore2] devstack-image-cache_base/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa.vmdk" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 898.681644] env[61806]: DEBUG nova.virt.vmwareapi.ds_util [None req-80f76c83-748f-43fe-b146-5517b19ec0ae tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61806) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 898.682358] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6ac00baf-27a0-4926-bd12-4028ed9f56a3 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 898.688942] env[61806]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-6b8876b3-6832-4db3-b6a4-e1be8dff9c77 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 898.691036] env[61806]: DEBUG oslo_vmware.api [None req-76aa021b-a2c9-4c97-bbb8-a4fc0140ddae tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Task: {'id': task-1294848, 'name': ReconfigVM_Task} progress is 10%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 898.697881] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3741e1cd-3c33-4cee-b423-1db25f6c6326 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 898.730507] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-48c36cd2-6557-4409-b9a4-dbd24d90ba3d {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 898.733595] env[61806]: DEBUG nova.virt.vmwareapi.ds_util [None req-80f76c83-748f-43fe-b146-5517b19ec0ae tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61806) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 898.733595] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-80f76c83-748f-43fe-b146-5517b19ec0ae tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61806) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 898.734071] env[61806]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-fc3c53c7-1346-4a67-b239-5adb39c0a4d9 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 898.742137] env[61806]: DEBUG oslo_vmware.api [None req-80f76c83-748f-43fe-b146-5517b19ec0ae tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] Waiting for the task: (returnval){ [ 898.742137] env[61806]: value = "session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]52d21fed-c935-8921-fa43-262fa63cb1a8" [ 898.742137] env[61806]: _type = "Task" [ 898.742137] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 898.743473] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-979536dd-92a1-41b6-8da3-4fed947a5115 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 898.756689] env[61806]: DEBUG oslo_vmware.api [None req-80f76c83-748f-43fe-b146-5517b19ec0ae tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] Task: {'id': session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]52d21fed-c935-8921-fa43-262fa63cb1a8, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 898.768093] env[61806]: DEBUG nova.compute.provider_tree [None req-18f7d4e8-522e-4e97-b6f4-8662acf8d0f8 tempest-ServerTagsTestJSON-1292926242 tempest-ServerTagsTestJSON-1292926242-project-member] Inventory has not changed in ProviderTree for provider: a2858be1-fd22-4e08-979e-87ad25293407 {{(pid=61806) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 898.873209] env[61806]: DEBUG oslo_concurrency.lockutils [req-b664b826-71c0-46bc-897e-681e8fff2d30 req-7e1d6e26-a5e0-4074-915b-e1de3ce3e431 service nova] Releasing lock "refresh_cache-ee002e7c-8965-494f-a1af-0b6f3b74eceb" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 898.887066] env[61806]: DEBUG oslo_vmware.api [None req-464e7d68-de9c-4f8c-9477-932cd9673dfd tempest-ImagesTestJSON-618852556 tempest-ImagesTestJSON-618852556-project-member] Task: {'id': task-1294843, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 898.906542] env[61806]: DEBUG oslo_vmware.api [None req-0bb2f11a-fff8-4c85-aa76-4f260c30bde9 tempest-AttachVolumeNegativeTest-1683033704 tempest-AttachVolumeNegativeTest-1683033704-project-member] Task: {'id': task-1294844, 'name': PowerOnVM_Task, 'duration_secs': 0.761141} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 898.907143] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [None req-0bb2f11a-fff8-4c85-aa76-4f260c30bde9 tempest-AttachVolumeNegativeTest-1683033704 tempest-AttachVolumeNegativeTest-1683033704-project-member] [instance: daff849e-eb4f-411e-92f8-9362e3bd91e3] Powered on the VM {{(pid=61806) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 898.907382] env[61806]: INFO nova.compute.manager [None req-0bb2f11a-fff8-4c85-aa76-4f260c30bde9 tempest-AttachVolumeNegativeTest-1683033704 tempest-AttachVolumeNegativeTest-1683033704-project-member] [instance: daff849e-eb4f-411e-92f8-9362e3bd91e3] Took 8.62 seconds to spawn the instance on the hypervisor. [ 898.907581] env[61806]: DEBUG nova.compute.manager [None req-0bb2f11a-fff8-4c85-aa76-4f260c30bde9 tempest-AttachVolumeNegativeTest-1683033704 tempest-AttachVolumeNegativeTest-1683033704-project-member] [instance: daff849e-eb4f-411e-92f8-9362e3bd91e3] Checking state {{(pid=61806) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 898.908327] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-25896b44-c021-4a34-9c31-a1e2502ae669 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 898.966269] env[61806]: DEBUG oslo_vmware.api [None req-095bcbe6-2af7-4761-a8ad-c73adfe5ab9d tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] Task: {'id': task-1294846, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.169779} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 898.969294] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-095bcbe6-2af7-4761-a8ad-c73adfe5ab9d tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] [instance: 6f6ba57a-e2d6-4749-a53a-e263861cb1c0] Extended root virtual disk {{(pid=61806) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 898.970280] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d3ca977d-afad-41b6-ac2c-b15598bad562 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 898.975971] env[61806]: DEBUG nova.compute.manager [req-5cb4bad8-c976-40ab-b3de-65a7e4c55e69 req-9797cb2e-4b29-44c7-91e4-c69b781ca16b service nova] [instance: 75f1aded-d1da-4ddc-be29-f1fb93799364] Received event network-vif-plugged-a06a82f0-c5d3-48b7-bd4b-d2725bd7eaec {{(pid=61806) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 898.976199] env[61806]: DEBUG oslo_concurrency.lockutils [req-5cb4bad8-c976-40ab-b3de-65a7e4c55e69 req-9797cb2e-4b29-44c7-91e4-c69b781ca16b service nova] Acquiring lock "75f1aded-d1da-4ddc-be29-f1fb93799364-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 898.976515] env[61806]: DEBUG oslo_concurrency.lockutils [req-5cb4bad8-c976-40ab-b3de-65a7e4c55e69 req-9797cb2e-4b29-44c7-91e4-c69b781ca16b service nova] Lock "75f1aded-d1da-4ddc-be29-f1fb93799364-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 898.976632] env[61806]: DEBUG oslo_concurrency.lockutils [req-5cb4bad8-c976-40ab-b3de-65a7e4c55e69 req-9797cb2e-4b29-44c7-91e4-c69b781ca16b service nova] Lock "75f1aded-d1da-4ddc-be29-f1fb93799364-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 898.976830] env[61806]: DEBUG nova.compute.manager [req-5cb4bad8-c976-40ab-b3de-65a7e4c55e69 req-9797cb2e-4b29-44c7-91e4-c69b781ca16b service nova] [instance: 75f1aded-d1da-4ddc-be29-f1fb93799364] No waiting events found dispatching network-vif-plugged-a06a82f0-c5d3-48b7-bd4b-d2725bd7eaec {{(pid=61806) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 898.977044] env[61806]: WARNING nova.compute.manager [req-5cb4bad8-c976-40ab-b3de-65a7e4c55e69 req-9797cb2e-4b29-44c7-91e4-c69b781ca16b service nova] [instance: 75f1aded-d1da-4ddc-be29-f1fb93799364] Received unexpected event network-vif-plugged-a06a82f0-c5d3-48b7-bd4b-d2725bd7eaec for instance with vm_state building and task_state spawning. [ 898.977294] env[61806]: DEBUG nova.compute.manager [req-5cb4bad8-c976-40ab-b3de-65a7e4c55e69 req-9797cb2e-4b29-44c7-91e4-c69b781ca16b service nova] [instance: 75f1aded-d1da-4ddc-be29-f1fb93799364] Received event network-changed-a06a82f0-c5d3-48b7-bd4b-d2725bd7eaec {{(pid=61806) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 898.977492] env[61806]: DEBUG nova.compute.manager [req-5cb4bad8-c976-40ab-b3de-65a7e4c55e69 req-9797cb2e-4b29-44c7-91e4-c69b781ca16b service nova] [instance: 75f1aded-d1da-4ddc-be29-f1fb93799364] Refreshing instance network info cache due to event network-changed-a06a82f0-c5d3-48b7-bd4b-d2725bd7eaec. {{(pid=61806) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 898.977706] env[61806]: DEBUG oslo_concurrency.lockutils [req-5cb4bad8-c976-40ab-b3de-65a7e4c55e69 req-9797cb2e-4b29-44c7-91e4-c69b781ca16b service nova] Acquiring lock "refresh_cache-75f1aded-d1da-4ddc-be29-f1fb93799364" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 898.977853] env[61806]: DEBUG oslo_concurrency.lockutils [req-5cb4bad8-c976-40ab-b3de-65a7e4c55e69 req-9797cb2e-4b29-44c7-91e4-c69b781ca16b service nova] Acquired lock "refresh_cache-75f1aded-d1da-4ddc-be29-f1fb93799364" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 898.978079] env[61806]: DEBUG nova.network.neutron [req-5cb4bad8-c976-40ab-b3de-65a7e4c55e69 req-9797cb2e-4b29-44c7-91e4-c69b781ca16b service nova] [instance: 75f1aded-d1da-4ddc-be29-f1fb93799364] Refreshing network info cache for port a06a82f0-c5d3-48b7-bd4b-d2725bd7eaec {{(pid=61806) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 899.000589] env[61806]: DEBUG nova.virt.vmwareapi.volumeops [None req-095bcbe6-2af7-4761-a8ad-c73adfe5ab9d tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] [instance: 6f6ba57a-e2d6-4749-a53a-e263861cb1c0] Reconfiguring VM instance instance-00000050 to attach disk [datastore2] 6f6ba57a-e2d6-4749-a53a-e263861cb1c0/6f6ba57a-e2d6-4749-a53a-e263861cb1c0.vmdk or device None with type sparse {{(pid=61806) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 899.000880] env[61806]: DEBUG oslo_vmware.api [None req-f9554f19-c70d-44f1-bec2-80b2a52a5572 tempest-ServerRescueTestJSONUnderV235-749414849 tempest-ServerRescueTestJSONUnderV235-749414849-project-member] Task: {'id': task-1294847, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 899.003832] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-7726291c-7fc2-43ab-b7d2-4124241f254f {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 899.027018] env[61806]: DEBUG oslo_vmware.api [None req-095bcbe6-2af7-4761-a8ad-c73adfe5ab9d tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] Waiting for the task: (returnval){ [ 899.027018] env[61806]: value = "task-1294849" [ 899.027018] env[61806]: _type = "Task" [ 899.027018] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 899.036620] env[61806]: DEBUG oslo_vmware.api [None req-095bcbe6-2af7-4761-a8ad-c73adfe5ab9d tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] Task: {'id': task-1294849, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 899.049466] env[61806]: DEBUG nova.network.neutron [req-5cb4bad8-c976-40ab-b3de-65a7e4c55e69 req-9797cb2e-4b29-44c7-91e4-c69b781ca16b service nova] [instance: 75f1aded-d1da-4ddc-be29-f1fb93799364] Instance cache missing network info. {{(pid=61806) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 899.160885] env[61806]: DEBUG nova.network.neutron [req-5cb4bad8-c976-40ab-b3de-65a7e4c55e69 req-9797cb2e-4b29-44c7-91e4-c69b781ca16b service nova] [instance: 75f1aded-d1da-4ddc-be29-f1fb93799364] Updating instance_info_cache with network_info: [] {{(pid=61806) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 899.184582] env[61806]: DEBUG oslo_vmware.api [None req-76aa021b-a2c9-4c97-bbb8-a4fc0140ddae tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Task: {'id': task-1294848, 'name': ReconfigVM_Task, 'duration_secs': 0.408573} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 899.184907] env[61806]: DEBUG nova.virt.vmwareapi.volumeops [None req-76aa021b-a2c9-4c97-bbb8-a4fc0140ddae tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] [instance: 59c22c16-3d1c-4e64-bfac-7f8886a1927e] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-277703', 'volume_id': '564a9f8e-1bec-410b-9d7d-ab9bf7b6d83b', 'name': 'volume-564a9f8e-1bec-410b-9d7d-ab9bf7b6d83b', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '59c22c16-3d1c-4e64-bfac-7f8886a1927e', 'attached_at': '', 'detached_at': '', 'volume_id': '564a9f8e-1bec-410b-9d7d-ab9bf7b6d83b', 'serial': '564a9f8e-1bec-410b-9d7d-ab9bf7b6d83b'} {{(pid=61806) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 899.187979] env[61806]: DEBUG oslo_concurrency.lockutils [None req-6d4e5237-b28c-45e9-8ae0-8ed79347c566 tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Acquiring lock "refresh_cache-75f1aded-d1da-4ddc-be29-f1fb93799364" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 899.257671] env[61806]: DEBUG oslo_vmware.api [None req-80f76c83-748f-43fe-b146-5517b19ec0ae tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] Task: {'id': session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]52d21fed-c935-8921-fa43-262fa63cb1a8, 'name': SearchDatastore_Task, 'duration_secs': 0.044678} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 899.258430] env[61806]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-4c7466bb-7fdb-4584-a436-0593d096661b {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 899.264060] env[61806]: DEBUG oslo_vmware.api [None req-80f76c83-748f-43fe-b146-5517b19ec0ae tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] Waiting for the task: (returnval){ [ 899.264060] env[61806]: value = "session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]5252c2d8-4dfc-093b-6730-06146d57300b" [ 899.264060] env[61806]: _type = "Task" [ 899.264060] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 899.272909] env[61806]: DEBUG nova.scheduler.client.report [None req-18f7d4e8-522e-4e97-b6f4-8662acf8d0f8 tempest-ServerTagsTestJSON-1292926242 tempest-ServerTagsTestJSON-1292926242-project-member] Inventory has not changed for provider a2858be1-fd22-4e08-979e-87ad25293407 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 138, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61806) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 899.276073] env[61806]: DEBUG oslo_vmware.api [None req-80f76c83-748f-43fe-b146-5517b19ec0ae tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] Task: {'id': session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]5252c2d8-4dfc-093b-6730-06146d57300b, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 899.387869] env[61806]: DEBUG oslo_vmware.api [None req-464e7d68-de9c-4f8c-9477-932cd9673dfd tempest-ImagesTestJSON-618852556 tempest-ImagesTestJSON-618852556-project-member] Task: {'id': task-1294843, 'name': CreateSnapshot_Task, 'duration_secs': 1.282284} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 899.388163] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-464e7d68-de9c-4f8c-9477-932cd9673dfd tempest-ImagesTestJSON-618852556 tempest-ImagesTestJSON-618852556-project-member] [instance: 1a73d994-5f21-4f80-8f33-5830d860a7ee] Created Snapshot of the VM instance {{(pid=61806) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 899.388923] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5704f6b6-caad-4bcb-aded-488ce72c7b09 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 899.427140] env[61806]: INFO nova.compute.manager [None req-0bb2f11a-fff8-4c85-aa76-4f260c30bde9 tempest-AttachVolumeNegativeTest-1683033704 tempest-AttachVolumeNegativeTest-1683033704-project-member] [instance: daff849e-eb4f-411e-92f8-9362e3bd91e3] Took 22.25 seconds to build instance. [ 899.476741] env[61806]: DEBUG oslo_vmware.api [None req-f9554f19-c70d-44f1-bec2-80b2a52a5572 tempest-ServerRescueTestJSONUnderV235-749414849 tempest-ServerRescueTestJSONUnderV235-749414849-project-member] Task: {'id': task-1294847, 'name': ReconfigVM_Task, 'duration_secs': 0.610547} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 899.477018] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [None req-f9554f19-c70d-44f1-bec2-80b2a52a5572 tempest-ServerRescueTestJSONUnderV235-749414849 tempest-ServerRescueTestJSONUnderV235-749414849-project-member] [instance: 5a526856-0dcd-4b3f-8359-13a5c3b9bfc4] Powering on the VM {{(pid=61806) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 899.477286] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-c1e74d7a-4e9e-4f0c-bfae-c09cccb54cc6 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 899.485079] env[61806]: DEBUG oslo_vmware.api [None req-f9554f19-c70d-44f1-bec2-80b2a52a5572 tempest-ServerRescueTestJSONUnderV235-749414849 tempest-ServerRescueTestJSONUnderV235-749414849-project-member] Waiting for the task: (returnval){ [ 899.485079] env[61806]: value = "task-1294850" [ 899.485079] env[61806]: _type = "Task" [ 899.485079] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 899.493084] env[61806]: DEBUG oslo_vmware.api [None req-f9554f19-c70d-44f1-bec2-80b2a52a5572 tempest-ServerRescueTestJSONUnderV235-749414849 tempest-ServerRescueTestJSONUnderV235-749414849-project-member] Task: {'id': task-1294850, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 899.538488] env[61806]: DEBUG oslo_vmware.api [None req-095bcbe6-2af7-4761-a8ad-c73adfe5ab9d tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] Task: {'id': task-1294849, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 899.664578] env[61806]: DEBUG oslo_concurrency.lockutils [req-5cb4bad8-c976-40ab-b3de-65a7e4c55e69 req-9797cb2e-4b29-44c7-91e4-c69b781ca16b service nova] Releasing lock "refresh_cache-75f1aded-d1da-4ddc-be29-f1fb93799364" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 899.665029] env[61806]: DEBUG oslo_concurrency.lockutils [None req-6d4e5237-b28c-45e9-8ae0-8ed79347c566 tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Acquired lock "refresh_cache-75f1aded-d1da-4ddc-be29-f1fb93799364" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 899.665176] env[61806]: DEBUG nova.network.neutron [None req-6d4e5237-b28c-45e9-8ae0-8ed79347c566 tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] [instance: 75f1aded-d1da-4ddc-be29-f1fb93799364] Building network info cache for instance {{(pid=61806) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 899.775059] env[61806]: DEBUG oslo_vmware.api [None req-80f76c83-748f-43fe-b146-5517b19ec0ae tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] Task: {'id': session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]5252c2d8-4dfc-093b-6730-06146d57300b, 'name': SearchDatastore_Task, 'duration_secs': 0.012713} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 899.775345] env[61806]: DEBUG oslo_concurrency.lockutils [None req-80f76c83-748f-43fe-b146-5517b19ec0ae tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] Releasing lock "[datastore2] devstack-image-cache_base/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa.vmdk" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 899.775605] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [None req-80f76c83-748f-43fe-b146-5517b19ec0ae tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa.vmdk to [datastore2] ee002e7c-8965-494f-a1af-0b6f3b74eceb/ee002e7c-8965-494f-a1af-0b6f3b74eceb.vmdk {{(pid=61806) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 899.776085] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-7a8aa5bf-1578-4ce8-aff5-f3296f7bf1c2 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 899.778549] env[61806]: DEBUG oslo_concurrency.lockutils [None req-18f7d4e8-522e-4e97-b6f4-8662acf8d0f8 tempest-ServerTagsTestJSON-1292926242 tempest-ServerTagsTestJSON-1292926242-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.814s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 899.786290] env[61806]: DEBUG oslo_vmware.api [None req-80f76c83-748f-43fe-b146-5517b19ec0ae tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] Waiting for the task: (returnval){ [ 899.786290] env[61806]: value = "task-1294851" [ 899.786290] env[61806]: _type = "Task" [ 899.786290] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 899.796830] env[61806]: DEBUG oslo_vmware.api [None req-80f76c83-748f-43fe-b146-5517b19ec0ae tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] Task: {'id': task-1294851, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 899.801981] env[61806]: INFO nova.scheduler.client.report [None req-18f7d4e8-522e-4e97-b6f4-8662acf8d0f8 tempest-ServerTagsTestJSON-1292926242 tempest-ServerTagsTestJSON-1292926242-project-member] Deleted allocations for instance 9c20d337-b8f5-4965-b707-0f87c4ab6fc4 [ 899.907401] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-464e7d68-de9c-4f8c-9477-932cd9673dfd tempest-ImagesTestJSON-618852556 tempest-ImagesTestJSON-618852556-project-member] [instance: 1a73d994-5f21-4f80-8f33-5830d860a7ee] Creating linked-clone VM from snapshot {{(pid=61806) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 899.907754] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-219239be-a5b8-44d8-89fa-8e3d6599b761 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 899.918584] env[61806]: DEBUG oslo_vmware.api [None req-464e7d68-de9c-4f8c-9477-932cd9673dfd tempest-ImagesTestJSON-618852556 tempest-ImagesTestJSON-618852556-project-member] Waiting for the task: (returnval){ [ 899.918584] env[61806]: value = "task-1294852" [ 899.918584] env[61806]: _type = "Task" [ 899.918584] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 899.927666] env[61806]: DEBUG oslo_vmware.api [None req-464e7d68-de9c-4f8c-9477-932cd9673dfd tempest-ImagesTestJSON-618852556 tempest-ImagesTestJSON-618852556-project-member] Task: {'id': task-1294852, 'name': CloneVM_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 899.929663] env[61806]: DEBUG oslo_concurrency.lockutils [None req-0bb2f11a-fff8-4c85-aa76-4f260c30bde9 tempest-AttachVolumeNegativeTest-1683033704 tempest-AttachVolumeNegativeTest-1683033704-project-member] Lock "daff849e-eb4f-411e-92f8-9362e3bd91e3" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 23.761s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 899.994917] env[61806]: DEBUG oslo_vmware.api [None req-f9554f19-c70d-44f1-bec2-80b2a52a5572 tempest-ServerRescueTestJSONUnderV235-749414849 tempest-ServerRescueTestJSONUnderV235-749414849-project-member] Task: {'id': task-1294850, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 900.040637] env[61806]: DEBUG oslo_vmware.api [None req-095bcbe6-2af7-4761-a8ad-c73adfe5ab9d tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] Task: {'id': task-1294849, 'name': ReconfigVM_Task, 'duration_secs': 0.788979} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 900.041020] env[61806]: DEBUG nova.virt.vmwareapi.volumeops [None req-095bcbe6-2af7-4761-a8ad-c73adfe5ab9d tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] [instance: 6f6ba57a-e2d6-4749-a53a-e263861cb1c0] Reconfigured VM instance instance-00000050 to attach disk [datastore2] 6f6ba57a-e2d6-4749-a53a-e263861cb1c0/6f6ba57a-e2d6-4749-a53a-e263861cb1c0.vmdk or device None with type sparse {{(pid=61806) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 900.041857] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-64448127-dff1-4da7-bf50-be954c033ddc {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 900.052081] env[61806]: DEBUG oslo_vmware.api [None req-095bcbe6-2af7-4761-a8ad-c73adfe5ab9d tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] Waiting for the task: (returnval){ [ 900.052081] env[61806]: value = "task-1294853" [ 900.052081] env[61806]: _type = "Task" [ 900.052081] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 900.062028] env[61806]: DEBUG oslo_vmware.api [None req-095bcbe6-2af7-4761-a8ad-c73adfe5ab9d tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] Task: {'id': task-1294853, 'name': Rename_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 900.224375] env[61806]: DEBUG nova.objects.instance [None req-76aa021b-a2c9-4c97-bbb8-a4fc0140ddae tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Lazy-loading 'flavor' on Instance uuid 59c22c16-3d1c-4e64-bfac-7f8886a1927e {{(pid=61806) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 900.245797] env[61806]: DEBUG nova.network.neutron [None req-6d4e5237-b28c-45e9-8ae0-8ed79347c566 tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] [instance: 75f1aded-d1da-4ddc-be29-f1fb93799364] Instance cache missing network info. {{(pid=61806) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 900.300726] env[61806]: DEBUG oslo_vmware.api [None req-80f76c83-748f-43fe-b146-5517b19ec0ae tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] Task: {'id': task-1294851, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 900.311110] env[61806]: DEBUG oslo_concurrency.lockutils [None req-18f7d4e8-522e-4e97-b6f4-8662acf8d0f8 tempest-ServerTagsTestJSON-1292926242 tempest-ServerTagsTestJSON-1292926242-project-member] Lock "9c20d337-b8f5-4965-b707-0f87c4ab6fc4" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 10.062s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 900.430597] env[61806]: DEBUG oslo_vmware.api [None req-464e7d68-de9c-4f8c-9477-932cd9673dfd tempest-ImagesTestJSON-618852556 tempest-ImagesTestJSON-618852556-project-member] Task: {'id': task-1294852, 'name': CloneVM_Task} progress is 94%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 900.503366] env[61806]: DEBUG oslo_vmware.api [None req-f9554f19-c70d-44f1-bec2-80b2a52a5572 tempest-ServerRescueTestJSONUnderV235-749414849 tempest-ServerRescueTestJSONUnderV235-749414849-project-member] Task: {'id': task-1294850, 'name': PowerOnVM_Task, 'duration_secs': 0.878827} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 900.505750] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [None req-f9554f19-c70d-44f1-bec2-80b2a52a5572 tempest-ServerRescueTestJSONUnderV235-749414849 tempest-ServerRescueTestJSONUnderV235-749414849-project-member] [instance: 5a526856-0dcd-4b3f-8359-13a5c3b9bfc4] Powered on the VM {{(pid=61806) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 900.508707] env[61806]: DEBUG nova.compute.manager [None req-f9554f19-c70d-44f1-bec2-80b2a52a5572 tempest-ServerRescueTestJSONUnderV235-749414849 tempest-ServerRescueTestJSONUnderV235-749414849-project-member] [instance: 5a526856-0dcd-4b3f-8359-13a5c3b9bfc4] Checking state {{(pid=61806) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 900.509558] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b3b437fc-3d1c-405c-9427-e290f726531c {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 900.563099] env[61806]: DEBUG oslo_vmware.api [None req-095bcbe6-2af7-4761-a8ad-c73adfe5ab9d tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] Task: {'id': task-1294853, 'name': Rename_Task, 'duration_secs': 0.351337} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 900.564123] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [None req-095bcbe6-2af7-4761-a8ad-c73adfe5ab9d tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] [instance: 6f6ba57a-e2d6-4749-a53a-e263861cb1c0] Powering on the VM {{(pid=61806) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 900.564409] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-2819dca7-8492-4d99-a4f7-0148be3961ca {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 900.571927] env[61806]: DEBUG oslo_vmware.api [None req-095bcbe6-2af7-4761-a8ad-c73adfe5ab9d tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] Waiting for the task: (returnval){ [ 900.571927] env[61806]: value = "task-1294854" [ 900.571927] env[61806]: _type = "Task" [ 900.571927] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 900.580573] env[61806]: DEBUG oslo_vmware.api [None req-095bcbe6-2af7-4761-a8ad-c73adfe5ab9d tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] Task: {'id': task-1294854, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 900.582771] env[61806]: DEBUG nova.network.neutron [None req-6d4e5237-b28c-45e9-8ae0-8ed79347c566 tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] [instance: 75f1aded-d1da-4ddc-be29-f1fb93799364] Updating instance_info_cache with network_info: [{"id": "a06a82f0-c5d3-48b7-bd4b-d2725bd7eaec", "address": "fa:16:3e:8f:1a:72", "network": {"id": "46515a4e-610f-42fb-a424-205a03df0640", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-1133623547-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4b9d94c3d684474b99e305fb6cd9b95f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "82ca17df-257e-40e6-9ec9-310ed6f05ccb", "external-id": "nsx-vlan-transportzone-778", "segmentation_id": 778, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa06a82f0-c5", "ovs_interfaceid": "a06a82f0-c5d3-48b7-bd4b-d2725bd7eaec", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61806) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 900.730381] env[61806]: DEBUG oslo_concurrency.lockutils [None req-76aa021b-a2c9-4c97-bbb8-a4fc0140ddae tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Lock "59c22c16-3d1c-4e64-bfac-7f8886a1927e" "released" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: held 7.843s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 900.799139] env[61806]: DEBUG oslo_vmware.api [None req-80f76c83-748f-43fe-b146-5517b19ec0ae tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] Task: {'id': task-1294851, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.595962} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 900.799433] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [None req-80f76c83-748f-43fe-b146-5517b19ec0ae tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa.vmdk to [datastore2] ee002e7c-8965-494f-a1af-0b6f3b74eceb/ee002e7c-8965-494f-a1af-0b6f3b74eceb.vmdk {{(pid=61806) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 900.799656] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-80f76c83-748f-43fe-b146-5517b19ec0ae tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] [instance: ee002e7c-8965-494f-a1af-0b6f3b74eceb] Extending root virtual disk to 1048576 {{(pid=61806) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 900.799915] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-62c9e9dc-2153-4383-8f40-b4e748a2f48b {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 900.807689] env[61806]: DEBUG oslo_vmware.api [None req-80f76c83-748f-43fe-b146-5517b19ec0ae tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] Waiting for the task: (returnval){ [ 900.807689] env[61806]: value = "task-1294855" [ 900.807689] env[61806]: _type = "Task" [ 900.807689] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 900.818181] env[61806]: DEBUG oslo_vmware.api [None req-80f76c83-748f-43fe-b146-5517b19ec0ae tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] Task: {'id': task-1294855, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 900.934519] env[61806]: DEBUG oslo_vmware.api [None req-464e7d68-de9c-4f8c-9477-932cd9673dfd tempest-ImagesTestJSON-618852556 tempest-ImagesTestJSON-618852556-project-member] Task: {'id': task-1294852, 'name': CloneVM_Task} progress is 94%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 901.075306] env[61806]: DEBUG nova.compute.manager [req-f33da80d-14dd-4ad4-831a-6a1423a44a94 req-0a612450-9cba-4130-9ec0-73a472c69673 service nova] [instance: daff849e-eb4f-411e-92f8-9362e3bd91e3] Received event network-changed-9fdf6139-082b-49e0-8ce8-7870749b2c7a {{(pid=61806) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 901.075306] env[61806]: DEBUG nova.compute.manager [req-f33da80d-14dd-4ad4-831a-6a1423a44a94 req-0a612450-9cba-4130-9ec0-73a472c69673 service nova] [instance: daff849e-eb4f-411e-92f8-9362e3bd91e3] Refreshing instance network info cache due to event network-changed-9fdf6139-082b-49e0-8ce8-7870749b2c7a. {{(pid=61806) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 901.075306] env[61806]: DEBUG oslo_concurrency.lockutils [req-f33da80d-14dd-4ad4-831a-6a1423a44a94 req-0a612450-9cba-4130-9ec0-73a472c69673 service nova] Acquiring lock "refresh_cache-daff849e-eb4f-411e-92f8-9362e3bd91e3" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 901.075306] env[61806]: DEBUG oslo_concurrency.lockutils [req-f33da80d-14dd-4ad4-831a-6a1423a44a94 req-0a612450-9cba-4130-9ec0-73a472c69673 service nova] Acquired lock "refresh_cache-daff849e-eb4f-411e-92f8-9362e3bd91e3" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 901.075306] env[61806]: DEBUG nova.network.neutron [req-f33da80d-14dd-4ad4-831a-6a1423a44a94 req-0a612450-9cba-4130-9ec0-73a472c69673 service nova] [instance: daff849e-eb4f-411e-92f8-9362e3bd91e3] Refreshing network info cache for port 9fdf6139-082b-49e0-8ce8-7870749b2c7a {{(pid=61806) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 901.086200] env[61806]: DEBUG oslo_concurrency.lockutils [None req-6d4e5237-b28c-45e9-8ae0-8ed79347c566 tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Releasing lock "refresh_cache-75f1aded-d1da-4ddc-be29-f1fb93799364" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 901.086200] env[61806]: DEBUG nova.compute.manager [None req-6d4e5237-b28c-45e9-8ae0-8ed79347c566 tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] [instance: 75f1aded-d1da-4ddc-be29-f1fb93799364] Instance network_info: |[{"id": "a06a82f0-c5d3-48b7-bd4b-d2725bd7eaec", "address": "fa:16:3e:8f:1a:72", "network": {"id": "46515a4e-610f-42fb-a424-205a03df0640", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-1133623547-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4b9d94c3d684474b99e305fb6cd9b95f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "82ca17df-257e-40e6-9ec9-310ed6f05ccb", "external-id": "nsx-vlan-transportzone-778", "segmentation_id": 778, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa06a82f0-c5", "ovs_interfaceid": "a06a82f0-c5d3-48b7-bd4b-d2725bd7eaec", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61806) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 901.086200] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-6d4e5237-b28c-45e9-8ae0-8ed79347c566 tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] [instance: 75f1aded-d1da-4ddc-be29-f1fb93799364] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:8f:1a:72', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '82ca17df-257e-40e6-9ec9-310ed6f05ccb', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'a06a82f0-c5d3-48b7-bd4b-d2725bd7eaec', 'vif_model': 'vmxnet3'}] {{(pid=61806) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 901.093978] env[61806]: DEBUG oslo.service.loopingcall [None req-6d4e5237-b28c-45e9-8ae0-8ed79347c566 tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61806) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 901.096029] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 75f1aded-d1da-4ddc-be29-f1fb93799364] Creating VM on the ESX host {{(pid=61806) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 901.096029] env[61806]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-14374667-87da-4aa1-9eb2-59f9f7b9307f {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 901.114983] env[61806]: DEBUG oslo_vmware.api [None req-095bcbe6-2af7-4761-a8ad-c73adfe5ab9d tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] Task: {'id': task-1294854, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 901.122348] env[61806]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 901.122348] env[61806]: value = "task-1294856" [ 901.122348] env[61806]: _type = "Task" [ 901.122348] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 901.131509] env[61806]: DEBUG oslo_vmware.api [-] Task: {'id': task-1294856, 'name': CreateVM_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 901.319201] env[61806]: DEBUG oslo_vmware.api [None req-80f76c83-748f-43fe-b146-5517b19ec0ae tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] Task: {'id': task-1294855, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.124876} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 901.319557] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-80f76c83-748f-43fe-b146-5517b19ec0ae tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] [instance: ee002e7c-8965-494f-a1af-0b6f3b74eceb] Extended root virtual disk {{(pid=61806) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 901.320292] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-620b8675-0384-4562-91d6-38e68a5553f7 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 901.344166] env[61806]: DEBUG nova.virt.vmwareapi.volumeops [None req-80f76c83-748f-43fe-b146-5517b19ec0ae tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] [instance: ee002e7c-8965-494f-a1af-0b6f3b74eceb] Reconfiguring VM instance instance-00000051 to attach disk [datastore2] ee002e7c-8965-494f-a1af-0b6f3b74eceb/ee002e7c-8965-494f-a1af-0b6f3b74eceb.vmdk or device None with type sparse {{(pid=61806) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 901.344695] env[61806]: DEBUG oslo_concurrency.lockutils [None req-1d9f7d6e-8006-4355-8395-1f9c0f916c73 tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Acquiring lock "59c22c16-3d1c-4e64-bfac-7f8886a1927e" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 901.344931] env[61806]: DEBUG oslo_concurrency.lockutils [None req-1d9f7d6e-8006-4355-8395-1f9c0f916c73 tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Lock "59c22c16-3d1c-4e64-bfac-7f8886a1927e" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 901.345158] env[61806]: DEBUG oslo_concurrency.lockutils [None req-1d9f7d6e-8006-4355-8395-1f9c0f916c73 tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Acquiring lock "59c22c16-3d1c-4e64-bfac-7f8886a1927e-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 901.345367] env[61806]: DEBUG oslo_concurrency.lockutils [None req-1d9f7d6e-8006-4355-8395-1f9c0f916c73 tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Lock "59c22c16-3d1c-4e64-bfac-7f8886a1927e-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 901.345545] env[61806]: DEBUG oslo_concurrency.lockutils [None req-1d9f7d6e-8006-4355-8395-1f9c0f916c73 tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Lock "59c22c16-3d1c-4e64-bfac-7f8886a1927e-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 901.347236] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-b5d43513-f7d9-4d78-bce2-ece9581b9ec7 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 901.362140] env[61806]: INFO nova.compute.manager [None req-1d9f7d6e-8006-4355-8395-1f9c0f916c73 tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] [instance: 59c22c16-3d1c-4e64-bfac-7f8886a1927e] Terminating instance [ 901.365079] env[61806]: DEBUG nova.compute.manager [None req-1d9f7d6e-8006-4355-8395-1f9c0f916c73 tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] [instance: 59c22c16-3d1c-4e64-bfac-7f8886a1927e] Start destroying the instance on the hypervisor. {{(pid=61806) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 901.365079] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [None req-1d9f7d6e-8006-4355-8395-1f9c0f916c73 tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] [instance: 59c22c16-3d1c-4e64-bfac-7f8886a1927e] Powering off the VM {{(pid=61806) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 901.365247] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-6940545b-578b-42dc-a656-6ce884172193 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 901.370018] env[61806]: DEBUG oslo_vmware.api [None req-80f76c83-748f-43fe-b146-5517b19ec0ae tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] Waiting for the task: (returnval){ [ 901.370018] env[61806]: value = "task-1294857" [ 901.370018] env[61806]: _type = "Task" [ 901.370018] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 901.375208] env[61806]: DEBUG oslo_vmware.api [None req-1d9f7d6e-8006-4355-8395-1f9c0f916c73 tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Waiting for the task: (returnval){ [ 901.375208] env[61806]: value = "task-1294858" [ 901.375208] env[61806]: _type = "Task" [ 901.375208] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 901.383520] env[61806]: DEBUG oslo_vmware.api [None req-80f76c83-748f-43fe-b146-5517b19ec0ae tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] Task: {'id': task-1294857, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 901.388256] env[61806]: DEBUG oslo_vmware.api [None req-1d9f7d6e-8006-4355-8395-1f9c0f916c73 tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Task: {'id': task-1294858, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 901.432383] env[61806]: DEBUG oslo_vmware.api [None req-464e7d68-de9c-4f8c-9477-932cd9673dfd tempest-ImagesTestJSON-618852556 tempest-ImagesTestJSON-618852556-project-member] Task: {'id': task-1294852, 'name': CloneVM_Task} progress is 94%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 901.583938] env[61806]: DEBUG oslo_vmware.api [None req-095bcbe6-2af7-4761-a8ad-c73adfe5ab9d tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] Task: {'id': task-1294854, 'name': PowerOnVM_Task, 'duration_secs': 0.732673} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 901.588025] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [None req-095bcbe6-2af7-4761-a8ad-c73adfe5ab9d tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] [instance: 6f6ba57a-e2d6-4749-a53a-e263861cb1c0] Powered on the VM {{(pid=61806) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 901.588025] env[61806]: INFO nova.compute.manager [None req-095bcbe6-2af7-4761-a8ad-c73adfe5ab9d tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] [instance: 6f6ba57a-e2d6-4749-a53a-e263861cb1c0] Took 8.96 seconds to spawn the instance on the hypervisor. [ 901.588025] env[61806]: DEBUG nova.compute.manager [None req-095bcbe6-2af7-4761-a8ad-c73adfe5ab9d tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] [instance: 6f6ba57a-e2d6-4749-a53a-e263861cb1c0] Checking state {{(pid=61806) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 901.589227] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0d8d5880-f497-4257-9ba5-6172b0a3a263 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 901.635719] env[61806]: DEBUG oslo_vmware.api [-] Task: {'id': task-1294856, 'name': CreateVM_Task} progress is 99%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 901.884422] env[61806]: DEBUG oslo_vmware.api [None req-80f76c83-748f-43fe-b146-5517b19ec0ae tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] Task: {'id': task-1294857, 'name': ReconfigVM_Task, 'duration_secs': 0.330427} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 901.885149] env[61806]: DEBUG nova.virt.vmwareapi.volumeops [None req-80f76c83-748f-43fe-b146-5517b19ec0ae tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] [instance: ee002e7c-8965-494f-a1af-0b6f3b74eceb] Reconfigured VM instance instance-00000051 to attach disk [datastore2] ee002e7c-8965-494f-a1af-0b6f3b74eceb/ee002e7c-8965-494f-a1af-0b6f3b74eceb.vmdk or device None with type sparse {{(pid=61806) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 901.885827] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-d86264b1-56ef-4a0e-8e24-ff1f8ce6c2ef {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 901.892960] env[61806]: DEBUG oslo_vmware.api [None req-1d9f7d6e-8006-4355-8395-1f9c0f916c73 tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Task: {'id': task-1294858, 'name': PowerOffVM_Task, 'duration_secs': 0.369602} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 901.892960] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [None req-1d9f7d6e-8006-4355-8395-1f9c0f916c73 tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] [instance: 59c22c16-3d1c-4e64-bfac-7f8886a1927e] Powered off the VM {{(pid=61806) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 901.892960] env[61806]: DEBUG nova.virt.vmwareapi.volumeops [None req-1d9f7d6e-8006-4355-8395-1f9c0f916c73 tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] [instance: 59c22c16-3d1c-4e64-bfac-7f8886a1927e] Volume detach. Driver type: vmdk {{(pid=61806) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 901.892960] env[61806]: DEBUG nova.virt.vmwareapi.volumeops [None req-1d9f7d6e-8006-4355-8395-1f9c0f916c73 tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] [instance: 59c22c16-3d1c-4e64-bfac-7f8886a1927e] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-277703', 'volume_id': '564a9f8e-1bec-410b-9d7d-ab9bf7b6d83b', 'name': 'volume-564a9f8e-1bec-410b-9d7d-ab9bf7b6d83b', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '59c22c16-3d1c-4e64-bfac-7f8886a1927e', 'attached_at': '', 'detached_at': '', 'volume_id': '564a9f8e-1bec-410b-9d7d-ab9bf7b6d83b', 'serial': '564a9f8e-1bec-410b-9d7d-ab9bf7b6d83b'} {{(pid=61806) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 901.892960] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d10d18d8-b11e-43ac-9077-23883787125d {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 901.897745] env[61806]: DEBUG oslo_vmware.api [None req-80f76c83-748f-43fe-b146-5517b19ec0ae tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] Waiting for the task: (returnval){ [ 901.897745] env[61806]: value = "task-1294859" [ 901.897745] env[61806]: _type = "Task" [ 901.897745] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 901.922320] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-af79d34d-4a38-4a53-8520-8cc9ce0b24ab {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 901.937019] env[61806]: DEBUG oslo_vmware.api [None req-80f76c83-748f-43fe-b146-5517b19ec0ae tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] Task: {'id': task-1294859, 'name': Rename_Task} progress is 14%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 901.941862] env[61806]: DEBUG oslo_vmware.api [None req-464e7d68-de9c-4f8c-9477-932cd9673dfd tempest-ImagesTestJSON-618852556 tempest-ImagesTestJSON-618852556-project-member] Task: {'id': task-1294852, 'name': CloneVM_Task} progress is 100%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 901.945668] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cea00b29-e1d5-43f4-b007-52022841aee6 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 901.970418] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b0b7b478-f2fa-439d-bfc4-baf6310e41cd {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 901.988648] env[61806]: DEBUG nova.virt.vmwareapi.volumeops [None req-1d9f7d6e-8006-4355-8395-1f9c0f916c73 tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] The volume has not been displaced from its original location: [datastore2] volume-564a9f8e-1bec-410b-9d7d-ab9bf7b6d83b/volume-564a9f8e-1bec-410b-9d7d-ab9bf7b6d83b.vmdk. No consolidation needed. {{(pid=61806) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 901.994112] env[61806]: DEBUG nova.virt.vmwareapi.volumeops [None req-1d9f7d6e-8006-4355-8395-1f9c0f916c73 tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] [instance: 59c22c16-3d1c-4e64-bfac-7f8886a1927e] Reconfiguring VM instance instance-0000004b to detach disk 2001 {{(pid=61806) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 901.994518] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-3effc3b8-656d-4d90-aebc-d058dedd714c {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 902.014935] env[61806]: DEBUG oslo_vmware.api [None req-1d9f7d6e-8006-4355-8395-1f9c0f916c73 tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Waiting for the task: (returnval){ [ 902.014935] env[61806]: value = "task-1294860" [ 902.014935] env[61806]: _type = "Task" [ 902.014935] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 902.023915] env[61806]: DEBUG oslo_vmware.api [None req-1d9f7d6e-8006-4355-8395-1f9c0f916c73 tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Task: {'id': task-1294860, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 902.114081] env[61806]: INFO nova.compute.manager [None req-095bcbe6-2af7-4761-a8ad-c73adfe5ab9d tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] [instance: 6f6ba57a-e2d6-4749-a53a-e263861cb1c0] Took 16.75 seconds to build instance. [ 902.134371] env[61806]: DEBUG oslo_vmware.api [-] Task: {'id': task-1294856, 'name': CreateVM_Task, 'duration_secs': 0.610661} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 902.135189] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 75f1aded-d1da-4ddc-be29-f1fb93799364] Created VM on the ESX host {{(pid=61806) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 902.135992] env[61806]: DEBUG oslo_concurrency.lockutils [None req-6d4e5237-b28c-45e9-8ae0-8ed79347c566 tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 902.136192] env[61806]: DEBUG oslo_concurrency.lockutils [None req-6d4e5237-b28c-45e9-8ae0-8ed79347c566 tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Acquired lock "[datastore1] devstack-image-cache_base/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 902.136526] env[61806]: DEBUG oslo_concurrency.lockutils [None req-6d4e5237-b28c-45e9-8ae0-8ed79347c566 tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 902.137058] env[61806]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-bfcf065f-89b2-4bde-97fa-85b308a48843 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 902.142538] env[61806]: DEBUG oslo_vmware.api [None req-6d4e5237-b28c-45e9-8ae0-8ed79347c566 tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Waiting for the task: (returnval){ [ 902.142538] env[61806]: value = "session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]5249b495-e332-2e5a-e8f4-43960cbca0ff" [ 902.142538] env[61806]: _type = "Task" [ 902.142538] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 902.151696] env[61806]: DEBUG oslo_vmware.api [None req-6d4e5237-b28c-45e9-8ae0-8ed79347c566 tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Task: {'id': session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]5249b495-e332-2e5a-e8f4-43960cbca0ff, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 902.161257] env[61806]: DEBUG nova.network.neutron [req-f33da80d-14dd-4ad4-831a-6a1423a44a94 req-0a612450-9cba-4130-9ec0-73a472c69673 service nova] [instance: daff849e-eb4f-411e-92f8-9362e3bd91e3] Updated VIF entry in instance network info cache for port 9fdf6139-082b-49e0-8ce8-7870749b2c7a. {{(pid=61806) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 902.161786] env[61806]: DEBUG nova.network.neutron [req-f33da80d-14dd-4ad4-831a-6a1423a44a94 req-0a612450-9cba-4130-9ec0-73a472c69673 service nova] [instance: daff849e-eb4f-411e-92f8-9362e3bd91e3] Updating instance_info_cache with network_info: [{"id": "9fdf6139-082b-49e0-8ce8-7870749b2c7a", "address": "fa:16:3e:a5:41:88", "network": {"id": "d636ca55-0737-48e4-8d9d-7fe0f797f49e", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-2090334349-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.140", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "22e47882c2c44ce492dff6b1c5d782e2", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6f493cd8-1cb4-42a1-8d56-bfa6ac7cf563", "external-id": "nsx-vlan-transportzone-931", "segmentation_id": 931, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9fdf6139-08", "ovs_interfaceid": "9fdf6139-082b-49e0-8ce8-7870749b2c7a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61806) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 902.413614] env[61806]: DEBUG oslo_vmware.api [None req-80f76c83-748f-43fe-b146-5517b19ec0ae tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] Task: {'id': task-1294859, 'name': Rename_Task, 'duration_secs': 0.164033} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 902.416147] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [None req-80f76c83-748f-43fe-b146-5517b19ec0ae tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] [instance: ee002e7c-8965-494f-a1af-0b6f3b74eceb] Powering on the VM {{(pid=61806) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 902.416147] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-1a558002-012e-489b-a24d-bf3082036bab {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 902.424028] env[61806]: DEBUG oslo_vmware.api [None req-80f76c83-748f-43fe-b146-5517b19ec0ae tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] Waiting for the task: (returnval){ [ 902.424028] env[61806]: value = "task-1294861" [ 902.424028] env[61806]: _type = "Task" [ 902.424028] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 902.437392] env[61806]: DEBUG oslo_vmware.api [None req-80f76c83-748f-43fe-b146-5517b19ec0ae tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] Task: {'id': task-1294861, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 902.441575] env[61806]: DEBUG oslo_vmware.api [None req-464e7d68-de9c-4f8c-9477-932cd9673dfd tempest-ImagesTestJSON-618852556 tempest-ImagesTestJSON-618852556-project-member] Task: {'id': task-1294852, 'name': CloneVM_Task, 'duration_secs': 2.030642} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 902.441975] env[61806]: INFO nova.virt.vmwareapi.vmops [None req-464e7d68-de9c-4f8c-9477-932cd9673dfd tempest-ImagesTestJSON-618852556 tempest-ImagesTestJSON-618852556-project-member] [instance: 1a73d994-5f21-4f80-8f33-5830d860a7ee] Created linked-clone VM from snapshot [ 902.442824] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0f4b5287-a544-4fd2-9cc5-7dafb59b846d {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 902.455447] env[61806]: DEBUG nova.virt.vmwareapi.images [None req-464e7d68-de9c-4f8c-9477-932cd9673dfd tempest-ImagesTestJSON-618852556 tempest-ImagesTestJSON-618852556-project-member] [instance: 1a73d994-5f21-4f80-8f33-5830d860a7ee] Uploading image d57e5203-fede-48a8-b0c9-622639a68303 {{(pid=61806) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:442}} [ 902.472730] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [None req-464e7d68-de9c-4f8c-9477-932cd9673dfd tempest-ImagesTestJSON-618852556 tempest-ImagesTestJSON-618852556-project-member] [instance: 1a73d994-5f21-4f80-8f33-5830d860a7ee] Destroying the VM {{(pid=61806) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1358}} [ 902.473062] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-acb45a88-e8b8-41f8-a407-c9bc935f50d6 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 902.483087] env[61806]: DEBUG oslo_vmware.api [None req-464e7d68-de9c-4f8c-9477-932cd9673dfd tempest-ImagesTestJSON-618852556 tempest-ImagesTestJSON-618852556-project-member] Waiting for the task: (returnval){ [ 902.483087] env[61806]: value = "task-1294862" [ 902.483087] env[61806]: _type = "Task" [ 902.483087] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 902.492378] env[61806]: DEBUG oslo_vmware.api [None req-464e7d68-de9c-4f8c-9477-932cd9673dfd tempest-ImagesTestJSON-618852556 tempest-ImagesTestJSON-618852556-project-member] Task: {'id': task-1294862, 'name': Destroy_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 902.525874] env[61806]: DEBUG oslo_vmware.api [None req-1d9f7d6e-8006-4355-8395-1f9c0f916c73 tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Task: {'id': task-1294860, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 902.616899] env[61806]: DEBUG oslo_concurrency.lockutils [None req-095bcbe6-2af7-4761-a8ad-c73adfe5ab9d tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] Lock "6f6ba57a-e2d6-4749-a53a-e263861cb1c0" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 18.260s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 902.653026] env[61806]: DEBUG oslo_vmware.api [None req-6d4e5237-b28c-45e9-8ae0-8ed79347c566 tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Task: {'id': session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]5249b495-e332-2e5a-e8f4-43960cbca0ff, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 902.663975] env[61806]: DEBUG oslo_concurrency.lockutils [req-f33da80d-14dd-4ad4-831a-6a1423a44a94 req-0a612450-9cba-4130-9ec0-73a472c69673 service nova] Releasing lock "refresh_cache-daff849e-eb4f-411e-92f8-9362e3bd91e3" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 902.941734] env[61806]: DEBUG oslo_vmware.api [None req-80f76c83-748f-43fe-b146-5517b19ec0ae tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] Task: {'id': task-1294861, 'name': PowerOnVM_Task} progress is 100%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 902.959667] env[61806]: DEBUG nova.compute.manager [req-9845851d-0bec-4c0d-bad9-d5d7862a80b1 req-26f8dde5-2169-4449-99b8-be8a7879aea1 service nova] [instance: 5a526856-0dcd-4b3f-8359-13a5c3b9bfc4] Received event network-changed-58782499-bc1a-4be1-9d48-655c6b2f11d7 {{(pid=61806) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 902.959927] env[61806]: DEBUG nova.compute.manager [req-9845851d-0bec-4c0d-bad9-d5d7862a80b1 req-26f8dde5-2169-4449-99b8-be8a7879aea1 service nova] [instance: 5a526856-0dcd-4b3f-8359-13a5c3b9bfc4] Refreshing instance network info cache due to event network-changed-58782499-bc1a-4be1-9d48-655c6b2f11d7. {{(pid=61806) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 902.961411] env[61806]: DEBUG oslo_concurrency.lockutils [req-9845851d-0bec-4c0d-bad9-d5d7862a80b1 req-26f8dde5-2169-4449-99b8-be8a7879aea1 service nova] Acquiring lock "refresh_cache-5a526856-0dcd-4b3f-8359-13a5c3b9bfc4" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 902.961411] env[61806]: DEBUG oslo_concurrency.lockutils [req-9845851d-0bec-4c0d-bad9-d5d7862a80b1 req-26f8dde5-2169-4449-99b8-be8a7879aea1 service nova] Acquired lock "refresh_cache-5a526856-0dcd-4b3f-8359-13a5c3b9bfc4" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 902.961411] env[61806]: DEBUG nova.network.neutron [req-9845851d-0bec-4c0d-bad9-d5d7862a80b1 req-26f8dde5-2169-4449-99b8-be8a7879aea1 service nova] [instance: 5a526856-0dcd-4b3f-8359-13a5c3b9bfc4] Refreshing network info cache for port 58782499-bc1a-4be1-9d48-655c6b2f11d7 {{(pid=61806) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 902.993913] env[61806]: DEBUG oslo_vmware.api [None req-464e7d68-de9c-4f8c-9477-932cd9673dfd tempest-ImagesTestJSON-618852556 tempest-ImagesTestJSON-618852556-project-member] Task: {'id': task-1294862, 'name': Destroy_Task} progress is 33%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 903.026982] env[61806]: DEBUG oslo_vmware.api [None req-1d9f7d6e-8006-4355-8395-1f9c0f916c73 tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Task: {'id': task-1294860, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 903.107391] env[61806]: DEBUG nova.compute.manager [req-16b0031f-bbd9-4768-b2ff-0881ac4502b0 req-c8677704-476b-4a0c-bfa6-be77d6aaf3f7 service nova] [instance: 5a526856-0dcd-4b3f-8359-13a5c3b9bfc4] Received event network-changed-58782499-bc1a-4be1-9d48-655c6b2f11d7 {{(pid=61806) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 903.107602] env[61806]: DEBUG nova.compute.manager [req-16b0031f-bbd9-4768-b2ff-0881ac4502b0 req-c8677704-476b-4a0c-bfa6-be77d6aaf3f7 service nova] [instance: 5a526856-0dcd-4b3f-8359-13a5c3b9bfc4] Refreshing instance network info cache due to event network-changed-58782499-bc1a-4be1-9d48-655c6b2f11d7. {{(pid=61806) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 903.107794] env[61806]: DEBUG oslo_concurrency.lockutils [req-16b0031f-bbd9-4768-b2ff-0881ac4502b0 req-c8677704-476b-4a0c-bfa6-be77d6aaf3f7 service nova] Acquiring lock "refresh_cache-5a526856-0dcd-4b3f-8359-13a5c3b9bfc4" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 903.153554] env[61806]: DEBUG oslo_vmware.api [None req-6d4e5237-b28c-45e9-8ae0-8ed79347c566 tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Task: {'id': session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]5249b495-e332-2e5a-e8f4-43960cbca0ff, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 903.444190] env[61806]: DEBUG oslo_vmware.api [None req-80f76c83-748f-43fe-b146-5517b19ec0ae tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] Task: {'id': task-1294861, 'name': PowerOnVM_Task, 'duration_secs': 0.522796} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 903.444190] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [None req-80f76c83-748f-43fe-b146-5517b19ec0ae tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] [instance: ee002e7c-8965-494f-a1af-0b6f3b74eceb] Powered on the VM {{(pid=61806) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 903.444190] env[61806]: INFO nova.compute.manager [None req-80f76c83-748f-43fe-b146-5517b19ec0ae tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] [instance: ee002e7c-8965-494f-a1af-0b6f3b74eceb] Took 7.53 seconds to spawn the instance on the hypervisor. [ 903.444190] env[61806]: DEBUG nova.compute.manager [None req-80f76c83-748f-43fe-b146-5517b19ec0ae tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] [instance: ee002e7c-8965-494f-a1af-0b6f3b74eceb] Checking state {{(pid=61806) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 903.444190] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-81c553de-52d4-4583-b56c-dd72ab823ee2 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 903.495807] env[61806]: DEBUG oslo_vmware.api [None req-464e7d68-de9c-4f8c-9477-932cd9673dfd tempest-ImagesTestJSON-618852556 tempest-ImagesTestJSON-618852556-project-member] Task: {'id': task-1294862, 'name': Destroy_Task, 'duration_secs': 0.599105} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 903.496188] env[61806]: INFO nova.virt.vmwareapi.vm_util [None req-464e7d68-de9c-4f8c-9477-932cd9673dfd tempest-ImagesTestJSON-618852556 tempest-ImagesTestJSON-618852556-project-member] [instance: 1a73d994-5f21-4f80-8f33-5830d860a7ee] Destroyed the VM [ 903.496355] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-464e7d68-de9c-4f8c-9477-932cd9673dfd tempest-ImagesTestJSON-618852556 tempest-ImagesTestJSON-618852556-project-member] [instance: 1a73d994-5f21-4f80-8f33-5830d860a7ee] Deleting Snapshot of the VM instance {{(pid=61806) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 903.496617] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-e5ab2de3-a545-43a3-a270-5bfc3a2ac841 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 903.504345] env[61806]: DEBUG oslo_vmware.api [None req-464e7d68-de9c-4f8c-9477-932cd9673dfd tempest-ImagesTestJSON-618852556 tempest-ImagesTestJSON-618852556-project-member] Waiting for the task: (returnval){ [ 903.504345] env[61806]: value = "task-1294863" [ 903.504345] env[61806]: _type = "Task" [ 903.504345] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 903.516219] env[61806]: DEBUG oslo_vmware.api [None req-464e7d68-de9c-4f8c-9477-932cd9673dfd tempest-ImagesTestJSON-618852556 tempest-ImagesTestJSON-618852556-project-member] Task: {'id': task-1294863, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 903.528469] env[61806]: DEBUG oslo_vmware.api [None req-1d9f7d6e-8006-4355-8395-1f9c0f916c73 tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Task: {'id': task-1294860, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 903.662649] env[61806]: DEBUG oslo_vmware.api [None req-6d4e5237-b28c-45e9-8ae0-8ed79347c566 tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Task: {'id': session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]5249b495-e332-2e5a-e8f4-43960cbca0ff, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 903.883823] env[61806]: DEBUG nova.network.neutron [req-9845851d-0bec-4c0d-bad9-d5d7862a80b1 req-26f8dde5-2169-4449-99b8-be8a7879aea1 service nova] [instance: 5a526856-0dcd-4b3f-8359-13a5c3b9bfc4] Updated VIF entry in instance network info cache for port 58782499-bc1a-4be1-9d48-655c6b2f11d7. {{(pid=61806) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 903.884245] env[61806]: DEBUG nova.network.neutron [req-9845851d-0bec-4c0d-bad9-d5d7862a80b1 req-26f8dde5-2169-4449-99b8-be8a7879aea1 service nova] [instance: 5a526856-0dcd-4b3f-8359-13a5c3b9bfc4] Updating instance_info_cache with network_info: [{"id": "58782499-bc1a-4be1-9d48-655c6b2f11d7", "address": "fa:16:3e:52:7a:2b", "network": {"id": "b4fa0830-265e-47bf-8029-5944534312e7", "bridge": "br-int", "label": "tempest-ServerRescueTestJSONUnderV235-1460638664-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "2ba59612579b47a4bde7f82d951ac3ff", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8a34aa30-95be-4b18-98ca-1f2d81f7e9e6", "external-id": "nsx-vlan-transportzone-234", "segmentation_id": 234, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap58782499-bc", "ovs_interfaceid": "58782499-bc1a-4be1-9d48-655c6b2f11d7", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61806) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 903.972705] env[61806]: INFO nova.compute.manager [None req-80f76c83-748f-43fe-b146-5517b19ec0ae tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] [instance: ee002e7c-8965-494f-a1af-0b6f3b74eceb] Took 16.10 seconds to build instance. [ 904.014594] env[61806]: DEBUG oslo_vmware.api [None req-464e7d68-de9c-4f8c-9477-932cd9673dfd tempest-ImagesTestJSON-618852556 tempest-ImagesTestJSON-618852556-project-member] Task: {'id': task-1294863, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 904.027737] env[61806]: DEBUG oslo_vmware.api [None req-1d9f7d6e-8006-4355-8395-1f9c0f916c73 tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Task: {'id': task-1294860, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 904.156121] env[61806]: DEBUG oslo_vmware.api [None req-6d4e5237-b28c-45e9-8ae0-8ed79347c566 tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Task: {'id': session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]5249b495-e332-2e5a-e8f4-43960cbca0ff, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 904.399019] env[61806]: DEBUG oslo_concurrency.lockutils [req-9845851d-0bec-4c0d-bad9-d5d7862a80b1 req-26f8dde5-2169-4449-99b8-be8a7879aea1 service nova] Releasing lock "refresh_cache-5a526856-0dcd-4b3f-8359-13a5c3b9bfc4" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 904.399019] env[61806]: DEBUG oslo_concurrency.lockutils [req-16b0031f-bbd9-4768-b2ff-0881ac4502b0 req-c8677704-476b-4a0c-bfa6-be77d6aaf3f7 service nova] Acquired lock "refresh_cache-5a526856-0dcd-4b3f-8359-13a5c3b9bfc4" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 904.399019] env[61806]: DEBUG nova.network.neutron [req-16b0031f-bbd9-4768-b2ff-0881ac4502b0 req-c8677704-476b-4a0c-bfa6-be77d6aaf3f7 service nova] [instance: 5a526856-0dcd-4b3f-8359-13a5c3b9bfc4] Refreshing network info cache for port 58782499-bc1a-4be1-9d48-655c6b2f11d7 {{(pid=61806) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 904.487813] env[61806]: DEBUG oslo_concurrency.lockutils [None req-80f76c83-748f-43fe-b146-5517b19ec0ae tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] Lock "ee002e7c-8965-494f-a1af-0b6f3b74eceb" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 17.609s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 904.519775] env[61806]: DEBUG oslo_vmware.api [None req-464e7d68-de9c-4f8c-9477-932cd9673dfd tempest-ImagesTestJSON-618852556 tempest-ImagesTestJSON-618852556-project-member] Task: {'id': task-1294863, 'name': RemoveSnapshot_Task} progress is 80%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 904.530876] env[61806]: DEBUG oslo_vmware.api [None req-1d9f7d6e-8006-4355-8395-1f9c0f916c73 tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Task: {'id': task-1294860, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 904.656987] env[61806]: DEBUG oslo_vmware.api [None req-6d4e5237-b28c-45e9-8ae0-8ed79347c566 tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Task: {'id': session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]5249b495-e332-2e5a-e8f4-43960cbca0ff, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 905.027799] env[61806]: DEBUG oslo_vmware.api [None req-464e7d68-de9c-4f8c-9477-932cd9673dfd tempest-ImagesTestJSON-618852556 tempest-ImagesTestJSON-618852556-project-member] Task: {'id': task-1294863, 'name': RemoveSnapshot_Task, 'duration_secs': 1.071037} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 905.033076] env[61806]: DEBUG oslo_concurrency.lockutils [None req-26d83e60-0148-4554-a6a0-baa143245ad5 tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] Acquiring lock "d9caee4a-a00c-4e50-b01c-99a8f796a09f" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 905.033312] env[61806]: DEBUG oslo_concurrency.lockutils [None req-26d83e60-0148-4554-a6a0-baa143245ad5 tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] Lock "d9caee4a-a00c-4e50-b01c-99a8f796a09f" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 905.034468] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-464e7d68-de9c-4f8c-9477-932cd9673dfd tempest-ImagesTestJSON-618852556 tempest-ImagesTestJSON-618852556-project-member] [instance: 1a73d994-5f21-4f80-8f33-5830d860a7ee] Deleted Snapshot of the VM instance {{(pid=61806) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 905.043955] env[61806]: DEBUG oslo_vmware.api [None req-1d9f7d6e-8006-4355-8395-1f9c0f916c73 tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Task: {'id': task-1294860, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 905.130600] env[61806]: DEBUG nova.compute.manager [req-6da7952d-a2fb-4bdc-ae7b-09f235083965 req-a2e752af-9874-4908-a495-143ab30adb40 service nova] [instance: 5a526856-0dcd-4b3f-8359-13a5c3b9bfc4] Received event network-changed-58782499-bc1a-4be1-9d48-655c6b2f11d7 {{(pid=61806) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 905.130869] env[61806]: DEBUG nova.compute.manager [req-6da7952d-a2fb-4bdc-ae7b-09f235083965 req-a2e752af-9874-4908-a495-143ab30adb40 service nova] [instance: 5a526856-0dcd-4b3f-8359-13a5c3b9bfc4] Refreshing instance network info cache due to event network-changed-58782499-bc1a-4be1-9d48-655c6b2f11d7. {{(pid=61806) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 905.131037] env[61806]: DEBUG oslo_concurrency.lockutils [req-6da7952d-a2fb-4bdc-ae7b-09f235083965 req-a2e752af-9874-4908-a495-143ab30adb40 service nova] Acquiring lock "refresh_cache-5a526856-0dcd-4b3f-8359-13a5c3b9bfc4" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 905.159209] env[61806]: DEBUG oslo_vmware.api [None req-6d4e5237-b28c-45e9-8ae0-8ed79347c566 tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Task: {'id': session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]5249b495-e332-2e5a-e8f4-43960cbca0ff, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 905.287435] env[61806]: DEBUG nova.network.neutron [req-16b0031f-bbd9-4768-b2ff-0881ac4502b0 req-c8677704-476b-4a0c-bfa6-be77d6aaf3f7 service nova] [instance: 5a526856-0dcd-4b3f-8359-13a5c3b9bfc4] Updated VIF entry in instance network info cache for port 58782499-bc1a-4be1-9d48-655c6b2f11d7. {{(pid=61806) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 905.287814] env[61806]: DEBUG nova.network.neutron [req-16b0031f-bbd9-4768-b2ff-0881ac4502b0 req-c8677704-476b-4a0c-bfa6-be77d6aaf3f7 service nova] [instance: 5a526856-0dcd-4b3f-8359-13a5c3b9bfc4] Updating instance_info_cache with network_info: [{"id": "58782499-bc1a-4be1-9d48-655c6b2f11d7", "address": "fa:16:3e:52:7a:2b", "network": {"id": "b4fa0830-265e-47bf-8029-5944534312e7", "bridge": "br-int", "label": "tempest-ServerRescueTestJSONUnderV235-1460638664-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.190", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "2ba59612579b47a4bde7f82d951ac3ff", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8a34aa30-95be-4b18-98ca-1f2d81f7e9e6", "external-id": "nsx-vlan-transportzone-234", "segmentation_id": 234, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap58782499-bc", "ovs_interfaceid": "58782499-bc1a-4be1-9d48-655c6b2f11d7", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61806) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 905.347434] env[61806]: DEBUG nova.compute.manager [req-3d2bd614-d2d8-4dfb-a6a1-10e5d1fba845 req-8ace91a9-a466-4eea-b4f2-448d5c81f5eb service nova] [instance: 6f6ba57a-e2d6-4749-a53a-e263861cb1c0] Received event network-changed-013c9f0d-65b1-4279-a823-a7f1c14dfb91 {{(pid=61806) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 905.347650] env[61806]: DEBUG nova.compute.manager [req-3d2bd614-d2d8-4dfb-a6a1-10e5d1fba845 req-8ace91a9-a466-4eea-b4f2-448d5c81f5eb service nova] [instance: 6f6ba57a-e2d6-4749-a53a-e263861cb1c0] Refreshing instance network info cache due to event network-changed-013c9f0d-65b1-4279-a823-a7f1c14dfb91. {{(pid=61806) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 905.347875] env[61806]: DEBUG oslo_concurrency.lockutils [req-3d2bd614-d2d8-4dfb-a6a1-10e5d1fba845 req-8ace91a9-a466-4eea-b4f2-448d5c81f5eb service nova] Acquiring lock "refresh_cache-6f6ba57a-e2d6-4749-a53a-e263861cb1c0" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 905.348035] env[61806]: DEBUG oslo_concurrency.lockutils [req-3d2bd614-d2d8-4dfb-a6a1-10e5d1fba845 req-8ace91a9-a466-4eea-b4f2-448d5c81f5eb service nova] Acquired lock "refresh_cache-6f6ba57a-e2d6-4749-a53a-e263861cb1c0" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 905.348208] env[61806]: DEBUG nova.network.neutron [req-3d2bd614-d2d8-4dfb-a6a1-10e5d1fba845 req-8ace91a9-a466-4eea-b4f2-448d5c81f5eb service nova] [instance: 6f6ba57a-e2d6-4749-a53a-e263861cb1c0] Refreshing network info cache for port 013c9f0d-65b1-4279-a823-a7f1c14dfb91 {{(pid=61806) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 905.537410] env[61806]: DEBUG oslo_vmware.api [None req-1d9f7d6e-8006-4355-8395-1f9c0f916c73 tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Task: {'id': task-1294860, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 905.541033] env[61806]: DEBUG nova.compute.manager [None req-26d83e60-0148-4554-a6a0-baa143245ad5 tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] [instance: d9caee4a-a00c-4e50-b01c-99a8f796a09f] Starting instance... {{(pid=61806) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 905.545941] env[61806]: WARNING nova.compute.manager [None req-464e7d68-de9c-4f8c-9477-932cd9673dfd tempest-ImagesTestJSON-618852556 tempest-ImagesTestJSON-618852556-project-member] [instance: 1a73d994-5f21-4f80-8f33-5830d860a7ee] Image not found during snapshot: nova.exception.ImageNotFound: Image d57e5203-fede-48a8-b0c9-622639a68303 could not be found. [ 905.602963] env[61806]: DEBUG oslo_concurrency.lockutils [None req-e34fcc99-6699-4d73-ba15-e4ab836a8262 tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] Acquiring lock "ee002e7c-8965-494f-a1af-0b6f3b74eceb" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 905.603458] env[61806]: DEBUG oslo_concurrency.lockutils [None req-e34fcc99-6699-4d73-ba15-e4ab836a8262 tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] Lock "ee002e7c-8965-494f-a1af-0b6f3b74eceb" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 905.603970] env[61806]: DEBUG oslo_concurrency.lockutils [None req-e34fcc99-6699-4d73-ba15-e4ab836a8262 tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] Acquiring lock "ee002e7c-8965-494f-a1af-0b6f3b74eceb-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 905.604069] env[61806]: DEBUG oslo_concurrency.lockutils [None req-e34fcc99-6699-4d73-ba15-e4ab836a8262 tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] Lock "ee002e7c-8965-494f-a1af-0b6f3b74eceb-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 905.604234] env[61806]: DEBUG oslo_concurrency.lockutils [None req-e34fcc99-6699-4d73-ba15-e4ab836a8262 tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] Lock "ee002e7c-8965-494f-a1af-0b6f3b74eceb-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 905.606464] env[61806]: INFO nova.compute.manager [None req-e34fcc99-6699-4d73-ba15-e4ab836a8262 tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] [instance: ee002e7c-8965-494f-a1af-0b6f3b74eceb] Terminating instance [ 905.608400] env[61806]: DEBUG nova.compute.manager [None req-e34fcc99-6699-4d73-ba15-e4ab836a8262 tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] [instance: ee002e7c-8965-494f-a1af-0b6f3b74eceb] Start destroying the instance on the hypervisor. {{(pid=61806) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 905.608620] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-e34fcc99-6699-4d73-ba15-e4ab836a8262 tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] [instance: ee002e7c-8965-494f-a1af-0b6f3b74eceb] Destroying instance {{(pid=61806) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 905.609462] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-473e8f55-5288-436e-af8f-d688b531d8fb {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 905.619074] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [None req-e34fcc99-6699-4d73-ba15-e4ab836a8262 tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] [instance: ee002e7c-8965-494f-a1af-0b6f3b74eceb] Powering off the VM {{(pid=61806) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 905.619400] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-d1b13439-92d2-4fde-bf43-c9df27e13c97 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 905.628812] env[61806]: DEBUG oslo_vmware.api [None req-e34fcc99-6699-4d73-ba15-e4ab836a8262 tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] Waiting for the task: (returnval){ [ 905.628812] env[61806]: value = "task-1294864" [ 905.628812] env[61806]: _type = "Task" [ 905.628812] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 905.638607] env[61806]: DEBUG oslo_vmware.api [None req-e34fcc99-6699-4d73-ba15-e4ab836a8262 tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] Task: {'id': task-1294864, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 905.660744] env[61806]: DEBUG oslo_vmware.api [None req-6d4e5237-b28c-45e9-8ae0-8ed79347c566 tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Task: {'id': session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]5249b495-e332-2e5a-e8f4-43960cbca0ff, 'name': SearchDatastore_Task, 'duration_secs': 3.367266} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 905.661437] env[61806]: DEBUG oslo_concurrency.lockutils [None req-6d4e5237-b28c-45e9-8ae0-8ed79347c566 tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Releasing lock "[datastore1] devstack-image-cache_base/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 905.661864] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-6d4e5237-b28c-45e9-8ae0-8ed79347c566 tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] [instance: 75f1aded-d1da-4ddc-be29-f1fb93799364] Processing image ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa {{(pid=61806) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 905.662378] env[61806]: DEBUG oslo_concurrency.lockutils [None req-6d4e5237-b28c-45e9-8ae0-8ed79347c566 tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa.vmdk" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 905.662502] env[61806]: DEBUG oslo_concurrency.lockutils [None req-6d4e5237-b28c-45e9-8ae0-8ed79347c566 tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Acquired lock "[datastore1] devstack-image-cache_base/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa.vmdk" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 905.662826] env[61806]: DEBUG nova.virt.vmwareapi.ds_util [None req-6d4e5237-b28c-45e9-8ae0-8ed79347c566 tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61806) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 905.663248] env[61806]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-2e8ea839-aa8a-4a35-b743-5e02caaa3c23 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 905.677838] env[61806]: DEBUG nova.virt.vmwareapi.ds_util [None req-6d4e5237-b28c-45e9-8ae0-8ed79347c566 tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61806) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 905.678050] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-6d4e5237-b28c-45e9-8ae0-8ed79347c566 tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61806) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 905.678817] env[61806]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-8f6a2186-0808-4c1a-bff8-750cd7e0e8a4 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 905.685767] env[61806]: DEBUG oslo_vmware.api [None req-6d4e5237-b28c-45e9-8ae0-8ed79347c566 tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Waiting for the task: (returnval){ [ 905.685767] env[61806]: value = "session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]52ea2d81-08b4-e61b-5a86-32986ac55612" [ 905.685767] env[61806]: _type = "Task" [ 905.685767] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 905.696262] env[61806]: DEBUG oslo_vmware.api [None req-6d4e5237-b28c-45e9-8ae0-8ed79347c566 tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Task: {'id': session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]52ea2d81-08b4-e61b-5a86-32986ac55612, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 905.790871] env[61806]: DEBUG oslo_concurrency.lockutils [req-16b0031f-bbd9-4768-b2ff-0881ac4502b0 req-c8677704-476b-4a0c-bfa6-be77d6aaf3f7 service nova] Releasing lock "refresh_cache-5a526856-0dcd-4b3f-8359-13a5c3b9bfc4" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 905.791452] env[61806]: DEBUG oslo_concurrency.lockutils [req-6da7952d-a2fb-4bdc-ae7b-09f235083965 req-a2e752af-9874-4908-a495-143ab30adb40 service nova] Acquired lock "refresh_cache-5a526856-0dcd-4b3f-8359-13a5c3b9bfc4" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 905.791669] env[61806]: DEBUG nova.network.neutron [req-6da7952d-a2fb-4bdc-ae7b-09f235083965 req-a2e752af-9874-4908-a495-143ab30adb40 service nova] [instance: 5a526856-0dcd-4b3f-8359-13a5c3b9bfc4] Refreshing network info cache for port 58782499-bc1a-4be1-9d48-655c6b2f11d7 {{(pid=61806) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 905.903670] env[61806]: DEBUG oslo_concurrency.lockutils [None req-9c3f23d6-436a-448c-a3fa-4250c69b6723 tempest-ImagesTestJSON-618852556 tempest-ImagesTestJSON-618852556-project-member] Acquiring lock "1a73d994-5f21-4f80-8f33-5830d860a7ee" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 905.904023] env[61806]: DEBUG oslo_concurrency.lockutils [None req-9c3f23d6-436a-448c-a3fa-4250c69b6723 tempest-ImagesTestJSON-618852556 tempest-ImagesTestJSON-618852556-project-member] Lock "1a73d994-5f21-4f80-8f33-5830d860a7ee" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 905.904250] env[61806]: DEBUG oslo_concurrency.lockutils [None req-9c3f23d6-436a-448c-a3fa-4250c69b6723 tempest-ImagesTestJSON-618852556 tempest-ImagesTestJSON-618852556-project-member] Acquiring lock "1a73d994-5f21-4f80-8f33-5830d860a7ee-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 905.904450] env[61806]: DEBUG oslo_concurrency.lockutils [None req-9c3f23d6-436a-448c-a3fa-4250c69b6723 tempest-ImagesTestJSON-618852556 tempest-ImagesTestJSON-618852556-project-member] Lock "1a73d994-5f21-4f80-8f33-5830d860a7ee-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 905.904685] env[61806]: DEBUG oslo_concurrency.lockutils [None req-9c3f23d6-436a-448c-a3fa-4250c69b6723 tempest-ImagesTestJSON-618852556 tempest-ImagesTestJSON-618852556-project-member] Lock "1a73d994-5f21-4f80-8f33-5830d860a7ee-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 905.909018] env[61806]: INFO nova.compute.manager [None req-9c3f23d6-436a-448c-a3fa-4250c69b6723 tempest-ImagesTestJSON-618852556 tempest-ImagesTestJSON-618852556-project-member] [instance: 1a73d994-5f21-4f80-8f33-5830d860a7ee] Terminating instance [ 905.910943] env[61806]: DEBUG nova.compute.manager [None req-9c3f23d6-436a-448c-a3fa-4250c69b6723 tempest-ImagesTestJSON-618852556 tempest-ImagesTestJSON-618852556-project-member] [instance: 1a73d994-5f21-4f80-8f33-5830d860a7ee] Start destroying the instance on the hypervisor. {{(pid=61806) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 905.911187] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-9c3f23d6-436a-448c-a3fa-4250c69b6723 tempest-ImagesTestJSON-618852556 tempest-ImagesTestJSON-618852556-project-member] [instance: 1a73d994-5f21-4f80-8f33-5830d860a7ee] Destroying instance {{(pid=61806) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 905.912068] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b2da07d0-de76-4b64-8ea8-2653dad6fd88 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 905.920116] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [None req-9c3f23d6-436a-448c-a3fa-4250c69b6723 tempest-ImagesTestJSON-618852556 tempest-ImagesTestJSON-618852556-project-member] [instance: 1a73d994-5f21-4f80-8f33-5830d860a7ee] Powering off the VM {{(pid=61806) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 905.920398] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-9288d3ba-ad76-4539-baae-3e01123c8e4b {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 905.927142] env[61806]: DEBUG oslo_vmware.api [None req-9c3f23d6-436a-448c-a3fa-4250c69b6723 tempest-ImagesTestJSON-618852556 tempest-ImagesTestJSON-618852556-project-member] Waiting for the task: (returnval){ [ 905.927142] env[61806]: value = "task-1294865" [ 905.927142] env[61806]: _type = "Task" [ 905.927142] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 905.936903] env[61806]: DEBUG oslo_vmware.api [None req-9c3f23d6-436a-448c-a3fa-4250c69b6723 tempest-ImagesTestJSON-618852556 tempest-ImagesTestJSON-618852556-project-member] Task: {'id': task-1294865, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 906.037405] env[61806]: DEBUG oslo_vmware.api [None req-1d9f7d6e-8006-4355-8395-1f9c0f916c73 tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Task: {'id': task-1294860, 'name': ReconfigVM_Task, 'duration_secs': 3.814105} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 906.037715] env[61806]: DEBUG nova.virt.vmwareapi.volumeops [None req-1d9f7d6e-8006-4355-8395-1f9c0f916c73 tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] [instance: 59c22c16-3d1c-4e64-bfac-7f8886a1927e] Reconfigured VM instance instance-0000004b to detach disk 2001 {{(pid=61806) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 906.045205] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-f82e6602-6a9e-4528-b619-e09863a94b06 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 906.067472] env[61806]: DEBUG oslo_vmware.api [None req-1d9f7d6e-8006-4355-8395-1f9c0f916c73 tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Waiting for the task: (returnval){ [ 906.067472] env[61806]: value = "task-1294866" [ 906.067472] env[61806]: _type = "Task" [ 906.067472] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 906.077841] env[61806]: DEBUG oslo_vmware.api [None req-1d9f7d6e-8006-4355-8395-1f9c0f916c73 tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Task: {'id': task-1294866, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 906.079584] env[61806]: DEBUG oslo_concurrency.lockutils [None req-26d83e60-0148-4554-a6a0-baa143245ad5 tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 906.079903] env[61806]: DEBUG oslo_concurrency.lockutils [None req-26d83e60-0148-4554-a6a0-baa143245ad5 tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.000s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 906.081713] env[61806]: INFO nova.compute.claims [None req-26d83e60-0148-4554-a6a0-baa143245ad5 tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] [instance: d9caee4a-a00c-4e50-b01c-99a8f796a09f] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 906.140734] env[61806]: DEBUG oslo_vmware.api [None req-e34fcc99-6699-4d73-ba15-e4ab836a8262 tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] Task: {'id': task-1294864, 'name': PowerOffVM_Task, 'duration_secs': 0.240034} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 906.141170] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [None req-e34fcc99-6699-4d73-ba15-e4ab836a8262 tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] [instance: ee002e7c-8965-494f-a1af-0b6f3b74eceb] Powered off the VM {{(pid=61806) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 906.141441] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-e34fcc99-6699-4d73-ba15-e4ab836a8262 tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] [instance: ee002e7c-8965-494f-a1af-0b6f3b74eceb] Unregistering the VM {{(pid=61806) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 906.141775] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-96324adf-3581-4621-8810-e39754486eab {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 906.173235] env[61806]: DEBUG nova.network.neutron [req-3d2bd614-d2d8-4dfb-a6a1-10e5d1fba845 req-8ace91a9-a466-4eea-b4f2-448d5c81f5eb service nova] [instance: 6f6ba57a-e2d6-4749-a53a-e263861cb1c0] Updated VIF entry in instance network info cache for port 013c9f0d-65b1-4279-a823-a7f1c14dfb91. {{(pid=61806) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 906.173235] env[61806]: DEBUG nova.network.neutron [req-3d2bd614-d2d8-4dfb-a6a1-10e5d1fba845 req-8ace91a9-a466-4eea-b4f2-448d5c81f5eb service nova] [instance: 6f6ba57a-e2d6-4749-a53a-e263861cb1c0] Updating instance_info_cache with network_info: [{"id": "013c9f0d-65b1-4279-a823-a7f1c14dfb91", "address": "fa:16:3e:15:56:4e", "network": {"id": "a7ffda72-e57c-41fa-a7ed-d7b12084bc9d", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-62794865-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.218", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "6e391e1267014fafbb8f5a3211299819", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8b3cfeb1-f262-4fd9-b506-8e9c0733e2d8", "external-id": "nsx-vlan-transportzone-119", "segmentation_id": 119, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap013c9f0d-65", "ovs_interfaceid": "013c9f0d-65b1-4279-a823-a7f1c14dfb91", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61806) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 906.201928] env[61806]: DEBUG oslo_vmware.api [None req-6d4e5237-b28c-45e9-8ae0-8ed79347c566 tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Task: {'id': session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]52ea2d81-08b4-e61b-5a86-32986ac55612, 'name': SearchDatastore_Task, 'duration_secs': 0.013793} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 906.204462] env[61806]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-243ec74b-0e7d-40da-b804-034ca716ade1 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 906.213727] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-e34fcc99-6699-4d73-ba15-e4ab836a8262 tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] [instance: ee002e7c-8965-494f-a1af-0b6f3b74eceb] Unregistered the VM {{(pid=61806) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 906.214192] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-e34fcc99-6699-4d73-ba15-e4ab836a8262 tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] [instance: ee002e7c-8965-494f-a1af-0b6f3b74eceb] Deleting contents of the VM from datastore datastore2 {{(pid=61806) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 906.214527] env[61806]: DEBUG nova.virt.vmwareapi.ds_util [None req-e34fcc99-6699-4d73-ba15-e4ab836a8262 tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] Deleting the datastore file [datastore2] ee002e7c-8965-494f-a1af-0b6f3b74eceb {{(pid=61806) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 906.215683] env[61806]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-96ea9c71-7800-4b2d-ae9f-468a2903e3b5 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 906.218779] env[61806]: DEBUG oslo_vmware.api [None req-6d4e5237-b28c-45e9-8ae0-8ed79347c566 tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Waiting for the task: (returnval){ [ 906.218779] env[61806]: value = "session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]5213eef7-60b7-6fee-12f1-34abb3713e29" [ 906.218779] env[61806]: _type = "Task" [ 906.218779] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 906.225555] env[61806]: DEBUG oslo_vmware.api [None req-e34fcc99-6699-4d73-ba15-e4ab836a8262 tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] Waiting for the task: (returnval){ [ 906.225555] env[61806]: value = "task-1294868" [ 906.225555] env[61806]: _type = "Task" [ 906.225555] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 906.235210] env[61806]: DEBUG oslo_vmware.api [None req-6d4e5237-b28c-45e9-8ae0-8ed79347c566 tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Task: {'id': session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]5213eef7-60b7-6fee-12f1-34abb3713e29, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 906.241730] env[61806]: DEBUG oslo_vmware.api [None req-e34fcc99-6699-4d73-ba15-e4ab836a8262 tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] Task: {'id': task-1294868, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 906.439371] env[61806]: DEBUG oslo_vmware.api [None req-9c3f23d6-436a-448c-a3fa-4250c69b6723 tempest-ImagesTestJSON-618852556 tempest-ImagesTestJSON-618852556-project-member] Task: {'id': task-1294865, 'name': PowerOffVM_Task, 'duration_secs': 0.338851} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 906.439719] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [None req-9c3f23d6-436a-448c-a3fa-4250c69b6723 tempest-ImagesTestJSON-618852556 tempest-ImagesTestJSON-618852556-project-member] [instance: 1a73d994-5f21-4f80-8f33-5830d860a7ee] Powered off the VM {{(pid=61806) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 906.439973] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-9c3f23d6-436a-448c-a3fa-4250c69b6723 tempest-ImagesTestJSON-618852556 tempest-ImagesTestJSON-618852556-project-member] [instance: 1a73d994-5f21-4f80-8f33-5830d860a7ee] Unregistering the VM {{(pid=61806) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 906.440351] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-d3d7d02c-38a0-4b7d-9f10-b5891ac48c42 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 906.520533] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-9c3f23d6-436a-448c-a3fa-4250c69b6723 tempest-ImagesTestJSON-618852556 tempest-ImagesTestJSON-618852556-project-member] [instance: 1a73d994-5f21-4f80-8f33-5830d860a7ee] Unregistered the VM {{(pid=61806) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 906.520791] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-9c3f23d6-436a-448c-a3fa-4250c69b6723 tempest-ImagesTestJSON-618852556 tempest-ImagesTestJSON-618852556-project-member] [instance: 1a73d994-5f21-4f80-8f33-5830d860a7ee] Deleting contents of the VM from datastore datastore2 {{(pid=61806) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 906.520982] env[61806]: DEBUG nova.virt.vmwareapi.ds_util [None req-9c3f23d6-436a-448c-a3fa-4250c69b6723 tempest-ImagesTestJSON-618852556 tempest-ImagesTestJSON-618852556-project-member] Deleting the datastore file [datastore2] 1a73d994-5f21-4f80-8f33-5830d860a7ee {{(pid=61806) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 906.521290] env[61806]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-50616650-bf62-4af4-b4e3-fe1f220b9964 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 906.529823] env[61806]: DEBUG oslo_vmware.api [None req-9c3f23d6-436a-448c-a3fa-4250c69b6723 tempest-ImagesTestJSON-618852556 tempest-ImagesTestJSON-618852556-project-member] Waiting for the task: (returnval){ [ 906.529823] env[61806]: value = "task-1294870" [ 906.529823] env[61806]: _type = "Task" [ 906.529823] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 906.538619] env[61806]: DEBUG oslo_vmware.api [None req-9c3f23d6-436a-448c-a3fa-4250c69b6723 tempest-ImagesTestJSON-618852556 tempest-ImagesTestJSON-618852556-project-member] Task: {'id': task-1294870, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 906.543751] env[61806]: DEBUG nova.network.neutron [req-6da7952d-a2fb-4bdc-ae7b-09f235083965 req-a2e752af-9874-4908-a495-143ab30adb40 service nova] [instance: 5a526856-0dcd-4b3f-8359-13a5c3b9bfc4] Updated VIF entry in instance network info cache for port 58782499-bc1a-4be1-9d48-655c6b2f11d7. {{(pid=61806) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 906.544140] env[61806]: DEBUG nova.network.neutron [req-6da7952d-a2fb-4bdc-ae7b-09f235083965 req-a2e752af-9874-4908-a495-143ab30adb40 service nova] [instance: 5a526856-0dcd-4b3f-8359-13a5c3b9bfc4] Updating instance_info_cache with network_info: [{"id": "58782499-bc1a-4be1-9d48-655c6b2f11d7", "address": "fa:16:3e:52:7a:2b", "network": {"id": "b4fa0830-265e-47bf-8029-5944534312e7", "bridge": "br-int", "label": "tempest-ServerRescueTestJSONUnderV235-1460638664-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "2ba59612579b47a4bde7f82d951ac3ff", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8a34aa30-95be-4b18-98ca-1f2d81f7e9e6", "external-id": "nsx-vlan-transportzone-234", "segmentation_id": 234, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap58782499-bc", "ovs_interfaceid": "58782499-bc1a-4be1-9d48-655c6b2f11d7", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61806) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 906.577881] env[61806]: DEBUG oslo_vmware.api [None req-1d9f7d6e-8006-4355-8395-1f9c0f916c73 tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Task: {'id': task-1294866, 'name': ReconfigVM_Task, 'duration_secs': 0.254434} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 906.578267] env[61806]: DEBUG nova.virt.vmwareapi.volumeops [None req-1d9f7d6e-8006-4355-8395-1f9c0f916c73 tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] [instance: 59c22c16-3d1c-4e64-bfac-7f8886a1927e] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-277703', 'volume_id': '564a9f8e-1bec-410b-9d7d-ab9bf7b6d83b', 'name': 'volume-564a9f8e-1bec-410b-9d7d-ab9bf7b6d83b', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '59c22c16-3d1c-4e64-bfac-7f8886a1927e', 'attached_at': '', 'detached_at': '', 'volume_id': '564a9f8e-1bec-410b-9d7d-ab9bf7b6d83b', 'serial': '564a9f8e-1bec-410b-9d7d-ab9bf7b6d83b'} {{(pid=61806) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 906.578623] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-1d9f7d6e-8006-4355-8395-1f9c0f916c73 tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] [instance: 59c22c16-3d1c-4e64-bfac-7f8886a1927e] Destroying instance {{(pid=61806) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 906.579437] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e85439c9-d209-4728-8c1c-7e5b2b31270d {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 906.589428] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-1d9f7d6e-8006-4355-8395-1f9c0f916c73 tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] [instance: 59c22c16-3d1c-4e64-bfac-7f8886a1927e] Unregistering the VM {{(pid=61806) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 906.590119] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-9b12147a-3c18-4ba1-a0e2-5c8142cc4361 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 906.662014] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-1d9f7d6e-8006-4355-8395-1f9c0f916c73 tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] [instance: 59c22c16-3d1c-4e64-bfac-7f8886a1927e] Unregistered the VM {{(pid=61806) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 906.662334] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-1d9f7d6e-8006-4355-8395-1f9c0f916c73 tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] [instance: 59c22c16-3d1c-4e64-bfac-7f8886a1927e] Deleting contents of the VM from datastore datastore1 {{(pid=61806) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 906.662592] env[61806]: DEBUG nova.virt.vmwareapi.ds_util [None req-1d9f7d6e-8006-4355-8395-1f9c0f916c73 tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Deleting the datastore file [datastore1] 59c22c16-3d1c-4e64-bfac-7f8886a1927e {{(pid=61806) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 906.663317] env[61806]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-946a1fca-5040-441f-8a96-01042ad3d6e3 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 906.671842] env[61806]: DEBUG oslo_vmware.api [None req-1d9f7d6e-8006-4355-8395-1f9c0f916c73 tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Waiting for the task: (returnval){ [ 906.671842] env[61806]: value = "task-1294872" [ 906.671842] env[61806]: _type = "Task" [ 906.671842] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 906.680782] env[61806]: DEBUG oslo_concurrency.lockutils [req-3d2bd614-d2d8-4dfb-a6a1-10e5d1fba845 req-8ace91a9-a466-4eea-b4f2-448d5c81f5eb service nova] Releasing lock "refresh_cache-6f6ba57a-e2d6-4749-a53a-e263861cb1c0" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 906.681164] env[61806]: DEBUG oslo_vmware.api [None req-1d9f7d6e-8006-4355-8395-1f9c0f916c73 tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Task: {'id': task-1294872, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 906.730586] env[61806]: DEBUG oslo_vmware.api [None req-6d4e5237-b28c-45e9-8ae0-8ed79347c566 tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Task: {'id': session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]5213eef7-60b7-6fee-12f1-34abb3713e29, 'name': SearchDatastore_Task, 'duration_secs': 0.026112} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 906.736908] env[61806]: DEBUG oslo_concurrency.lockutils [None req-6d4e5237-b28c-45e9-8ae0-8ed79347c566 tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Releasing lock "[datastore1] devstack-image-cache_base/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa.vmdk" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 906.737232] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [None req-6d4e5237-b28c-45e9-8ae0-8ed79347c566 tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa.vmdk to [datastore1] 75f1aded-d1da-4ddc-be29-f1fb93799364/75f1aded-d1da-4ddc-be29-f1fb93799364.vmdk {{(pid=61806) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 906.737584] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-d1daceee-4045-46b0-b557-09e5ebeca581 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 906.747289] env[61806]: DEBUG oslo_vmware.api [None req-e34fcc99-6699-4d73-ba15-e4ab836a8262 tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] Task: {'id': task-1294868, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.325003} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 906.748699] env[61806]: DEBUG nova.virt.vmwareapi.ds_util [None req-e34fcc99-6699-4d73-ba15-e4ab836a8262 tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] Deleted the datastore file {{(pid=61806) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 906.748909] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-e34fcc99-6699-4d73-ba15-e4ab836a8262 tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] [instance: ee002e7c-8965-494f-a1af-0b6f3b74eceb] Deleted contents of the VM from datastore datastore2 {{(pid=61806) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 906.749105] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-e34fcc99-6699-4d73-ba15-e4ab836a8262 tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] [instance: ee002e7c-8965-494f-a1af-0b6f3b74eceb] Instance destroyed {{(pid=61806) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 906.749284] env[61806]: INFO nova.compute.manager [None req-e34fcc99-6699-4d73-ba15-e4ab836a8262 tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] [instance: ee002e7c-8965-494f-a1af-0b6f3b74eceb] Took 1.14 seconds to destroy the instance on the hypervisor. [ 906.749533] env[61806]: DEBUG oslo.service.loopingcall [None req-e34fcc99-6699-4d73-ba15-e4ab836a8262 tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61806) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 906.749814] env[61806]: DEBUG oslo_vmware.api [None req-6d4e5237-b28c-45e9-8ae0-8ed79347c566 tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Waiting for the task: (returnval){ [ 906.749814] env[61806]: value = "task-1294873" [ 906.749814] env[61806]: _type = "Task" [ 906.749814] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 906.750008] env[61806]: DEBUG nova.compute.manager [-] [instance: ee002e7c-8965-494f-a1af-0b6f3b74eceb] Deallocating network for instance {{(pid=61806) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 906.750115] env[61806]: DEBUG nova.network.neutron [-] [instance: ee002e7c-8965-494f-a1af-0b6f3b74eceb] deallocate_for_instance() {{(pid=61806) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 906.755862] env[61806]: DEBUG oslo_concurrency.lockutils [None req-dab50ec4-3405-470b-a203-8c80641dde31 tempest-ServerRescueTestJSONUnderV235-749414849 tempest-ServerRescueTestJSONUnderV235-749414849-project-member] Acquiring lock "5a526856-0dcd-4b3f-8359-13a5c3b9bfc4" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 906.756129] env[61806]: DEBUG oslo_concurrency.lockutils [None req-dab50ec4-3405-470b-a203-8c80641dde31 tempest-ServerRescueTestJSONUnderV235-749414849 tempest-ServerRescueTestJSONUnderV235-749414849-project-member] Lock "5a526856-0dcd-4b3f-8359-13a5c3b9bfc4" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 906.756422] env[61806]: DEBUG oslo_concurrency.lockutils [None req-dab50ec4-3405-470b-a203-8c80641dde31 tempest-ServerRescueTestJSONUnderV235-749414849 tempest-ServerRescueTestJSONUnderV235-749414849-project-member] Acquiring lock "5a526856-0dcd-4b3f-8359-13a5c3b9bfc4-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 906.756627] env[61806]: DEBUG oslo_concurrency.lockutils [None req-dab50ec4-3405-470b-a203-8c80641dde31 tempest-ServerRescueTestJSONUnderV235-749414849 tempest-ServerRescueTestJSONUnderV235-749414849-project-member] Lock "5a526856-0dcd-4b3f-8359-13a5c3b9bfc4-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 906.756894] env[61806]: DEBUG oslo_concurrency.lockutils [None req-dab50ec4-3405-470b-a203-8c80641dde31 tempest-ServerRescueTestJSONUnderV235-749414849 tempest-ServerRescueTestJSONUnderV235-749414849-project-member] Lock "5a526856-0dcd-4b3f-8359-13a5c3b9bfc4-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 906.759219] env[61806]: INFO nova.compute.manager [None req-dab50ec4-3405-470b-a203-8c80641dde31 tempest-ServerRescueTestJSONUnderV235-749414849 tempest-ServerRescueTestJSONUnderV235-749414849-project-member] [instance: 5a526856-0dcd-4b3f-8359-13a5c3b9bfc4] Terminating instance [ 906.763472] env[61806]: DEBUG oslo_vmware.api [None req-6d4e5237-b28c-45e9-8ae0-8ed79347c566 tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Task: {'id': task-1294873, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 906.764059] env[61806]: DEBUG nova.compute.manager [None req-dab50ec4-3405-470b-a203-8c80641dde31 tempest-ServerRescueTestJSONUnderV235-749414849 tempest-ServerRescueTestJSONUnderV235-749414849-project-member] [instance: 5a526856-0dcd-4b3f-8359-13a5c3b9bfc4] Start destroying the instance on the hypervisor. {{(pid=61806) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 906.764259] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-dab50ec4-3405-470b-a203-8c80641dde31 tempest-ServerRescueTestJSONUnderV235-749414849 tempest-ServerRescueTestJSONUnderV235-749414849-project-member] [instance: 5a526856-0dcd-4b3f-8359-13a5c3b9bfc4] Destroying instance {{(pid=61806) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 906.765034] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-445c982f-f753-48c8-ac87-ae857c0a6bc4 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 906.772838] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [None req-dab50ec4-3405-470b-a203-8c80641dde31 tempest-ServerRescueTestJSONUnderV235-749414849 tempest-ServerRescueTestJSONUnderV235-749414849-project-member] [instance: 5a526856-0dcd-4b3f-8359-13a5c3b9bfc4] Powering off the VM {{(pid=61806) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 906.773108] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-0ac1c99c-ab22-4fb2-85dc-e3ed00d32460 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 906.779945] env[61806]: DEBUG oslo_vmware.api [None req-dab50ec4-3405-470b-a203-8c80641dde31 tempest-ServerRescueTestJSONUnderV235-749414849 tempest-ServerRescueTestJSONUnderV235-749414849-project-member] Waiting for the task: (returnval){ [ 906.779945] env[61806]: value = "task-1294874" [ 906.779945] env[61806]: _type = "Task" [ 906.779945] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 906.788552] env[61806]: DEBUG oslo_vmware.api [None req-dab50ec4-3405-470b-a203-8c80641dde31 tempest-ServerRescueTestJSONUnderV235-749414849 tempest-ServerRescueTestJSONUnderV235-749414849-project-member] Task: {'id': task-1294874, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 907.045431] env[61806]: DEBUG oslo_vmware.api [None req-9c3f23d6-436a-448c-a3fa-4250c69b6723 tempest-ImagesTestJSON-618852556 tempest-ImagesTestJSON-618852556-project-member] Task: {'id': task-1294870, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.220708} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 907.045800] env[61806]: DEBUG nova.virt.vmwareapi.ds_util [None req-9c3f23d6-436a-448c-a3fa-4250c69b6723 tempest-ImagesTestJSON-618852556 tempest-ImagesTestJSON-618852556-project-member] Deleted the datastore file {{(pid=61806) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 907.045997] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-9c3f23d6-436a-448c-a3fa-4250c69b6723 tempest-ImagesTestJSON-618852556 tempest-ImagesTestJSON-618852556-project-member] [instance: 1a73d994-5f21-4f80-8f33-5830d860a7ee] Deleted contents of the VM from datastore datastore2 {{(pid=61806) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 907.046205] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-9c3f23d6-436a-448c-a3fa-4250c69b6723 tempest-ImagesTestJSON-618852556 tempest-ImagesTestJSON-618852556-project-member] [instance: 1a73d994-5f21-4f80-8f33-5830d860a7ee] Instance destroyed {{(pid=61806) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 907.046393] env[61806]: INFO nova.compute.manager [None req-9c3f23d6-436a-448c-a3fa-4250c69b6723 tempest-ImagesTestJSON-618852556 tempest-ImagesTestJSON-618852556-project-member] [instance: 1a73d994-5f21-4f80-8f33-5830d860a7ee] Took 1.14 seconds to destroy the instance on the hypervisor. [ 907.046665] env[61806]: DEBUG oslo.service.loopingcall [None req-9c3f23d6-436a-448c-a3fa-4250c69b6723 tempest-ImagesTestJSON-618852556 tempest-ImagesTestJSON-618852556-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61806) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 907.048238] env[61806]: DEBUG oslo_concurrency.lockutils [req-6da7952d-a2fb-4bdc-ae7b-09f235083965 req-a2e752af-9874-4908-a495-143ab30adb40 service nova] Releasing lock "refresh_cache-5a526856-0dcd-4b3f-8359-13a5c3b9bfc4" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 907.049032] env[61806]: DEBUG nova.compute.manager [-] [instance: 1a73d994-5f21-4f80-8f33-5830d860a7ee] Deallocating network for instance {{(pid=61806) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 907.049129] env[61806]: DEBUG nova.network.neutron [-] [instance: 1a73d994-5f21-4f80-8f33-5830d860a7ee] deallocate_for_instance() {{(pid=61806) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 907.185125] env[61806]: DEBUG oslo_vmware.api [None req-1d9f7d6e-8006-4355-8395-1f9c0f916c73 tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Task: {'id': task-1294872, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.206102} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 907.185873] env[61806]: DEBUG nova.virt.vmwareapi.ds_util [None req-1d9f7d6e-8006-4355-8395-1f9c0f916c73 tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Deleted the datastore file {{(pid=61806) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 907.185873] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-1d9f7d6e-8006-4355-8395-1f9c0f916c73 tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] [instance: 59c22c16-3d1c-4e64-bfac-7f8886a1927e] Deleted contents of the VM from datastore datastore1 {{(pid=61806) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 907.186043] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-1d9f7d6e-8006-4355-8395-1f9c0f916c73 tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] [instance: 59c22c16-3d1c-4e64-bfac-7f8886a1927e] Instance destroyed {{(pid=61806) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 907.186191] env[61806]: INFO nova.compute.manager [None req-1d9f7d6e-8006-4355-8395-1f9c0f916c73 tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] [instance: 59c22c16-3d1c-4e64-bfac-7f8886a1927e] Took 5.82 seconds to destroy the instance on the hypervisor. [ 907.186510] env[61806]: DEBUG oslo.service.loopingcall [None req-1d9f7d6e-8006-4355-8395-1f9c0f916c73 tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61806) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 907.186754] env[61806]: DEBUG nova.compute.manager [-] [instance: 59c22c16-3d1c-4e64-bfac-7f8886a1927e] Deallocating network for instance {{(pid=61806) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 907.186847] env[61806]: DEBUG nova.network.neutron [-] [instance: 59c22c16-3d1c-4e64-bfac-7f8886a1927e] deallocate_for_instance() {{(pid=61806) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 907.264509] env[61806]: DEBUG oslo_vmware.api [None req-6d4e5237-b28c-45e9-8ae0-8ed79347c566 tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Task: {'id': task-1294873, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 907.271538] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-55d22cb7-bb9c-4a54-83e8-8a904a3681f7 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 907.280014] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cf415f38-1875-4ffc-bcc1-e86b736d2206 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 907.293293] env[61806]: DEBUG oslo_vmware.api [None req-dab50ec4-3405-470b-a203-8c80641dde31 tempest-ServerRescueTestJSONUnderV235-749414849 tempest-ServerRescueTestJSONUnderV235-749414849-project-member] Task: {'id': task-1294874, 'name': PowerOffVM_Task, 'duration_secs': 0.275277} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 907.320815] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [None req-dab50ec4-3405-470b-a203-8c80641dde31 tempest-ServerRescueTestJSONUnderV235-749414849 tempest-ServerRescueTestJSONUnderV235-749414849-project-member] [instance: 5a526856-0dcd-4b3f-8359-13a5c3b9bfc4] Powered off the VM {{(pid=61806) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 907.321034] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-dab50ec4-3405-470b-a203-8c80641dde31 tempest-ServerRescueTestJSONUnderV235-749414849 tempest-ServerRescueTestJSONUnderV235-749414849-project-member] [instance: 5a526856-0dcd-4b3f-8359-13a5c3b9bfc4] Unregistering the VM {{(pid=61806) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 907.321893] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-1c05b7a4-072a-43ed-a956-ca3de0f3e930 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 907.324152] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b91041aa-7e95-44d0-a9a5-77c0ade98c4c {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 907.334256] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ff3dbdb4-c6d5-4e45-89ee-a617141ef7dc {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 907.351577] env[61806]: DEBUG nova.compute.provider_tree [None req-26d83e60-0148-4554-a6a0-baa143245ad5 tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] Inventory has not changed in ProviderTree for provider: a2858be1-fd22-4e08-979e-87ad25293407 {{(pid=61806) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 907.389213] env[61806]: DEBUG nova.compute.manager [req-3a6967a4-fccd-4b9d-8438-18f0715a35fd req-2158630e-4917-4058-8b93-304ffe7fed2c service nova] [instance: 5a526856-0dcd-4b3f-8359-13a5c3b9bfc4] Received event network-changed-58782499-bc1a-4be1-9d48-655c6b2f11d7 {{(pid=61806) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 907.389524] env[61806]: DEBUG nova.compute.manager [req-3a6967a4-fccd-4b9d-8438-18f0715a35fd req-2158630e-4917-4058-8b93-304ffe7fed2c service nova] [instance: 5a526856-0dcd-4b3f-8359-13a5c3b9bfc4] Refreshing instance network info cache due to event network-changed-58782499-bc1a-4be1-9d48-655c6b2f11d7. {{(pid=61806) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 907.389820] env[61806]: DEBUG oslo_concurrency.lockutils [req-3a6967a4-fccd-4b9d-8438-18f0715a35fd req-2158630e-4917-4058-8b93-304ffe7fed2c service nova] Acquiring lock "refresh_cache-5a526856-0dcd-4b3f-8359-13a5c3b9bfc4" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 907.390096] env[61806]: DEBUG oslo_concurrency.lockutils [req-3a6967a4-fccd-4b9d-8438-18f0715a35fd req-2158630e-4917-4058-8b93-304ffe7fed2c service nova] Acquired lock "refresh_cache-5a526856-0dcd-4b3f-8359-13a5c3b9bfc4" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 907.392516] env[61806]: DEBUG nova.network.neutron [req-3a6967a4-fccd-4b9d-8438-18f0715a35fd req-2158630e-4917-4058-8b93-304ffe7fed2c service nova] [instance: 5a526856-0dcd-4b3f-8359-13a5c3b9bfc4] Refreshing network info cache for port 58782499-bc1a-4be1-9d48-655c6b2f11d7 {{(pid=61806) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 907.423130] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-dab50ec4-3405-470b-a203-8c80641dde31 tempest-ServerRescueTestJSONUnderV235-749414849 tempest-ServerRescueTestJSONUnderV235-749414849-project-member] [instance: 5a526856-0dcd-4b3f-8359-13a5c3b9bfc4] Unregistered the VM {{(pid=61806) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 907.423375] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-dab50ec4-3405-470b-a203-8c80641dde31 tempest-ServerRescueTestJSONUnderV235-749414849 tempest-ServerRescueTestJSONUnderV235-749414849-project-member] [instance: 5a526856-0dcd-4b3f-8359-13a5c3b9bfc4] Deleting contents of the VM from datastore datastore2 {{(pid=61806) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 907.423634] env[61806]: DEBUG nova.virt.vmwareapi.ds_util [None req-dab50ec4-3405-470b-a203-8c80641dde31 tempest-ServerRescueTestJSONUnderV235-749414849 tempest-ServerRescueTestJSONUnderV235-749414849-project-member] Deleting the datastore file [datastore2] 5a526856-0dcd-4b3f-8359-13a5c3b9bfc4 {{(pid=61806) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 907.424389] env[61806]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-5717427b-a621-47fe-af19-d4f0ffcd18c3 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 907.434818] env[61806]: DEBUG oslo_vmware.api [None req-dab50ec4-3405-470b-a203-8c80641dde31 tempest-ServerRescueTestJSONUnderV235-749414849 tempest-ServerRescueTestJSONUnderV235-749414849-project-member] Waiting for the task: (returnval){ [ 907.434818] env[61806]: value = "task-1294876" [ 907.434818] env[61806]: _type = "Task" [ 907.434818] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 907.444682] env[61806]: DEBUG oslo_vmware.api [None req-dab50ec4-3405-470b-a203-8c80641dde31 tempest-ServerRescueTestJSONUnderV235-749414849 tempest-ServerRescueTestJSONUnderV235-749414849-project-member] Task: {'id': task-1294876, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 907.516571] env[61806]: DEBUG nova.compute.manager [req-5426e90f-2d62-45f9-b35f-e3ff57b0b44b req-c6919727-676b-4406-abd1-ee5e73170245 service nova] [instance: 59c22c16-3d1c-4e64-bfac-7f8886a1927e] Received event network-vif-deleted-a3901838-4afa-473f-a673-faac7dee3e75 {{(pid=61806) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 907.517178] env[61806]: INFO nova.compute.manager [req-5426e90f-2d62-45f9-b35f-e3ff57b0b44b req-c6919727-676b-4406-abd1-ee5e73170245 service nova] [instance: 59c22c16-3d1c-4e64-bfac-7f8886a1927e] Neutron deleted interface a3901838-4afa-473f-a673-faac7dee3e75; detaching it from the instance and deleting it from the info cache [ 907.517389] env[61806]: DEBUG nova.network.neutron [req-5426e90f-2d62-45f9-b35f-e3ff57b0b44b req-c6919727-676b-4406-abd1-ee5e73170245 service nova] [instance: 59c22c16-3d1c-4e64-bfac-7f8886a1927e] Updating instance_info_cache with network_info: [] {{(pid=61806) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 907.552403] env[61806]: DEBUG nova.network.neutron [-] [instance: ee002e7c-8965-494f-a1af-0b6f3b74eceb] Updating instance_info_cache with network_info: [] {{(pid=61806) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 907.763964] env[61806]: DEBUG oslo_vmware.api [None req-6d4e5237-b28c-45e9-8ae0-8ed79347c566 tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Task: {'id': task-1294873, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.676093} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 907.764171] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [None req-6d4e5237-b28c-45e9-8ae0-8ed79347c566 tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa.vmdk to [datastore1] 75f1aded-d1da-4ddc-be29-f1fb93799364/75f1aded-d1da-4ddc-be29-f1fb93799364.vmdk {{(pid=61806) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 907.764410] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-6d4e5237-b28c-45e9-8ae0-8ed79347c566 tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] [instance: 75f1aded-d1da-4ddc-be29-f1fb93799364] Extending root virtual disk to 1048576 {{(pid=61806) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 907.764694] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-b26fdf4b-8371-4c45-bc1c-ff945cc96542 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 907.771878] env[61806]: DEBUG oslo_vmware.api [None req-6d4e5237-b28c-45e9-8ae0-8ed79347c566 tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Waiting for the task: (returnval){ [ 907.771878] env[61806]: value = "task-1294877" [ 907.771878] env[61806]: _type = "Task" [ 907.771878] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 907.780720] env[61806]: DEBUG oslo_vmware.api [None req-6d4e5237-b28c-45e9-8ae0-8ed79347c566 tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Task: {'id': task-1294877, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 907.822323] env[61806]: DEBUG nova.network.neutron [-] [instance: 1a73d994-5f21-4f80-8f33-5830d860a7ee] Updating instance_info_cache with network_info: [] {{(pid=61806) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 907.854585] env[61806]: DEBUG nova.scheduler.client.report [None req-26d83e60-0148-4554-a6a0-baa143245ad5 tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] Inventory has not changed for provider a2858be1-fd22-4e08-979e-87ad25293407 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 138, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61806) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 907.946019] env[61806]: DEBUG nova.network.neutron [-] [instance: 59c22c16-3d1c-4e64-bfac-7f8886a1927e] Updating instance_info_cache with network_info: [] {{(pid=61806) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 907.947423] env[61806]: DEBUG oslo_vmware.api [None req-dab50ec4-3405-470b-a203-8c80641dde31 tempest-ServerRescueTestJSONUnderV235-749414849 tempest-ServerRescueTestJSONUnderV235-749414849-project-member] Task: {'id': task-1294876, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.425388} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 907.947614] env[61806]: DEBUG nova.virt.vmwareapi.ds_util [None req-dab50ec4-3405-470b-a203-8c80641dde31 tempest-ServerRescueTestJSONUnderV235-749414849 tempest-ServerRescueTestJSONUnderV235-749414849-project-member] Deleted the datastore file {{(pid=61806) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 907.947803] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-dab50ec4-3405-470b-a203-8c80641dde31 tempest-ServerRescueTestJSONUnderV235-749414849 tempest-ServerRescueTestJSONUnderV235-749414849-project-member] [instance: 5a526856-0dcd-4b3f-8359-13a5c3b9bfc4] Deleted contents of the VM from datastore datastore2 {{(pid=61806) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 907.947985] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-dab50ec4-3405-470b-a203-8c80641dde31 tempest-ServerRescueTestJSONUnderV235-749414849 tempest-ServerRescueTestJSONUnderV235-749414849-project-member] [instance: 5a526856-0dcd-4b3f-8359-13a5c3b9bfc4] Instance destroyed {{(pid=61806) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 907.948281] env[61806]: INFO nova.compute.manager [None req-dab50ec4-3405-470b-a203-8c80641dde31 tempest-ServerRescueTestJSONUnderV235-749414849 tempest-ServerRescueTestJSONUnderV235-749414849-project-member] [instance: 5a526856-0dcd-4b3f-8359-13a5c3b9bfc4] Took 1.18 seconds to destroy the instance on the hypervisor. [ 907.948562] env[61806]: DEBUG oslo.service.loopingcall [None req-dab50ec4-3405-470b-a203-8c80641dde31 tempest-ServerRescueTestJSONUnderV235-749414849 tempest-ServerRescueTestJSONUnderV235-749414849-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61806) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 907.949071] env[61806]: DEBUG nova.compute.manager [-] [instance: 5a526856-0dcd-4b3f-8359-13a5c3b9bfc4] Deallocating network for instance {{(pid=61806) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 907.949167] env[61806]: DEBUG nova.network.neutron [-] [instance: 5a526856-0dcd-4b3f-8359-13a5c3b9bfc4] deallocate_for_instance() {{(pid=61806) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 908.020107] env[61806]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-2fd63341-e442-4672-8fa6-be914b880520 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 908.031051] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-37876f44-9667-461c-a7f9-a30a10bb71d1 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 908.060169] env[61806]: INFO nova.compute.manager [-] [instance: ee002e7c-8965-494f-a1af-0b6f3b74eceb] Took 1.31 seconds to deallocate network for instance. [ 908.060586] env[61806]: DEBUG nova.compute.manager [req-5426e90f-2d62-45f9-b35f-e3ff57b0b44b req-c6919727-676b-4406-abd1-ee5e73170245 service nova] [instance: 59c22c16-3d1c-4e64-bfac-7f8886a1927e] Detach interface failed, port_id=a3901838-4afa-473f-a673-faac7dee3e75, reason: Instance 59c22c16-3d1c-4e64-bfac-7f8886a1927e could not be found. {{(pid=61806) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10963}} [ 908.104961] env[61806]: DEBUG nova.network.neutron [req-3a6967a4-fccd-4b9d-8438-18f0715a35fd req-2158630e-4917-4058-8b93-304ffe7fed2c service nova] [instance: 5a526856-0dcd-4b3f-8359-13a5c3b9bfc4] Updated VIF entry in instance network info cache for port 58782499-bc1a-4be1-9d48-655c6b2f11d7. {{(pid=61806) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 908.105328] env[61806]: DEBUG nova.network.neutron [req-3a6967a4-fccd-4b9d-8438-18f0715a35fd req-2158630e-4917-4058-8b93-304ffe7fed2c service nova] [instance: 5a526856-0dcd-4b3f-8359-13a5c3b9bfc4] Updating instance_info_cache with network_info: [{"id": "58782499-bc1a-4be1-9d48-655c6b2f11d7", "address": "fa:16:3e:52:7a:2b", "network": {"id": "b4fa0830-265e-47bf-8029-5944534312e7", "bridge": "br-int", "label": "tempest-ServerRescueTestJSONUnderV235-1460638664-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "2ba59612579b47a4bde7f82d951ac3ff", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8a34aa30-95be-4b18-98ca-1f2d81f7e9e6", "external-id": "nsx-vlan-transportzone-234", "segmentation_id": 234, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap58782499-bc", "ovs_interfaceid": "58782499-bc1a-4be1-9d48-655c6b2f11d7", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61806) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 908.282460] env[61806]: DEBUG oslo_vmware.api [None req-6d4e5237-b28c-45e9-8ae0-8ed79347c566 tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Task: {'id': task-1294877, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.06909} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 908.282699] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-6d4e5237-b28c-45e9-8ae0-8ed79347c566 tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] [instance: 75f1aded-d1da-4ddc-be29-f1fb93799364] Extended root virtual disk {{(pid=61806) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 908.283496] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5a0cf8d6-d7a3-44d4-b2e8-c2a4f8874a92 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 908.305122] env[61806]: DEBUG nova.virt.vmwareapi.volumeops [None req-6d4e5237-b28c-45e9-8ae0-8ed79347c566 tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] [instance: 75f1aded-d1da-4ddc-be29-f1fb93799364] Reconfiguring VM instance instance-00000052 to attach disk [datastore1] 75f1aded-d1da-4ddc-be29-f1fb93799364/75f1aded-d1da-4ddc-be29-f1fb93799364.vmdk or device None with type sparse {{(pid=61806) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 908.305372] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-c18c1e17-60e3-47bb-87f6-b8f44bda431d {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 908.324787] env[61806]: INFO nova.compute.manager [-] [instance: 1a73d994-5f21-4f80-8f33-5830d860a7ee] Took 1.28 seconds to deallocate network for instance. [ 908.325124] env[61806]: DEBUG oslo_vmware.api [None req-6d4e5237-b28c-45e9-8ae0-8ed79347c566 tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Waiting for the task: (returnval){ [ 908.325124] env[61806]: value = "task-1294878" [ 908.325124] env[61806]: _type = "Task" [ 908.325124] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 908.337972] env[61806]: DEBUG oslo_vmware.api [None req-6d4e5237-b28c-45e9-8ae0-8ed79347c566 tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Task: {'id': task-1294878, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 908.360129] env[61806]: DEBUG oslo_concurrency.lockutils [None req-26d83e60-0148-4554-a6a0-baa143245ad5 tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.280s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 908.360665] env[61806]: DEBUG nova.compute.manager [None req-26d83e60-0148-4554-a6a0-baa143245ad5 tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] [instance: d9caee4a-a00c-4e50-b01c-99a8f796a09f] Start building networks asynchronously for instance. {{(pid=61806) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 908.450167] env[61806]: INFO nova.compute.manager [-] [instance: 59c22c16-3d1c-4e64-bfac-7f8886a1927e] Took 1.26 seconds to deallocate network for instance. [ 908.569484] env[61806]: DEBUG oslo_concurrency.lockutils [None req-e34fcc99-6699-4d73-ba15-e4ab836a8262 tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 908.569815] env[61806]: DEBUG oslo_concurrency.lockutils [None req-e34fcc99-6699-4d73-ba15-e4ab836a8262 tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.001s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 908.570130] env[61806]: DEBUG nova.objects.instance [None req-e34fcc99-6699-4d73-ba15-e4ab836a8262 tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] Lazy-loading 'resources' on Instance uuid ee002e7c-8965-494f-a1af-0b6f3b74eceb {{(pid=61806) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 908.608019] env[61806]: DEBUG oslo_concurrency.lockutils [req-3a6967a4-fccd-4b9d-8438-18f0715a35fd req-2158630e-4917-4058-8b93-304ffe7fed2c service nova] Releasing lock "refresh_cache-5a526856-0dcd-4b3f-8359-13a5c3b9bfc4" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 908.608335] env[61806]: DEBUG nova.compute.manager [req-3a6967a4-fccd-4b9d-8438-18f0715a35fd req-2158630e-4917-4058-8b93-304ffe7fed2c service nova] [instance: ee002e7c-8965-494f-a1af-0b6f3b74eceb] Received event network-vif-deleted-5c3f2277-5769-4c2e-8eb9-df0a99f35008 {{(pid=61806) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 908.608550] env[61806]: INFO nova.compute.manager [req-3a6967a4-fccd-4b9d-8438-18f0715a35fd req-2158630e-4917-4058-8b93-304ffe7fed2c service nova] [instance: ee002e7c-8965-494f-a1af-0b6f3b74eceb] Neutron deleted interface 5c3f2277-5769-4c2e-8eb9-df0a99f35008; detaching it from the instance and deleting it from the info cache [ 908.608728] env[61806]: DEBUG nova.network.neutron [req-3a6967a4-fccd-4b9d-8438-18f0715a35fd req-2158630e-4917-4058-8b93-304ffe7fed2c service nova] [instance: ee002e7c-8965-494f-a1af-0b6f3b74eceb] Updating instance_info_cache with network_info: [] {{(pid=61806) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 908.656062] env[61806]: DEBUG nova.network.neutron [-] [instance: 5a526856-0dcd-4b3f-8359-13a5c3b9bfc4] Updating instance_info_cache with network_info: [] {{(pid=61806) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 908.834130] env[61806]: DEBUG oslo_concurrency.lockutils [None req-9c3f23d6-436a-448c-a3fa-4250c69b6723 tempest-ImagesTestJSON-618852556 tempest-ImagesTestJSON-618852556-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 908.837857] env[61806]: DEBUG oslo_vmware.api [None req-6d4e5237-b28c-45e9-8ae0-8ed79347c566 tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Task: {'id': task-1294878, 'name': ReconfigVM_Task, 'duration_secs': 0.337234} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 908.838138] env[61806]: DEBUG nova.virt.vmwareapi.volumeops [None req-6d4e5237-b28c-45e9-8ae0-8ed79347c566 tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] [instance: 75f1aded-d1da-4ddc-be29-f1fb93799364] Reconfigured VM instance instance-00000052 to attach disk [datastore1] 75f1aded-d1da-4ddc-be29-f1fb93799364/75f1aded-d1da-4ddc-be29-f1fb93799364.vmdk or device None with type sparse {{(pid=61806) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 908.838746] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-57d59539-0fa6-4701-95d6-1cc4e0d5e1f6 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 908.845616] env[61806]: DEBUG oslo_vmware.api [None req-6d4e5237-b28c-45e9-8ae0-8ed79347c566 tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Waiting for the task: (returnval){ [ 908.845616] env[61806]: value = "task-1294879" [ 908.845616] env[61806]: _type = "Task" [ 908.845616] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 908.853476] env[61806]: DEBUG oslo_vmware.api [None req-6d4e5237-b28c-45e9-8ae0-8ed79347c566 tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Task: {'id': task-1294879, 'name': Rename_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 908.866069] env[61806]: DEBUG nova.compute.utils [None req-26d83e60-0148-4554-a6a0-baa143245ad5 tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] Using /dev/sd instead of None {{(pid=61806) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 908.868125] env[61806]: DEBUG nova.compute.manager [None req-26d83e60-0148-4554-a6a0-baa143245ad5 tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] [instance: d9caee4a-a00c-4e50-b01c-99a8f796a09f] Allocating IP information in the background. {{(pid=61806) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 908.868320] env[61806]: DEBUG nova.network.neutron [None req-26d83e60-0148-4554-a6a0-baa143245ad5 tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] [instance: d9caee4a-a00c-4e50-b01c-99a8f796a09f] allocate_for_instance() {{(pid=61806) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 908.915685] env[61806]: DEBUG nova.policy [None req-26d83e60-0148-4554-a6a0-baa143245ad5 tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'b1eac9072ccb4b3bbd23f240e941a76e', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '6e391e1267014fafbb8f5a3211299819', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61806) authorize /opt/stack/nova/nova/policy.py:201}} [ 909.000416] env[61806]: INFO nova.compute.manager [None req-1d9f7d6e-8006-4355-8395-1f9c0f916c73 tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] [instance: 59c22c16-3d1c-4e64-bfac-7f8886a1927e] Took 0.55 seconds to detach 1 volumes for instance. [ 909.111092] env[61806]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-09145b6e-2735-4290-9f53-973d7a6e6a5e {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 909.122034] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5c08817a-07ab-4a05-83d0-cebbca40a027 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 909.157317] env[61806]: DEBUG nova.compute.manager [req-3a6967a4-fccd-4b9d-8438-18f0715a35fd req-2158630e-4917-4058-8b93-304ffe7fed2c service nova] [instance: ee002e7c-8965-494f-a1af-0b6f3b74eceb] Detach interface failed, port_id=5c3f2277-5769-4c2e-8eb9-df0a99f35008, reason: Instance ee002e7c-8965-494f-a1af-0b6f3b74eceb could not be found. {{(pid=61806) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10963}} [ 909.157588] env[61806]: DEBUG nova.compute.manager [req-3a6967a4-fccd-4b9d-8438-18f0715a35fd req-2158630e-4917-4058-8b93-304ffe7fed2c service nova] [instance: 1a73d994-5f21-4f80-8f33-5830d860a7ee] Received event network-vif-deleted-4017a509-57e9-4526-a107-3d54b7ed6684 {{(pid=61806) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 909.160392] env[61806]: INFO nova.compute.manager [req-3a6967a4-fccd-4b9d-8438-18f0715a35fd req-2158630e-4917-4058-8b93-304ffe7fed2c service nova] [instance: 1a73d994-5f21-4f80-8f33-5830d860a7ee] Neutron deleted interface 4017a509-57e9-4526-a107-3d54b7ed6684; detaching it from the instance and deleting it from the info cache [ 909.160392] env[61806]: DEBUG nova.network.neutron [req-3a6967a4-fccd-4b9d-8438-18f0715a35fd req-2158630e-4917-4058-8b93-304ffe7fed2c service nova] [instance: 1a73d994-5f21-4f80-8f33-5830d860a7ee] Updating instance_info_cache with network_info: [] {{(pid=61806) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 909.160392] env[61806]: INFO nova.compute.manager [-] [instance: 5a526856-0dcd-4b3f-8359-13a5c3b9bfc4] Took 1.21 seconds to deallocate network for instance. [ 909.187884] env[61806]: DEBUG nova.network.neutron [None req-26d83e60-0148-4554-a6a0-baa143245ad5 tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] [instance: d9caee4a-a00c-4e50-b01c-99a8f796a09f] Successfully created port: c725ee28-d7a8-4b31-92bc-f63174564f5e {{(pid=61806) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 909.252309] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e418698e-ecf7-4fb3-8b4b-cffd1dbb0b33 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 909.262664] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f83e1bff-7855-4ebb-9d63-09f9064470fa {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 909.294319] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-58009d53-a201-4f35-aaff-ec789bba0a36 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 909.302876] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bcad5b1d-a0e1-402c-b49d-a9d55c80f570 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 909.317213] env[61806]: DEBUG nova.compute.provider_tree [None req-e34fcc99-6699-4d73-ba15-e4ab836a8262 tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] Inventory has not changed in ProviderTree for provider: a2858be1-fd22-4e08-979e-87ad25293407 {{(pid=61806) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 909.356087] env[61806]: DEBUG oslo_vmware.api [None req-6d4e5237-b28c-45e9-8ae0-8ed79347c566 tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Task: {'id': task-1294879, 'name': Rename_Task, 'duration_secs': 0.161293} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 909.356464] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [None req-6d4e5237-b28c-45e9-8ae0-8ed79347c566 tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] [instance: 75f1aded-d1da-4ddc-be29-f1fb93799364] Powering on the VM {{(pid=61806) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 909.356687] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-542d3143-84b9-4c01-bf59-8b49e517f104 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 909.363478] env[61806]: DEBUG oslo_vmware.api [None req-6d4e5237-b28c-45e9-8ae0-8ed79347c566 tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Waiting for the task: (returnval){ [ 909.363478] env[61806]: value = "task-1294880" [ 909.363478] env[61806]: _type = "Task" [ 909.363478] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 909.371499] env[61806]: DEBUG nova.compute.manager [None req-26d83e60-0148-4554-a6a0-baa143245ad5 tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] [instance: d9caee4a-a00c-4e50-b01c-99a8f796a09f] Start building block device mappings for instance. {{(pid=61806) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 909.374072] env[61806]: DEBUG oslo_vmware.api [None req-6d4e5237-b28c-45e9-8ae0-8ed79347c566 tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Task: {'id': task-1294880, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 909.420872] env[61806]: DEBUG nova.compute.manager [req-a9f5f1ae-35cc-4809-9531-85c86ec31138 req-13f66889-2e94-4eb0-bd7f-f5873f02acb8 service nova] [instance: 5a526856-0dcd-4b3f-8359-13a5c3b9bfc4] Received event network-vif-deleted-58782499-bc1a-4be1-9d48-655c6b2f11d7 {{(pid=61806) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 909.507993] env[61806]: DEBUG oslo_concurrency.lockutils [None req-1d9f7d6e-8006-4355-8395-1f9c0f916c73 tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 909.661238] env[61806]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-19440cbe-4518-4037-ae93-d745cb647aec {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 909.670854] env[61806]: DEBUG oslo_concurrency.lockutils [None req-dab50ec4-3405-470b-a203-8c80641dde31 tempest-ServerRescueTestJSONUnderV235-749414849 tempest-ServerRescueTestJSONUnderV235-749414849-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 909.674392] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4224a810-9726-48a6-928e-7a07b9a4e56e {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 909.704897] env[61806]: DEBUG nova.compute.manager [req-3a6967a4-fccd-4b9d-8438-18f0715a35fd req-2158630e-4917-4058-8b93-304ffe7fed2c service nova] [instance: 1a73d994-5f21-4f80-8f33-5830d860a7ee] Detach interface failed, port_id=4017a509-57e9-4526-a107-3d54b7ed6684, reason: Instance 1a73d994-5f21-4f80-8f33-5830d860a7ee could not be found. {{(pid=61806) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10963}} [ 909.821157] env[61806]: DEBUG nova.scheduler.client.report [None req-e34fcc99-6699-4d73-ba15-e4ab836a8262 tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] Inventory has not changed for provider a2858be1-fd22-4e08-979e-87ad25293407 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 138, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61806) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 909.879646] env[61806]: DEBUG oslo_vmware.api [None req-6d4e5237-b28c-45e9-8ae0-8ed79347c566 tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Task: {'id': task-1294880, 'name': PowerOnVM_Task} progress is 89%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 910.325737] env[61806]: DEBUG oslo_concurrency.lockutils [None req-e34fcc99-6699-4d73-ba15-e4ab836a8262 tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.756s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 910.328157] env[61806]: DEBUG oslo_concurrency.lockutils [None req-9c3f23d6-436a-448c-a3fa-4250c69b6723 tempest-ImagesTestJSON-618852556 tempest-ImagesTestJSON-618852556-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 1.494s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 910.328406] env[61806]: DEBUG nova.objects.instance [None req-9c3f23d6-436a-448c-a3fa-4250c69b6723 tempest-ImagesTestJSON-618852556 tempest-ImagesTestJSON-618852556-project-member] Lazy-loading 'resources' on Instance uuid 1a73d994-5f21-4f80-8f33-5830d860a7ee {{(pid=61806) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 910.353582] env[61806]: INFO nova.scheduler.client.report [None req-e34fcc99-6699-4d73-ba15-e4ab836a8262 tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] Deleted allocations for instance ee002e7c-8965-494f-a1af-0b6f3b74eceb [ 910.375337] env[61806]: DEBUG oslo_vmware.api [None req-6d4e5237-b28c-45e9-8ae0-8ed79347c566 tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Task: {'id': task-1294880, 'name': PowerOnVM_Task, 'duration_secs': 0.585172} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 910.375602] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [None req-6d4e5237-b28c-45e9-8ae0-8ed79347c566 tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] [instance: 75f1aded-d1da-4ddc-be29-f1fb93799364] Powered on the VM {{(pid=61806) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 910.375827] env[61806]: INFO nova.compute.manager [None req-6d4e5237-b28c-45e9-8ae0-8ed79347c566 tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] [instance: 75f1aded-d1da-4ddc-be29-f1fb93799364] Took 12.22 seconds to spawn the instance on the hypervisor. [ 910.376019] env[61806]: DEBUG nova.compute.manager [None req-6d4e5237-b28c-45e9-8ae0-8ed79347c566 tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] [instance: 75f1aded-d1da-4ddc-be29-f1fb93799364] Checking state {{(pid=61806) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 910.376808] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dd203a54-c437-40ae-aa65-ca3c8a3d7bb0 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 910.381999] env[61806]: DEBUG nova.compute.manager [None req-26d83e60-0148-4554-a6a0-baa143245ad5 tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] [instance: d9caee4a-a00c-4e50-b01c-99a8f796a09f] Start spawning the instance on the hypervisor. {{(pid=61806) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 910.409534] env[61806]: DEBUG nova.virt.hardware [None req-26d83e60-0148-4554-a6a0-baa143245ad5 tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-15T18:05:10Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-15T18:04:52Z,direct_url=,disk_format='vmdk',id=ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='84ba83fce2c349a6988173c5d6fc3b07',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-15T18:04:53Z,virtual_size=,visibility=), allow threads: False {{(pid=61806) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 910.409871] env[61806]: DEBUG nova.virt.hardware [None req-26d83e60-0148-4554-a6a0-baa143245ad5 tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] Flavor limits 0:0:0 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 910.410080] env[61806]: DEBUG nova.virt.hardware [None req-26d83e60-0148-4554-a6a0-baa143245ad5 tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] Image limits 0:0:0 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 910.410310] env[61806]: DEBUG nova.virt.hardware [None req-26d83e60-0148-4554-a6a0-baa143245ad5 tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] Flavor pref 0:0:0 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 910.410543] env[61806]: DEBUG nova.virt.hardware [None req-26d83e60-0148-4554-a6a0-baa143245ad5 tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] Image pref 0:0:0 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 910.410710] env[61806]: DEBUG nova.virt.hardware [None req-26d83e60-0148-4554-a6a0-baa143245ad5 tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 910.410967] env[61806]: DEBUG nova.virt.hardware [None req-26d83e60-0148-4554-a6a0-baa143245ad5 tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61806) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 910.411055] env[61806]: DEBUG nova.virt.hardware [None req-26d83e60-0148-4554-a6a0-baa143245ad5 tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61806) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 910.411273] env[61806]: DEBUG nova.virt.hardware [None req-26d83e60-0148-4554-a6a0-baa143245ad5 tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] Got 1 possible topologies {{(pid=61806) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 910.411579] env[61806]: DEBUG nova.virt.hardware [None req-26d83e60-0148-4554-a6a0-baa143245ad5 tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61806) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 910.411693] env[61806]: DEBUG nova.virt.hardware [None req-26d83e60-0148-4554-a6a0-baa143245ad5 tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61806) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 910.412916] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1c0af7fd-2c6c-4b90-8d0c-f741ae45ca67 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 910.424591] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-99a3cb6c-2936-4aae-a224-d9f18ec5832e {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 910.709545] env[61806]: DEBUG nova.network.neutron [None req-26d83e60-0148-4554-a6a0-baa143245ad5 tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] [instance: d9caee4a-a00c-4e50-b01c-99a8f796a09f] Successfully updated port: c725ee28-d7a8-4b31-92bc-f63174564f5e {{(pid=61806) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 910.862628] env[61806]: DEBUG oslo_concurrency.lockutils [None req-e34fcc99-6699-4d73-ba15-e4ab836a8262 tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] Lock "ee002e7c-8965-494f-a1af-0b6f3b74eceb" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.259s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 910.899866] env[61806]: INFO nova.compute.manager [None req-6d4e5237-b28c-45e9-8ae0-8ed79347c566 tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] [instance: 75f1aded-d1da-4ddc-be29-f1fb93799364] Took 21.10 seconds to build instance. [ 911.005789] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a1661756-5164-4bd4-bfac-a428a98d7317 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 911.015021] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1a3332f0-254c-4be4-b436-236979befed1 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 911.048955] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8011c189-da5a-47f8-b1ab-14fac769ba53 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 911.058904] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b900c537-3cfa-4f02-b56a-c398e6c44e4c {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 911.074657] env[61806]: DEBUG nova.compute.provider_tree [None req-9c3f23d6-436a-448c-a3fa-4250c69b6723 tempest-ImagesTestJSON-618852556 tempest-ImagesTestJSON-618852556-project-member] Inventory has not changed in ProviderTree for provider: a2858be1-fd22-4e08-979e-87ad25293407 {{(pid=61806) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 911.213974] env[61806]: DEBUG oslo_concurrency.lockutils [None req-26d83e60-0148-4554-a6a0-baa143245ad5 tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] Acquiring lock "refresh_cache-d9caee4a-a00c-4e50-b01c-99a8f796a09f" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 911.213974] env[61806]: DEBUG oslo_concurrency.lockutils [None req-26d83e60-0148-4554-a6a0-baa143245ad5 tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] Acquired lock "refresh_cache-d9caee4a-a00c-4e50-b01c-99a8f796a09f" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 911.213974] env[61806]: DEBUG nova.network.neutron [None req-26d83e60-0148-4554-a6a0-baa143245ad5 tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] [instance: d9caee4a-a00c-4e50-b01c-99a8f796a09f] Building network info cache for instance {{(pid=61806) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 911.401640] env[61806]: DEBUG oslo_concurrency.lockutils [None req-6d4e5237-b28c-45e9-8ae0-8ed79347c566 tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Lock "75f1aded-d1da-4ddc-be29-f1fb93799364" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 22.618s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 911.456510] env[61806]: DEBUG nova.compute.manager [req-612c591d-a1a4-4beb-b084-702dade7ffe9 req-0f1b29be-376d-4ed6-b9e9-80a381fcc943 service nova] [instance: d9caee4a-a00c-4e50-b01c-99a8f796a09f] Received event network-vif-plugged-c725ee28-d7a8-4b31-92bc-f63174564f5e {{(pid=61806) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 911.456723] env[61806]: DEBUG oslo_concurrency.lockutils [req-612c591d-a1a4-4beb-b084-702dade7ffe9 req-0f1b29be-376d-4ed6-b9e9-80a381fcc943 service nova] Acquiring lock "d9caee4a-a00c-4e50-b01c-99a8f796a09f-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 911.456867] env[61806]: DEBUG oslo_concurrency.lockutils [req-612c591d-a1a4-4beb-b084-702dade7ffe9 req-0f1b29be-376d-4ed6-b9e9-80a381fcc943 service nova] Lock "d9caee4a-a00c-4e50-b01c-99a8f796a09f-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 911.457061] env[61806]: DEBUG oslo_concurrency.lockutils [req-612c591d-a1a4-4beb-b084-702dade7ffe9 req-0f1b29be-376d-4ed6-b9e9-80a381fcc943 service nova] Lock "d9caee4a-a00c-4e50-b01c-99a8f796a09f-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 911.457244] env[61806]: DEBUG nova.compute.manager [req-612c591d-a1a4-4beb-b084-702dade7ffe9 req-0f1b29be-376d-4ed6-b9e9-80a381fcc943 service nova] [instance: d9caee4a-a00c-4e50-b01c-99a8f796a09f] No waiting events found dispatching network-vif-plugged-c725ee28-d7a8-4b31-92bc-f63174564f5e {{(pid=61806) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 911.457420] env[61806]: WARNING nova.compute.manager [req-612c591d-a1a4-4beb-b084-702dade7ffe9 req-0f1b29be-376d-4ed6-b9e9-80a381fcc943 service nova] [instance: d9caee4a-a00c-4e50-b01c-99a8f796a09f] Received unexpected event network-vif-plugged-c725ee28-d7a8-4b31-92bc-f63174564f5e for instance with vm_state building and task_state spawning. [ 911.457595] env[61806]: DEBUG nova.compute.manager [req-612c591d-a1a4-4beb-b084-702dade7ffe9 req-0f1b29be-376d-4ed6-b9e9-80a381fcc943 service nova] [instance: d9caee4a-a00c-4e50-b01c-99a8f796a09f] Received event network-changed-c725ee28-d7a8-4b31-92bc-f63174564f5e {{(pid=61806) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 911.457803] env[61806]: DEBUG nova.compute.manager [req-612c591d-a1a4-4beb-b084-702dade7ffe9 req-0f1b29be-376d-4ed6-b9e9-80a381fcc943 service nova] [instance: d9caee4a-a00c-4e50-b01c-99a8f796a09f] Refreshing instance network info cache due to event network-changed-c725ee28-d7a8-4b31-92bc-f63174564f5e. {{(pid=61806) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 911.457983] env[61806]: DEBUG oslo_concurrency.lockutils [req-612c591d-a1a4-4beb-b084-702dade7ffe9 req-0f1b29be-376d-4ed6-b9e9-80a381fcc943 service nova] Acquiring lock "refresh_cache-d9caee4a-a00c-4e50-b01c-99a8f796a09f" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 911.578585] env[61806]: DEBUG nova.scheduler.client.report [None req-9c3f23d6-436a-448c-a3fa-4250c69b6723 tempest-ImagesTestJSON-618852556 tempest-ImagesTestJSON-618852556-project-member] Inventory has not changed for provider a2858be1-fd22-4e08-979e-87ad25293407 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 138, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61806) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 911.754485] env[61806]: DEBUG nova.network.neutron [None req-26d83e60-0148-4554-a6a0-baa143245ad5 tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] [instance: d9caee4a-a00c-4e50-b01c-99a8f796a09f] Instance cache missing network info. {{(pid=61806) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 911.792169] env[61806]: DEBUG oslo_concurrency.lockutils [None req-531998d5-4ee2-42e3-ac57-d5e3f0c96508 tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] Acquiring lock "c486c00d-61fc-4795-9299-e6548b686ee8" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 911.792418] env[61806]: DEBUG oslo_concurrency.lockutils [None req-531998d5-4ee2-42e3-ac57-d5e3f0c96508 tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] Lock "c486c00d-61fc-4795-9299-e6548b686ee8" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 911.905903] env[61806]: DEBUG nova.network.neutron [None req-26d83e60-0148-4554-a6a0-baa143245ad5 tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] [instance: d9caee4a-a00c-4e50-b01c-99a8f796a09f] Updating instance_info_cache with network_info: [{"id": "c725ee28-d7a8-4b31-92bc-f63174564f5e", "address": "fa:16:3e:3a:18:8e", "network": {"id": "a7ffda72-e57c-41fa-a7ed-d7b12084bc9d", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-62794865-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "6e391e1267014fafbb8f5a3211299819", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8b3cfeb1-f262-4fd9-b506-8e9c0733e2d8", "external-id": "nsx-vlan-transportzone-119", "segmentation_id": 119, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc725ee28-d7", "ovs_interfaceid": "c725ee28-d7a8-4b31-92bc-f63174564f5e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61806) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 912.083791] env[61806]: DEBUG oslo_concurrency.lockutils [None req-9c3f23d6-436a-448c-a3fa-4250c69b6723 tempest-ImagesTestJSON-618852556 tempest-ImagesTestJSON-618852556-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.755s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 912.087081] env[61806]: DEBUG oslo_concurrency.lockutils [None req-1d9f7d6e-8006-4355-8395-1f9c0f916c73 tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 2.578s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 912.087422] env[61806]: DEBUG nova.objects.instance [None req-1d9f7d6e-8006-4355-8395-1f9c0f916c73 tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Lazy-loading 'resources' on Instance uuid 59c22c16-3d1c-4e64-bfac-7f8886a1927e {{(pid=61806) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 912.105659] env[61806]: INFO nova.scheduler.client.report [None req-9c3f23d6-436a-448c-a3fa-4250c69b6723 tempest-ImagesTestJSON-618852556 tempest-ImagesTestJSON-618852556-project-member] Deleted allocations for instance 1a73d994-5f21-4f80-8f33-5830d860a7ee [ 912.106791] env[61806]: INFO nova.compute.manager [None req-70a8784a-38a8-4a3e-8a9f-16278c289170 tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] [instance: 75f1aded-d1da-4ddc-be29-f1fb93799364] Rebuilding instance [ 912.144257] env[61806]: DEBUG nova.compute.manager [None req-70a8784a-38a8-4a3e-8a9f-16278c289170 tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] [instance: 75f1aded-d1da-4ddc-be29-f1fb93799364] Checking state {{(pid=61806) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 912.145170] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1dbff7f5-0783-427d-916e-0a204cce7feb {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 912.295362] env[61806]: DEBUG nova.compute.manager [None req-531998d5-4ee2-42e3-ac57-d5e3f0c96508 tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] [instance: c486c00d-61fc-4795-9299-e6548b686ee8] Starting instance... {{(pid=61806) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 912.408777] env[61806]: DEBUG oslo_concurrency.lockutils [None req-26d83e60-0148-4554-a6a0-baa143245ad5 tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] Releasing lock "refresh_cache-d9caee4a-a00c-4e50-b01c-99a8f796a09f" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 912.408990] env[61806]: DEBUG nova.compute.manager [None req-26d83e60-0148-4554-a6a0-baa143245ad5 tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] [instance: d9caee4a-a00c-4e50-b01c-99a8f796a09f] Instance network_info: |[{"id": "c725ee28-d7a8-4b31-92bc-f63174564f5e", "address": "fa:16:3e:3a:18:8e", "network": {"id": "a7ffda72-e57c-41fa-a7ed-d7b12084bc9d", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-62794865-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "6e391e1267014fafbb8f5a3211299819", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8b3cfeb1-f262-4fd9-b506-8e9c0733e2d8", "external-id": "nsx-vlan-transportzone-119", "segmentation_id": 119, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc725ee28-d7", "ovs_interfaceid": "c725ee28-d7a8-4b31-92bc-f63174564f5e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61806) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 912.409392] env[61806]: DEBUG oslo_concurrency.lockutils [req-612c591d-a1a4-4beb-b084-702dade7ffe9 req-0f1b29be-376d-4ed6-b9e9-80a381fcc943 service nova] Acquired lock "refresh_cache-d9caee4a-a00c-4e50-b01c-99a8f796a09f" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 912.409687] env[61806]: DEBUG nova.network.neutron [req-612c591d-a1a4-4beb-b084-702dade7ffe9 req-0f1b29be-376d-4ed6-b9e9-80a381fcc943 service nova] [instance: d9caee4a-a00c-4e50-b01c-99a8f796a09f] Refreshing network info cache for port c725ee28-d7a8-4b31-92bc-f63174564f5e {{(pid=61806) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 912.410906] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-26d83e60-0148-4554-a6a0-baa143245ad5 tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] [instance: d9caee4a-a00c-4e50-b01c-99a8f796a09f] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:3a:18:8e', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '8b3cfeb1-f262-4fd9-b506-8e9c0733e2d8', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'c725ee28-d7a8-4b31-92bc-f63174564f5e', 'vif_model': 'vmxnet3'}] {{(pid=61806) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 912.419020] env[61806]: DEBUG oslo.service.loopingcall [None req-26d83e60-0148-4554-a6a0-baa143245ad5 tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61806) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 912.422735] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: d9caee4a-a00c-4e50-b01c-99a8f796a09f] Creating VM on the ESX host {{(pid=61806) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 912.423287] env[61806]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-ac604467-9854-40db-b7f9-be3c24c31c97 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 912.445433] env[61806]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 912.445433] env[61806]: value = "task-1294881" [ 912.445433] env[61806]: _type = "Task" [ 912.445433] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 912.455024] env[61806]: DEBUG oslo_vmware.api [-] Task: {'id': task-1294881, 'name': CreateVM_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 912.620086] env[61806]: DEBUG oslo_concurrency.lockutils [None req-9c3f23d6-436a-448c-a3fa-4250c69b6723 tempest-ImagesTestJSON-618852556 tempest-ImagesTestJSON-618852556-project-member] Lock "1a73d994-5f21-4f80-8f33-5830d860a7ee" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 6.716s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 912.657983] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [None req-70a8784a-38a8-4a3e-8a9f-16278c289170 tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] [instance: 75f1aded-d1da-4ddc-be29-f1fb93799364] Powering off the VM {{(pid=61806) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 912.658572] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-6400eaeb-ed82-4903-9ced-e7d6f6f5428f {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 912.667658] env[61806]: DEBUG oslo_vmware.api [None req-70a8784a-38a8-4a3e-8a9f-16278c289170 tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Waiting for the task: (returnval){ [ 912.667658] env[61806]: value = "task-1294882" [ 912.667658] env[61806]: _type = "Task" [ 912.667658] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 912.671831] env[61806]: DEBUG nova.network.neutron [req-612c591d-a1a4-4beb-b084-702dade7ffe9 req-0f1b29be-376d-4ed6-b9e9-80a381fcc943 service nova] [instance: d9caee4a-a00c-4e50-b01c-99a8f796a09f] Updated VIF entry in instance network info cache for port c725ee28-d7a8-4b31-92bc-f63174564f5e. {{(pid=61806) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 912.672224] env[61806]: DEBUG nova.network.neutron [req-612c591d-a1a4-4beb-b084-702dade7ffe9 req-0f1b29be-376d-4ed6-b9e9-80a381fcc943 service nova] [instance: d9caee4a-a00c-4e50-b01c-99a8f796a09f] Updating instance_info_cache with network_info: [{"id": "c725ee28-d7a8-4b31-92bc-f63174564f5e", "address": "fa:16:3e:3a:18:8e", "network": {"id": "a7ffda72-e57c-41fa-a7ed-d7b12084bc9d", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-62794865-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "6e391e1267014fafbb8f5a3211299819", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8b3cfeb1-f262-4fd9-b506-8e9c0733e2d8", "external-id": "nsx-vlan-transportzone-119", "segmentation_id": 119, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc725ee28-d7", "ovs_interfaceid": "c725ee28-d7a8-4b31-92bc-f63174564f5e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61806) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 912.684160] env[61806]: DEBUG oslo_vmware.api [None req-70a8784a-38a8-4a3e-8a9f-16278c289170 tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Task: {'id': task-1294882, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 912.759526] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-83cf9475-cd54-476d-944d-ad245d419c23 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 912.767430] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f09c0eaa-2450-4607-9d63-c952317242b3 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 912.803598] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a273a01a-4462-4767-857b-70f675238804 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 912.814830] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-81c2f5ee-7518-4853-8a96-8d566dd6b6fb {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 912.824542] env[61806]: DEBUG oslo_concurrency.lockutils [None req-531998d5-4ee2-42e3-ac57-d5e3f0c96508 tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 912.837546] env[61806]: DEBUG nova.compute.provider_tree [None req-1d9f7d6e-8006-4355-8395-1f9c0f916c73 tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Inventory has not changed in ProviderTree for provider: a2858be1-fd22-4e08-979e-87ad25293407 {{(pid=61806) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 912.956780] env[61806]: DEBUG oslo_vmware.api [-] Task: {'id': task-1294881, 'name': CreateVM_Task, 'duration_secs': 0.349006} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 912.956958] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: d9caee4a-a00c-4e50-b01c-99a8f796a09f] Created VM on the ESX host {{(pid=61806) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 912.957738] env[61806]: DEBUG oslo_concurrency.lockutils [None req-26d83e60-0148-4554-a6a0-baa143245ad5 tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 912.957920] env[61806]: DEBUG oslo_concurrency.lockutils [None req-26d83e60-0148-4554-a6a0-baa143245ad5 tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] Acquired lock "[datastore1] devstack-image-cache_base/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 912.958309] env[61806]: DEBUG oslo_concurrency.lockutils [None req-26d83e60-0148-4554-a6a0-baa143245ad5 tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 912.958559] env[61806]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-bf22fb23-23a3-4658-a98d-62712d154ec4 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 912.964158] env[61806]: DEBUG oslo_vmware.api [None req-26d83e60-0148-4554-a6a0-baa143245ad5 tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] Waiting for the task: (returnval){ [ 912.964158] env[61806]: value = "session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]526574a4-3e97-b5c9-f5a4-869a60d568bc" [ 912.964158] env[61806]: _type = "Task" [ 912.964158] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 912.973208] env[61806]: DEBUG oslo_vmware.api [None req-26d83e60-0148-4554-a6a0-baa143245ad5 tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] Task: {'id': session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]526574a4-3e97-b5c9-f5a4-869a60d568bc, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 913.181760] env[61806]: DEBUG oslo_concurrency.lockutils [req-612c591d-a1a4-4beb-b084-702dade7ffe9 req-0f1b29be-376d-4ed6-b9e9-80a381fcc943 service nova] Releasing lock "refresh_cache-d9caee4a-a00c-4e50-b01c-99a8f796a09f" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 913.182210] env[61806]: DEBUG oslo_vmware.api [None req-70a8784a-38a8-4a3e-8a9f-16278c289170 tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Task: {'id': task-1294882, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 913.342023] env[61806]: DEBUG nova.scheduler.client.report [None req-1d9f7d6e-8006-4355-8395-1f9c0f916c73 tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Inventory has not changed for provider a2858be1-fd22-4e08-979e-87ad25293407 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 138, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61806) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 913.475769] env[61806]: DEBUG oslo_vmware.api [None req-26d83e60-0148-4554-a6a0-baa143245ad5 tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] Task: {'id': session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]526574a4-3e97-b5c9-f5a4-869a60d568bc, 'name': SearchDatastore_Task, 'duration_secs': 0.011521} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 913.476597] env[61806]: DEBUG oslo_concurrency.lockutils [None req-26d83e60-0148-4554-a6a0-baa143245ad5 tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] Releasing lock "[datastore1] devstack-image-cache_base/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 913.477015] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-26d83e60-0148-4554-a6a0-baa143245ad5 tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] [instance: d9caee4a-a00c-4e50-b01c-99a8f796a09f] Processing image ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa {{(pid=61806) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 913.477155] env[61806]: DEBUG oslo_concurrency.lockutils [None req-26d83e60-0148-4554-a6a0-baa143245ad5 tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa.vmdk" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 913.477330] env[61806]: DEBUG oslo_concurrency.lockutils [None req-26d83e60-0148-4554-a6a0-baa143245ad5 tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] Acquired lock "[datastore1] devstack-image-cache_base/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa.vmdk" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 913.477537] env[61806]: DEBUG nova.virt.vmwareapi.ds_util [None req-26d83e60-0148-4554-a6a0-baa143245ad5 tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61806) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 913.477796] env[61806]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-d14ec599-cbe8-4de9-9335-5ed145b38f60 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 913.499327] env[61806]: DEBUG nova.virt.vmwareapi.ds_util [None req-26d83e60-0148-4554-a6a0-baa143245ad5 tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61806) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 913.499579] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-26d83e60-0148-4554-a6a0-baa143245ad5 tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61806) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 913.500286] env[61806]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-da58fe3c-df22-4d14-bc21-95aec202e0df {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 913.506236] env[61806]: DEBUG oslo_vmware.api [None req-26d83e60-0148-4554-a6a0-baa143245ad5 tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] Waiting for the task: (returnval){ [ 913.506236] env[61806]: value = "session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]529a157d-ba46-cc9f-2ec0-2eae5381ee92" [ 913.506236] env[61806]: _type = "Task" [ 913.506236] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 913.514108] env[61806]: DEBUG oslo_vmware.api [None req-26d83e60-0148-4554-a6a0-baa143245ad5 tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] Task: {'id': session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]529a157d-ba46-cc9f-2ec0-2eae5381ee92, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 913.682150] env[61806]: DEBUG oslo_vmware.api [None req-70a8784a-38a8-4a3e-8a9f-16278c289170 tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Task: {'id': task-1294882, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 913.846049] env[61806]: DEBUG oslo_concurrency.lockutils [None req-1d9f7d6e-8006-4355-8395-1f9c0f916c73 tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.759s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 913.849034] env[61806]: DEBUG oslo_concurrency.lockutils [None req-dab50ec4-3405-470b-a203-8c80641dde31 tempest-ServerRescueTestJSONUnderV235-749414849 tempest-ServerRescueTestJSONUnderV235-749414849-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 4.178s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 913.849790] env[61806]: DEBUG nova.objects.instance [None req-dab50ec4-3405-470b-a203-8c80641dde31 tempest-ServerRescueTestJSONUnderV235-749414849 tempest-ServerRescueTestJSONUnderV235-749414849-project-member] Lazy-loading 'resources' on Instance uuid 5a526856-0dcd-4b3f-8359-13a5c3b9bfc4 {{(pid=61806) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 913.874026] env[61806]: INFO nova.scheduler.client.report [None req-1d9f7d6e-8006-4355-8395-1f9c0f916c73 tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Deleted allocations for instance 59c22c16-3d1c-4e64-bfac-7f8886a1927e [ 914.016769] env[61806]: DEBUG oslo_vmware.api [None req-26d83e60-0148-4554-a6a0-baa143245ad5 tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] Task: {'id': session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]529a157d-ba46-cc9f-2ec0-2eae5381ee92, 'name': SearchDatastore_Task, 'duration_secs': 0.020688} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 914.017561] env[61806]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c23fdb89-9da8-4d66-9862-19e0a005dd0f {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 914.023221] env[61806]: DEBUG oslo_vmware.api [None req-26d83e60-0148-4554-a6a0-baa143245ad5 tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] Waiting for the task: (returnval){ [ 914.023221] env[61806]: value = "session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]52edb28e-5dda-ad26-4245-c294857ac40f" [ 914.023221] env[61806]: _type = "Task" [ 914.023221] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 914.030789] env[61806]: DEBUG oslo_vmware.api [None req-26d83e60-0148-4554-a6a0-baa143245ad5 tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] Task: {'id': session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]52edb28e-5dda-ad26-4245-c294857ac40f, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 914.181904] env[61806]: DEBUG oslo_vmware.api [None req-70a8784a-38a8-4a3e-8a9f-16278c289170 tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Task: {'id': task-1294882, 'name': PowerOffVM_Task, 'duration_secs': 1.089295} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 914.182229] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [None req-70a8784a-38a8-4a3e-8a9f-16278c289170 tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] [instance: 75f1aded-d1da-4ddc-be29-f1fb93799364] Powered off the VM {{(pid=61806) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 914.182438] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-70a8784a-38a8-4a3e-8a9f-16278c289170 tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] [instance: 75f1aded-d1da-4ddc-be29-f1fb93799364] Destroying instance {{(pid=61806) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 914.183189] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6db5c971-7962-4160-b001-ecac5388678e {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 914.189850] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-70a8784a-38a8-4a3e-8a9f-16278c289170 tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] [instance: 75f1aded-d1da-4ddc-be29-f1fb93799364] Unregistering the VM {{(pid=61806) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 914.190100] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-37f0cd09-04ba-42af-922f-9e5441236407 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 914.255095] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-70a8784a-38a8-4a3e-8a9f-16278c289170 tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] [instance: 75f1aded-d1da-4ddc-be29-f1fb93799364] Unregistered the VM {{(pid=61806) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 914.255293] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-70a8784a-38a8-4a3e-8a9f-16278c289170 tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] [instance: 75f1aded-d1da-4ddc-be29-f1fb93799364] Deleting contents of the VM from datastore datastore1 {{(pid=61806) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 914.255337] env[61806]: DEBUG nova.virt.vmwareapi.ds_util [None req-70a8784a-38a8-4a3e-8a9f-16278c289170 tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Deleting the datastore file [datastore1] 75f1aded-d1da-4ddc-be29-f1fb93799364 {{(pid=61806) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 914.256029] env[61806]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-af255cf4-a954-4f06-9e94-e2c9a3148ec9 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 914.264227] env[61806]: DEBUG oslo_vmware.api [None req-70a8784a-38a8-4a3e-8a9f-16278c289170 tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Waiting for the task: (returnval){ [ 914.264227] env[61806]: value = "task-1294884" [ 914.264227] env[61806]: _type = "Task" [ 914.264227] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 914.272411] env[61806]: DEBUG oslo_vmware.api [None req-70a8784a-38a8-4a3e-8a9f-16278c289170 tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Task: {'id': task-1294884, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 914.381996] env[61806]: DEBUG oslo_concurrency.lockutils [None req-1d9f7d6e-8006-4355-8395-1f9c0f916c73 tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Lock "59c22c16-3d1c-4e64-bfac-7f8886a1927e" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 13.037s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 914.508038] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e817ecff-fc61-4805-a8e8-7db21cb46e5d {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 914.516197] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-58ce324e-4d6a-44d7-a662-4a0ce85034d8 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 914.552965] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dbdec41e-81a3-47eb-9359-b6a5a5d0f7ec {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 914.559232] env[61806]: DEBUG oslo_vmware.api [None req-26d83e60-0148-4554-a6a0-baa143245ad5 tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] Task: {'id': session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]52edb28e-5dda-ad26-4245-c294857ac40f, 'name': SearchDatastore_Task, 'duration_secs': 0.012982} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 914.560288] env[61806]: DEBUG oslo_concurrency.lockutils [None req-26d83e60-0148-4554-a6a0-baa143245ad5 tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] Releasing lock "[datastore1] devstack-image-cache_base/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa.vmdk" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 914.560596] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [None req-26d83e60-0148-4554-a6a0-baa143245ad5 tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa.vmdk to [datastore1] d9caee4a-a00c-4e50-b01c-99a8f796a09f/d9caee4a-a00c-4e50-b01c-99a8f796a09f.vmdk {{(pid=61806) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 914.560872] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-1ac94074-1e57-437e-8cc5-f6a3872ac23e {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 914.567147] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a5ac133c-a831-4099-8169-cb44f06ae514 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 914.572330] env[61806]: DEBUG oslo_vmware.api [None req-26d83e60-0148-4554-a6a0-baa143245ad5 tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] Waiting for the task: (returnval){ [ 914.572330] env[61806]: value = "task-1294885" [ 914.572330] env[61806]: _type = "Task" [ 914.572330] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 914.583522] env[61806]: DEBUG nova.compute.provider_tree [None req-dab50ec4-3405-470b-a203-8c80641dde31 tempest-ServerRescueTestJSONUnderV235-749414849 tempest-ServerRescueTestJSONUnderV235-749414849-project-member] Inventory has not changed in ProviderTree for provider: a2858be1-fd22-4e08-979e-87ad25293407 {{(pid=61806) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 914.590616] env[61806]: DEBUG oslo_vmware.api [None req-26d83e60-0148-4554-a6a0-baa143245ad5 tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] Task: {'id': task-1294885, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 914.775546] env[61806]: DEBUG oslo_vmware.api [None req-70a8784a-38a8-4a3e-8a9f-16278c289170 tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Task: {'id': task-1294884, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.197052} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 914.775816] env[61806]: DEBUG nova.virt.vmwareapi.ds_util [None req-70a8784a-38a8-4a3e-8a9f-16278c289170 tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Deleted the datastore file {{(pid=61806) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 914.776021] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-70a8784a-38a8-4a3e-8a9f-16278c289170 tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] [instance: 75f1aded-d1da-4ddc-be29-f1fb93799364] Deleted contents of the VM from datastore datastore1 {{(pid=61806) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 914.776207] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-70a8784a-38a8-4a3e-8a9f-16278c289170 tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] [instance: 75f1aded-d1da-4ddc-be29-f1fb93799364] Instance destroyed {{(pid=61806) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 915.085628] env[61806]: DEBUG oslo_vmware.api [None req-26d83e60-0148-4554-a6a0-baa143245ad5 tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] Task: {'id': task-1294885, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 915.086839] env[61806]: DEBUG nova.scheduler.client.report [None req-dab50ec4-3405-470b-a203-8c80641dde31 tempest-ServerRescueTestJSONUnderV235-749414849 tempest-ServerRescueTestJSONUnderV235-749414849-project-member] Inventory has not changed for provider a2858be1-fd22-4e08-979e-87ad25293407 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 138, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61806) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 915.481274] env[61806]: DEBUG oslo_concurrency.lockutils [None req-018e57c8-1b55-4515-9361-baa3401dd22f tempest-ServersAaction247Test-1786617390 tempest-ServersAaction247Test-1786617390-project-member] Acquiring lock "3f4fa1ba-a6ad-4e0f-aebc-973816ee09d5" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 915.481537] env[61806]: DEBUG oslo_concurrency.lockutils [None req-018e57c8-1b55-4515-9361-baa3401dd22f tempest-ServersAaction247Test-1786617390 tempest-ServersAaction247Test-1786617390-project-member] Lock "3f4fa1ba-a6ad-4e0f-aebc-973816ee09d5" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 915.585081] env[61806]: DEBUG oslo_vmware.api [None req-26d83e60-0148-4554-a6a0-baa143245ad5 tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] Task: {'id': task-1294885, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.690194} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 915.585390] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [None req-26d83e60-0148-4554-a6a0-baa143245ad5 tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa.vmdk to [datastore1] d9caee4a-a00c-4e50-b01c-99a8f796a09f/d9caee4a-a00c-4e50-b01c-99a8f796a09f.vmdk {{(pid=61806) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 915.585620] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-26d83e60-0148-4554-a6a0-baa143245ad5 tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] [instance: d9caee4a-a00c-4e50-b01c-99a8f796a09f] Extending root virtual disk to 1048576 {{(pid=61806) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 915.585901] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-53206b0d-acd0-41c4-8e30-2caa284eec3a {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 915.590985] env[61806]: DEBUG oslo_concurrency.lockutils [None req-dab50ec4-3405-470b-a203-8c80641dde31 tempest-ServerRescueTestJSONUnderV235-749414849 tempest-ServerRescueTestJSONUnderV235-749414849-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.742s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 915.593165] env[61806]: DEBUG oslo_concurrency.lockutils [None req-531998d5-4ee2-42e3-ac57-d5e3f0c96508 tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 2.769s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 915.594761] env[61806]: INFO nova.compute.claims [None req-531998d5-4ee2-42e3-ac57-d5e3f0c96508 tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] [instance: c486c00d-61fc-4795-9299-e6548b686ee8] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 915.599381] env[61806]: DEBUG oslo_vmware.api [None req-26d83e60-0148-4554-a6a0-baa143245ad5 tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] Waiting for the task: (returnval){ [ 915.599381] env[61806]: value = "task-1294886" [ 915.599381] env[61806]: _type = "Task" [ 915.599381] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 915.612300] env[61806]: DEBUG oslo_vmware.api [None req-26d83e60-0148-4554-a6a0-baa143245ad5 tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] Task: {'id': task-1294886, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 915.613708] env[61806]: INFO nova.scheduler.client.report [None req-dab50ec4-3405-470b-a203-8c80641dde31 tempest-ServerRescueTestJSONUnderV235-749414849 tempest-ServerRescueTestJSONUnderV235-749414849-project-member] Deleted allocations for instance 5a526856-0dcd-4b3f-8359-13a5c3b9bfc4 [ 915.812058] env[61806]: DEBUG nova.virt.hardware [None req-70a8784a-38a8-4a3e-8a9f-16278c289170 tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-15T18:05:10Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-15T18:04:52Z,direct_url=,disk_format='vmdk',id=ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='84ba83fce2c349a6988173c5d6fc3b07',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-15T18:04:53Z,virtual_size=,visibility=), allow threads: False {{(pid=61806) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 915.812344] env[61806]: DEBUG nova.virt.hardware [None req-70a8784a-38a8-4a3e-8a9f-16278c289170 tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Flavor limits 0:0:0 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 915.812516] env[61806]: DEBUG nova.virt.hardware [None req-70a8784a-38a8-4a3e-8a9f-16278c289170 tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Image limits 0:0:0 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 915.812730] env[61806]: DEBUG nova.virt.hardware [None req-70a8784a-38a8-4a3e-8a9f-16278c289170 tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Flavor pref 0:0:0 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 915.812882] env[61806]: DEBUG nova.virt.hardware [None req-70a8784a-38a8-4a3e-8a9f-16278c289170 tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Image pref 0:0:0 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 915.813033] env[61806]: DEBUG nova.virt.hardware [None req-70a8784a-38a8-4a3e-8a9f-16278c289170 tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 915.813259] env[61806]: DEBUG nova.virt.hardware [None req-70a8784a-38a8-4a3e-8a9f-16278c289170 tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61806) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 915.813428] env[61806]: DEBUG nova.virt.hardware [None req-70a8784a-38a8-4a3e-8a9f-16278c289170 tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61806) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 915.813660] env[61806]: DEBUG nova.virt.hardware [None req-70a8784a-38a8-4a3e-8a9f-16278c289170 tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Got 1 possible topologies {{(pid=61806) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 915.813843] env[61806]: DEBUG nova.virt.hardware [None req-70a8784a-38a8-4a3e-8a9f-16278c289170 tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61806) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 915.814036] env[61806]: DEBUG nova.virt.hardware [None req-70a8784a-38a8-4a3e-8a9f-16278c289170 tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61806) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 915.814964] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-02748f4a-232b-4e74-b1a5-53ce053d4fb9 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 915.823778] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4e299468-4a1d-4a9a-a642-90dc1ac3b112 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 915.837790] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-70a8784a-38a8-4a3e-8a9f-16278c289170 tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] [instance: 75f1aded-d1da-4ddc-be29-f1fb93799364] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:8f:1a:72', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '82ca17df-257e-40e6-9ec9-310ed6f05ccb', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'a06a82f0-c5d3-48b7-bd4b-d2725bd7eaec', 'vif_model': 'vmxnet3'}] {{(pid=61806) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 915.845168] env[61806]: DEBUG oslo.service.loopingcall [None req-70a8784a-38a8-4a3e-8a9f-16278c289170 tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61806) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 915.845428] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 75f1aded-d1da-4ddc-be29-f1fb93799364] Creating VM on the ESX host {{(pid=61806) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 915.845643] env[61806]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-cc344808-415d-4c4b-87f4-c954b5c5846a {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 915.865897] env[61806]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 915.865897] env[61806]: value = "task-1294887" [ 915.865897] env[61806]: _type = "Task" [ 915.865897] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 915.874282] env[61806]: DEBUG oslo_vmware.api [-] Task: {'id': task-1294887, 'name': CreateVM_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 915.983808] env[61806]: DEBUG nova.compute.manager [None req-018e57c8-1b55-4515-9361-baa3401dd22f tempest-ServersAaction247Test-1786617390 tempest-ServersAaction247Test-1786617390-project-member] [instance: 3f4fa1ba-a6ad-4e0f-aebc-973816ee09d5] Starting instance... {{(pid=61806) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 916.106038] env[61806]: DEBUG oslo_concurrency.lockutils [None req-d76979a9-290f-44f9-bf19-39a08fc72d03 tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Acquiring lock "380aa63f-46ba-437d-bc29-0fc141ad1cf3" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 916.106038] env[61806]: DEBUG oslo_concurrency.lockutils [None req-d76979a9-290f-44f9-bf19-39a08fc72d03 tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Lock "380aa63f-46ba-437d-bc29-0fc141ad1cf3" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 916.121729] env[61806]: DEBUG oslo_vmware.api [None req-26d83e60-0148-4554-a6a0-baa143245ad5 tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] Task: {'id': task-1294886, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.086829} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 916.122012] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-26d83e60-0148-4554-a6a0-baa143245ad5 tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] [instance: d9caee4a-a00c-4e50-b01c-99a8f796a09f] Extended root virtual disk {{(pid=61806) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 916.122860] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a0850634-ae75-4e5b-9fed-3f9798f95b18 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 916.126692] env[61806]: DEBUG oslo_concurrency.lockutils [None req-dab50ec4-3405-470b-a203-8c80641dde31 tempest-ServerRescueTestJSONUnderV235-749414849 tempest-ServerRescueTestJSONUnderV235-749414849-project-member] Lock "5a526856-0dcd-4b3f-8359-13a5c3b9bfc4" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 9.370s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 916.152085] env[61806]: DEBUG nova.virt.vmwareapi.volumeops [None req-26d83e60-0148-4554-a6a0-baa143245ad5 tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] [instance: d9caee4a-a00c-4e50-b01c-99a8f796a09f] Reconfiguring VM instance instance-00000053 to attach disk [datastore1] d9caee4a-a00c-4e50-b01c-99a8f796a09f/d9caee4a-a00c-4e50-b01c-99a8f796a09f.vmdk or device None with type sparse {{(pid=61806) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 916.152278] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-abe5dc3f-a96d-4da0-a063-e4d100a56efe {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 916.174229] env[61806]: DEBUG oslo_vmware.api [None req-26d83e60-0148-4554-a6a0-baa143245ad5 tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] Waiting for the task: (returnval){ [ 916.174229] env[61806]: value = "task-1294888" [ 916.174229] env[61806]: _type = "Task" [ 916.174229] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 916.185370] env[61806]: DEBUG oslo_vmware.api [None req-26d83e60-0148-4554-a6a0-baa143245ad5 tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] Task: {'id': task-1294888, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 916.377126] env[61806]: DEBUG oslo_vmware.api [-] Task: {'id': task-1294887, 'name': CreateVM_Task, 'duration_secs': 0.420043} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 916.377305] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 75f1aded-d1da-4ddc-be29-f1fb93799364] Created VM on the ESX host {{(pid=61806) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 916.378012] env[61806]: DEBUG oslo_concurrency.lockutils [None req-70a8784a-38a8-4a3e-8a9f-16278c289170 tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 916.378191] env[61806]: DEBUG oslo_concurrency.lockutils [None req-70a8784a-38a8-4a3e-8a9f-16278c289170 tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Acquired lock "[datastore1] devstack-image-cache_base/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 916.378522] env[61806]: DEBUG oslo_concurrency.lockutils [None req-70a8784a-38a8-4a3e-8a9f-16278c289170 tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 916.379123] env[61806]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-6e5a5720-8db1-473f-8e49-6a3e75c8ac2e {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 916.385020] env[61806]: DEBUG oslo_vmware.api [None req-70a8784a-38a8-4a3e-8a9f-16278c289170 tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Waiting for the task: (returnval){ [ 916.385020] env[61806]: value = "session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]528152f2-e615-4156-0ec6-5e69f208a078" [ 916.385020] env[61806]: _type = "Task" [ 916.385020] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 916.393229] env[61806]: DEBUG oslo_vmware.api [None req-70a8784a-38a8-4a3e-8a9f-16278c289170 tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Task: {'id': session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]528152f2-e615-4156-0ec6-5e69f208a078, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 916.507855] env[61806]: DEBUG oslo_concurrency.lockutils [None req-018e57c8-1b55-4515-9361-baa3401dd22f tempest-ServersAaction247Test-1786617390 tempest-ServersAaction247Test-1786617390-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 916.608724] env[61806]: DEBUG nova.compute.manager [None req-d76979a9-290f-44f9-bf19-39a08fc72d03 tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] [instance: 380aa63f-46ba-437d-bc29-0fc141ad1cf3] Starting instance... {{(pid=61806) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 916.685717] env[61806]: DEBUG oslo_vmware.api [None req-26d83e60-0148-4554-a6a0-baa143245ad5 tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] Task: {'id': task-1294888, 'name': ReconfigVM_Task, 'duration_secs': 0.361724} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 916.686043] env[61806]: DEBUG nova.virt.vmwareapi.volumeops [None req-26d83e60-0148-4554-a6a0-baa143245ad5 tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] [instance: d9caee4a-a00c-4e50-b01c-99a8f796a09f] Reconfigured VM instance instance-00000053 to attach disk [datastore1] d9caee4a-a00c-4e50-b01c-99a8f796a09f/d9caee4a-a00c-4e50-b01c-99a8f796a09f.vmdk or device None with type sparse {{(pid=61806) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 916.686803] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-f9476879-a546-4963-bf88-71dbf0504a9d {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 916.697230] env[61806]: DEBUG oslo_vmware.api [None req-26d83e60-0148-4554-a6a0-baa143245ad5 tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] Waiting for the task: (returnval){ [ 916.697230] env[61806]: value = "task-1294889" [ 916.697230] env[61806]: _type = "Task" [ 916.697230] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 916.713859] env[61806]: DEBUG oslo_vmware.api [None req-26d83e60-0148-4554-a6a0-baa143245ad5 tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] Task: {'id': task-1294889, 'name': Rename_Task} progress is 6%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 916.752474] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b5267d59-8997-4400-a06e-039375033f44 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 916.760682] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b93ec46e-7294-4374-ae34-96de61409ecd {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 916.796042] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1ed97e7f-62af-47e9-8dcc-52fed430078e {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 916.804257] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-506d3e0d-a1de-4527-9b9d-da68ed2d2b29 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 916.819422] env[61806]: DEBUG nova.compute.provider_tree [None req-531998d5-4ee2-42e3-ac57-d5e3f0c96508 tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] Inventory has not changed in ProviderTree for provider: a2858be1-fd22-4e08-979e-87ad25293407 {{(pid=61806) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 916.895714] env[61806]: DEBUG oslo_vmware.api [None req-70a8784a-38a8-4a3e-8a9f-16278c289170 tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Task: {'id': session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]528152f2-e615-4156-0ec6-5e69f208a078, 'name': SearchDatastore_Task, 'duration_secs': 0.021715} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 916.896055] env[61806]: DEBUG oslo_concurrency.lockutils [None req-70a8784a-38a8-4a3e-8a9f-16278c289170 tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Releasing lock "[datastore1] devstack-image-cache_base/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 916.896304] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-70a8784a-38a8-4a3e-8a9f-16278c289170 tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] [instance: 75f1aded-d1da-4ddc-be29-f1fb93799364] Processing image ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa {{(pid=61806) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 916.896554] env[61806]: DEBUG oslo_concurrency.lockutils [None req-70a8784a-38a8-4a3e-8a9f-16278c289170 tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa.vmdk" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 916.896713] env[61806]: DEBUG oslo_concurrency.lockutils [None req-70a8784a-38a8-4a3e-8a9f-16278c289170 tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Acquired lock "[datastore1] devstack-image-cache_base/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa.vmdk" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 916.896898] env[61806]: DEBUG nova.virt.vmwareapi.ds_util [None req-70a8784a-38a8-4a3e-8a9f-16278c289170 tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61806) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 916.897187] env[61806]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-d9299b1f-f294-4f6a-a362-f37090c344b0 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 916.915770] env[61806]: DEBUG nova.virt.vmwareapi.ds_util [None req-70a8784a-38a8-4a3e-8a9f-16278c289170 tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61806) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 916.915976] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-70a8784a-38a8-4a3e-8a9f-16278c289170 tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61806) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 916.916841] env[61806]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-85f91960-316a-415b-9f90-aedb893b870f {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 916.923880] env[61806]: DEBUG oslo_vmware.api [None req-70a8784a-38a8-4a3e-8a9f-16278c289170 tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Waiting for the task: (returnval){ [ 916.923880] env[61806]: value = "session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]52dcd047-5304-f4b5-b142-275b324db627" [ 916.923880] env[61806]: _type = "Task" [ 916.923880] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 916.933078] env[61806]: DEBUG oslo_vmware.api [None req-70a8784a-38a8-4a3e-8a9f-16278c289170 tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Task: {'id': session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]52dcd047-5304-f4b5-b142-275b324db627, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 917.128937] env[61806]: DEBUG oslo_concurrency.lockutils [None req-d76979a9-290f-44f9-bf19-39a08fc72d03 tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 917.208409] env[61806]: DEBUG oslo_vmware.api [None req-26d83e60-0148-4554-a6a0-baa143245ad5 tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] Task: {'id': task-1294889, 'name': Rename_Task, 'duration_secs': 0.189899} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 917.208708] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [None req-26d83e60-0148-4554-a6a0-baa143245ad5 tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] [instance: d9caee4a-a00c-4e50-b01c-99a8f796a09f] Powering on the VM {{(pid=61806) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 917.208960] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-469b7deb-8484-4546-82a4-a467ae7bc949 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 917.216113] env[61806]: DEBUG oslo_vmware.api [None req-26d83e60-0148-4554-a6a0-baa143245ad5 tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] Waiting for the task: (returnval){ [ 917.216113] env[61806]: value = "task-1294890" [ 917.216113] env[61806]: _type = "Task" [ 917.216113] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 917.224672] env[61806]: DEBUG oslo_vmware.api [None req-26d83e60-0148-4554-a6a0-baa143245ad5 tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] Task: {'id': task-1294890, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 917.323109] env[61806]: DEBUG nova.scheduler.client.report [None req-531998d5-4ee2-42e3-ac57-d5e3f0c96508 tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] Inventory has not changed for provider a2858be1-fd22-4e08-979e-87ad25293407 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 138, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61806) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 917.437131] env[61806]: DEBUG oslo_vmware.api [None req-70a8784a-38a8-4a3e-8a9f-16278c289170 tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Task: {'id': session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]52dcd047-5304-f4b5-b142-275b324db627, 'name': SearchDatastore_Task, 'duration_secs': 0.030962} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 917.438172] env[61806]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-4a66d38f-89e4-4474-92e8-20d5f6987cf1 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 917.444970] env[61806]: DEBUG oslo_vmware.api [None req-70a8784a-38a8-4a3e-8a9f-16278c289170 tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Waiting for the task: (returnval){ [ 917.444970] env[61806]: value = "session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]52b86b65-0767-782c-a1b8-be16aa0fdd11" [ 917.444970] env[61806]: _type = "Task" [ 917.444970] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 917.455720] env[61806]: DEBUG oslo_vmware.api [None req-70a8784a-38a8-4a3e-8a9f-16278c289170 tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Task: {'id': session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]52b86b65-0767-782c-a1b8-be16aa0fdd11, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 917.730514] env[61806]: DEBUG oslo_vmware.api [None req-26d83e60-0148-4554-a6a0-baa143245ad5 tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] Task: {'id': task-1294890, 'name': PowerOnVM_Task} progress is 100%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 917.828781] env[61806]: DEBUG oslo_concurrency.lockutils [None req-531998d5-4ee2-42e3-ac57-d5e3f0c96508 tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.236s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 917.829385] env[61806]: DEBUG nova.compute.manager [None req-531998d5-4ee2-42e3-ac57-d5e3f0c96508 tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] [instance: c486c00d-61fc-4795-9299-e6548b686ee8] Start building networks asynchronously for instance. {{(pid=61806) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 917.832390] env[61806]: DEBUG oslo_concurrency.lockutils [None req-018e57c8-1b55-4515-9361-baa3401dd22f tempest-ServersAaction247Test-1786617390 tempest-ServersAaction247Test-1786617390-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 1.325s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 917.834074] env[61806]: INFO nova.compute.claims [None req-018e57c8-1b55-4515-9361-baa3401dd22f tempest-ServersAaction247Test-1786617390 tempest-ServersAaction247Test-1786617390-project-member] [instance: 3f4fa1ba-a6ad-4e0f-aebc-973816ee09d5] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 917.956277] env[61806]: DEBUG oslo_vmware.api [None req-70a8784a-38a8-4a3e-8a9f-16278c289170 tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Task: {'id': session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]52b86b65-0767-782c-a1b8-be16aa0fdd11, 'name': SearchDatastore_Task, 'duration_secs': 0.012262} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 917.956579] env[61806]: DEBUG oslo_concurrency.lockutils [None req-70a8784a-38a8-4a3e-8a9f-16278c289170 tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Releasing lock "[datastore1] devstack-image-cache_base/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa.vmdk" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 917.956901] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [None req-70a8784a-38a8-4a3e-8a9f-16278c289170 tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa.vmdk to [datastore1] 75f1aded-d1da-4ddc-be29-f1fb93799364/75f1aded-d1da-4ddc-be29-f1fb93799364.vmdk {{(pid=61806) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 917.957209] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-3c6dbffa-5963-4f49-905e-a48abba97170 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 917.964906] env[61806]: DEBUG oslo_vmware.api [None req-70a8784a-38a8-4a3e-8a9f-16278c289170 tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Waiting for the task: (returnval){ [ 917.964906] env[61806]: value = "task-1294891" [ 917.964906] env[61806]: _type = "Task" [ 917.964906] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 917.973177] env[61806]: DEBUG oslo_vmware.api [None req-70a8784a-38a8-4a3e-8a9f-16278c289170 tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Task: {'id': task-1294891, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 918.044030] env[61806]: DEBUG oslo_concurrency.lockutils [None req-d718d0b7-b777-42ad-8e19-32c28e9d031c tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Acquiring lock "67d1bdd8-cf8f-4fb5-bd3c-42008e4d60e3" by "nova.compute.manager.ComputeManager.reboot_instance..do_reboot_instance" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 918.044171] env[61806]: DEBUG oslo_concurrency.lockutils [None req-d718d0b7-b777-42ad-8e19-32c28e9d031c tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Lock "67d1bdd8-cf8f-4fb5-bd3c-42008e4d60e3" acquired by "nova.compute.manager.ComputeManager.reboot_instance..do_reboot_instance" :: waited 0.001s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 918.044342] env[61806]: INFO nova.compute.manager [None req-d718d0b7-b777-42ad-8e19-32c28e9d031c tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] [instance: 67d1bdd8-cf8f-4fb5-bd3c-42008e4d60e3] Rebooting instance [ 918.226610] env[61806]: DEBUG oslo_vmware.api [None req-26d83e60-0148-4554-a6a0-baa143245ad5 tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] Task: {'id': task-1294890, 'name': PowerOnVM_Task, 'duration_secs': 0.526696} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 918.226997] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [None req-26d83e60-0148-4554-a6a0-baa143245ad5 tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] [instance: d9caee4a-a00c-4e50-b01c-99a8f796a09f] Powered on the VM {{(pid=61806) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 918.227144] env[61806]: INFO nova.compute.manager [None req-26d83e60-0148-4554-a6a0-baa143245ad5 tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] [instance: d9caee4a-a00c-4e50-b01c-99a8f796a09f] Took 7.84 seconds to spawn the instance on the hypervisor. [ 918.227341] env[61806]: DEBUG nova.compute.manager [None req-26d83e60-0148-4554-a6a0-baa143245ad5 tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] [instance: d9caee4a-a00c-4e50-b01c-99a8f796a09f] Checking state {{(pid=61806) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 918.228126] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2c2a5183-8031-421c-9fde-59ef2b033577 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 918.339345] env[61806]: DEBUG nova.compute.utils [None req-531998d5-4ee2-42e3-ac57-d5e3f0c96508 tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] Using /dev/sd instead of None {{(pid=61806) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 918.343116] env[61806]: DEBUG nova.compute.manager [None req-531998d5-4ee2-42e3-ac57-d5e3f0c96508 tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] [instance: c486c00d-61fc-4795-9299-e6548b686ee8] Allocating IP information in the background. {{(pid=61806) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 918.343222] env[61806]: DEBUG nova.network.neutron [None req-531998d5-4ee2-42e3-ac57-d5e3f0c96508 tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] [instance: c486c00d-61fc-4795-9299-e6548b686ee8] allocate_for_instance() {{(pid=61806) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 918.418353] env[61806]: DEBUG nova.policy [None req-531998d5-4ee2-42e3-ac57-d5e3f0c96508 tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '4c5d4c6e1ad844a585c08c4644a60bbf', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '73ebcc4d79d248efb653a084a5e44302', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61806) authorize /opt/stack/nova/nova/policy.py:201}} [ 918.476717] env[61806]: DEBUG oslo_vmware.api [None req-70a8784a-38a8-4a3e-8a9f-16278c289170 tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Task: {'id': task-1294891, 'name': CopyVirtualDisk_Task} progress is 25%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 918.562636] env[61806]: DEBUG oslo_concurrency.lockutils [None req-d718d0b7-b777-42ad-8e19-32c28e9d031c tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Acquiring lock "refresh_cache-67d1bdd8-cf8f-4fb5-bd3c-42008e4d60e3" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 918.562878] env[61806]: DEBUG oslo_concurrency.lockutils [None req-d718d0b7-b777-42ad-8e19-32c28e9d031c tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Acquired lock "refresh_cache-67d1bdd8-cf8f-4fb5-bd3c-42008e4d60e3" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 918.563326] env[61806]: DEBUG nova.network.neutron [None req-d718d0b7-b777-42ad-8e19-32c28e9d031c tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] [instance: 67d1bdd8-cf8f-4fb5-bd3c-42008e4d60e3] Building network info cache for instance {{(pid=61806) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 918.748262] env[61806]: INFO nova.compute.manager [None req-26d83e60-0148-4554-a6a0-baa143245ad5 tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] [instance: d9caee4a-a00c-4e50-b01c-99a8f796a09f] Took 12.69 seconds to build instance. [ 918.843458] env[61806]: DEBUG nova.compute.manager [None req-531998d5-4ee2-42e3-ac57-d5e3f0c96508 tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] [instance: c486c00d-61fc-4795-9299-e6548b686ee8] Start building block device mappings for instance. {{(pid=61806) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 918.972037] env[61806]: DEBUG nova.network.neutron [None req-531998d5-4ee2-42e3-ac57-d5e3f0c96508 tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] [instance: c486c00d-61fc-4795-9299-e6548b686ee8] Successfully created port: df6e9523-3657-4405-8d44-5b35cfc6fa9f {{(pid=61806) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 918.984398] env[61806]: DEBUG oslo_vmware.api [None req-70a8784a-38a8-4a3e-8a9f-16278c289170 tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Task: {'id': task-1294891, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.76807} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 918.984719] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [None req-70a8784a-38a8-4a3e-8a9f-16278c289170 tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa.vmdk to [datastore1] 75f1aded-d1da-4ddc-be29-f1fb93799364/75f1aded-d1da-4ddc-be29-f1fb93799364.vmdk {{(pid=61806) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 918.984971] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-70a8784a-38a8-4a3e-8a9f-16278c289170 tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] [instance: 75f1aded-d1da-4ddc-be29-f1fb93799364] Extending root virtual disk to 1048576 {{(pid=61806) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 918.985351] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-3af7c273-977e-46db-b598-6a5fe153a0b9 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 918.994503] env[61806]: DEBUG oslo_vmware.api [None req-70a8784a-38a8-4a3e-8a9f-16278c289170 tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Waiting for the task: (returnval){ [ 918.994503] env[61806]: value = "task-1294892" [ 918.994503] env[61806]: _type = "Task" [ 918.994503] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 919.004011] env[61806]: DEBUG oslo_vmware.api [None req-70a8784a-38a8-4a3e-8a9f-16278c289170 tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Task: {'id': task-1294892, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 919.019431] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e82c61ba-3f7d-4f6f-a784-3ea43252d511 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 919.028015] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-93b5e945-4b0d-4e18-927d-3c9db4f9ab75 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 919.064019] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-989e2dcc-fa1f-4767-89a6-964720bbd6f3 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 919.075156] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-56220aa4-6e49-4947-b9cd-5911e3b2d8d7 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 919.092385] env[61806]: DEBUG nova.compute.provider_tree [None req-018e57c8-1b55-4515-9361-baa3401dd22f tempest-ServersAaction247Test-1786617390 tempest-ServersAaction247Test-1786617390-project-member] Inventory has not changed in ProviderTree for provider: a2858be1-fd22-4e08-979e-87ad25293407 {{(pid=61806) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 919.249918] env[61806]: DEBUG oslo_concurrency.lockutils [None req-26d83e60-0148-4554-a6a0-baa143245ad5 tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] Lock "d9caee4a-a00c-4e50-b01c-99a8f796a09f" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 14.216s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 919.319380] env[61806]: DEBUG nova.network.neutron [None req-d718d0b7-b777-42ad-8e19-32c28e9d031c tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] [instance: 67d1bdd8-cf8f-4fb5-bd3c-42008e4d60e3] Updating instance_info_cache with network_info: [{"id": "859fb723-4a1c-4ea8-bd7c-0ed308f3a9e3", "address": "fa:16:3e:3c:40:d6", "network": {"id": "746a03b6-c1a6-4d27-98b9-0bee6376634f", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-1971382587-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.245", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c1f2d161c50f4b70987b90e6364c3779", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6a31b224-77d7-48a4-af87-312758611b8e", "external-id": "nsx-vlan-transportzone-761", "segmentation_id": 761, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap859fb723-4a", "ovs_interfaceid": "859fb723-4a1c-4ea8-bd7c-0ed308f3a9e3", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61806) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 919.504961] env[61806]: DEBUG oslo_vmware.api [None req-70a8784a-38a8-4a3e-8a9f-16278c289170 tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Task: {'id': task-1294892, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.079699} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 919.505264] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-70a8784a-38a8-4a3e-8a9f-16278c289170 tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] [instance: 75f1aded-d1da-4ddc-be29-f1fb93799364] Extended root virtual disk {{(pid=61806) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 919.506061] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0d46c550-c455-4247-8b36-20d3842d7575 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 919.527806] env[61806]: DEBUG nova.virt.vmwareapi.volumeops [None req-70a8784a-38a8-4a3e-8a9f-16278c289170 tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] [instance: 75f1aded-d1da-4ddc-be29-f1fb93799364] Reconfiguring VM instance instance-00000052 to attach disk [datastore1] 75f1aded-d1da-4ddc-be29-f1fb93799364/75f1aded-d1da-4ddc-be29-f1fb93799364.vmdk or device None with type sparse {{(pid=61806) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 919.528419] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-95a46ca3-7722-4936-ab84-6303a3068e82 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 919.550206] env[61806]: DEBUG oslo_vmware.api [None req-70a8784a-38a8-4a3e-8a9f-16278c289170 tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Waiting for the task: (returnval){ [ 919.550206] env[61806]: value = "task-1294893" [ 919.550206] env[61806]: _type = "Task" [ 919.550206] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 919.558832] env[61806]: DEBUG oslo_vmware.api [None req-70a8784a-38a8-4a3e-8a9f-16278c289170 tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Task: {'id': task-1294893, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 919.596163] env[61806]: DEBUG nova.scheduler.client.report [None req-018e57c8-1b55-4515-9361-baa3401dd22f tempest-ServersAaction247Test-1786617390 tempest-ServersAaction247Test-1786617390-project-member] Inventory has not changed for provider a2858be1-fd22-4e08-979e-87ad25293407 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 138, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61806) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 919.822670] env[61806]: DEBUG oslo_concurrency.lockutils [None req-d718d0b7-b777-42ad-8e19-32c28e9d031c tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Releasing lock "refresh_cache-67d1bdd8-cf8f-4fb5-bd3c-42008e4d60e3" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 919.832614] env[61806]: DEBUG nova.compute.manager [None req-d718d0b7-b777-42ad-8e19-32c28e9d031c tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] [instance: 67d1bdd8-cf8f-4fb5-bd3c-42008e4d60e3] Checking state {{(pid=61806) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 919.833649] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-97ced3c7-eb90-48c1-b2d2-f7404591d89e {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 919.855770] env[61806]: DEBUG nova.compute.manager [None req-531998d5-4ee2-42e3-ac57-d5e3f0c96508 tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] [instance: c486c00d-61fc-4795-9299-e6548b686ee8] Start spawning the instance on the hypervisor. {{(pid=61806) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 919.889429] env[61806]: DEBUG nova.virt.hardware [None req-531998d5-4ee2-42e3-ac57-d5e3f0c96508 tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-15T18:05:10Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-15T18:04:52Z,direct_url=,disk_format='vmdk',id=ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='84ba83fce2c349a6988173c5d6fc3b07',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-15T18:04:53Z,virtual_size=,visibility=), allow threads: False {{(pid=61806) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 919.890552] env[61806]: DEBUG nova.virt.hardware [None req-531998d5-4ee2-42e3-ac57-d5e3f0c96508 tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] Flavor limits 0:0:0 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 919.890552] env[61806]: DEBUG nova.virt.hardware [None req-531998d5-4ee2-42e3-ac57-d5e3f0c96508 tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] Image limits 0:0:0 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 919.890552] env[61806]: DEBUG nova.virt.hardware [None req-531998d5-4ee2-42e3-ac57-d5e3f0c96508 tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] Flavor pref 0:0:0 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 919.890552] env[61806]: DEBUG nova.virt.hardware [None req-531998d5-4ee2-42e3-ac57-d5e3f0c96508 tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] Image pref 0:0:0 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 919.890552] env[61806]: DEBUG nova.virt.hardware [None req-531998d5-4ee2-42e3-ac57-d5e3f0c96508 tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 919.890761] env[61806]: DEBUG nova.virt.hardware [None req-531998d5-4ee2-42e3-ac57-d5e3f0c96508 tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61806) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 919.890794] env[61806]: DEBUG nova.virt.hardware [None req-531998d5-4ee2-42e3-ac57-d5e3f0c96508 tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61806) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 919.891532] env[61806]: DEBUG nova.virt.hardware [None req-531998d5-4ee2-42e3-ac57-d5e3f0c96508 tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] Got 1 possible topologies {{(pid=61806) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 919.891532] env[61806]: DEBUG nova.virt.hardware [None req-531998d5-4ee2-42e3-ac57-d5e3f0c96508 tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61806) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 919.891532] env[61806]: DEBUG nova.virt.hardware [None req-531998d5-4ee2-42e3-ac57-d5e3f0c96508 tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61806) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 919.892251] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6b166420-5c27-43e6-b3ea-c58fd9f80d40 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 919.902033] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f3003037-896a-4498-bb8c-74c868de0729 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 920.066933] env[61806]: DEBUG oslo_vmware.api [None req-70a8784a-38a8-4a3e-8a9f-16278c289170 tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Task: {'id': task-1294893, 'name': ReconfigVM_Task, 'duration_secs': 0.32137} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 920.067465] env[61806]: DEBUG nova.virt.vmwareapi.volumeops [None req-70a8784a-38a8-4a3e-8a9f-16278c289170 tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] [instance: 75f1aded-d1da-4ddc-be29-f1fb93799364] Reconfigured VM instance instance-00000052 to attach disk [datastore1] 75f1aded-d1da-4ddc-be29-f1fb93799364/75f1aded-d1da-4ddc-be29-f1fb93799364.vmdk or device None with type sparse {{(pid=61806) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 920.068484] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-82090742-49d1-400a-8412-c8a7b2d14b48 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 920.077827] env[61806]: DEBUG oslo_vmware.api [None req-70a8784a-38a8-4a3e-8a9f-16278c289170 tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Waiting for the task: (returnval){ [ 920.077827] env[61806]: value = "task-1294894" [ 920.077827] env[61806]: _type = "Task" [ 920.077827] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 920.089669] env[61806]: DEBUG oslo_vmware.api [None req-70a8784a-38a8-4a3e-8a9f-16278c289170 tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Task: {'id': task-1294894, 'name': Rename_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 920.101937] env[61806]: DEBUG oslo_concurrency.lockutils [None req-018e57c8-1b55-4515-9361-baa3401dd22f tempest-ServersAaction247Test-1786617390 tempest-ServersAaction247Test-1786617390-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.269s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 920.104192] env[61806]: DEBUG nova.compute.manager [None req-018e57c8-1b55-4515-9361-baa3401dd22f tempest-ServersAaction247Test-1786617390 tempest-ServersAaction247Test-1786617390-project-member] [instance: 3f4fa1ba-a6ad-4e0f-aebc-973816ee09d5] Start building networks asynchronously for instance. {{(pid=61806) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 920.107496] env[61806]: DEBUG oslo_concurrency.lockutils [None req-d76979a9-290f-44f9-bf19-39a08fc72d03 tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 2.979s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 920.109577] env[61806]: INFO nova.compute.claims [None req-d76979a9-290f-44f9-bf19-39a08fc72d03 tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] [instance: 380aa63f-46ba-437d-bc29-0fc141ad1cf3] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 920.409735] env[61806]: DEBUG nova.compute.manager [req-32dc92ee-5715-45fc-b145-d7fb1097b284 req-c4a89cb5-2e46-4213-a2e2-36372102d181 service nova] [instance: 6f6ba57a-e2d6-4749-a53a-e263861cb1c0] Received event network-changed-013c9f0d-65b1-4279-a823-a7f1c14dfb91 {{(pid=61806) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 920.410212] env[61806]: DEBUG nova.compute.manager [req-32dc92ee-5715-45fc-b145-d7fb1097b284 req-c4a89cb5-2e46-4213-a2e2-36372102d181 service nova] [instance: 6f6ba57a-e2d6-4749-a53a-e263861cb1c0] Refreshing instance network info cache due to event network-changed-013c9f0d-65b1-4279-a823-a7f1c14dfb91. {{(pid=61806) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 920.410460] env[61806]: DEBUG oslo_concurrency.lockutils [req-32dc92ee-5715-45fc-b145-d7fb1097b284 req-c4a89cb5-2e46-4213-a2e2-36372102d181 service nova] Acquiring lock "refresh_cache-6f6ba57a-e2d6-4749-a53a-e263861cb1c0" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 920.410614] env[61806]: DEBUG oslo_concurrency.lockutils [req-32dc92ee-5715-45fc-b145-d7fb1097b284 req-c4a89cb5-2e46-4213-a2e2-36372102d181 service nova] Acquired lock "refresh_cache-6f6ba57a-e2d6-4749-a53a-e263861cb1c0" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 920.410814] env[61806]: DEBUG nova.network.neutron [req-32dc92ee-5715-45fc-b145-d7fb1097b284 req-c4a89cb5-2e46-4213-a2e2-36372102d181 service nova] [instance: 6f6ba57a-e2d6-4749-a53a-e263861cb1c0] Refreshing network info cache for port 013c9f0d-65b1-4279-a823-a7f1c14dfb91 {{(pid=61806) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 920.591793] env[61806]: DEBUG oslo_vmware.api [None req-70a8784a-38a8-4a3e-8a9f-16278c289170 tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Task: {'id': task-1294894, 'name': Rename_Task, 'duration_secs': 0.172344} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 920.592525] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [None req-70a8784a-38a8-4a3e-8a9f-16278c289170 tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] [instance: 75f1aded-d1da-4ddc-be29-f1fb93799364] Powering on the VM {{(pid=61806) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 920.595094] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-1fff626a-be5c-4da4-92b2-304b12c49c4c {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 920.603185] env[61806]: DEBUG oslo_vmware.api [None req-70a8784a-38a8-4a3e-8a9f-16278c289170 tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Waiting for the task: (returnval){ [ 920.603185] env[61806]: value = "task-1294895" [ 920.603185] env[61806]: _type = "Task" [ 920.603185] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 920.613163] env[61806]: DEBUG oslo_vmware.api [None req-70a8784a-38a8-4a3e-8a9f-16278c289170 tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Task: {'id': task-1294895, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 920.615891] env[61806]: DEBUG nova.compute.utils [None req-018e57c8-1b55-4515-9361-baa3401dd22f tempest-ServersAaction247Test-1786617390 tempest-ServersAaction247Test-1786617390-project-member] Using /dev/sd instead of None {{(pid=61806) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 920.621208] env[61806]: DEBUG nova.compute.manager [None req-018e57c8-1b55-4515-9361-baa3401dd22f tempest-ServersAaction247Test-1786617390 tempest-ServersAaction247Test-1786617390-project-member] [instance: 3f4fa1ba-a6ad-4e0f-aebc-973816ee09d5] Not allocating networking since 'none' was specified. {{(pid=61806) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1977}} [ 920.859696] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4d30dbbf-c037-4b8c-b52e-64e5e4897f6a {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 920.869074] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-d718d0b7-b777-42ad-8e19-32c28e9d031c tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] [instance: 67d1bdd8-cf8f-4fb5-bd3c-42008e4d60e3] Doing hard reboot of VM {{(pid=61806) reboot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1064}} [ 920.869074] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ResetVM_Task with opID=oslo.vmware-608081ff-8c81-41c2-945d-bdab4771eea6 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 920.876279] env[61806]: DEBUG oslo_vmware.api [None req-d718d0b7-b777-42ad-8e19-32c28e9d031c tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Waiting for the task: (returnval){ [ 920.876279] env[61806]: value = "task-1294896" [ 920.876279] env[61806]: _type = "Task" [ 920.876279] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 920.889024] env[61806]: DEBUG oslo_vmware.api [None req-d718d0b7-b777-42ad-8e19-32c28e9d031c tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Task: {'id': task-1294896, 'name': ResetVM_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 921.119662] env[61806]: DEBUG oslo_vmware.api [None req-70a8784a-38a8-4a3e-8a9f-16278c289170 tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Task: {'id': task-1294895, 'name': PowerOnVM_Task} progress is 100%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 921.121545] env[61806]: DEBUG nova.compute.manager [None req-018e57c8-1b55-4515-9361-baa3401dd22f tempest-ServersAaction247Test-1786617390 tempest-ServersAaction247Test-1786617390-project-member] [instance: 3f4fa1ba-a6ad-4e0f-aebc-973816ee09d5] Start building block device mappings for instance. {{(pid=61806) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 921.219322] env[61806]: DEBUG nova.network.neutron [req-32dc92ee-5715-45fc-b145-d7fb1097b284 req-c4a89cb5-2e46-4213-a2e2-36372102d181 service nova] [instance: 6f6ba57a-e2d6-4749-a53a-e263861cb1c0] Updated VIF entry in instance network info cache for port 013c9f0d-65b1-4279-a823-a7f1c14dfb91. {{(pid=61806) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 921.219669] env[61806]: DEBUG nova.network.neutron [req-32dc92ee-5715-45fc-b145-d7fb1097b284 req-c4a89cb5-2e46-4213-a2e2-36372102d181 service nova] [instance: 6f6ba57a-e2d6-4749-a53a-e263861cb1c0] Updating instance_info_cache with network_info: [{"id": "013c9f0d-65b1-4279-a823-a7f1c14dfb91", "address": "fa:16:3e:15:56:4e", "network": {"id": "a7ffda72-e57c-41fa-a7ed-d7b12084bc9d", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-62794865-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "6e391e1267014fafbb8f5a3211299819", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8b3cfeb1-f262-4fd9-b506-8e9c0733e2d8", "external-id": "nsx-vlan-transportzone-119", "segmentation_id": 119, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap013c9f0d-65", "ovs_interfaceid": "013c9f0d-65b1-4279-a823-a7f1c14dfb91", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61806) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 921.300454] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-89206fa9-b514-4ab8-ab09-752e93d4144e {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 921.315830] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6700137f-2809-4a8f-a55c-6df132d4beed {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 921.350457] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c1fabdbb-78ee-4693-8335-bd6ae226ab8a {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 921.359583] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-81386b90-0f74-4d55-9af7-612aba50f048 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 921.374481] env[61806]: DEBUG nova.compute.provider_tree [None req-d76979a9-290f-44f9-bf19-39a08fc72d03 tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Inventory has not changed in ProviderTree for provider: a2858be1-fd22-4e08-979e-87ad25293407 {{(pid=61806) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 921.387616] env[61806]: DEBUG oslo_vmware.api [None req-d718d0b7-b777-42ad-8e19-32c28e9d031c tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Task: {'id': task-1294896, 'name': ResetVM_Task, 'duration_secs': 0.126705} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 921.388621] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-d718d0b7-b777-42ad-8e19-32c28e9d031c tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] [instance: 67d1bdd8-cf8f-4fb5-bd3c-42008e4d60e3] Did hard reboot of VM {{(pid=61806) reboot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1068}} [ 921.388826] env[61806]: DEBUG nova.compute.manager [None req-d718d0b7-b777-42ad-8e19-32c28e9d031c tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] [instance: 67d1bdd8-cf8f-4fb5-bd3c-42008e4d60e3] Checking state {{(pid=61806) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 921.391730] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8e87f62b-0ab7-4df7-8dfe-4afb4e7083ae {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 921.563066] env[61806]: DEBUG nova.network.neutron [None req-531998d5-4ee2-42e3-ac57-d5e3f0c96508 tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] [instance: c486c00d-61fc-4795-9299-e6548b686ee8] Successfully updated port: df6e9523-3657-4405-8d44-5b35cfc6fa9f {{(pid=61806) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 921.617837] env[61806]: DEBUG oslo_vmware.api [None req-70a8784a-38a8-4a3e-8a9f-16278c289170 tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Task: {'id': task-1294895, 'name': PowerOnVM_Task, 'duration_secs': 0.527253} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 921.617837] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [None req-70a8784a-38a8-4a3e-8a9f-16278c289170 tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] [instance: 75f1aded-d1da-4ddc-be29-f1fb93799364] Powered on the VM {{(pid=61806) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 921.617837] env[61806]: DEBUG nova.compute.manager [None req-70a8784a-38a8-4a3e-8a9f-16278c289170 tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] [instance: 75f1aded-d1da-4ddc-be29-f1fb93799364] Checking state {{(pid=61806) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 921.617837] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-afbbbaf3-f7d4-4cf2-bb76-5c332ea4c953 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 921.726620] env[61806]: DEBUG oslo_concurrency.lockutils [req-32dc92ee-5715-45fc-b145-d7fb1097b284 req-c4a89cb5-2e46-4213-a2e2-36372102d181 service nova] Releasing lock "refresh_cache-6f6ba57a-e2d6-4749-a53a-e263861cb1c0" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 921.878091] env[61806]: DEBUG nova.scheduler.client.report [None req-d76979a9-290f-44f9-bf19-39a08fc72d03 tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Inventory has not changed for provider a2858be1-fd22-4e08-979e-87ad25293407 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 138, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61806) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 921.902371] env[61806]: DEBUG oslo_concurrency.lockutils [None req-d718d0b7-b777-42ad-8e19-32c28e9d031c tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Lock "67d1bdd8-cf8f-4fb5-bd3c-42008e4d60e3" "released" by "nova.compute.manager.ComputeManager.reboot_instance..do_reboot_instance" :: held 3.858s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 922.066352] env[61806]: DEBUG oslo_concurrency.lockutils [None req-531998d5-4ee2-42e3-ac57-d5e3f0c96508 tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] Acquiring lock "refresh_cache-c486c00d-61fc-4795-9299-e6548b686ee8" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 922.066459] env[61806]: DEBUG oslo_concurrency.lockutils [None req-531998d5-4ee2-42e3-ac57-d5e3f0c96508 tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] Acquired lock "refresh_cache-c486c00d-61fc-4795-9299-e6548b686ee8" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 922.066759] env[61806]: DEBUG nova.network.neutron [None req-531998d5-4ee2-42e3-ac57-d5e3f0c96508 tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] [instance: c486c00d-61fc-4795-9299-e6548b686ee8] Building network info cache for instance {{(pid=61806) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 922.138450] env[61806]: DEBUG nova.compute.manager [None req-018e57c8-1b55-4515-9361-baa3401dd22f tempest-ServersAaction247Test-1786617390 tempest-ServersAaction247Test-1786617390-project-member] [instance: 3f4fa1ba-a6ad-4e0f-aebc-973816ee09d5] Start spawning the instance on the hypervisor. {{(pid=61806) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 922.145242] env[61806]: DEBUG oslo_concurrency.lockutils [None req-70a8784a-38a8-4a3e-8a9f-16278c289170 tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 922.162201] env[61806]: DEBUG nova.virt.hardware [None req-018e57c8-1b55-4515-9361-baa3401dd22f tempest-ServersAaction247Test-1786617390 tempest-ServersAaction247Test-1786617390-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-15T18:05:10Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-15T18:04:52Z,direct_url=,disk_format='vmdk',id=ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='84ba83fce2c349a6988173c5d6fc3b07',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-15T18:04:53Z,virtual_size=,visibility=), allow threads: False {{(pid=61806) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 922.162520] env[61806]: DEBUG nova.virt.hardware [None req-018e57c8-1b55-4515-9361-baa3401dd22f tempest-ServersAaction247Test-1786617390 tempest-ServersAaction247Test-1786617390-project-member] Flavor limits 0:0:0 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 922.162698] env[61806]: DEBUG nova.virt.hardware [None req-018e57c8-1b55-4515-9361-baa3401dd22f tempest-ServersAaction247Test-1786617390 tempest-ServersAaction247Test-1786617390-project-member] Image limits 0:0:0 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 922.162896] env[61806]: DEBUG nova.virt.hardware [None req-018e57c8-1b55-4515-9361-baa3401dd22f tempest-ServersAaction247Test-1786617390 tempest-ServersAaction247Test-1786617390-project-member] Flavor pref 0:0:0 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 922.163062] env[61806]: DEBUG nova.virt.hardware [None req-018e57c8-1b55-4515-9361-baa3401dd22f tempest-ServersAaction247Test-1786617390 tempest-ServersAaction247Test-1786617390-project-member] Image pref 0:0:0 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 922.163287] env[61806]: DEBUG nova.virt.hardware [None req-018e57c8-1b55-4515-9361-baa3401dd22f tempest-ServersAaction247Test-1786617390 tempest-ServersAaction247Test-1786617390-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 922.163437] env[61806]: DEBUG nova.virt.hardware [None req-018e57c8-1b55-4515-9361-baa3401dd22f tempest-ServersAaction247Test-1786617390 tempest-ServersAaction247Test-1786617390-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61806) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 922.163635] env[61806]: DEBUG nova.virt.hardware [None req-018e57c8-1b55-4515-9361-baa3401dd22f tempest-ServersAaction247Test-1786617390 tempest-ServersAaction247Test-1786617390-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61806) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 922.163829] env[61806]: DEBUG nova.virt.hardware [None req-018e57c8-1b55-4515-9361-baa3401dd22f tempest-ServersAaction247Test-1786617390 tempest-ServersAaction247Test-1786617390-project-member] Got 1 possible topologies {{(pid=61806) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 922.164028] env[61806]: DEBUG nova.virt.hardware [None req-018e57c8-1b55-4515-9361-baa3401dd22f tempest-ServersAaction247Test-1786617390 tempest-ServersAaction247Test-1786617390-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61806) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 922.164216] env[61806]: DEBUG nova.virt.hardware [None req-018e57c8-1b55-4515-9361-baa3401dd22f tempest-ServersAaction247Test-1786617390 tempest-ServersAaction247Test-1786617390-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61806) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 922.165125] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a75603f3-8c58-40bc-b9d4-24acb47cf16f {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 922.173927] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-243ed441-b44f-4a24-baa2-97b96245915d {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 922.188989] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-018e57c8-1b55-4515-9361-baa3401dd22f tempest-ServersAaction247Test-1786617390 tempest-ServersAaction247Test-1786617390-project-member] [instance: 3f4fa1ba-a6ad-4e0f-aebc-973816ee09d5] Instance VIF info [] {{(pid=61806) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 922.194801] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [None req-018e57c8-1b55-4515-9361-baa3401dd22f tempest-ServersAaction247Test-1786617390 tempest-ServersAaction247Test-1786617390-project-member] Creating folder: Project (65f85064536f41b89925f9988ca56034). Parent ref: group-v277609. {{(pid=61806) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 922.195161] env[61806]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-6da97ee2-0cf1-4ca5-85c6-72797817d770 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 922.208658] env[61806]: INFO nova.virt.vmwareapi.vm_util [None req-018e57c8-1b55-4515-9361-baa3401dd22f tempest-ServersAaction247Test-1786617390 tempest-ServersAaction247Test-1786617390-project-member] Created folder: Project (65f85064536f41b89925f9988ca56034) in parent group-v277609. [ 922.208954] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [None req-018e57c8-1b55-4515-9361-baa3401dd22f tempest-ServersAaction247Test-1786617390 tempest-ServersAaction247Test-1786617390-project-member] Creating folder: Instances. Parent ref: group-v277712. {{(pid=61806) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 922.209258] env[61806]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-a16bc3ce-2ba1-4119-add5-071cf9bdb129 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 922.220536] env[61806]: INFO nova.virt.vmwareapi.vm_util [None req-018e57c8-1b55-4515-9361-baa3401dd22f tempest-ServersAaction247Test-1786617390 tempest-ServersAaction247Test-1786617390-project-member] Created folder: Instances in parent group-v277712. [ 922.220800] env[61806]: DEBUG oslo.service.loopingcall [None req-018e57c8-1b55-4515-9361-baa3401dd22f tempest-ServersAaction247Test-1786617390 tempest-ServersAaction247Test-1786617390-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61806) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 922.221063] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 3f4fa1ba-a6ad-4e0f-aebc-973816ee09d5] Creating VM on the ESX host {{(pid=61806) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 922.221591] env[61806]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-89c85efa-daf4-4e23-b59a-cc0f1879c694 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 922.241085] env[61806]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 922.241085] env[61806]: value = "task-1294899" [ 922.241085] env[61806]: _type = "Task" [ 922.241085] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 922.251128] env[61806]: DEBUG oslo_vmware.api [-] Task: {'id': task-1294899, 'name': CreateVM_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 922.388019] env[61806]: DEBUG oslo_concurrency.lockutils [None req-d76979a9-290f-44f9-bf19-39a08fc72d03 tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.278s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 922.388019] env[61806]: DEBUG nova.compute.manager [None req-d76979a9-290f-44f9-bf19-39a08fc72d03 tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] [instance: 380aa63f-46ba-437d-bc29-0fc141ad1cf3] Start building networks asynchronously for instance. {{(pid=61806) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 922.391421] env[61806]: DEBUG oslo_concurrency.lockutils [None req-70a8784a-38a8-4a3e-8a9f-16278c289170 tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: waited 0.246s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 922.394979] env[61806]: DEBUG nova.objects.instance [None req-70a8784a-38a8-4a3e-8a9f-16278c289170 tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] [instance: 75f1aded-d1da-4ddc-be29-f1fb93799364] Trying to apply a migration context that does not seem to be set for this instance {{(pid=61806) apply_migration_context /opt/stack/nova/nova/objects/instance.py:1067}} [ 922.609599] env[61806]: DEBUG nova.network.neutron [None req-531998d5-4ee2-42e3-ac57-d5e3f0c96508 tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] [instance: c486c00d-61fc-4795-9299-e6548b686ee8] Instance cache missing network info. {{(pid=61806) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 922.628747] env[61806]: DEBUG nova.compute.manager [req-529e30ad-8000-42cc-9de5-9ecbcddf93e0 req-d52a861f-2d11-466a-9884-c20e5f3ba61c service nova] [instance: d9caee4a-a00c-4e50-b01c-99a8f796a09f] Received event network-changed-c725ee28-d7a8-4b31-92bc-f63174564f5e {{(pid=61806) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 922.629085] env[61806]: DEBUG nova.compute.manager [req-529e30ad-8000-42cc-9de5-9ecbcddf93e0 req-d52a861f-2d11-466a-9884-c20e5f3ba61c service nova] [instance: d9caee4a-a00c-4e50-b01c-99a8f796a09f] Refreshing instance network info cache due to event network-changed-c725ee28-d7a8-4b31-92bc-f63174564f5e. {{(pid=61806) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 922.629751] env[61806]: DEBUG oslo_concurrency.lockutils [req-529e30ad-8000-42cc-9de5-9ecbcddf93e0 req-d52a861f-2d11-466a-9884-c20e5f3ba61c service nova] Acquiring lock "refresh_cache-d9caee4a-a00c-4e50-b01c-99a8f796a09f" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 922.629949] env[61806]: DEBUG oslo_concurrency.lockutils [req-529e30ad-8000-42cc-9de5-9ecbcddf93e0 req-d52a861f-2d11-466a-9884-c20e5f3ba61c service nova] Acquired lock "refresh_cache-d9caee4a-a00c-4e50-b01c-99a8f796a09f" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 922.630176] env[61806]: DEBUG nova.network.neutron [req-529e30ad-8000-42cc-9de5-9ecbcddf93e0 req-d52a861f-2d11-466a-9884-c20e5f3ba61c service nova] [instance: d9caee4a-a00c-4e50-b01c-99a8f796a09f] Refreshing network info cache for port c725ee28-d7a8-4b31-92bc-f63174564f5e {{(pid=61806) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 922.719965] env[61806]: DEBUG oslo_concurrency.lockutils [None req-33516f06-7375-4e3f-a28f-63fccd97ac4e tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Acquiring lock "75f1aded-d1da-4ddc-be29-f1fb93799364" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 922.719965] env[61806]: DEBUG oslo_concurrency.lockutils [None req-33516f06-7375-4e3f-a28f-63fccd97ac4e tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Lock "75f1aded-d1da-4ddc-be29-f1fb93799364" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 922.719965] env[61806]: DEBUG oslo_concurrency.lockutils [None req-33516f06-7375-4e3f-a28f-63fccd97ac4e tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Acquiring lock "75f1aded-d1da-4ddc-be29-f1fb93799364-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 922.719965] env[61806]: DEBUG oslo_concurrency.lockutils [None req-33516f06-7375-4e3f-a28f-63fccd97ac4e tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Lock "75f1aded-d1da-4ddc-be29-f1fb93799364-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 922.719965] env[61806]: DEBUG oslo_concurrency.lockutils [None req-33516f06-7375-4e3f-a28f-63fccd97ac4e tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Lock "75f1aded-d1da-4ddc-be29-f1fb93799364-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 922.722647] env[61806]: INFO nova.compute.manager [None req-33516f06-7375-4e3f-a28f-63fccd97ac4e tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] [instance: 75f1aded-d1da-4ddc-be29-f1fb93799364] Terminating instance [ 922.729750] env[61806]: DEBUG nova.compute.manager [None req-33516f06-7375-4e3f-a28f-63fccd97ac4e tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] [instance: 75f1aded-d1da-4ddc-be29-f1fb93799364] Start destroying the instance on the hypervisor. {{(pid=61806) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 922.729878] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-33516f06-7375-4e3f-a28f-63fccd97ac4e tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] [instance: 75f1aded-d1da-4ddc-be29-f1fb93799364] Destroying instance {{(pid=61806) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 922.730723] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-87001422-d5fd-40e1-9b17-93722a4fb48a {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 922.740778] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [None req-33516f06-7375-4e3f-a28f-63fccd97ac4e tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] [instance: 75f1aded-d1da-4ddc-be29-f1fb93799364] Powering off the VM {{(pid=61806) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 922.741214] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-25b59c3e-4054-41b9-be89-2601de30af48 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 922.758966] env[61806]: DEBUG oslo_vmware.api [-] Task: {'id': task-1294899, 'name': CreateVM_Task, 'duration_secs': 0.402703} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 922.762708] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 3f4fa1ba-a6ad-4e0f-aebc-973816ee09d5] Created VM on the ESX host {{(pid=61806) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 922.762708] env[61806]: DEBUG oslo_vmware.api [None req-33516f06-7375-4e3f-a28f-63fccd97ac4e tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Waiting for the task: (returnval){ [ 922.762708] env[61806]: value = "task-1294900" [ 922.762708] env[61806]: _type = "Task" [ 922.762708] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 922.763159] env[61806]: DEBUG oslo_concurrency.lockutils [None req-018e57c8-1b55-4515-9361-baa3401dd22f tempest-ServersAaction247Test-1786617390 tempest-ServersAaction247Test-1786617390-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 922.763248] env[61806]: DEBUG oslo_concurrency.lockutils [None req-018e57c8-1b55-4515-9361-baa3401dd22f tempest-ServersAaction247Test-1786617390 tempest-ServersAaction247Test-1786617390-project-member] Acquired lock "[datastore1] devstack-image-cache_base/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 922.763681] env[61806]: DEBUG oslo_concurrency.lockutils [None req-018e57c8-1b55-4515-9361-baa3401dd22f tempest-ServersAaction247Test-1786617390 tempest-ServersAaction247Test-1786617390-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 922.764915] env[61806]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-2f82ae76-7e03-4042-9c94-a59597157389 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 922.774587] env[61806]: DEBUG oslo_vmware.api [None req-018e57c8-1b55-4515-9361-baa3401dd22f tempest-ServersAaction247Test-1786617390 tempest-ServersAaction247Test-1786617390-project-member] Waiting for the task: (returnval){ [ 922.774587] env[61806]: value = "session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]520fdf7b-585a-f4b8-cca5-4a5aed5d979b" [ 922.774587] env[61806]: _type = "Task" [ 922.774587] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 922.782083] env[61806]: DEBUG oslo_vmware.api [None req-33516f06-7375-4e3f-a28f-63fccd97ac4e tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Task: {'id': task-1294900, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 922.790884] env[61806]: DEBUG oslo_vmware.api [None req-018e57c8-1b55-4515-9361-baa3401dd22f tempest-ServersAaction247Test-1786617390 tempest-ServersAaction247Test-1786617390-project-member] Task: {'id': session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]520fdf7b-585a-f4b8-cca5-4a5aed5d979b, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 922.790884] env[61806]: DEBUG nova.network.neutron [None req-531998d5-4ee2-42e3-ac57-d5e3f0c96508 tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] [instance: c486c00d-61fc-4795-9299-e6548b686ee8] Updating instance_info_cache with network_info: [{"id": "df6e9523-3657-4405-8d44-5b35cfc6fa9f", "address": "fa:16:3e:85:ad:44", "network": {"id": "f280a01b-7000-4e54-ab9a-20eab156a410", "bridge": "br-int", "label": "tempest-ServersTestJSON-2080718146-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "73ebcc4d79d248efb653a084a5e44302", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "43798f54-0c07-4417-a23f-58bb6b7e204b", "external-id": "nsx-vlan-transportzone-571", "segmentation_id": 571, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapdf6e9523-36", "ovs_interfaceid": "df6e9523-3657-4405-8d44-5b35cfc6fa9f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61806) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 922.900049] env[61806]: DEBUG nova.compute.utils [None req-d76979a9-290f-44f9-bf19-39a08fc72d03 tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Using /dev/sd instead of None {{(pid=61806) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 922.904390] env[61806]: DEBUG nova.compute.manager [None req-d76979a9-290f-44f9-bf19-39a08fc72d03 tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] [instance: 380aa63f-46ba-437d-bc29-0fc141ad1cf3] Allocating IP information in the background. {{(pid=61806) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 922.904614] env[61806]: DEBUG nova.network.neutron [None req-d76979a9-290f-44f9-bf19-39a08fc72d03 tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] [instance: 380aa63f-46ba-437d-bc29-0fc141ad1cf3] allocate_for_instance() {{(pid=61806) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 922.951254] env[61806]: DEBUG oslo_concurrency.lockutils [None req-bdb89e04-5c89-42d4-937f-829d8aee12b6 tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] Acquiring lock "fd9e520a-2e7b-43d4-b84d-8933c32cc6a6" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 922.951518] env[61806]: DEBUG oslo_concurrency.lockutils [None req-bdb89e04-5c89-42d4-937f-829d8aee12b6 tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] Lock "fd9e520a-2e7b-43d4-b84d-8933c32cc6a6" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 922.954398] env[61806]: DEBUG nova.policy [None req-d76979a9-290f-44f9-bf19-39a08fc72d03 tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '2b86d596e89645c2ab31332afdcbc3c2', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '7ab95fb72b5d46c3b6c7bebbccf897cd', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61806) authorize /opt/stack/nova/nova/policy.py:201}} [ 923.219447] env[61806]: DEBUG nova.network.neutron [None req-d76979a9-290f-44f9-bf19-39a08fc72d03 tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] [instance: 380aa63f-46ba-437d-bc29-0fc141ad1cf3] Successfully created port: 2c4c1295-84da-4bdf-8e6d-742b6358064c {{(pid=61806) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 923.277141] env[61806]: DEBUG oslo_vmware.api [None req-33516f06-7375-4e3f-a28f-63fccd97ac4e tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Task: {'id': task-1294900, 'name': PowerOffVM_Task, 'duration_secs': 0.212576} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 923.281709] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [None req-33516f06-7375-4e3f-a28f-63fccd97ac4e tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] [instance: 75f1aded-d1da-4ddc-be29-f1fb93799364] Powered off the VM {{(pid=61806) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 923.281949] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-33516f06-7375-4e3f-a28f-63fccd97ac4e tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] [instance: 75f1aded-d1da-4ddc-be29-f1fb93799364] Unregistering the VM {{(pid=61806) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 923.285229] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-acc1ce91-e1d0-4aa3-8aa2-13408f945474 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 923.287216] env[61806]: DEBUG oslo_concurrency.lockutils [None req-0c4ed53e-d274-4427-995e-e2d8d9d3b2cc tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Acquiring lock "676f18b3-3da1-4a53-b96d-fe64ee8f9101" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 923.287216] env[61806]: DEBUG oslo_concurrency.lockutils [None req-0c4ed53e-d274-4427-995e-e2d8d9d3b2cc tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Lock "676f18b3-3da1-4a53-b96d-fe64ee8f9101" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 923.299450] env[61806]: DEBUG oslo_concurrency.lockutils [None req-531998d5-4ee2-42e3-ac57-d5e3f0c96508 tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] Releasing lock "refresh_cache-c486c00d-61fc-4795-9299-e6548b686ee8" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 923.299670] env[61806]: DEBUG nova.compute.manager [None req-531998d5-4ee2-42e3-ac57-d5e3f0c96508 tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] [instance: c486c00d-61fc-4795-9299-e6548b686ee8] Instance network_info: |[{"id": "df6e9523-3657-4405-8d44-5b35cfc6fa9f", "address": "fa:16:3e:85:ad:44", "network": {"id": "f280a01b-7000-4e54-ab9a-20eab156a410", "bridge": "br-int", "label": "tempest-ServersTestJSON-2080718146-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "73ebcc4d79d248efb653a084a5e44302", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "43798f54-0c07-4417-a23f-58bb6b7e204b", "external-id": "nsx-vlan-transportzone-571", "segmentation_id": 571, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapdf6e9523-36", "ovs_interfaceid": "df6e9523-3657-4405-8d44-5b35cfc6fa9f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61806) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 923.299976] env[61806]: DEBUG oslo_vmware.api [None req-018e57c8-1b55-4515-9361-baa3401dd22f tempest-ServersAaction247Test-1786617390 tempest-ServersAaction247Test-1786617390-project-member] Task: {'id': session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]520fdf7b-585a-f4b8-cca5-4a5aed5d979b, 'name': SearchDatastore_Task, 'duration_secs': 0.013784} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 923.300477] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-531998d5-4ee2-42e3-ac57-d5e3f0c96508 tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] [instance: c486c00d-61fc-4795-9299-e6548b686ee8] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:85:ad:44', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '43798f54-0c07-4417-a23f-58bb6b7e204b', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'df6e9523-3657-4405-8d44-5b35cfc6fa9f', 'vif_model': 'vmxnet3'}] {{(pid=61806) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 923.313900] env[61806]: DEBUG oslo.service.loopingcall [None req-531998d5-4ee2-42e3-ac57-d5e3f0c96508 tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61806) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 923.314220] env[61806]: DEBUG oslo_concurrency.lockutils [None req-018e57c8-1b55-4515-9361-baa3401dd22f tempest-ServersAaction247Test-1786617390 tempest-ServersAaction247Test-1786617390-project-member] Releasing lock "[datastore1] devstack-image-cache_base/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 923.314452] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-018e57c8-1b55-4515-9361-baa3401dd22f tempest-ServersAaction247Test-1786617390 tempest-ServersAaction247Test-1786617390-project-member] [instance: 3f4fa1ba-a6ad-4e0f-aebc-973816ee09d5] Processing image ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa {{(pid=61806) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 923.314730] env[61806]: DEBUG oslo_concurrency.lockutils [None req-018e57c8-1b55-4515-9361-baa3401dd22f tempest-ServersAaction247Test-1786617390 tempest-ServersAaction247Test-1786617390-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa.vmdk" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 923.314897] env[61806]: DEBUG oslo_concurrency.lockutils [None req-018e57c8-1b55-4515-9361-baa3401dd22f tempest-ServersAaction247Test-1786617390 tempest-ServersAaction247Test-1786617390-project-member] Acquired lock "[datastore1] devstack-image-cache_base/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa.vmdk" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 923.315171] env[61806]: DEBUG nova.virt.vmwareapi.ds_util [None req-018e57c8-1b55-4515-9361-baa3401dd22f tempest-ServersAaction247Test-1786617390 tempest-ServersAaction247Test-1786617390-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61806) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 923.319069] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: c486c00d-61fc-4795-9299-e6548b686ee8] Creating VM on the ESX host {{(pid=61806) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 923.319357] env[61806]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-7219e904-650d-430e-813f-3f331ac5101e {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 923.321892] env[61806]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-29f958f5-b773-41c1-a637-d46b8c78d766 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 923.344589] env[61806]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 923.344589] env[61806]: value = "task-1294902" [ 923.344589] env[61806]: _type = "Task" [ 923.344589] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 923.353224] env[61806]: DEBUG nova.virt.vmwareapi.ds_util [None req-018e57c8-1b55-4515-9361-baa3401dd22f tempest-ServersAaction247Test-1786617390 tempest-ServersAaction247Test-1786617390-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61806) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 923.353444] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-018e57c8-1b55-4515-9361-baa3401dd22f tempest-ServersAaction247Test-1786617390 tempest-ServersAaction247Test-1786617390-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61806) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 923.354710] env[61806]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-6611e522-f634-4a71-844f-91d9d45b29d3 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 923.359197] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-33516f06-7375-4e3f-a28f-63fccd97ac4e tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] [instance: 75f1aded-d1da-4ddc-be29-f1fb93799364] Unregistered the VM {{(pid=61806) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 923.359501] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-33516f06-7375-4e3f-a28f-63fccd97ac4e tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] [instance: 75f1aded-d1da-4ddc-be29-f1fb93799364] Deleting contents of the VM from datastore datastore1 {{(pid=61806) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 923.359691] env[61806]: DEBUG nova.virt.vmwareapi.ds_util [None req-33516f06-7375-4e3f-a28f-63fccd97ac4e tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Deleting the datastore file [datastore1] 75f1aded-d1da-4ddc-be29-f1fb93799364 {{(pid=61806) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 923.364034] env[61806]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-650fb717-208b-414b-af65-ca46c776c21c {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 923.365024] env[61806]: DEBUG oslo_vmware.api [-] Task: {'id': task-1294902, 'name': CreateVM_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 923.366352] env[61806]: DEBUG oslo_vmware.api [None req-018e57c8-1b55-4515-9361-baa3401dd22f tempest-ServersAaction247Test-1786617390 tempest-ServersAaction247Test-1786617390-project-member] Waiting for the task: (returnval){ [ 923.366352] env[61806]: value = "session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]52d40a09-622f-4453-9bf1-847b6d3f3cc1" [ 923.366352] env[61806]: _type = "Task" [ 923.366352] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 923.373132] env[61806]: DEBUG oslo_vmware.api [None req-33516f06-7375-4e3f-a28f-63fccd97ac4e tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Waiting for the task: (returnval){ [ 923.373132] env[61806]: value = "task-1294903" [ 923.373132] env[61806]: _type = "Task" [ 923.373132] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 923.379824] env[61806]: DEBUG nova.network.neutron [req-529e30ad-8000-42cc-9de5-9ecbcddf93e0 req-d52a861f-2d11-466a-9884-c20e5f3ba61c service nova] [instance: d9caee4a-a00c-4e50-b01c-99a8f796a09f] Updated VIF entry in instance network info cache for port c725ee28-d7a8-4b31-92bc-f63174564f5e. {{(pid=61806) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 923.380154] env[61806]: DEBUG nova.network.neutron [req-529e30ad-8000-42cc-9de5-9ecbcddf93e0 req-d52a861f-2d11-466a-9884-c20e5f3ba61c service nova] [instance: d9caee4a-a00c-4e50-b01c-99a8f796a09f] Updating instance_info_cache with network_info: [{"id": "c725ee28-d7a8-4b31-92bc-f63174564f5e", "address": "fa:16:3e:3a:18:8e", "network": {"id": "a7ffda72-e57c-41fa-a7ed-d7b12084bc9d", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-62794865-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "6e391e1267014fafbb8f5a3211299819", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8b3cfeb1-f262-4fd9-b506-8e9c0733e2d8", "external-id": "nsx-vlan-transportzone-119", "segmentation_id": 119, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc725ee28-d7", "ovs_interfaceid": "c725ee28-d7a8-4b31-92bc-f63174564f5e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61806) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 923.381416] env[61806]: DEBUG oslo_vmware.api [None req-018e57c8-1b55-4515-9361-baa3401dd22f tempest-ServersAaction247Test-1786617390 tempest-ServersAaction247Test-1786617390-project-member] Task: {'id': session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]52d40a09-622f-4453-9bf1-847b6d3f3cc1, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 923.387175] env[61806]: DEBUG oslo_vmware.api [None req-33516f06-7375-4e3f-a28f-63fccd97ac4e tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Task: {'id': task-1294903, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 923.408934] env[61806]: DEBUG oslo_concurrency.lockutils [None req-70a8784a-38a8-4a3e-8a9f-16278c289170 tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: held 1.017s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 923.410178] env[61806]: DEBUG nova.compute.manager [None req-d76979a9-290f-44f9-bf19-39a08fc72d03 tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] [instance: 380aa63f-46ba-437d-bc29-0fc141ad1cf3] Start building block device mappings for instance. {{(pid=61806) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 923.440328] env[61806]: DEBUG oslo_service.periodic_task [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=61806) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 923.440661] env[61806]: DEBUG oslo_service.periodic_task [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=61806) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 923.444036] env[61806]: DEBUG nova.compute.manager [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Starting heal instance info cache {{(pid=61806) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9926}} [ 923.457568] env[61806]: DEBUG nova.compute.manager [None req-bdb89e04-5c89-42d4-937f-829d8aee12b6 tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] [instance: fd9e520a-2e7b-43d4-b84d-8933c32cc6a6] Starting instance... {{(pid=61806) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 923.794131] env[61806]: DEBUG nova.compute.manager [None req-0c4ed53e-d274-4427-995e-e2d8d9d3b2cc tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] [instance: 676f18b3-3da1-4a53-b96d-fe64ee8f9101] Starting instance... {{(pid=61806) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 923.855330] env[61806]: DEBUG oslo_vmware.api [-] Task: {'id': task-1294902, 'name': CreateVM_Task, 'duration_secs': 0.491659} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 923.855475] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: c486c00d-61fc-4795-9299-e6548b686ee8] Created VM on the ESX host {{(pid=61806) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 923.856158] env[61806]: DEBUG oslo_concurrency.lockutils [None req-531998d5-4ee2-42e3-ac57-d5e3f0c96508 tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 923.856341] env[61806]: DEBUG oslo_concurrency.lockutils [None req-531998d5-4ee2-42e3-ac57-d5e3f0c96508 tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] Acquired lock "[datastore1] devstack-image-cache_base/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 923.856682] env[61806]: DEBUG oslo_concurrency.lockutils [None req-531998d5-4ee2-42e3-ac57-d5e3f0c96508 tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 923.856940] env[61806]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-47cea22e-35fd-4f25-a806-28e398a17707 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 923.862298] env[61806]: DEBUG oslo_vmware.api [None req-531998d5-4ee2-42e3-ac57-d5e3f0c96508 tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] Waiting for the task: (returnval){ [ 923.862298] env[61806]: value = "session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]52d6cffc-3475-9043-81a1-fd83d058cdec" [ 923.862298] env[61806]: _type = "Task" [ 923.862298] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 923.873097] env[61806]: DEBUG oslo_vmware.api [None req-531998d5-4ee2-42e3-ac57-d5e3f0c96508 tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] Task: {'id': session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]52d6cffc-3475-9043-81a1-fd83d058cdec, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 923.879752] env[61806]: DEBUG oslo_vmware.api [None req-018e57c8-1b55-4515-9361-baa3401dd22f tempest-ServersAaction247Test-1786617390 tempest-ServersAaction247Test-1786617390-project-member] Task: {'id': session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]52d40a09-622f-4453-9bf1-847b6d3f3cc1, 'name': SearchDatastore_Task, 'duration_secs': 0.017309} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 923.880835] env[61806]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-aad58fc7-bb20-425a-88ab-1f517d299f11 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 923.885973] env[61806]: DEBUG oslo_concurrency.lockutils [req-529e30ad-8000-42cc-9de5-9ecbcddf93e0 req-d52a861f-2d11-466a-9884-c20e5f3ba61c service nova] Releasing lock "refresh_cache-d9caee4a-a00c-4e50-b01c-99a8f796a09f" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 923.886238] env[61806]: DEBUG nova.compute.manager [req-529e30ad-8000-42cc-9de5-9ecbcddf93e0 req-d52a861f-2d11-466a-9884-c20e5f3ba61c service nova] [instance: c486c00d-61fc-4795-9299-e6548b686ee8] Received event network-vif-plugged-df6e9523-3657-4405-8d44-5b35cfc6fa9f {{(pid=61806) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 923.886437] env[61806]: DEBUG oslo_concurrency.lockutils [req-529e30ad-8000-42cc-9de5-9ecbcddf93e0 req-d52a861f-2d11-466a-9884-c20e5f3ba61c service nova] Acquiring lock "c486c00d-61fc-4795-9299-e6548b686ee8-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 923.886643] env[61806]: DEBUG oslo_concurrency.lockutils [req-529e30ad-8000-42cc-9de5-9ecbcddf93e0 req-d52a861f-2d11-466a-9884-c20e5f3ba61c service nova] Lock "c486c00d-61fc-4795-9299-e6548b686ee8-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 923.886811] env[61806]: DEBUG oslo_concurrency.lockutils [req-529e30ad-8000-42cc-9de5-9ecbcddf93e0 req-d52a861f-2d11-466a-9884-c20e5f3ba61c service nova] Lock "c486c00d-61fc-4795-9299-e6548b686ee8-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 923.886984] env[61806]: DEBUG nova.compute.manager [req-529e30ad-8000-42cc-9de5-9ecbcddf93e0 req-d52a861f-2d11-466a-9884-c20e5f3ba61c service nova] [instance: c486c00d-61fc-4795-9299-e6548b686ee8] No waiting events found dispatching network-vif-plugged-df6e9523-3657-4405-8d44-5b35cfc6fa9f {{(pid=61806) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 923.887176] env[61806]: WARNING nova.compute.manager [req-529e30ad-8000-42cc-9de5-9ecbcddf93e0 req-d52a861f-2d11-466a-9884-c20e5f3ba61c service nova] [instance: c486c00d-61fc-4795-9299-e6548b686ee8] Received unexpected event network-vif-plugged-df6e9523-3657-4405-8d44-5b35cfc6fa9f for instance with vm_state building and task_state spawning. [ 923.887350] env[61806]: DEBUG nova.compute.manager [req-529e30ad-8000-42cc-9de5-9ecbcddf93e0 req-d52a861f-2d11-466a-9884-c20e5f3ba61c service nova] [instance: c486c00d-61fc-4795-9299-e6548b686ee8] Received event network-changed-df6e9523-3657-4405-8d44-5b35cfc6fa9f {{(pid=61806) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 923.887514] env[61806]: DEBUG nova.compute.manager [req-529e30ad-8000-42cc-9de5-9ecbcddf93e0 req-d52a861f-2d11-466a-9884-c20e5f3ba61c service nova] [instance: c486c00d-61fc-4795-9299-e6548b686ee8] Refreshing instance network info cache due to event network-changed-df6e9523-3657-4405-8d44-5b35cfc6fa9f. {{(pid=61806) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 923.887704] env[61806]: DEBUG oslo_concurrency.lockutils [req-529e30ad-8000-42cc-9de5-9ecbcddf93e0 req-d52a861f-2d11-466a-9884-c20e5f3ba61c service nova] Acquiring lock "refresh_cache-c486c00d-61fc-4795-9299-e6548b686ee8" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 923.887851] env[61806]: DEBUG oslo_concurrency.lockutils [req-529e30ad-8000-42cc-9de5-9ecbcddf93e0 req-d52a861f-2d11-466a-9884-c20e5f3ba61c service nova] Acquired lock "refresh_cache-c486c00d-61fc-4795-9299-e6548b686ee8" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 923.888021] env[61806]: DEBUG nova.network.neutron [req-529e30ad-8000-42cc-9de5-9ecbcddf93e0 req-d52a861f-2d11-466a-9884-c20e5f3ba61c service nova] [instance: c486c00d-61fc-4795-9299-e6548b686ee8] Refreshing network info cache for port df6e9523-3657-4405-8d44-5b35cfc6fa9f {{(pid=61806) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 923.889128] env[61806]: DEBUG oslo_vmware.api [None req-33516f06-7375-4e3f-a28f-63fccd97ac4e tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Task: {'id': task-1294903, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.355351} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 923.889892] env[61806]: DEBUG nova.virt.vmwareapi.ds_util [None req-33516f06-7375-4e3f-a28f-63fccd97ac4e tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Deleted the datastore file {{(pid=61806) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 923.890116] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-33516f06-7375-4e3f-a28f-63fccd97ac4e tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] [instance: 75f1aded-d1da-4ddc-be29-f1fb93799364] Deleted contents of the VM from datastore datastore1 {{(pid=61806) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 923.890307] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-33516f06-7375-4e3f-a28f-63fccd97ac4e tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] [instance: 75f1aded-d1da-4ddc-be29-f1fb93799364] Instance destroyed {{(pid=61806) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 923.890487] env[61806]: INFO nova.compute.manager [None req-33516f06-7375-4e3f-a28f-63fccd97ac4e tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] [instance: 75f1aded-d1da-4ddc-be29-f1fb93799364] Took 1.16 seconds to destroy the instance on the hypervisor. [ 923.890732] env[61806]: DEBUG oslo.service.loopingcall [None req-33516f06-7375-4e3f-a28f-63fccd97ac4e tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61806) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 923.890930] env[61806]: DEBUG nova.compute.manager [-] [instance: 75f1aded-d1da-4ddc-be29-f1fb93799364] Deallocating network for instance {{(pid=61806) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 923.891035] env[61806]: DEBUG nova.network.neutron [-] [instance: 75f1aded-d1da-4ddc-be29-f1fb93799364] deallocate_for_instance() {{(pid=61806) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 923.894144] env[61806]: DEBUG oslo_vmware.api [None req-018e57c8-1b55-4515-9361-baa3401dd22f tempest-ServersAaction247Test-1786617390 tempest-ServersAaction247Test-1786617390-project-member] Waiting for the task: (returnval){ [ 923.894144] env[61806]: value = "session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]523c57b4-5a8d-f320-b538-4b3eb528df32" [ 923.894144] env[61806]: _type = "Task" [ 923.894144] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 923.904261] env[61806]: DEBUG oslo_vmware.api [None req-018e57c8-1b55-4515-9361-baa3401dd22f tempest-ServersAaction247Test-1786617390 tempest-ServersAaction247Test-1786617390-project-member] Task: {'id': session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]523c57b4-5a8d-f320-b538-4b3eb528df32, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 923.978965] env[61806]: DEBUG oslo_concurrency.lockutils [None req-bdb89e04-5c89-42d4-937f-829d8aee12b6 tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 923.979267] env[61806]: DEBUG oslo_concurrency.lockutils [None req-bdb89e04-5c89-42d4-937f-829d8aee12b6 tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.000s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 923.980875] env[61806]: INFO nova.compute.claims [None req-bdb89e04-5c89-42d4-937f-829d8aee12b6 tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] [instance: fd9e520a-2e7b-43d4-b84d-8933c32cc6a6] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 924.136191] env[61806]: DEBUG nova.compute.manager [req-3f07e10e-8ba6-4aa8-9622-7ba300ecf554 req-da2e56bd-660c-4a5e-84ae-fd5224437985 service nova] [instance: 75f1aded-d1da-4ddc-be29-f1fb93799364] Received event network-vif-deleted-a06a82f0-c5d3-48b7-bd4b-d2725bd7eaec {{(pid=61806) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 924.136403] env[61806]: INFO nova.compute.manager [req-3f07e10e-8ba6-4aa8-9622-7ba300ecf554 req-da2e56bd-660c-4a5e-84ae-fd5224437985 service nova] [instance: 75f1aded-d1da-4ddc-be29-f1fb93799364] Neutron deleted interface a06a82f0-c5d3-48b7-bd4b-d2725bd7eaec; detaching it from the instance and deleting it from the info cache [ 924.136581] env[61806]: DEBUG nova.network.neutron [req-3f07e10e-8ba6-4aa8-9622-7ba300ecf554 req-da2e56bd-660c-4a5e-84ae-fd5224437985 service nova] [instance: 75f1aded-d1da-4ddc-be29-f1fb93799364] Updating instance_info_cache with network_info: [] {{(pid=61806) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 924.316701] env[61806]: DEBUG oslo_concurrency.lockutils [None req-0c4ed53e-d274-4427-995e-e2d8d9d3b2cc tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 924.373925] env[61806]: DEBUG oslo_vmware.api [None req-531998d5-4ee2-42e3-ac57-d5e3f0c96508 tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] Task: {'id': session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]52d6cffc-3475-9043-81a1-fd83d058cdec, 'name': SearchDatastore_Task, 'duration_secs': 0.022067} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 924.374442] env[61806]: DEBUG oslo_concurrency.lockutils [None req-531998d5-4ee2-42e3-ac57-d5e3f0c96508 tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] Releasing lock "[datastore1] devstack-image-cache_base/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 924.374850] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-531998d5-4ee2-42e3-ac57-d5e3f0c96508 tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] [instance: c486c00d-61fc-4795-9299-e6548b686ee8] Processing image ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa {{(pid=61806) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 924.375179] env[61806]: DEBUG oslo_concurrency.lockutils [None req-531998d5-4ee2-42e3-ac57-d5e3f0c96508 tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa.vmdk" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 924.407680] env[61806]: DEBUG oslo_vmware.api [None req-018e57c8-1b55-4515-9361-baa3401dd22f tempest-ServersAaction247Test-1786617390 tempest-ServersAaction247Test-1786617390-project-member] Task: {'id': session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]523c57b4-5a8d-f320-b538-4b3eb528df32, 'name': SearchDatastore_Task, 'duration_secs': 0.021816} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 924.407929] env[61806]: DEBUG oslo_concurrency.lockutils [None req-018e57c8-1b55-4515-9361-baa3401dd22f tempest-ServersAaction247Test-1786617390 tempest-ServersAaction247Test-1786617390-project-member] Releasing lock "[datastore1] devstack-image-cache_base/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa.vmdk" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 924.408286] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [None req-018e57c8-1b55-4515-9361-baa3401dd22f tempest-ServersAaction247Test-1786617390 tempest-ServersAaction247Test-1786617390-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa.vmdk to [datastore1] 3f4fa1ba-a6ad-4e0f-aebc-973816ee09d5/3f4fa1ba-a6ad-4e0f-aebc-973816ee09d5.vmdk {{(pid=61806) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 924.408661] env[61806]: DEBUG oslo_concurrency.lockutils [None req-531998d5-4ee2-42e3-ac57-d5e3f0c96508 tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] Acquired lock "[datastore1] devstack-image-cache_base/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa.vmdk" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 924.408916] env[61806]: DEBUG nova.virt.vmwareapi.ds_util [None req-531998d5-4ee2-42e3-ac57-d5e3f0c96508 tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61806) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 924.409355] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-ea929e86-1668-4b1c-ad50-88a4b75359eb {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 924.411684] env[61806]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-87af46b8-0a69-4483-8e44-e679314304a2 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 924.422480] env[61806]: DEBUG nova.compute.manager [None req-d76979a9-290f-44f9-bf19-39a08fc72d03 tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] [instance: 380aa63f-46ba-437d-bc29-0fc141ad1cf3] Start spawning the instance on the hypervisor. {{(pid=61806) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 924.425256] env[61806]: DEBUG oslo_vmware.api [None req-018e57c8-1b55-4515-9361-baa3401dd22f tempest-ServersAaction247Test-1786617390 tempest-ServersAaction247Test-1786617390-project-member] Waiting for the task: (returnval){ [ 924.425256] env[61806]: value = "task-1294904" [ 924.425256] env[61806]: _type = "Task" [ 924.425256] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 924.430350] env[61806]: DEBUG nova.virt.vmwareapi.ds_util [None req-531998d5-4ee2-42e3-ac57-d5e3f0c96508 tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61806) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 924.431023] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-531998d5-4ee2-42e3-ac57-d5e3f0c96508 tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61806) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 924.435882] env[61806]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-732272da-4ecf-40d4-9972-355c87d5bede {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 924.448006] env[61806]: DEBUG oslo_vmware.api [None req-018e57c8-1b55-4515-9361-baa3401dd22f tempest-ServersAaction247Test-1786617390 tempest-ServersAaction247Test-1786617390-project-member] Task: {'id': task-1294904, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 924.451578] env[61806]: DEBUG oslo_vmware.api [None req-531998d5-4ee2-42e3-ac57-d5e3f0c96508 tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] Waiting for the task: (returnval){ [ 924.451578] env[61806]: value = "session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]52ef7bd7-7d7f-3828-f9f2-2e3311e06853" [ 924.451578] env[61806]: _type = "Task" [ 924.451578] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 924.459417] env[61806]: DEBUG nova.virt.hardware [None req-d76979a9-290f-44f9-bf19-39a08fc72d03 tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-15T18:05:10Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-15T18:04:52Z,direct_url=,disk_format='vmdk',id=ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='84ba83fce2c349a6988173c5d6fc3b07',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-15T18:04:53Z,virtual_size=,visibility=), allow threads: False {{(pid=61806) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 924.459736] env[61806]: DEBUG nova.virt.hardware [None req-d76979a9-290f-44f9-bf19-39a08fc72d03 tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Flavor limits 0:0:0 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 924.459955] env[61806]: DEBUG nova.virt.hardware [None req-d76979a9-290f-44f9-bf19-39a08fc72d03 tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Image limits 0:0:0 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 924.460224] env[61806]: DEBUG nova.virt.hardware [None req-d76979a9-290f-44f9-bf19-39a08fc72d03 tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Flavor pref 0:0:0 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 924.460434] env[61806]: DEBUG nova.virt.hardware [None req-d76979a9-290f-44f9-bf19-39a08fc72d03 tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Image pref 0:0:0 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 924.460645] env[61806]: DEBUG nova.virt.hardware [None req-d76979a9-290f-44f9-bf19-39a08fc72d03 tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 924.460950] env[61806]: DEBUG nova.virt.hardware [None req-d76979a9-290f-44f9-bf19-39a08fc72d03 tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61806) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 924.461364] env[61806]: DEBUG nova.virt.hardware [None req-d76979a9-290f-44f9-bf19-39a08fc72d03 tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61806) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 924.461649] env[61806]: DEBUG nova.virt.hardware [None req-d76979a9-290f-44f9-bf19-39a08fc72d03 tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Got 1 possible topologies {{(pid=61806) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 924.461869] env[61806]: DEBUG nova.virt.hardware [None req-d76979a9-290f-44f9-bf19-39a08fc72d03 tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61806) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 924.462129] env[61806]: DEBUG nova.virt.hardware [None req-d76979a9-290f-44f9-bf19-39a08fc72d03 tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61806) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 924.463216] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-69bcb1c2-9d5d-4bad-aeb9-e2a7dd25f77e {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 924.472324] env[61806]: DEBUG oslo_vmware.api [None req-531998d5-4ee2-42e3-ac57-d5e3f0c96508 tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] Task: {'id': session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]52ef7bd7-7d7f-3828-f9f2-2e3311e06853, 'name': SearchDatastore_Task, 'duration_secs': 0.011456} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 924.474031] env[61806]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f2f4ea3a-a15d-495e-93d7-5e90001a23d8 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 924.481353] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-09bbfa85-f6a6-4748-bf05-910c9c2da8a7 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 924.493439] env[61806]: DEBUG oslo_vmware.api [None req-531998d5-4ee2-42e3-ac57-d5e3f0c96508 tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] Waiting for the task: (returnval){ [ 924.493439] env[61806]: value = "session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]5256befc-ce1d-d9d9-e0d2-cd433669537a" [ 924.493439] env[61806]: _type = "Task" [ 924.493439] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 924.516368] env[61806]: DEBUG oslo_vmware.api [None req-531998d5-4ee2-42e3-ac57-d5e3f0c96508 tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] Task: {'id': session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]5256befc-ce1d-d9d9-e0d2-cd433669537a, 'name': SearchDatastore_Task, 'duration_secs': 0.017192} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 924.519481] env[61806]: DEBUG oslo_concurrency.lockutils [None req-531998d5-4ee2-42e3-ac57-d5e3f0c96508 tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] Releasing lock "[datastore1] devstack-image-cache_base/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa.vmdk" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 924.519828] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [None req-531998d5-4ee2-42e3-ac57-d5e3f0c96508 tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa.vmdk to [datastore1] c486c00d-61fc-4795-9299-e6548b686ee8/c486c00d-61fc-4795-9299-e6548b686ee8.vmdk {{(pid=61806) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 924.520504] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-96d9aa02-2699-4e8b-b1c1-1f42ac3ef208 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 924.530749] env[61806]: DEBUG oslo_vmware.api [None req-531998d5-4ee2-42e3-ac57-d5e3f0c96508 tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] Waiting for the task: (returnval){ [ 924.530749] env[61806]: value = "task-1294905" [ 924.530749] env[61806]: _type = "Task" [ 924.530749] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 924.540433] env[61806]: DEBUG oslo_vmware.api [None req-531998d5-4ee2-42e3-ac57-d5e3f0c96508 tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] Task: {'id': task-1294905, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 924.613887] env[61806]: DEBUG nova.network.neutron [-] [instance: 75f1aded-d1da-4ddc-be29-f1fb93799364] Updating instance_info_cache with network_info: [] {{(pid=61806) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 924.639703] env[61806]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-7c7d8fc8-902c-4152-92c7-1f15ba27ebf2 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 924.649675] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-92d46ad1-0a6b-4974-a461-579c67372268 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 924.661943] env[61806]: DEBUG nova.network.neutron [req-529e30ad-8000-42cc-9de5-9ecbcddf93e0 req-d52a861f-2d11-466a-9884-c20e5f3ba61c service nova] [instance: c486c00d-61fc-4795-9299-e6548b686ee8] Updated VIF entry in instance network info cache for port df6e9523-3657-4405-8d44-5b35cfc6fa9f. {{(pid=61806) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 924.662305] env[61806]: DEBUG nova.network.neutron [req-529e30ad-8000-42cc-9de5-9ecbcddf93e0 req-d52a861f-2d11-466a-9884-c20e5f3ba61c service nova] [instance: c486c00d-61fc-4795-9299-e6548b686ee8] Updating instance_info_cache with network_info: [{"id": "df6e9523-3657-4405-8d44-5b35cfc6fa9f", "address": "fa:16:3e:85:ad:44", "network": {"id": "f280a01b-7000-4e54-ab9a-20eab156a410", "bridge": "br-int", "label": "tempest-ServersTestJSON-2080718146-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "73ebcc4d79d248efb653a084a5e44302", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "43798f54-0c07-4417-a23f-58bb6b7e204b", "external-id": "nsx-vlan-transportzone-571", "segmentation_id": 571, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapdf6e9523-36", "ovs_interfaceid": "df6e9523-3657-4405-8d44-5b35cfc6fa9f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61806) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 924.682815] env[61806]: DEBUG nova.compute.manager [req-3f07e10e-8ba6-4aa8-9622-7ba300ecf554 req-da2e56bd-660c-4a5e-84ae-fd5224437985 service nova] [instance: 75f1aded-d1da-4ddc-be29-f1fb93799364] Detach interface failed, port_id=a06a82f0-c5d3-48b7-bd4b-d2725bd7eaec, reason: Instance 75f1aded-d1da-4ddc-be29-f1fb93799364 could not be found. {{(pid=61806) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10963}} [ 924.728910] env[61806]: DEBUG nova.compute.manager [req-c77934a0-b242-48ca-b5d6-5212241051ae req-317fa15e-224e-4c89-acc9-9daed818ec35 service nova] [instance: 380aa63f-46ba-437d-bc29-0fc141ad1cf3] Received event network-vif-plugged-2c4c1295-84da-4bdf-8e6d-742b6358064c {{(pid=61806) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 924.729179] env[61806]: DEBUG oslo_concurrency.lockutils [req-c77934a0-b242-48ca-b5d6-5212241051ae req-317fa15e-224e-4c89-acc9-9daed818ec35 service nova] Acquiring lock "380aa63f-46ba-437d-bc29-0fc141ad1cf3-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 924.729377] env[61806]: DEBUG oslo_concurrency.lockutils [req-c77934a0-b242-48ca-b5d6-5212241051ae req-317fa15e-224e-4c89-acc9-9daed818ec35 service nova] Lock "380aa63f-46ba-437d-bc29-0fc141ad1cf3-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 924.729533] env[61806]: DEBUG oslo_concurrency.lockutils [req-c77934a0-b242-48ca-b5d6-5212241051ae req-317fa15e-224e-4c89-acc9-9daed818ec35 service nova] Lock "380aa63f-46ba-437d-bc29-0fc141ad1cf3-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 924.729707] env[61806]: DEBUG nova.compute.manager [req-c77934a0-b242-48ca-b5d6-5212241051ae req-317fa15e-224e-4c89-acc9-9daed818ec35 service nova] [instance: 380aa63f-46ba-437d-bc29-0fc141ad1cf3] No waiting events found dispatching network-vif-plugged-2c4c1295-84da-4bdf-8e6d-742b6358064c {{(pid=61806) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 924.729878] env[61806]: WARNING nova.compute.manager [req-c77934a0-b242-48ca-b5d6-5212241051ae req-317fa15e-224e-4c89-acc9-9daed818ec35 service nova] [instance: 380aa63f-46ba-437d-bc29-0fc141ad1cf3] Received unexpected event network-vif-plugged-2c4c1295-84da-4bdf-8e6d-742b6358064c for instance with vm_state building and task_state spawning. [ 924.827428] env[61806]: DEBUG nova.network.neutron [None req-d76979a9-290f-44f9-bf19-39a08fc72d03 tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] [instance: 380aa63f-46ba-437d-bc29-0fc141ad1cf3] Successfully updated port: 2c4c1295-84da-4bdf-8e6d-742b6358064c {{(pid=61806) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 924.940901] env[61806]: DEBUG oslo_vmware.api [None req-018e57c8-1b55-4515-9361-baa3401dd22f tempest-ServersAaction247Test-1786617390 tempest-ServersAaction247Test-1786617390-project-member] Task: {'id': task-1294904, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 925.042195] env[61806]: DEBUG oslo_vmware.api [None req-531998d5-4ee2-42e3-ac57-d5e3f0c96508 tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] Task: {'id': task-1294905, 'name': CopyVirtualDisk_Task} progress is 4%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 925.119707] env[61806]: INFO nova.compute.manager [-] [instance: 75f1aded-d1da-4ddc-be29-f1fb93799364] Took 1.23 seconds to deallocate network for instance. [ 925.165274] env[61806]: DEBUG oslo_concurrency.lockutils [req-529e30ad-8000-42cc-9de5-9ecbcddf93e0 req-d52a861f-2d11-466a-9884-c20e5f3ba61c service nova] Releasing lock "refresh_cache-c486c00d-61fc-4795-9299-e6548b686ee8" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 925.165583] env[61806]: DEBUG nova.compute.manager [req-529e30ad-8000-42cc-9de5-9ecbcddf93e0 req-d52a861f-2d11-466a-9884-c20e5f3ba61c service nova] [instance: d9caee4a-a00c-4e50-b01c-99a8f796a09f] Received event network-changed-c725ee28-d7a8-4b31-92bc-f63174564f5e {{(pid=61806) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 925.165768] env[61806]: DEBUG nova.compute.manager [req-529e30ad-8000-42cc-9de5-9ecbcddf93e0 req-d52a861f-2d11-466a-9884-c20e5f3ba61c service nova] [instance: d9caee4a-a00c-4e50-b01c-99a8f796a09f] Refreshing instance network info cache due to event network-changed-c725ee28-d7a8-4b31-92bc-f63174564f5e. {{(pid=61806) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 925.166074] env[61806]: DEBUG oslo_concurrency.lockutils [req-529e30ad-8000-42cc-9de5-9ecbcddf93e0 req-d52a861f-2d11-466a-9884-c20e5f3ba61c service nova] Acquiring lock "refresh_cache-d9caee4a-a00c-4e50-b01c-99a8f796a09f" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 925.166250] env[61806]: DEBUG oslo_concurrency.lockutils [req-529e30ad-8000-42cc-9de5-9ecbcddf93e0 req-d52a861f-2d11-466a-9884-c20e5f3ba61c service nova] Acquired lock "refresh_cache-d9caee4a-a00c-4e50-b01c-99a8f796a09f" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 925.166426] env[61806]: DEBUG nova.network.neutron [req-529e30ad-8000-42cc-9de5-9ecbcddf93e0 req-d52a861f-2d11-466a-9884-c20e5f3ba61c service nova] [instance: d9caee4a-a00c-4e50-b01c-99a8f796a09f] Refreshing network info cache for port c725ee28-d7a8-4b31-92bc-f63174564f5e {{(pid=61806) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 925.174875] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-32cb0dea-7424-474f-abe2-316cccff0a45 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 925.183822] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-192b8ebb-c017-47f2-aed9-a0588581e80c {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 925.216787] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2b112781-88f4-4f4c-a823-256ee52a92e4 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 925.225977] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-89f6d7dd-a432-4932-a95e-b96b88bd87ac {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 925.244679] env[61806]: DEBUG nova.compute.provider_tree [None req-bdb89e04-5c89-42d4-937f-829d8aee12b6 tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] Inventory has not changed in ProviderTree for provider: a2858be1-fd22-4e08-979e-87ad25293407 {{(pid=61806) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 925.330297] env[61806]: DEBUG oslo_concurrency.lockutils [None req-d76979a9-290f-44f9-bf19-39a08fc72d03 tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Acquiring lock "refresh_cache-380aa63f-46ba-437d-bc29-0fc141ad1cf3" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 925.330504] env[61806]: DEBUG oslo_concurrency.lockutils [None req-d76979a9-290f-44f9-bf19-39a08fc72d03 tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Acquired lock "refresh_cache-380aa63f-46ba-437d-bc29-0fc141ad1cf3" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 925.330599] env[61806]: DEBUG nova.network.neutron [None req-d76979a9-290f-44f9-bf19-39a08fc72d03 tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] [instance: 380aa63f-46ba-437d-bc29-0fc141ad1cf3] Building network info cache for instance {{(pid=61806) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 925.442104] env[61806]: DEBUG oslo_vmware.api [None req-018e57c8-1b55-4515-9361-baa3401dd22f tempest-ServersAaction247Test-1786617390 tempest-ServersAaction247Test-1786617390-project-member] Task: {'id': task-1294904, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.67849} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 925.442373] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [None req-018e57c8-1b55-4515-9361-baa3401dd22f tempest-ServersAaction247Test-1786617390 tempest-ServersAaction247Test-1786617390-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa.vmdk to [datastore1] 3f4fa1ba-a6ad-4e0f-aebc-973816ee09d5/3f4fa1ba-a6ad-4e0f-aebc-973816ee09d5.vmdk {{(pid=61806) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 925.442569] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-018e57c8-1b55-4515-9361-baa3401dd22f tempest-ServersAaction247Test-1786617390 tempest-ServersAaction247Test-1786617390-project-member] [instance: 3f4fa1ba-a6ad-4e0f-aebc-973816ee09d5] Extending root virtual disk to 1048576 {{(pid=61806) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 925.442856] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-94093aa4-98b7-4a03-a9d6-16eefdaef420 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 925.450293] env[61806]: DEBUG oslo_vmware.api [None req-018e57c8-1b55-4515-9361-baa3401dd22f tempest-ServersAaction247Test-1786617390 tempest-ServersAaction247Test-1786617390-project-member] Waiting for the task: (returnval){ [ 925.450293] env[61806]: value = "task-1294906" [ 925.450293] env[61806]: _type = "Task" [ 925.450293] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 925.459394] env[61806]: DEBUG oslo_vmware.api [None req-018e57c8-1b55-4515-9361-baa3401dd22f tempest-ServersAaction247Test-1786617390 tempest-ServersAaction247Test-1786617390-project-member] Task: {'id': task-1294906, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 925.542690] env[61806]: DEBUG oslo_vmware.api [None req-531998d5-4ee2-42e3-ac57-d5e3f0c96508 tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] Task: {'id': task-1294905, 'name': CopyVirtualDisk_Task} progress is 89%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 925.627507] env[61806]: DEBUG oslo_concurrency.lockutils [None req-33516f06-7375-4e3f-a28f-63fccd97ac4e tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 925.748834] env[61806]: DEBUG nova.scheduler.client.report [None req-bdb89e04-5c89-42d4-937f-829d8aee12b6 tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] Inventory has not changed for provider a2858be1-fd22-4e08-979e-87ad25293407 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 138, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61806) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 925.864551] env[61806]: DEBUG nova.network.neutron [None req-d76979a9-290f-44f9-bf19-39a08fc72d03 tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] [instance: 380aa63f-46ba-437d-bc29-0fc141ad1cf3] Instance cache missing network info. {{(pid=61806) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 925.884893] env[61806]: DEBUG nova.network.neutron [req-529e30ad-8000-42cc-9de5-9ecbcddf93e0 req-d52a861f-2d11-466a-9884-c20e5f3ba61c service nova] [instance: d9caee4a-a00c-4e50-b01c-99a8f796a09f] Updated VIF entry in instance network info cache for port c725ee28-d7a8-4b31-92bc-f63174564f5e. {{(pid=61806) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 925.885267] env[61806]: DEBUG nova.network.neutron [req-529e30ad-8000-42cc-9de5-9ecbcddf93e0 req-d52a861f-2d11-466a-9884-c20e5f3ba61c service nova] [instance: d9caee4a-a00c-4e50-b01c-99a8f796a09f] Updating instance_info_cache with network_info: [{"id": "c725ee28-d7a8-4b31-92bc-f63174564f5e", "address": "fa:16:3e:3a:18:8e", "network": {"id": "a7ffda72-e57c-41fa-a7ed-d7b12084bc9d", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-62794865-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "6e391e1267014fafbb8f5a3211299819", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8b3cfeb1-f262-4fd9-b506-8e9c0733e2d8", "external-id": "nsx-vlan-transportzone-119", "segmentation_id": 119, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc725ee28-d7", "ovs_interfaceid": "c725ee28-d7a8-4b31-92bc-f63174564f5e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61806) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 925.964363] env[61806]: DEBUG oslo_vmware.api [None req-018e57c8-1b55-4515-9361-baa3401dd22f tempest-ServersAaction247Test-1786617390 tempest-ServersAaction247Test-1786617390-project-member] Task: {'id': task-1294906, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.20925} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 925.965457] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-018e57c8-1b55-4515-9361-baa3401dd22f tempest-ServersAaction247Test-1786617390 tempest-ServersAaction247Test-1786617390-project-member] [instance: 3f4fa1ba-a6ad-4e0f-aebc-973816ee09d5] Extended root virtual disk {{(pid=61806) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 925.966389] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a22dcade-2b59-49d6-b759-05fcb207b663 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 925.986272] env[61806]: DEBUG nova.virt.vmwareapi.volumeops [None req-018e57c8-1b55-4515-9361-baa3401dd22f tempest-ServersAaction247Test-1786617390 tempest-ServersAaction247Test-1786617390-project-member] [instance: 3f4fa1ba-a6ad-4e0f-aebc-973816ee09d5] Reconfiguring VM instance instance-00000055 to attach disk [datastore1] 3f4fa1ba-a6ad-4e0f-aebc-973816ee09d5/3f4fa1ba-a6ad-4e0f-aebc-973816ee09d5.vmdk or device None with type sparse {{(pid=61806) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 925.987147] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-7d1dfcc2-90a6-4726-abcc-9ccb2b8f8150 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 926.001748] env[61806]: DEBUG nova.network.neutron [None req-d76979a9-290f-44f9-bf19-39a08fc72d03 tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] [instance: 380aa63f-46ba-437d-bc29-0fc141ad1cf3] Updating instance_info_cache with network_info: [{"id": "2c4c1295-84da-4bdf-8e6d-742b6358064c", "address": "fa:16:3e:14:c5:ea", "network": {"id": "b7e78559-d885-4b59-9c7f-9d966b424d04", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-1456721354-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7ab95fb72b5d46c3b6c7bebbccf897cd", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d891a662-7da0-4be5-8d0d-01b4ff70552a", "external-id": "nsx-vlan-transportzone-815", "segmentation_id": 815, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2c4c1295-84", "ovs_interfaceid": "2c4c1295-84da-4bdf-8e6d-742b6358064c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61806) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 926.009884] env[61806]: DEBUG oslo_vmware.api [None req-018e57c8-1b55-4515-9361-baa3401dd22f tempest-ServersAaction247Test-1786617390 tempest-ServersAaction247Test-1786617390-project-member] Waiting for the task: (returnval){ [ 926.009884] env[61806]: value = "task-1294907" [ 926.009884] env[61806]: _type = "Task" [ 926.009884] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 926.018862] env[61806]: DEBUG oslo_vmware.api [None req-018e57c8-1b55-4515-9361-baa3401dd22f tempest-ServersAaction247Test-1786617390 tempest-ServersAaction247Test-1786617390-project-member] Task: {'id': task-1294907, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 926.042230] env[61806]: DEBUG oslo_vmware.api [None req-531998d5-4ee2-42e3-ac57-d5e3f0c96508 tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] Task: {'id': task-1294905, 'name': CopyVirtualDisk_Task, 'duration_secs': 1.024618} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 926.042504] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [None req-531998d5-4ee2-42e3-ac57-d5e3f0c96508 tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa.vmdk to [datastore1] c486c00d-61fc-4795-9299-e6548b686ee8/c486c00d-61fc-4795-9299-e6548b686ee8.vmdk {{(pid=61806) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 926.042722] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-531998d5-4ee2-42e3-ac57-d5e3f0c96508 tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] [instance: c486c00d-61fc-4795-9299-e6548b686ee8] Extending root virtual disk to 1048576 {{(pid=61806) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 926.042975] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-ad039b27-0456-4a09-bf4f-07d863c7e512 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 926.049882] env[61806]: DEBUG oslo_vmware.api [None req-531998d5-4ee2-42e3-ac57-d5e3f0c96508 tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] Waiting for the task: (returnval){ [ 926.049882] env[61806]: value = "task-1294908" [ 926.049882] env[61806]: _type = "Task" [ 926.049882] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 926.059722] env[61806]: DEBUG oslo_vmware.api [None req-531998d5-4ee2-42e3-ac57-d5e3f0c96508 tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] Task: {'id': task-1294908, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 926.253020] env[61806]: DEBUG oslo_concurrency.lockutils [None req-bdb89e04-5c89-42d4-937f-829d8aee12b6 tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.274s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 926.253604] env[61806]: DEBUG nova.compute.manager [None req-bdb89e04-5c89-42d4-937f-829d8aee12b6 tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] [instance: fd9e520a-2e7b-43d4-b84d-8933c32cc6a6] Start building networks asynchronously for instance. {{(pid=61806) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 926.256654] env[61806]: DEBUG oslo_concurrency.lockutils [None req-0c4ed53e-d274-4427-995e-e2d8d9d3b2cc tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 1.940s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 926.258030] env[61806]: INFO nova.compute.claims [None req-0c4ed53e-d274-4427-995e-e2d8d9d3b2cc tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] [instance: 676f18b3-3da1-4a53-b96d-fe64ee8f9101] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 926.387604] env[61806]: DEBUG oslo_concurrency.lockutils [req-529e30ad-8000-42cc-9de5-9ecbcddf93e0 req-d52a861f-2d11-466a-9884-c20e5f3ba61c service nova] Releasing lock "refresh_cache-d9caee4a-a00c-4e50-b01c-99a8f796a09f" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 926.388063] env[61806]: DEBUG nova.compute.manager [req-529e30ad-8000-42cc-9de5-9ecbcddf93e0 req-d52a861f-2d11-466a-9884-c20e5f3ba61c service nova] [instance: 6f6ba57a-e2d6-4749-a53a-e263861cb1c0] Received event network-changed-013c9f0d-65b1-4279-a823-a7f1c14dfb91 {{(pid=61806) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 926.388154] env[61806]: DEBUG nova.compute.manager [req-529e30ad-8000-42cc-9de5-9ecbcddf93e0 req-d52a861f-2d11-466a-9884-c20e5f3ba61c service nova] [instance: 6f6ba57a-e2d6-4749-a53a-e263861cb1c0] Refreshing instance network info cache due to event network-changed-013c9f0d-65b1-4279-a823-a7f1c14dfb91. {{(pid=61806) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 926.388426] env[61806]: DEBUG oslo_concurrency.lockutils [req-529e30ad-8000-42cc-9de5-9ecbcddf93e0 req-d52a861f-2d11-466a-9884-c20e5f3ba61c service nova] Acquiring lock "refresh_cache-6f6ba57a-e2d6-4749-a53a-e263861cb1c0" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 926.388614] env[61806]: DEBUG oslo_concurrency.lockutils [req-529e30ad-8000-42cc-9de5-9ecbcddf93e0 req-d52a861f-2d11-466a-9884-c20e5f3ba61c service nova] Acquired lock "refresh_cache-6f6ba57a-e2d6-4749-a53a-e263861cb1c0" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 926.388796] env[61806]: DEBUG nova.network.neutron [req-529e30ad-8000-42cc-9de5-9ecbcddf93e0 req-d52a861f-2d11-466a-9884-c20e5f3ba61c service nova] [instance: 6f6ba57a-e2d6-4749-a53a-e263861cb1c0] Refreshing network info cache for port 013c9f0d-65b1-4279-a823-a7f1c14dfb91 {{(pid=61806) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 926.504323] env[61806]: DEBUG oslo_concurrency.lockutils [None req-d76979a9-290f-44f9-bf19-39a08fc72d03 tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Releasing lock "refresh_cache-380aa63f-46ba-437d-bc29-0fc141ad1cf3" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 926.504651] env[61806]: DEBUG nova.compute.manager [None req-d76979a9-290f-44f9-bf19-39a08fc72d03 tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] [instance: 380aa63f-46ba-437d-bc29-0fc141ad1cf3] Instance network_info: |[{"id": "2c4c1295-84da-4bdf-8e6d-742b6358064c", "address": "fa:16:3e:14:c5:ea", "network": {"id": "b7e78559-d885-4b59-9c7f-9d966b424d04", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-1456721354-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7ab95fb72b5d46c3b6c7bebbccf897cd", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d891a662-7da0-4be5-8d0d-01b4ff70552a", "external-id": "nsx-vlan-transportzone-815", "segmentation_id": 815, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2c4c1295-84", "ovs_interfaceid": "2c4c1295-84da-4bdf-8e6d-742b6358064c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61806) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 926.505104] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-d76979a9-290f-44f9-bf19-39a08fc72d03 tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] [instance: 380aa63f-46ba-437d-bc29-0fc141ad1cf3] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:14:c5:ea', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'd891a662-7da0-4be5-8d0d-01b4ff70552a', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '2c4c1295-84da-4bdf-8e6d-742b6358064c', 'vif_model': 'vmxnet3'}] {{(pid=61806) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 926.512450] env[61806]: DEBUG oslo.service.loopingcall [None req-d76979a9-290f-44f9-bf19-39a08fc72d03 tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61806) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 926.512965] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 380aa63f-46ba-437d-bc29-0fc141ad1cf3] Creating VM on the ESX host {{(pid=61806) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 926.516510] env[61806]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-dc8991ea-24ff-4b18-b6c1-cf7b47978042 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 926.535915] env[61806]: DEBUG oslo_vmware.api [None req-018e57c8-1b55-4515-9361-baa3401dd22f tempest-ServersAaction247Test-1786617390 tempest-ServersAaction247Test-1786617390-project-member] Task: {'id': task-1294907, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 926.537071] env[61806]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 926.537071] env[61806]: value = "task-1294909" [ 926.537071] env[61806]: _type = "Task" [ 926.537071] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 926.544242] env[61806]: DEBUG oslo_vmware.api [-] Task: {'id': task-1294909, 'name': CreateVM_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 926.558538] env[61806]: DEBUG oslo_vmware.api [None req-531998d5-4ee2-42e3-ac57-d5e3f0c96508 tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] Task: {'id': task-1294908, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.103681} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 926.558818] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-531998d5-4ee2-42e3-ac57-d5e3f0c96508 tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] [instance: c486c00d-61fc-4795-9299-e6548b686ee8] Extended root virtual disk {{(pid=61806) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 926.559490] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-885bb317-3d7f-4054-8037-2757f489edc3 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 926.583071] env[61806]: DEBUG nova.virt.vmwareapi.volumeops [None req-531998d5-4ee2-42e3-ac57-d5e3f0c96508 tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] [instance: c486c00d-61fc-4795-9299-e6548b686ee8] Reconfiguring VM instance instance-00000054 to attach disk [datastore1] c486c00d-61fc-4795-9299-e6548b686ee8/c486c00d-61fc-4795-9299-e6548b686ee8.vmdk or device None with type sparse {{(pid=61806) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 926.583404] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-cf241b22-c0cf-428f-9b5e-2549978ca39c {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 926.604376] env[61806]: DEBUG oslo_vmware.api [None req-531998d5-4ee2-42e3-ac57-d5e3f0c96508 tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] Waiting for the task: (returnval){ [ 926.604376] env[61806]: value = "task-1294910" [ 926.604376] env[61806]: _type = "Task" [ 926.604376] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 926.617389] env[61806]: DEBUG oslo_vmware.api [None req-531998d5-4ee2-42e3-ac57-d5e3f0c96508 tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] Task: {'id': task-1294910, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 926.757719] env[61806]: DEBUG nova.compute.manager [req-3aff7894-183f-4207-92b4-41937fb1d1a8 req-dcfbdabf-7ca7-4448-b329-60dde05ca662 service nova] [instance: 380aa63f-46ba-437d-bc29-0fc141ad1cf3] Received event network-changed-2c4c1295-84da-4bdf-8e6d-742b6358064c {{(pid=61806) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 926.757965] env[61806]: DEBUG nova.compute.manager [req-3aff7894-183f-4207-92b4-41937fb1d1a8 req-dcfbdabf-7ca7-4448-b329-60dde05ca662 service nova] [instance: 380aa63f-46ba-437d-bc29-0fc141ad1cf3] Refreshing instance network info cache due to event network-changed-2c4c1295-84da-4bdf-8e6d-742b6358064c. {{(pid=61806) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 926.758237] env[61806]: DEBUG oslo_concurrency.lockutils [req-3aff7894-183f-4207-92b4-41937fb1d1a8 req-dcfbdabf-7ca7-4448-b329-60dde05ca662 service nova] Acquiring lock "refresh_cache-380aa63f-46ba-437d-bc29-0fc141ad1cf3" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 926.758442] env[61806]: DEBUG oslo_concurrency.lockutils [req-3aff7894-183f-4207-92b4-41937fb1d1a8 req-dcfbdabf-7ca7-4448-b329-60dde05ca662 service nova] Acquired lock "refresh_cache-380aa63f-46ba-437d-bc29-0fc141ad1cf3" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 926.758657] env[61806]: DEBUG nova.network.neutron [req-3aff7894-183f-4207-92b4-41937fb1d1a8 req-dcfbdabf-7ca7-4448-b329-60dde05ca662 service nova] [instance: 380aa63f-46ba-437d-bc29-0fc141ad1cf3] Refreshing network info cache for port 2c4c1295-84da-4bdf-8e6d-742b6358064c {{(pid=61806) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 926.761685] env[61806]: DEBUG nova.compute.utils [None req-bdb89e04-5c89-42d4-937f-829d8aee12b6 tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] Using /dev/sd instead of None {{(pid=61806) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 926.766055] env[61806]: DEBUG nova.compute.manager [None req-bdb89e04-5c89-42d4-937f-829d8aee12b6 tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] [instance: fd9e520a-2e7b-43d4-b84d-8933c32cc6a6] Allocating IP information in the background. {{(pid=61806) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 926.766284] env[61806]: DEBUG nova.network.neutron [None req-bdb89e04-5c89-42d4-937f-829d8aee12b6 tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] [instance: fd9e520a-2e7b-43d4-b84d-8933c32cc6a6] allocate_for_instance() {{(pid=61806) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 926.808330] env[61806]: DEBUG nova.policy [None req-bdb89e04-5c89-42d4-937f-829d8aee12b6 tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '50a051926b2b4ea9b22828762d548ce6', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'bbbb09a49203421d807f733616c854df', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61806) authorize /opt/stack/nova/nova/policy.py:201}} [ 927.025651] env[61806]: DEBUG oslo_vmware.api [None req-018e57c8-1b55-4515-9361-baa3401dd22f tempest-ServersAaction247Test-1786617390 tempest-ServersAaction247Test-1786617390-project-member] Task: {'id': task-1294907, 'name': ReconfigVM_Task, 'duration_secs': 1.005615} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 927.026032] env[61806]: DEBUG nova.virt.vmwareapi.volumeops [None req-018e57c8-1b55-4515-9361-baa3401dd22f tempest-ServersAaction247Test-1786617390 tempest-ServersAaction247Test-1786617390-project-member] [instance: 3f4fa1ba-a6ad-4e0f-aebc-973816ee09d5] Reconfigured VM instance instance-00000055 to attach disk [datastore1] 3f4fa1ba-a6ad-4e0f-aebc-973816ee09d5/3f4fa1ba-a6ad-4e0f-aebc-973816ee09d5.vmdk or device None with type sparse {{(pid=61806) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 927.026552] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-9dbe6c58-b025-46c6-867d-65ba2b9ee8d6 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 927.033098] env[61806]: DEBUG oslo_vmware.api [None req-018e57c8-1b55-4515-9361-baa3401dd22f tempest-ServersAaction247Test-1786617390 tempest-ServersAaction247Test-1786617390-project-member] Waiting for the task: (returnval){ [ 927.033098] env[61806]: value = "task-1294911" [ 927.033098] env[61806]: _type = "Task" [ 927.033098] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 927.044492] env[61806]: DEBUG oslo_vmware.api [None req-018e57c8-1b55-4515-9361-baa3401dd22f tempest-ServersAaction247Test-1786617390 tempest-ServersAaction247Test-1786617390-project-member] Task: {'id': task-1294911, 'name': Rename_Task} progress is 5%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 927.049848] env[61806]: DEBUG oslo_vmware.api [-] Task: {'id': task-1294909, 'name': CreateVM_Task, 'duration_secs': 0.37899} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 927.050011] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 380aa63f-46ba-437d-bc29-0fc141ad1cf3] Created VM on the ESX host {{(pid=61806) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 927.050710] env[61806]: DEBUG oslo_concurrency.lockutils [None req-d76979a9-290f-44f9-bf19-39a08fc72d03 tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 927.050884] env[61806]: DEBUG oslo_concurrency.lockutils [None req-d76979a9-290f-44f9-bf19-39a08fc72d03 tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Acquired lock "[datastore1] devstack-image-cache_base/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 927.051219] env[61806]: DEBUG oslo_concurrency.lockutils [None req-d76979a9-290f-44f9-bf19-39a08fc72d03 tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 927.051471] env[61806]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-262f195c-9d90-4dc9-adb4-ce197bff4981 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 927.056266] env[61806]: DEBUG oslo_vmware.api [None req-d76979a9-290f-44f9-bf19-39a08fc72d03 tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Waiting for the task: (returnval){ [ 927.056266] env[61806]: value = "session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]529e09f4-a856-40c3-7af7-86ac3f3e3238" [ 927.056266] env[61806]: _type = "Task" [ 927.056266] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 927.064764] env[61806]: DEBUG oslo_vmware.api [None req-d76979a9-290f-44f9-bf19-39a08fc72d03 tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Task: {'id': session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]529e09f4-a856-40c3-7af7-86ac3f3e3238, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 927.117200] env[61806]: DEBUG oslo_vmware.api [None req-531998d5-4ee2-42e3-ac57-d5e3f0c96508 tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] Task: {'id': task-1294910, 'name': ReconfigVM_Task, 'duration_secs': 0.291161} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 927.118468] env[61806]: DEBUG nova.network.neutron [None req-bdb89e04-5c89-42d4-937f-829d8aee12b6 tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] [instance: fd9e520a-2e7b-43d4-b84d-8933c32cc6a6] Successfully created port: 670f75e3-58f9-493b-9148-416e2e9cea5d {{(pid=61806) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 927.120518] env[61806]: DEBUG nova.virt.vmwareapi.volumeops [None req-531998d5-4ee2-42e3-ac57-d5e3f0c96508 tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] [instance: c486c00d-61fc-4795-9299-e6548b686ee8] Reconfigured VM instance instance-00000054 to attach disk [datastore1] c486c00d-61fc-4795-9299-e6548b686ee8/c486c00d-61fc-4795-9299-e6548b686ee8.vmdk or device None with type sparse {{(pid=61806) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 927.121161] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-0fb7ba01-9d60-42d5-b2e2-c55f66c05013 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 927.130150] env[61806]: DEBUG oslo_vmware.api [None req-531998d5-4ee2-42e3-ac57-d5e3f0c96508 tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] Waiting for the task: (returnval){ [ 927.130150] env[61806]: value = "task-1294912" [ 927.130150] env[61806]: _type = "Task" [ 927.130150] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 927.140437] env[61806]: DEBUG oslo_vmware.api [None req-531998d5-4ee2-42e3-ac57-d5e3f0c96508 tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] Task: {'id': task-1294912, 'name': Rename_Task} progress is 5%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 927.148300] env[61806]: DEBUG nova.network.neutron [req-529e30ad-8000-42cc-9de5-9ecbcddf93e0 req-d52a861f-2d11-466a-9884-c20e5f3ba61c service nova] [instance: 6f6ba57a-e2d6-4749-a53a-e263861cb1c0] Updated VIF entry in instance network info cache for port 013c9f0d-65b1-4279-a823-a7f1c14dfb91. {{(pid=61806) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 927.148675] env[61806]: DEBUG nova.network.neutron [req-529e30ad-8000-42cc-9de5-9ecbcddf93e0 req-d52a861f-2d11-466a-9884-c20e5f3ba61c service nova] [instance: 6f6ba57a-e2d6-4749-a53a-e263861cb1c0] Updating instance_info_cache with network_info: [{"id": "013c9f0d-65b1-4279-a823-a7f1c14dfb91", "address": "fa:16:3e:15:56:4e", "network": {"id": "a7ffda72-e57c-41fa-a7ed-d7b12084bc9d", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-62794865-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.218", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "6e391e1267014fafbb8f5a3211299819", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8b3cfeb1-f262-4fd9-b506-8e9c0733e2d8", "external-id": "nsx-vlan-transportzone-119", "segmentation_id": 119, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap013c9f0d-65", "ovs_interfaceid": "013c9f0d-65b1-4279-a823-a7f1c14dfb91", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61806) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 927.264251] env[61806]: DEBUG nova.compute.manager [None req-bdb89e04-5c89-42d4-937f-829d8aee12b6 tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] [instance: fd9e520a-2e7b-43d4-b84d-8933c32cc6a6] Start building block device mappings for instance. {{(pid=61806) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 927.444196] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5a0e2663-d559-4795-b8d5-df293a9940b6 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 927.453322] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-620a3a32-7eb4-4cd5-85e8-42c3fbe99e09 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 927.487526] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ca0a11ea-988f-40fd-8112-6feafec10073 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 927.506637] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-82b17f23-59a1-4e89-85da-e24aed41f27d {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 927.522103] env[61806]: DEBUG nova.compute.provider_tree [None req-0c4ed53e-d274-4427-995e-e2d8d9d3b2cc tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Inventory has not changed in ProviderTree for provider: a2858be1-fd22-4e08-979e-87ad25293407 {{(pid=61806) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 927.523975] env[61806]: DEBUG nova.network.neutron [req-3aff7894-183f-4207-92b4-41937fb1d1a8 req-dcfbdabf-7ca7-4448-b329-60dde05ca662 service nova] [instance: 380aa63f-46ba-437d-bc29-0fc141ad1cf3] Updated VIF entry in instance network info cache for port 2c4c1295-84da-4bdf-8e6d-742b6358064c. {{(pid=61806) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 927.524255] env[61806]: DEBUG nova.network.neutron [req-3aff7894-183f-4207-92b4-41937fb1d1a8 req-dcfbdabf-7ca7-4448-b329-60dde05ca662 service nova] [instance: 380aa63f-46ba-437d-bc29-0fc141ad1cf3] Updating instance_info_cache with network_info: [{"id": "2c4c1295-84da-4bdf-8e6d-742b6358064c", "address": "fa:16:3e:14:c5:ea", "network": {"id": "b7e78559-d885-4b59-9c7f-9d966b424d04", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-1456721354-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7ab95fb72b5d46c3b6c7bebbccf897cd", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d891a662-7da0-4be5-8d0d-01b4ff70552a", "external-id": "nsx-vlan-transportzone-815", "segmentation_id": 815, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2c4c1295-84", "ovs_interfaceid": "2c4c1295-84da-4bdf-8e6d-742b6358064c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61806) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 927.543581] env[61806]: DEBUG oslo_vmware.api [None req-018e57c8-1b55-4515-9361-baa3401dd22f tempest-ServersAaction247Test-1786617390 tempest-ServersAaction247Test-1786617390-project-member] Task: {'id': task-1294911, 'name': Rename_Task, 'duration_secs': 0.375887} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 927.543868] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [None req-018e57c8-1b55-4515-9361-baa3401dd22f tempest-ServersAaction247Test-1786617390 tempest-ServersAaction247Test-1786617390-project-member] [instance: 3f4fa1ba-a6ad-4e0f-aebc-973816ee09d5] Powering on the VM {{(pid=61806) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 927.544135] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-f0aeaf17-873b-4092-a538-6cd0d2dae422 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 927.550980] env[61806]: DEBUG oslo_vmware.api [None req-018e57c8-1b55-4515-9361-baa3401dd22f tempest-ServersAaction247Test-1786617390 tempest-ServersAaction247Test-1786617390-project-member] Waiting for the task: (returnval){ [ 927.550980] env[61806]: value = "task-1294913" [ 927.550980] env[61806]: _type = "Task" [ 927.550980] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 927.559276] env[61806]: DEBUG oslo_vmware.api [None req-018e57c8-1b55-4515-9361-baa3401dd22f tempest-ServersAaction247Test-1786617390 tempest-ServersAaction247Test-1786617390-project-member] Task: {'id': task-1294913, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 927.567531] env[61806]: DEBUG oslo_vmware.api [None req-d76979a9-290f-44f9-bf19-39a08fc72d03 tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Task: {'id': session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]529e09f4-a856-40c3-7af7-86ac3f3e3238, 'name': SearchDatastore_Task, 'duration_secs': 0.06143} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 927.568049] env[61806]: DEBUG oslo_concurrency.lockutils [None req-d76979a9-290f-44f9-bf19-39a08fc72d03 tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Releasing lock "[datastore1] devstack-image-cache_base/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 927.568049] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-d76979a9-290f-44f9-bf19-39a08fc72d03 tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] [instance: 380aa63f-46ba-437d-bc29-0fc141ad1cf3] Processing image ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa {{(pid=61806) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 927.568272] env[61806]: DEBUG oslo_concurrency.lockutils [None req-d76979a9-290f-44f9-bf19-39a08fc72d03 tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa.vmdk" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 927.568451] env[61806]: DEBUG oslo_concurrency.lockutils [None req-d76979a9-290f-44f9-bf19-39a08fc72d03 tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Acquired lock "[datastore1] devstack-image-cache_base/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa.vmdk" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 927.568605] env[61806]: DEBUG nova.virt.vmwareapi.ds_util [None req-d76979a9-290f-44f9-bf19-39a08fc72d03 tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61806) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 927.568859] env[61806]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-e4166322-6e64-44c6-a75e-947fd563f0a1 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 927.640876] env[61806]: DEBUG oslo_vmware.api [None req-531998d5-4ee2-42e3-ac57-d5e3f0c96508 tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] Task: {'id': task-1294912, 'name': Rename_Task} progress is 99%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 927.651631] env[61806]: DEBUG oslo_concurrency.lockutils [req-529e30ad-8000-42cc-9de5-9ecbcddf93e0 req-d52a861f-2d11-466a-9884-c20e5f3ba61c service nova] Releasing lock "refresh_cache-6f6ba57a-e2d6-4749-a53a-e263861cb1c0" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 928.030034] env[61806]: DEBUG nova.scheduler.client.report [None req-0c4ed53e-d274-4427-995e-e2d8d9d3b2cc tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Inventory has not changed for provider a2858be1-fd22-4e08-979e-87ad25293407 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 138, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61806) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 928.031258] env[61806]: DEBUG oslo_concurrency.lockutils [req-3aff7894-183f-4207-92b4-41937fb1d1a8 req-dcfbdabf-7ca7-4448-b329-60dde05ca662 service nova] Releasing lock "refresh_cache-380aa63f-46ba-437d-bc29-0fc141ad1cf3" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 928.062152] env[61806]: DEBUG oslo_vmware.api [None req-018e57c8-1b55-4515-9361-baa3401dd22f tempest-ServersAaction247Test-1786617390 tempest-ServersAaction247Test-1786617390-project-member] Task: {'id': task-1294913, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 928.142373] env[61806]: DEBUG oslo_vmware.api [None req-531998d5-4ee2-42e3-ac57-d5e3f0c96508 tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] Task: {'id': task-1294912, 'name': Rename_Task} progress is 99%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 928.222402] env[61806]: DEBUG oslo_concurrency.lockutils [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Acquiring lock "refresh_cache-43dcce0d-4fdb-4ee1-8309-92e92d2331a9" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 928.222594] env[61806]: DEBUG oslo_concurrency.lockutils [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Acquired lock "refresh_cache-43dcce0d-4fdb-4ee1-8309-92e92d2331a9" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 928.222749] env[61806]: DEBUG nova.network.neutron [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] [instance: 43dcce0d-4fdb-4ee1-8309-92e92d2331a9] Forcefully refreshing network info cache for instance {{(pid=61806) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2004}} [ 928.273703] env[61806]: DEBUG nova.compute.manager [None req-bdb89e04-5c89-42d4-937f-829d8aee12b6 tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] [instance: fd9e520a-2e7b-43d4-b84d-8933c32cc6a6] Start spawning the instance on the hypervisor. {{(pid=61806) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 928.299332] env[61806]: DEBUG nova.virt.hardware [None req-bdb89e04-5c89-42d4-937f-829d8aee12b6 tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-15T18:05:10Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-15T18:04:52Z,direct_url=,disk_format='vmdk',id=ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='84ba83fce2c349a6988173c5d6fc3b07',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-15T18:04:53Z,virtual_size=,visibility=), allow threads: False {{(pid=61806) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 928.299597] env[61806]: DEBUG nova.virt.hardware [None req-bdb89e04-5c89-42d4-937f-829d8aee12b6 tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] Flavor limits 0:0:0 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 928.299759] env[61806]: DEBUG nova.virt.hardware [None req-bdb89e04-5c89-42d4-937f-829d8aee12b6 tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] Image limits 0:0:0 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 928.299967] env[61806]: DEBUG nova.virt.hardware [None req-bdb89e04-5c89-42d4-937f-829d8aee12b6 tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] Flavor pref 0:0:0 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 928.300141] env[61806]: DEBUG nova.virt.hardware [None req-bdb89e04-5c89-42d4-937f-829d8aee12b6 tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] Image pref 0:0:0 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 928.300300] env[61806]: DEBUG nova.virt.hardware [None req-bdb89e04-5c89-42d4-937f-829d8aee12b6 tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 928.300513] env[61806]: DEBUG nova.virt.hardware [None req-bdb89e04-5c89-42d4-937f-829d8aee12b6 tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61806) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 928.300679] env[61806]: DEBUG nova.virt.hardware [None req-bdb89e04-5c89-42d4-937f-829d8aee12b6 tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61806) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 928.300856] env[61806]: DEBUG nova.virt.hardware [None req-bdb89e04-5c89-42d4-937f-829d8aee12b6 tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] Got 1 possible topologies {{(pid=61806) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 928.301037] env[61806]: DEBUG nova.virt.hardware [None req-bdb89e04-5c89-42d4-937f-829d8aee12b6 tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61806) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 928.301227] env[61806]: DEBUG nova.virt.hardware [None req-bdb89e04-5c89-42d4-937f-829d8aee12b6 tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61806) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 928.302388] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-16d4ec4a-2637-454f-8539-67e7fb1b9e60 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 928.310461] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-01c78cfc-e918-4365-bcaf-18658193ea95 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 928.510557] env[61806]: DEBUG nova.compute.manager [req-7fb6815f-577f-47ea-88c7-c08dbe41c444 req-0e74b263-3bd4-4941-b0b4-4ba96d4677af service nova] [instance: fd9e520a-2e7b-43d4-b84d-8933c32cc6a6] Received event network-vif-plugged-670f75e3-58f9-493b-9148-416e2e9cea5d {{(pid=61806) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 928.510916] env[61806]: DEBUG oslo_concurrency.lockutils [req-7fb6815f-577f-47ea-88c7-c08dbe41c444 req-0e74b263-3bd4-4941-b0b4-4ba96d4677af service nova] Acquiring lock "fd9e520a-2e7b-43d4-b84d-8933c32cc6a6-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 928.511057] env[61806]: DEBUG oslo_concurrency.lockutils [req-7fb6815f-577f-47ea-88c7-c08dbe41c444 req-0e74b263-3bd4-4941-b0b4-4ba96d4677af service nova] Lock "fd9e520a-2e7b-43d4-b84d-8933c32cc6a6-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 928.511220] env[61806]: DEBUG oslo_concurrency.lockutils [req-7fb6815f-577f-47ea-88c7-c08dbe41c444 req-0e74b263-3bd4-4941-b0b4-4ba96d4677af service nova] Lock "fd9e520a-2e7b-43d4-b84d-8933c32cc6a6-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 928.511398] env[61806]: DEBUG nova.compute.manager [req-7fb6815f-577f-47ea-88c7-c08dbe41c444 req-0e74b263-3bd4-4941-b0b4-4ba96d4677af service nova] [instance: fd9e520a-2e7b-43d4-b84d-8933c32cc6a6] No waiting events found dispatching network-vif-plugged-670f75e3-58f9-493b-9148-416e2e9cea5d {{(pid=61806) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 928.511579] env[61806]: WARNING nova.compute.manager [req-7fb6815f-577f-47ea-88c7-c08dbe41c444 req-0e74b263-3bd4-4941-b0b4-4ba96d4677af service nova] [instance: fd9e520a-2e7b-43d4-b84d-8933c32cc6a6] Received unexpected event network-vif-plugged-670f75e3-58f9-493b-9148-416e2e9cea5d for instance with vm_state building and task_state spawning. [ 928.531667] env[61806]: DEBUG oslo_concurrency.lockutils [None req-0c4ed53e-d274-4427-995e-e2d8d9d3b2cc tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.275s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 928.532240] env[61806]: DEBUG nova.compute.manager [None req-0c4ed53e-d274-4427-995e-e2d8d9d3b2cc tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] [instance: 676f18b3-3da1-4a53-b96d-fe64ee8f9101] Start building networks asynchronously for instance. {{(pid=61806) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 928.534811] env[61806]: DEBUG oslo_concurrency.lockutils [None req-33516f06-7375-4e3f-a28f-63fccd97ac4e tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 2.907s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 928.535073] env[61806]: DEBUG nova.objects.instance [None req-33516f06-7375-4e3f-a28f-63fccd97ac4e tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Lazy-loading 'resources' on Instance uuid 75f1aded-d1da-4ddc-be29-f1fb93799364 {{(pid=61806) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 928.561634] env[61806]: DEBUG oslo_vmware.api [None req-018e57c8-1b55-4515-9361-baa3401dd22f tempest-ServersAaction247Test-1786617390 tempest-ServersAaction247Test-1786617390-project-member] Task: {'id': task-1294913, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 928.601153] env[61806]: DEBUG nova.network.neutron [None req-bdb89e04-5c89-42d4-937f-829d8aee12b6 tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] [instance: fd9e520a-2e7b-43d4-b84d-8933c32cc6a6] Successfully updated port: 670f75e3-58f9-493b-9148-416e2e9cea5d {{(pid=61806) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 928.642427] env[61806]: DEBUG oslo_vmware.api [None req-531998d5-4ee2-42e3-ac57-d5e3f0c96508 tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] Task: {'id': task-1294912, 'name': Rename_Task, 'duration_secs': 1.225705} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 928.642427] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [None req-531998d5-4ee2-42e3-ac57-d5e3f0c96508 tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] [instance: c486c00d-61fc-4795-9299-e6548b686ee8] Powering on the VM {{(pid=61806) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 928.642652] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-a926ceef-5365-48e9-a9c0-22e056a6c2b9 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 928.649812] env[61806]: DEBUG oslo_vmware.api [None req-531998d5-4ee2-42e3-ac57-d5e3f0c96508 tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] Waiting for the task: (returnval){ [ 928.649812] env[61806]: value = "task-1294914" [ 928.649812] env[61806]: _type = "Task" [ 928.649812] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 928.657835] env[61806]: DEBUG oslo_vmware.api [None req-531998d5-4ee2-42e3-ac57-d5e3f0c96508 tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] Task: {'id': task-1294914, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 929.040172] env[61806]: DEBUG nova.compute.utils [None req-0c4ed53e-d274-4427-995e-e2d8d9d3b2cc tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Using /dev/sd instead of None {{(pid=61806) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 929.044783] env[61806]: DEBUG nova.compute.manager [None req-0c4ed53e-d274-4427-995e-e2d8d9d3b2cc tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] [instance: 676f18b3-3da1-4a53-b96d-fe64ee8f9101] Allocating IP information in the background. {{(pid=61806) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 929.045452] env[61806]: DEBUG nova.network.neutron [None req-0c4ed53e-d274-4427-995e-e2d8d9d3b2cc tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] [instance: 676f18b3-3da1-4a53-b96d-fe64ee8f9101] allocate_for_instance() {{(pid=61806) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 929.064348] env[61806]: DEBUG oslo_vmware.api [None req-018e57c8-1b55-4515-9361-baa3401dd22f tempest-ServersAaction247Test-1786617390 tempest-ServersAaction247Test-1786617390-project-member] Task: {'id': task-1294913, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 929.087292] env[61806]: DEBUG nova.policy [None req-0c4ed53e-d274-4427-995e-e2d8d9d3b2cc tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'b2631ce1f6dd40a4bfc5da7f90cde260', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '4b9d94c3d684474b99e305fb6cd9b95f', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61806) authorize /opt/stack/nova/nova/policy.py:201}} [ 929.107597] env[61806]: DEBUG oslo_concurrency.lockutils [None req-bdb89e04-5c89-42d4-937f-829d8aee12b6 tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] Acquiring lock "refresh_cache-fd9e520a-2e7b-43d4-b84d-8933c32cc6a6" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 929.107797] env[61806]: DEBUG oslo_concurrency.lockutils [None req-bdb89e04-5c89-42d4-937f-829d8aee12b6 tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] Acquired lock "refresh_cache-fd9e520a-2e7b-43d4-b84d-8933c32cc6a6" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 929.108021] env[61806]: DEBUG nova.network.neutron [None req-bdb89e04-5c89-42d4-937f-829d8aee12b6 tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] [instance: fd9e520a-2e7b-43d4-b84d-8933c32cc6a6] Building network info cache for instance {{(pid=61806) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 929.162393] env[61806]: DEBUG oslo_vmware.api [None req-531998d5-4ee2-42e3-ac57-d5e3f0c96508 tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] Task: {'id': task-1294914, 'name': PowerOnVM_Task, 'duration_secs': 0.460944} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 929.162750] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [None req-531998d5-4ee2-42e3-ac57-d5e3f0c96508 tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] [instance: c486c00d-61fc-4795-9299-e6548b686ee8] Powered on the VM {{(pid=61806) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 929.163013] env[61806]: INFO nova.compute.manager [None req-531998d5-4ee2-42e3-ac57-d5e3f0c96508 tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] [instance: c486c00d-61fc-4795-9299-e6548b686ee8] Took 9.31 seconds to spawn the instance on the hypervisor. [ 929.163520] env[61806]: DEBUG nova.compute.manager [None req-531998d5-4ee2-42e3-ac57-d5e3f0c96508 tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] [instance: c486c00d-61fc-4795-9299-e6548b686ee8] Checking state {{(pid=61806) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 929.164180] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a3aff18b-a434-4d53-b2f2-cc92dd3d56f0 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 929.222681] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4be7bd72-6c62-45be-932e-e931eb539400 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 929.231970] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f5fcf2f8-36e7-4220-b143-8bdccac60034 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 929.275629] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9654f2a8-0d7e-4e99-9596-157d8d885194 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 929.284632] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-92e5ac71-2c94-4408-b106-0b9007da5d94 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 929.300139] env[61806]: DEBUG nova.compute.provider_tree [None req-33516f06-7375-4e3f-a28f-63fccd97ac4e tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Inventory has not changed in ProviderTree for provider: a2858be1-fd22-4e08-979e-87ad25293407 {{(pid=61806) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 929.349821] env[61806]: DEBUG nova.virt.vmwareapi.ds_util [None req-d76979a9-290f-44f9-bf19-39a08fc72d03 tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61806) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 929.350038] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-d76979a9-290f-44f9-bf19-39a08fc72d03 tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61806) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 929.350807] env[61806]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-8cfab71a-5c86-4416-92a3-7c8a0bebafef {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 929.358852] env[61806]: DEBUG oslo_vmware.api [None req-d76979a9-290f-44f9-bf19-39a08fc72d03 tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Waiting for the task: (returnval){ [ 929.358852] env[61806]: value = "session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]5225aaa8-d85b-f128-76cf-4dcb6db62dd4" [ 929.358852] env[61806]: _type = "Task" [ 929.358852] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 929.368916] env[61806]: DEBUG oslo_vmware.api [None req-d76979a9-290f-44f9-bf19-39a08fc72d03 tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Task: {'id': session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]5225aaa8-d85b-f128-76cf-4dcb6db62dd4, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 929.393700] env[61806]: DEBUG nova.network.neutron [None req-0c4ed53e-d274-4427-995e-e2d8d9d3b2cc tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] [instance: 676f18b3-3da1-4a53-b96d-fe64ee8f9101] Successfully created port: d3a921a7-f1fa-4352-ae53-9f2996576fef {{(pid=61806) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 929.489602] env[61806]: DEBUG nova.network.neutron [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] [instance: 43dcce0d-4fdb-4ee1-8309-92e92d2331a9] Updating instance_info_cache with network_info: [{"id": "6361e800-fbf2-48e0-a11d-b00d9a9955b4", "address": "fa:16:3e:db:ef:a2", "network": {"id": "d636ca55-0737-48e4-8d9d-7fe0f797f49e", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-2090334349-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.164", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "22e47882c2c44ce492dff6b1c5d782e2", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6f493cd8-1cb4-42a1-8d56-bfa6ac7cf563", "external-id": "nsx-vlan-transportzone-931", "segmentation_id": 931, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap6361e800-fb", "ovs_interfaceid": "6361e800-fbf2-48e0-a11d-b00d9a9955b4", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61806) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 929.546266] env[61806]: DEBUG nova.compute.manager [None req-0c4ed53e-d274-4427-995e-e2d8d9d3b2cc tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] [instance: 676f18b3-3da1-4a53-b96d-fe64ee8f9101] Start building block device mappings for instance. {{(pid=61806) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 929.562483] env[61806]: DEBUG oslo_vmware.api [None req-018e57c8-1b55-4515-9361-baa3401dd22f tempest-ServersAaction247Test-1786617390 tempest-ServersAaction247Test-1786617390-project-member] Task: {'id': task-1294913, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 929.641404] env[61806]: DEBUG nova.network.neutron [None req-bdb89e04-5c89-42d4-937f-829d8aee12b6 tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] [instance: fd9e520a-2e7b-43d4-b84d-8933c32cc6a6] Instance cache missing network info. {{(pid=61806) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 929.684185] env[61806]: INFO nova.compute.manager [None req-531998d5-4ee2-42e3-ac57-d5e3f0c96508 tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] [instance: c486c00d-61fc-4795-9299-e6548b686ee8] Took 16.88 seconds to build instance. [ 929.804989] env[61806]: DEBUG nova.scheduler.client.report [None req-33516f06-7375-4e3f-a28f-63fccd97ac4e tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Inventory has not changed for provider a2858be1-fd22-4e08-979e-87ad25293407 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 138, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61806) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 929.809342] env[61806]: DEBUG nova.network.neutron [None req-bdb89e04-5c89-42d4-937f-829d8aee12b6 tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] [instance: fd9e520a-2e7b-43d4-b84d-8933c32cc6a6] Updating instance_info_cache with network_info: [{"id": "670f75e3-58f9-493b-9148-416e2e9cea5d", "address": "fa:16:3e:76:be:e1", "network": {"id": "a9d993bc-0817-4ee6-ab6e-1ef9432a110f", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-939246736-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "bbbb09a49203421d807f733616c854df", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ad4c015b-4a8b-46ca-9556-74bad8db9fb3", "external-id": "nsx-vlan-transportzone-246", "segmentation_id": 246, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap670f75e3-58", "ovs_interfaceid": "670f75e3-58f9-493b-9148-416e2e9cea5d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61806) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 929.869234] env[61806]: DEBUG oslo_vmware.api [None req-d76979a9-290f-44f9-bf19-39a08fc72d03 tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Task: {'id': session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]5225aaa8-d85b-f128-76cf-4dcb6db62dd4, 'name': SearchDatastore_Task, 'duration_secs': 0.013165} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 929.870084] env[61806]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-44fb1433-b2ce-40f4-92aa-8fd1fbb4669d {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 929.879775] env[61806]: DEBUG oslo_vmware.api [None req-d76979a9-290f-44f9-bf19-39a08fc72d03 tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Waiting for the task: (returnval){ [ 929.879775] env[61806]: value = "session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]529db794-c9cd-4bb7-bd77-240d13c46c73" [ 929.879775] env[61806]: _type = "Task" [ 929.879775] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 929.890317] env[61806]: DEBUG oslo_vmware.api [None req-d76979a9-290f-44f9-bf19-39a08fc72d03 tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Task: {'id': session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]529db794-c9cd-4bb7-bd77-240d13c46c73, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 929.992396] env[61806]: DEBUG oslo_concurrency.lockutils [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Releasing lock "refresh_cache-43dcce0d-4fdb-4ee1-8309-92e92d2331a9" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 929.992396] env[61806]: DEBUG nova.compute.manager [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] [instance: 43dcce0d-4fdb-4ee1-8309-92e92d2331a9] Updated the network info_cache for instance {{(pid=61806) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9997}} [ 929.992594] env[61806]: DEBUG oslo_service.periodic_task [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=61806) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 929.992748] env[61806]: DEBUG oslo_service.periodic_task [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=61806) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 929.992897] env[61806]: DEBUG oslo_service.periodic_task [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=61806) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 929.993056] env[61806]: DEBUG oslo_service.periodic_task [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=61806) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 929.993208] env[61806]: DEBUG oslo_service.periodic_task [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=61806) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 929.993350] env[61806]: DEBUG oslo_service.periodic_task [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=61806) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 929.993477] env[61806]: DEBUG nova.compute.manager [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=61806) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10545}} [ 929.993619] env[61806]: DEBUG oslo_service.periodic_task [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Running periodic task ComputeManager.update_available_resource {{(pid=61806) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 930.064401] env[61806]: DEBUG oslo_vmware.api [None req-018e57c8-1b55-4515-9361-baa3401dd22f tempest-ServersAaction247Test-1786617390 tempest-ServersAaction247Test-1786617390-project-member] Task: {'id': task-1294913, 'name': PowerOnVM_Task, 'duration_secs': 2.356672} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 930.064918] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [None req-018e57c8-1b55-4515-9361-baa3401dd22f tempest-ServersAaction247Test-1786617390 tempest-ServersAaction247Test-1786617390-project-member] [instance: 3f4fa1ba-a6ad-4e0f-aebc-973816ee09d5] Powered on the VM {{(pid=61806) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 930.065033] env[61806]: INFO nova.compute.manager [None req-018e57c8-1b55-4515-9361-baa3401dd22f tempest-ServersAaction247Test-1786617390 tempest-ServersAaction247Test-1786617390-project-member] [instance: 3f4fa1ba-a6ad-4e0f-aebc-973816ee09d5] Took 7.93 seconds to spawn the instance on the hypervisor. [ 930.065150] env[61806]: DEBUG nova.compute.manager [None req-018e57c8-1b55-4515-9361-baa3401dd22f tempest-ServersAaction247Test-1786617390 tempest-ServersAaction247Test-1786617390-project-member] [instance: 3f4fa1ba-a6ad-4e0f-aebc-973816ee09d5] Checking state {{(pid=61806) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 930.065988] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-04f3b094-0bc6-4f17-8419-70dd291fa08c {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 930.186867] env[61806]: DEBUG oslo_concurrency.lockutils [None req-531998d5-4ee2-42e3-ac57-d5e3f0c96508 tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] Lock "c486c00d-61fc-4795-9299-e6548b686ee8" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 18.394s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 930.308472] env[61806]: DEBUG oslo_concurrency.lockutils [None req-33516f06-7375-4e3f-a28f-63fccd97ac4e tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.774s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 930.311982] env[61806]: DEBUG oslo_concurrency.lockutils [None req-bdb89e04-5c89-42d4-937f-829d8aee12b6 tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] Releasing lock "refresh_cache-fd9e520a-2e7b-43d4-b84d-8933c32cc6a6" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 930.312293] env[61806]: DEBUG nova.compute.manager [None req-bdb89e04-5c89-42d4-937f-829d8aee12b6 tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] [instance: fd9e520a-2e7b-43d4-b84d-8933c32cc6a6] Instance network_info: |[{"id": "670f75e3-58f9-493b-9148-416e2e9cea5d", "address": "fa:16:3e:76:be:e1", "network": {"id": "a9d993bc-0817-4ee6-ab6e-1ef9432a110f", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-939246736-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "bbbb09a49203421d807f733616c854df", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ad4c015b-4a8b-46ca-9556-74bad8db9fb3", "external-id": "nsx-vlan-transportzone-246", "segmentation_id": 246, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap670f75e3-58", "ovs_interfaceid": "670f75e3-58f9-493b-9148-416e2e9cea5d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61806) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 930.312692] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-bdb89e04-5c89-42d4-937f-829d8aee12b6 tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] [instance: fd9e520a-2e7b-43d4-b84d-8933c32cc6a6] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:76:be:e1', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'ad4c015b-4a8b-46ca-9556-74bad8db9fb3', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '670f75e3-58f9-493b-9148-416e2e9cea5d', 'vif_model': 'vmxnet3'}] {{(pid=61806) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 930.320451] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [None req-bdb89e04-5c89-42d4-937f-829d8aee12b6 tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] Creating folder: Project (bbbb09a49203421d807f733616c854df). Parent ref: group-v277609. {{(pid=61806) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 930.320737] env[61806]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-c684048f-09c2-4faa-971e-00285540faa5 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 930.328257] env[61806]: INFO nova.scheduler.client.report [None req-33516f06-7375-4e3f-a28f-63fccd97ac4e tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Deleted allocations for instance 75f1aded-d1da-4ddc-be29-f1fb93799364 [ 930.333594] env[61806]: INFO nova.virt.vmwareapi.vm_util [None req-bdb89e04-5c89-42d4-937f-829d8aee12b6 tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] Created folder: Project (bbbb09a49203421d807f733616c854df) in parent group-v277609. [ 930.333941] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [None req-bdb89e04-5c89-42d4-937f-829d8aee12b6 tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] Creating folder: Instances. Parent ref: group-v277717. {{(pid=61806) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 930.334087] env[61806]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-261705fe-5c3e-4017-a21d-0eefd5ad57d4 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 930.344914] env[61806]: INFO nova.virt.vmwareapi.vm_util [None req-bdb89e04-5c89-42d4-937f-829d8aee12b6 tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] Created folder: Instances in parent group-v277717. [ 930.345187] env[61806]: DEBUG oslo.service.loopingcall [None req-bdb89e04-5c89-42d4-937f-829d8aee12b6 tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61806) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 930.345390] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: fd9e520a-2e7b-43d4-b84d-8933c32cc6a6] Creating VM on the ESX host {{(pid=61806) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 930.345601] env[61806]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-e261cd3e-5d8c-498a-a331-0771c872202e {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 930.366337] env[61806]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 930.366337] env[61806]: value = "task-1294917" [ 930.366337] env[61806]: _type = "Task" [ 930.366337] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 930.374779] env[61806]: DEBUG oslo_vmware.api [-] Task: {'id': task-1294917, 'name': CreateVM_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 930.389501] env[61806]: DEBUG oslo_vmware.api [None req-d76979a9-290f-44f9-bf19-39a08fc72d03 tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Task: {'id': session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]529db794-c9cd-4bb7-bd77-240d13c46c73, 'name': SearchDatastore_Task, 'duration_secs': 0.015185} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 930.389775] env[61806]: DEBUG oslo_concurrency.lockutils [None req-d76979a9-290f-44f9-bf19-39a08fc72d03 tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Releasing lock "[datastore1] devstack-image-cache_base/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa.vmdk" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 930.390045] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [None req-d76979a9-290f-44f9-bf19-39a08fc72d03 tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa.vmdk to [datastore1] 380aa63f-46ba-437d-bc29-0fc141ad1cf3/380aa63f-46ba-437d-bc29-0fc141ad1cf3.vmdk {{(pid=61806) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 930.390342] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-b74604b9-400c-42b3-9045-ff20bd4e3d54 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 930.398472] env[61806]: DEBUG oslo_vmware.api [None req-d76979a9-290f-44f9-bf19-39a08fc72d03 tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Waiting for the task: (returnval){ [ 930.398472] env[61806]: value = "task-1294918" [ 930.398472] env[61806]: _type = "Task" [ 930.398472] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 930.406749] env[61806]: DEBUG oslo_vmware.api [None req-d76979a9-290f-44f9-bf19-39a08fc72d03 tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Task: {'id': task-1294918, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 930.498799] env[61806]: DEBUG oslo_concurrency.lockutils [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 930.499124] env[61806]: DEBUG oslo_concurrency.lockutils [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 0.001s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 930.499379] env[61806]: DEBUG oslo_concurrency.lockutils [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 930.499617] env[61806]: DEBUG nova.compute.resource_tracker [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=61806) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 930.500875] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b04ac6cb-34dc-472a-966e-2c9129ae22d4 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 930.517901] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-55c5c096-4b50-41c8-baed-42ba91a9e052 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 930.535730] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-635748a1-e43c-485d-b899-9b2c9dd82534 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 930.541360] env[61806]: DEBUG nova.compute.manager [req-fbcafcf8-6ed6-44a3-ab72-39f829d54786 req-2abc4b40-63a2-4075-a8ed-b6c312bc9732 service nova] [instance: fd9e520a-2e7b-43d4-b84d-8933c32cc6a6] Received event network-changed-670f75e3-58f9-493b-9148-416e2e9cea5d {{(pid=61806) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 930.541618] env[61806]: DEBUG nova.compute.manager [req-fbcafcf8-6ed6-44a3-ab72-39f829d54786 req-2abc4b40-63a2-4075-a8ed-b6c312bc9732 service nova] [instance: fd9e520a-2e7b-43d4-b84d-8933c32cc6a6] Refreshing instance network info cache due to event network-changed-670f75e3-58f9-493b-9148-416e2e9cea5d. {{(pid=61806) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 930.541787] env[61806]: DEBUG oslo_concurrency.lockutils [req-fbcafcf8-6ed6-44a3-ab72-39f829d54786 req-2abc4b40-63a2-4075-a8ed-b6c312bc9732 service nova] Acquiring lock "refresh_cache-fd9e520a-2e7b-43d4-b84d-8933c32cc6a6" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 930.541937] env[61806]: DEBUG oslo_concurrency.lockutils [req-fbcafcf8-6ed6-44a3-ab72-39f829d54786 req-2abc4b40-63a2-4075-a8ed-b6c312bc9732 service nova] Acquired lock "refresh_cache-fd9e520a-2e7b-43d4-b84d-8933c32cc6a6" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 930.542122] env[61806]: DEBUG nova.network.neutron [req-fbcafcf8-6ed6-44a3-ab72-39f829d54786 req-2abc4b40-63a2-4075-a8ed-b6c312bc9732 service nova] [instance: fd9e520a-2e7b-43d4-b84d-8933c32cc6a6] Refreshing network info cache for port 670f75e3-58f9-493b-9148-416e2e9cea5d {{(pid=61806) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 930.547217] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-650f557f-275b-444b-a29f-00ed1bab3b26 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 930.579581] env[61806]: DEBUG nova.compute.manager [None req-0c4ed53e-d274-4427-995e-e2d8d9d3b2cc tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] [instance: 676f18b3-3da1-4a53-b96d-fe64ee8f9101] Start spawning the instance on the hypervisor. {{(pid=61806) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 930.585062] env[61806]: DEBUG nova.compute.resource_tracker [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=180218MB free_disk=138GB free_vcpus=48 pci_devices=None {{(pid=61806) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 930.585231] env[61806]: DEBUG oslo_concurrency.lockutils [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 930.585424] env[61806]: DEBUG oslo_concurrency.lockutils [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 0.000s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 930.589771] env[61806]: INFO nova.compute.manager [None req-018e57c8-1b55-4515-9361-baa3401dd22f tempest-ServersAaction247Test-1786617390 tempest-ServersAaction247Test-1786617390-project-member] [instance: 3f4fa1ba-a6ad-4e0f-aebc-973816ee09d5] Took 14.10 seconds to build instance. [ 930.617633] env[61806]: DEBUG nova.virt.hardware [None req-0c4ed53e-d274-4427-995e-e2d8d9d3b2cc tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-15T18:05:10Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-15T18:04:52Z,direct_url=,disk_format='vmdk',id=ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='84ba83fce2c349a6988173c5d6fc3b07',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-15T18:04:53Z,virtual_size=,visibility=), allow threads: False {{(pid=61806) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 930.617903] env[61806]: DEBUG nova.virt.hardware [None req-0c4ed53e-d274-4427-995e-e2d8d9d3b2cc tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Flavor limits 0:0:0 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 930.618100] env[61806]: DEBUG nova.virt.hardware [None req-0c4ed53e-d274-4427-995e-e2d8d9d3b2cc tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Image limits 0:0:0 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 930.618298] env[61806]: DEBUG nova.virt.hardware [None req-0c4ed53e-d274-4427-995e-e2d8d9d3b2cc tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Flavor pref 0:0:0 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 930.618454] env[61806]: DEBUG nova.virt.hardware [None req-0c4ed53e-d274-4427-995e-e2d8d9d3b2cc tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Image pref 0:0:0 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 930.618607] env[61806]: DEBUG nova.virt.hardware [None req-0c4ed53e-d274-4427-995e-e2d8d9d3b2cc tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 930.618817] env[61806]: DEBUG nova.virt.hardware [None req-0c4ed53e-d274-4427-995e-e2d8d9d3b2cc tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61806) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 930.618984] env[61806]: DEBUG nova.virt.hardware [None req-0c4ed53e-d274-4427-995e-e2d8d9d3b2cc tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61806) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 930.619173] env[61806]: DEBUG nova.virt.hardware [None req-0c4ed53e-d274-4427-995e-e2d8d9d3b2cc tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Got 1 possible topologies {{(pid=61806) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 930.619343] env[61806]: DEBUG nova.virt.hardware [None req-0c4ed53e-d274-4427-995e-e2d8d9d3b2cc tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61806) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 930.619526] env[61806]: DEBUG nova.virt.hardware [None req-0c4ed53e-d274-4427-995e-e2d8d9d3b2cc tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61806) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 930.620441] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d897088e-eaa7-4d02-b3a0-c44a3c6b88e7 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 930.630889] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6a4745b8-edef-4fe4-a8fb-9e94540d2490 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 930.835699] env[61806]: DEBUG oslo_concurrency.lockutils [None req-33516f06-7375-4e3f-a28f-63fccd97ac4e tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Lock "75f1aded-d1da-4ddc-be29-f1fb93799364" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 8.117s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 930.876340] env[61806]: DEBUG oslo_vmware.api [-] Task: {'id': task-1294917, 'name': CreateVM_Task, 'duration_secs': 0.352229} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 930.876689] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: fd9e520a-2e7b-43d4-b84d-8933c32cc6a6] Created VM on the ESX host {{(pid=61806) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 930.877275] env[61806]: DEBUG oslo_concurrency.lockutils [None req-bdb89e04-5c89-42d4-937f-829d8aee12b6 tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 930.877448] env[61806]: DEBUG oslo_concurrency.lockutils [None req-bdb89e04-5c89-42d4-937f-829d8aee12b6 tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] Acquired lock "[datastore2] devstack-image-cache_base/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 930.877802] env[61806]: DEBUG oslo_concurrency.lockutils [None req-bdb89e04-5c89-42d4-937f-829d8aee12b6 tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 930.878078] env[61806]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-1efd20a1-3566-4f15-8eb9-c9b3d5a575a6 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 930.883989] env[61806]: DEBUG oslo_vmware.api [None req-bdb89e04-5c89-42d4-937f-829d8aee12b6 tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] Waiting for the task: (returnval){ [ 930.883989] env[61806]: value = "session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]5257305a-a93f-585b-1f53-4c3042f8b3eb" [ 930.883989] env[61806]: _type = "Task" [ 930.883989] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 930.892800] env[61806]: DEBUG oslo_vmware.api [None req-bdb89e04-5c89-42d4-937f-829d8aee12b6 tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] Task: {'id': session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]5257305a-a93f-585b-1f53-4c3042f8b3eb, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 930.910027] env[61806]: DEBUG oslo_vmware.api [None req-d76979a9-290f-44f9-bf19-39a08fc72d03 tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Task: {'id': task-1294918, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 930.959704] env[61806]: DEBUG nova.network.neutron [None req-0c4ed53e-d274-4427-995e-e2d8d9d3b2cc tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] [instance: 676f18b3-3da1-4a53-b96d-fe64ee8f9101] Successfully updated port: d3a921a7-f1fa-4352-ae53-9f2996576fef {{(pid=61806) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 931.096777] env[61806]: DEBUG oslo_concurrency.lockutils [None req-018e57c8-1b55-4515-9361-baa3401dd22f tempest-ServersAaction247Test-1786617390 tempest-ServersAaction247Test-1786617390-project-member] Lock "3f4fa1ba-a6ad-4e0f-aebc-973816ee09d5" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 15.615s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 931.142020] env[61806]: DEBUG oslo_concurrency.lockutils [None req-9dda5c0d-d86c-4217-825a-c2d7c73e302d tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] Acquiring lock "c486c00d-61fc-4795-9299-e6548b686ee8" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 931.142020] env[61806]: DEBUG oslo_concurrency.lockutils [None req-9dda5c0d-d86c-4217-825a-c2d7c73e302d tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] Lock "c486c00d-61fc-4795-9299-e6548b686ee8" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 931.142020] env[61806]: DEBUG oslo_concurrency.lockutils [None req-9dda5c0d-d86c-4217-825a-c2d7c73e302d tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] Acquiring lock "c486c00d-61fc-4795-9299-e6548b686ee8-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 931.142020] env[61806]: DEBUG oslo_concurrency.lockutils [None req-9dda5c0d-d86c-4217-825a-c2d7c73e302d tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] Lock "c486c00d-61fc-4795-9299-e6548b686ee8-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.001s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 931.142020] env[61806]: DEBUG oslo_concurrency.lockutils [None req-9dda5c0d-d86c-4217-825a-c2d7c73e302d tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] Lock "c486c00d-61fc-4795-9299-e6548b686ee8-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 931.143530] env[61806]: INFO nova.compute.manager [None req-9dda5c0d-d86c-4217-825a-c2d7c73e302d tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] [instance: c486c00d-61fc-4795-9299-e6548b686ee8] Terminating instance [ 931.146147] env[61806]: DEBUG nova.compute.manager [None req-9dda5c0d-d86c-4217-825a-c2d7c73e302d tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] [instance: c486c00d-61fc-4795-9299-e6548b686ee8] Start destroying the instance on the hypervisor. {{(pid=61806) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 931.146501] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-9dda5c0d-d86c-4217-825a-c2d7c73e302d tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] [instance: c486c00d-61fc-4795-9299-e6548b686ee8] Destroying instance {{(pid=61806) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 931.147614] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-34357fd8-36c7-4777-b388-2b6dacc496e2 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 931.157414] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [None req-9dda5c0d-d86c-4217-825a-c2d7c73e302d tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] [instance: c486c00d-61fc-4795-9299-e6548b686ee8] Powering off the VM {{(pid=61806) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 931.157701] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-a9a8f564-b334-48ad-9b2f-81f00f19b2e1 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 931.163408] env[61806]: DEBUG oslo_vmware.api [None req-9dda5c0d-d86c-4217-825a-c2d7c73e302d tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] Waiting for the task: (returnval){ [ 931.163408] env[61806]: value = "task-1294919" [ 931.163408] env[61806]: _type = "Task" [ 931.163408] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 931.171074] env[61806]: DEBUG oslo_vmware.api [None req-9dda5c0d-d86c-4217-825a-c2d7c73e302d tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] Task: {'id': task-1294919, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 931.305731] env[61806]: DEBUG nova.network.neutron [req-fbcafcf8-6ed6-44a3-ab72-39f829d54786 req-2abc4b40-63a2-4075-a8ed-b6c312bc9732 service nova] [instance: fd9e520a-2e7b-43d4-b84d-8933c32cc6a6] Updated VIF entry in instance network info cache for port 670f75e3-58f9-493b-9148-416e2e9cea5d. {{(pid=61806) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 931.306164] env[61806]: DEBUG nova.network.neutron [req-fbcafcf8-6ed6-44a3-ab72-39f829d54786 req-2abc4b40-63a2-4075-a8ed-b6c312bc9732 service nova] [instance: fd9e520a-2e7b-43d4-b84d-8933c32cc6a6] Updating instance_info_cache with network_info: [{"id": "670f75e3-58f9-493b-9148-416e2e9cea5d", "address": "fa:16:3e:76:be:e1", "network": {"id": "a9d993bc-0817-4ee6-ab6e-1ef9432a110f", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-939246736-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "bbbb09a49203421d807f733616c854df", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ad4c015b-4a8b-46ca-9556-74bad8db9fb3", "external-id": "nsx-vlan-transportzone-246", "segmentation_id": 246, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap670f75e3-58", "ovs_interfaceid": "670f75e3-58f9-493b-9148-416e2e9cea5d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61806) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 931.396337] env[61806]: DEBUG oslo_vmware.api [None req-bdb89e04-5c89-42d4-937f-829d8aee12b6 tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] Task: {'id': session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]5257305a-a93f-585b-1f53-4c3042f8b3eb, 'name': SearchDatastore_Task, 'duration_secs': 0.024393} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 931.396639] env[61806]: DEBUG oslo_concurrency.lockutils [None req-bdb89e04-5c89-42d4-937f-829d8aee12b6 tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] Releasing lock "[datastore2] devstack-image-cache_base/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 931.396951] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-bdb89e04-5c89-42d4-937f-829d8aee12b6 tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] [instance: fd9e520a-2e7b-43d4-b84d-8933c32cc6a6] Processing image ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa {{(pid=61806) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 931.397243] env[61806]: DEBUG oslo_concurrency.lockutils [None req-bdb89e04-5c89-42d4-937f-829d8aee12b6 tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa.vmdk" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 931.397413] env[61806]: DEBUG oslo_concurrency.lockutils [None req-bdb89e04-5c89-42d4-937f-829d8aee12b6 tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] Acquired lock "[datastore2] devstack-image-cache_base/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa.vmdk" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 931.397639] env[61806]: DEBUG nova.virt.vmwareapi.ds_util [None req-bdb89e04-5c89-42d4-937f-829d8aee12b6 tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61806) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 931.397931] env[61806]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-ffdb0646-8a84-47b1-8134-fb6bed477505 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 931.407619] env[61806]: DEBUG oslo_vmware.api [None req-d76979a9-290f-44f9-bf19-39a08fc72d03 tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Task: {'id': task-1294918, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.632149} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 931.407875] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [None req-d76979a9-290f-44f9-bf19-39a08fc72d03 tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa.vmdk to [datastore1] 380aa63f-46ba-437d-bc29-0fc141ad1cf3/380aa63f-46ba-437d-bc29-0fc141ad1cf3.vmdk {{(pid=61806) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 931.408079] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-d76979a9-290f-44f9-bf19-39a08fc72d03 tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] [instance: 380aa63f-46ba-437d-bc29-0fc141ad1cf3] Extending root virtual disk to 1048576 {{(pid=61806) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 931.409117] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-e59dafe1-bf4f-4e13-a4d7-c660ca946a18 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 931.410860] env[61806]: DEBUG nova.virt.vmwareapi.ds_util [None req-bdb89e04-5c89-42d4-937f-829d8aee12b6 tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61806) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 931.411079] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-bdb89e04-5c89-42d4-937f-829d8aee12b6 tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61806) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 931.411749] env[61806]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-889a35ab-c900-4cf5-b1da-2f41c04f3d57 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 931.417166] env[61806]: DEBUG oslo_vmware.api [None req-bdb89e04-5c89-42d4-937f-829d8aee12b6 tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] Waiting for the task: (returnval){ [ 931.417166] env[61806]: value = "session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]522d6f64-955e-986e-ae12-cb484c0b8701" [ 931.417166] env[61806]: _type = "Task" [ 931.417166] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 931.418347] env[61806]: DEBUG oslo_vmware.api [None req-d76979a9-290f-44f9-bf19-39a08fc72d03 tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Waiting for the task: (returnval){ [ 931.418347] env[61806]: value = "task-1294920" [ 931.418347] env[61806]: _type = "Task" [ 931.418347] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 931.429879] env[61806]: DEBUG oslo_vmware.api [None req-bdb89e04-5c89-42d4-937f-829d8aee12b6 tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] Task: {'id': session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]522d6f64-955e-986e-ae12-cb484c0b8701, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 931.432999] env[61806]: DEBUG oslo_vmware.api [None req-d76979a9-290f-44f9-bf19-39a08fc72d03 tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Task: {'id': task-1294920, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 931.462413] env[61806]: DEBUG oslo_concurrency.lockutils [None req-0c4ed53e-d274-4427-995e-e2d8d9d3b2cc tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Acquiring lock "refresh_cache-676f18b3-3da1-4a53-b96d-fe64ee8f9101" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 931.462695] env[61806]: DEBUG oslo_concurrency.lockutils [None req-0c4ed53e-d274-4427-995e-e2d8d9d3b2cc tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Acquired lock "refresh_cache-676f18b3-3da1-4a53-b96d-fe64ee8f9101" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 931.462907] env[61806]: DEBUG nova.network.neutron [None req-0c4ed53e-d274-4427-995e-e2d8d9d3b2cc tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] [instance: 676f18b3-3da1-4a53-b96d-fe64ee8f9101] Building network info cache for instance {{(pid=61806) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 931.621239] env[61806]: DEBUG nova.compute.resource_tracker [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Instance 43dcce0d-4fdb-4ee1-8309-92e92d2331a9 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61806) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 931.621405] env[61806]: DEBUG nova.compute.resource_tracker [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Instance 67d1bdd8-cf8f-4fb5-bd3c-42008e4d60e3 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61806) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 931.621532] env[61806]: DEBUG nova.compute.resource_tracker [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Instance d1802da4-a670-427d-84d7-0fcb4717e18f actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61806) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 931.621669] env[61806]: DEBUG nova.compute.resource_tracker [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Instance daff849e-eb4f-411e-92f8-9362e3bd91e3 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61806) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 931.621787] env[61806]: DEBUG nova.compute.resource_tracker [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Instance 6f6ba57a-e2d6-4749-a53a-e263861cb1c0 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61806) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 931.621906] env[61806]: DEBUG nova.compute.resource_tracker [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Instance d9caee4a-a00c-4e50-b01c-99a8f796a09f actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61806) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 931.622032] env[61806]: DEBUG nova.compute.resource_tracker [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Instance c486c00d-61fc-4795-9299-e6548b686ee8 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61806) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 931.622153] env[61806]: DEBUG nova.compute.resource_tracker [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Instance 3f4fa1ba-a6ad-4e0f-aebc-973816ee09d5 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61806) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 931.622267] env[61806]: DEBUG nova.compute.resource_tracker [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Instance 380aa63f-46ba-437d-bc29-0fc141ad1cf3 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61806) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 931.622380] env[61806]: DEBUG nova.compute.resource_tracker [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Instance fd9e520a-2e7b-43d4-b84d-8933c32cc6a6 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61806) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 931.622492] env[61806]: DEBUG nova.compute.resource_tracker [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Instance 676f18b3-3da1-4a53-b96d-fe64ee8f9101 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61806) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 931.622679] env[61806]: DEBUG nova.compute.resource_tracker [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Total usable vcpus: 48, total allocated vcpus: 11 {{(pid=61806) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 931.622819] env[61806]: DEBUG nova.compute.resource_tracker [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=2624MB phys_disk=200GB used_disk=11GB total_vcpus=48 used_vcpus=11 pci_stats=[] {{(pid=61806) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 931.674060] env[61806]: DEBUG oslo_vmware.api [None req-9dda5c0d-d86c-4217-825a-c2d7c73e302d tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] Task: {'id': task-1294919, 'name': PowerOffVM_Task, 'duration_secs': 0.181179} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 931.674367] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [None req-9dda5c0d-d86c-4217-825a-c2d7c73e302d tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] [instance: c486c00d-61fc-4795-9299-e6548b686ee8] Powered off the VM {{(pid=61806) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 931.674540] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-9dda5c0d-d86c-4217-825a-c2d7c73e302d tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] [instance: c486c00d-61fc-4795-9299-e6548b686ee8] Unregistering the VM {{(pid=61806) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 931.674824] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-a8a4cbf1-3a9e-46f0-a897-8221722d0f1f {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 931.734870] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-9dda5c0d-d86c-4217-825a-c2d7c73e302d tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] [instance: c486c00d-61fc-4795-9299-e6548b686ee8] Unregistered the VM {{(pid=61806) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 931.735121] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-9dda5c0d-d86c-4217-825a-c2d7c73e302d tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] [instance: c486c00d-61fc-4795-9299-e6548b686ee8] Deleting contents of the VM from datastore datastore1 {{(pid=61806) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 931.735342] env[61806]: DEBUG nova.virt.vmwareapi.ds_util [None req-9dda5c0d-d86c-4217-825a-c2d7c73e302d tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] Deleting the datastore file [datastore1] c486c00d-61fc-4795-9299-e6548b686ee8 {{(pid=61806) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 931.735631] env[61806]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-6773a8b5-bb9c-40b3-9872-0221bce53492 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 931.743010] env[61806]: DEBUG oslo_vmware.api [None req-9dda5c0d-d86c-4217-825a-c2d7c73e302d tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] Waiting for the task: (returnval){ [ 931.743010] env[61806]: value = "task-1294922" [ 931.743010] env[61806]: _type = "Task" [ 931.743010] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 931.750322] env[61806]: DEBUG oslo_vmware.api [None req-9dda5c0d-d86c-4217-825a-c2d7c73e302d tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] Task: {'id': task-1294922, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 931.751774] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e4ff481c-7ed8-4f7b-bf9f-270ae709cb65 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 931.757938] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-54be1892-e4aa-43eb-b5e1-20d6458931c0 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 931.788289] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-05e9e77b-26b7-4d17-8eea-b20bafaf6e7d {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 931.794831] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aa5b43f7-6557-4714-918e-347669718ef4 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 931.807783] env[61806]: DEBUG nova.compute.provider_tree [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Inventory has not changed in ProviderTree for provider: a2858be1-fd22-4e08-979e-87ad25293407 {{(pid=61806) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 931.809165] env[61806]: DEBUG oslo_concurrency.lockutils [req-fbcafcf8-6ed6-44a3-ab72-39f829d54786 req-2abc4b40-63a2-4075-a8ed-b6c312bc9732 service nova] Releasing lock "refresh_cache-fd9e520a-2e7b-43d4-b84d-8933c32cc6a6" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 931.930816] env[61806]: DEBUG oslo_vmware.api [None req-d76979a9-290f-44f9-bf19-39a08fc72d03 tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Task: {'id': task-1294920, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.228074} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 931.934403] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-d76979a9-290f-44f9-bf19-39a08fc72d03 tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] [instance: 380aa63f-46ba-437d-bc29-0fc141ad1cf3] Extended root virtual disk {{(pid=61806) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 931.934740] env[61806]: DEBUG oslo_vmware.api [None req-bdb89e04-5c89-42d4-937f-829d8aee12b6 tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] Task: {'id': session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]522d6f64-955e-986e-ae12-cb484c0b8701, 'name': SearchDatastore_Task, 'duration_secs': 0.015121} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 931.935458] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-859c0cfb-bedc-4638-adfb-8392c185d984 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 931.938288] env[61806]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-dadc4173-3c44-40b4-9cdb-323a2637d936 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 931.960099] env[61806]: DEBUG nova.virt.vmwareapi.volumeops [None req-d76979a9-290f-44f9-bf19-39a08fc72d03 tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] [instance: 380aa63f-46ba-437d-bc29-0fc141ad1cf3] Reconfiguring VM instance instance-00000056 to attach disk [datastore1] 380aa63f-46ba-437d-bc29-0fc141ad1cf3/380aa63f-46ba-437d-bc29-0fc141ad1cf3.vmdk or device None with type sparse {{(pid=61806) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 931.961464] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-e5d35a2e-29ce-4924-bfda-b18abc0148ab {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 931.977101] env[61806]: DEBUG oslo_vmware.api [None req-bdb89e04-5c89-42d4-937f-829d8aee12b6 tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] Waiting for the task: (returnval){ [ 931.977101] env[61806]: value = "session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]52db0fa6-9cdd-5542-4146-433421a14c53" [ 931.977101] env[61806]: _type = "Task" [ 931.977101] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 931.979096] env[61806]: DEBUG nova.compute.manager [None req-e9736ad2-6387-42e9-9138-c30403fdc0ce tempest-ServersAaction247Test-1786617390 tempest-ServersAaction247Test-1786617390-project-member] [instance: 3f4fa1ba-a6ad-4e0f-aebc-973816ee09d5] Checking state {{(pid=61806) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 931.979937] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8552ccf9-c9f5-4f8c-874c-f25115a4a7ba {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 931.987629] env[61806]: DEBUG oslo_vmware.api [None req-d76979a9-290f-44f9-bf19-39a08fc72d03 tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Waiting for the task: (returnval){ [ 931.987629] env[61806]: value = "task-1294923" [ 931.987629] env[61806]: _type = "Task" [ 931.987629] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 931.998066] env[61806]: DEBUG oslo_vmware.api [None req-bdb89e04-5c89-42d4-937f-829d8aee12b6 tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] Task: {'id': session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]52db0fa6-9cdd-5542-4146-433421a14c53, 'name': SearchDatastore_Task, 'duration_secs': 0.03315} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 931.998270] env[61806]: DEBUG oslo_concurrency.lockutils [None req-bdb89e04-5c89-42d4-937f-829d8aee12b6 tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] Releasing lock "[datastore2] devstack-image-cache_base/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa.vmdk" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 931.998601] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [None req-bdb89e04-5c89-42d4-937f-829d8aee12b6 tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa.vmdk to [datastore2] fd9e520a-2e7b-43d4-b84d-8933c32cc6a6/fd9e520a-2e7b-43d4-b84d-8933c32cc6a6.vmdk {{(pid=61806) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 931.999238] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-d5b30b59-7ad5-497e-854d-2802a2e284cf {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 932.004932] env[61806]: DEBUG oslo_vmware.api [None req-d76979a9-290f-44f9-bf19-39a08fc72d03 tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Task: {'id': task-1294923, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 932.006210] env[61806]: DEBUG oslo_vmware.api [None req-bdb89e04-5c89-42d4-937f-829d8aee12b6 tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] Waiting for the task: (returnval){ [ 932.006210] env[61806]: value = "task-1294924" [ 932.006210] env[61806]: _type = "Task" [ 932.006210] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 932.015468] env[61806]: DEBUG oslo_vmware.api [None req-bdb89e04-5c89-42d4-937f-829d8aee12b6 tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] Task: {'id': task-1294924, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 932.027485] env[61806]: DEBUG oslo_concurrency.lockutils [None req-d547963d-0e28-4458-8b96-eaec3d80a721 tempest-ServersAaction247Test-1786617390 tempest-ServersAaction247Test-1786617390-project-member] Acquiring lock "3f4fa1ba-a6ad-4e0f-aebc-973816ee09d5" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 932.027781] env[61806]: DEBUG oslo_concurrency.lockutils [None req-d547963d-0e28-4458-8b96-eaec3d80a721 tempest-ServersAaction247Test-1786617390 tempest-ServersAaction247Test-1786617390-project-member] Lock "3f4fa1ba-a6ad-4e0f-aebc-973816ee09d5" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 932.028048] env[61806]: DEBUG oslo_concurrency.lockutils [None req-d547963d-0e28-4458-8b96-eaec3d80a721 tempest-ServersAaction247Test-1786617390 tempest-ServersAaction247Test-1786617390-project-member] Acquiring lock "3f4fa1ba-a6ad-4e0f-aebc-973816ee09d5-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 932.028286] env[61806]: DEBUG oslo_concurrency.lockutils [None req-d547963d-0e28-4458-8b96-eaec3d80a721 tempest-ServersAaction247Test-1786617390 tempest-ServersAaction247Test-1786617390-project-member] Lock "3f4fa1ba-a6ad-4e0f-aebc-973816ee09d5-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 932.028508] env[61806]: DEBUG oslo_concurrency.lockutils [None req-d547963d-0e28-4458-8b96-eaec3d80a721 tempest-ServersAaction247Test-1786617390 tempest-ServersAaction247Test-1786617390-project-member] Lock "3f4fa1ba-a6ad-4e0f-aebc-973816ee09d5-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 932.030931] env[61806]: DEBUG nova.network.neutron [None req-0c4ed53e-d274-4427-995e-e2d8d9d3b2cc tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] [instance: 676f18b3-3da1-4a53-b96d-fe64ee8f9101] Instance cache missing network info. {{(pid=61806) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 932.033471] env[61806]: INFO nova.compute.manager [None req-d547963d-0e28-4458-8b96-eaec3d80a721 tempest-ServersAaction247Test-1786617390 tempest-ServersAaction247Test-1786617390-project-member] [instance: 3f4fa1ba-a6ad-4e0f-aebc-973816ee09d5] Terminating instance [ 932.035497] env[61806]: DEBUG oslo_concurrency.lockutils [None req-d547963d-0e28-4458-8b96-eaec3d80a721 tempest-ServersAaction247Test-1786617390 tempest-ServersAaction247Test-1786617390-project-member] Acquiring lock "refresh_cache-3f4fa1ba-a6ad-4e0f-aebc-973816ee09d5" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 932.035670] env[61806]: DEBUG oslo_concurrency.lockutils [None req-d547963d-0e28-4458-8b96-eaec3d80a721 tempest-ServersAaction247Test-1786617390 tempest-ServersAaction247Test-1786617390-project-member] Acquired lock "refresh_cache-3f4fa1ba-a6ad-4e0f-aebc-973816ee09d5" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 932.035842] env[61806]: DEBUG nova.network.neutron [None req-d547963d-0e28-4458-8b96-eaec3d80a721 tempest-ServersAaction247Test-1786617390 tempest-ServersAaction247Test-1786617390-project-member] [instance: 3f4fa1ba-a6ad-4e0f-aebc-973816ee09d5] Building network info cache for instance {{(pid=61806) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 932.193291] env[61806]: DEBUG nova.network.neutron [None req-0c4ed53e-d274-4427-995e-e2d8d9d3b2cc tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] [instance: 676f18b3-3da1-4a53-b96d-fe64ee8f9101] Updating instance_info_cache with network_info: [{"id": "d3a921a7-f1fa-4352-ae53-9f2996576fef", "address": "fa:16:3e:66:40:96", "network": {"id": "46515a4e-610f-42fb-a424-205a03df0640", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-1133623547-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4b9d94c3d684474b99e305fb6cd9b95f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "82ca17df-257e-40e6-9ec9-310ed6f05ccb", "external-id": "nsx-vlan-transportzone-778", "segmentation_id": 778, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd3a921a7-f1", "ovs_interfaceid": "d3a921a7-f1fa-4352-ae53-9f2996576fef", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61806) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 932.252793] env[61806]: DEBUG oslo_vmware.api [None req-9dda5c0d-d86c-4217-825a-c2d7c73e302d tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] Task: {'id': task-1294922, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.485189} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 932.253079] env[61806]: DEBUG nova.virt.vmwareapi.ds_util [None req-9dda5c0d-d86c-4217-825a-c2d7c73e302d tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] Deleted the datastore file {{(pid=61806) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 932.253272] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-9dda5c0d-d86c-4217-825a-c2d7c73e302d tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] [instance: c486c00d-61fc-4795-9299-e6548b686ee8] Deleted contents of the VM from datastore datastore1 {{(pid=61806) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 932.253455] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-9dda5c0d-d86c-4217-825a-c2d7c73e302d tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] [instance: c486c00d-61fc-4795-9299-e6548b686ee8] Instance destroyed {{(pid=61806) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 932.253720] env[61806]: INFO nova.compute.manager [None req-9dda5c0d-d86c-4217-825a-c2d7c73e302d tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] [instance: c486c00d-61fc-4795-9299-e6548b686ee8] Took 1.11 seconds to destroy the instance on the hypervisor. [ 932.254109] env[61806]: DEBUG oslo.service.loopingcall [None req-9dda5c0d-d86c-4217-825a-c2d7c73e302d tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61806) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 932.254384] env[61806]: DEBUG nova.compute.manager [-] [instance: c486c00d-61fc-4795-9299-e6548b686ee8] Deallocating network for instance {{(pid=61806) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 932.254532] env[61806]: DEBUG nova.network.neutron [-] [instance: c486c00d-61fc-4795-9299-e6548b686ee8] deallocate_for_instance() {{(pid=61806) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 932.310909] env[61806]: DEBUG nova.scheduler.client.report [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Inventory has not changed for provider a2858be1-fd22-4e08-979e-87ad25293407 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 138, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61806) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 932.499445] env[61806]: INFO nova.compute.manager [None req-e9736ad2-6387-42e9-9138-c30403fdc0ce tempest-ServersAaction247Test-1786617390 tempest-ServersAaction247Test-1786617390-project-member] [instance: 3f4fa1ba-a6ad-4e0f-aebc-973816ee09d5] instance snapshotting [ 932.500309] env[61806]: DEBUG nova.objects.instance [None req-e9736ad2-6387-42e9-9138-c30403fdc0ce tempest-ServersAaction247Test-1786617390 tempest-ServersAaction247Test-1786617390-project-member] Lazy-loading 'flavor' on Instance uuid 3f4fa1ba-a6ad-4e0f-aebc-973816ee09d5 {{(pid=61806) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 932.506662] env[61806]: DEBUG oslo_vmware.api [None req-d76979a9-290f-44f9-bf19-39a08fc72d03 tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Task: {'id': task-1294923, 'name': ReconfigVM_Task, 'duration_secs': 0.294738} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 932.506894] env[61806]: DEBUG nova.virt.vmwareapi.volumeops [None req-d76979a9-290f-44f9-bf19-39a08fc72d03 tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] [instance: 380aa63f-46ba-437d-bc29-0fc141ad1cf3] Reconfigured VM instance instance-00000056 to attach disk [datastore1] 380aa63f-46ba-437d-bc29-0fc141ad1cf3/380aa63f-46ba-437d-bc29-0fc141ad1cf3.vmdk or device None with type sparse {{(pid=61806) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 932.511118] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-0f7c1cc6-46f3-4e84-a582-dd5596397910 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 932.520481] env[61806]: DEBUG oslo_vmware.api [None req-bdb89e04-5c89-42d4-937f-829d8aee12b6 tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] Task: {'id': task-1294924, 'name': CopyVirtualDisk_Task} progress is 25%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 932.521945] env[61806]: DEBUG oslo_vmware.api [None req-d76979a9-290f-44f9-bf19-39a08fc72d03 tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Waiting for the task: (returnval){ [ 932.521945] env[61806]: value = "task-1294925" [ 932.521945] env[61806]: _type = "Task" [ 932.521945] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 932.532865] env[61806]: DEBUG oslo_vmware.api [None req-d76979a9-290f-44f9-bf19-39a08fc72d03 tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Task: {'id': task-1294925, 'name': Rename_Task} progress is 5%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 932.560673] env[61806]: DEBUG nova.network.neutron [None req-d547963d-0e28-4458-8b96-eaec3d80a721 tempest-ServersAaction247Test-1786617390 tempest-ServersAaction247Test-1786617390-project-member] [instance: 3f4fa1ba-a6ad-4e0f-aebc-973816ee09d5] Instance cache missing network info. {{(pid=61806) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 932.576472] env[61806]: DEBUG nova.compute.manager [req-302fce67-d7cd-4027-bf8e-66ad1c968b7e req-7cbde3ec-592b-43b2-82ba-0b41bb5791d8 service nova] [instance: 676f18b3-3da1-4a53-b96d-fe64ee8f9101] Received event network-vif-plugged-d3a921a7-f1fa-4352-ae53-9f2996576fef {{(pid=61806) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 932.576947] env[61806]: DEBUG oslo_concurrency.lockutils [req-302fce67-d7cd-4027-bf8e-66ad1c968b7e req-7cbde3ec-592b-43b2-82ba-0b41bb5791d8 service nova] Acquiring lock "676f18b3-3da1-4a53-b96d-fe64ee8f9101-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 932.577013] env[61806]: DEBUG oslo_concurrency.lockutils [req-302fce67-d7cd-4027-bf8e-66ad1c968b7e req-7cbde3ec-592b-43b2-82ba-0b41bb5791d8 service nova] Lock "676f18b3-3da1-4a53-b96d-fe64ee8f9101-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 932.577178] env[61806]: DEBUG oslo_concurrency.lockutils [req-302fce67-d7cd-4027-bf8e-66ad1c968b7e req-7cbde3ec-592b-43b2-82ba-0b41bb5791d8 service nova] Lock "676f18b3-3da1-4a53-b96d-fe64ee8f9101-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 932.577418] env[61806]: DEBUG nova.compute.manager [req-302fce67-d7cd-4027-bf8e-66ad1c968b7e req-7cbde3ec-592b-43b2-82ba-0b41bb5791d8 service nova] [instance: 676f18b3-3da1-4a53-b96d-fe64ee8f9101] No waiting events found dispatching network-vif-plugged-d3a921a7-f1fa-4352-ae53-9f2996576fef {{(pid=61806) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 932.577566] env[61806]: WARNING nova.compute.manager [req-302fce67-d7cd-4027-bf8e-66ad1c968b7e req-7cbde3ec-592b-43b2-82ba-0b41bb5791d8 service nova] [instance: 676f18b3-3da1-4a53-b96d-fe64ee8f9101] Received unexpected event network-vif-plugged-d3a921a7-f1fa-4352-ae53-9f2996576fef for instance with vm_state building and task_state spawning. [ 932.577738] env[61806]: DEBUG nova.compute.manager [req-302fce67-d7cd-4027-bf8e-66ad1c968b7e req-7cbde3ec-592b-43b2-82ba-0b41bb5791d8 service nova] [instance: 676f18b3-3da1-4a53-b96d-fe64ee8f9101] Received event network-changed-d3a921a7-f1fa-4352-ae53-9f2996576fef {{(pid=61806) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 932.578218] env[61806]: DEBUG nova.compute.manager [req-302fce67-d7cd-4027-bf8e-66ad1c968b7e req-7cbde3ec-592b-43b2-82ba-0b41bb5791d8 service nova] [instance: 676f18b3-3da1-4a53-b96d-fe64ee8f9101] Refreshing instance network info cache due to event network-changed-d3a921a7-f1fa-4352-ae53-9f2996576fef. {{(pid=61806) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 932.578358] env[61806]: DEBUG oslo_concurrency.lockutils [req-302fce67-d7cd-4027-bf8e-66ad1c968b7e req-7cbde3ec-592b-43b2-82ba-0b41bb5791d8 service nova] Acquiring lock "refresh_cache-676f18b3-3da1-4a53-b96d-fe64ee8f9101" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 932.657139] env[61806]: DEBUG nova.network.neutron [None req-d547963d-0e28-4458-8b96-eaec3d80a721 tempest-ServersAaction247Test-1786617390 tempest-ServersAaction247Test-1786617390-project-member] [instance: 3f4fa1ba-a6ad-4e0f-aebc-973816ee09d5] Updating instance_info_cache with network_info: [] {{(pid=61806) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 932.698715] env[61806]: DEBUG oslo_concurrency.lockutils [None req-0c4ed53e-d274-4427-995e-e2d8d9d3b2cc tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Releasing lock "refresh_cache-676f18b3-3da1-4a53-b96d-fe64ee8f9101" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 932.699079] env[61806]: DEBUG nova.compute.manager [None req-0c4ed53e-d274-4427-995e-e2d8d9d3b2cc tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] [instance: 676f18b3-3da1-4a53-b96d-fe64ee8f9101] Instance network_info: |[{"id": "d3a921a7-f1fa-4352-ae53-9f2996576fef", "address": "fa:16:3e:66:40:96", "network": {"id": "46515a4e-610f-42fb-a424-205a03df0640", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-1133623547-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4b9d94c3d684474b99e305fb6cd9b95f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "82ca17df-257e-40e6-9ec9-310ed6f05ccb", "external-id": "nsx-vlan-transportzone-778", "segmentation_id": 778, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd3a921a7-f1", "ovs_interfaceid": "d3a921a7-f1fa-4352-ae53-9f2996576fef", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61806) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 932.699444] env[61806]: DEBUG oslo_concurrency.lockutils [req-302fce67-d7cd-4027-bf8e-66ad1c968b7e req-7cbde3ec-592b-43b2-82ba-0b41bb5791d8 service nova] Acquired lock "refresh_cache-676f18b3-3da1-4a53-b96d-fe64ee8f9101" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 932.699623] env[61806]: DEBUG nova.network.neutron [req-302fce67-d7cd-4027-bf8e-66ad1c968b7e req-7cbde3ec-592b-43b2-82ba-0b41bb5791d8 service nova] [instance: 676f18b3-3da1-4a53-b96d-fe64ee8f9101] Refreshing network info cache for port d3a921a7-f1fa-4352-ae53-9f2996576fef {{(pid=61806) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 932.701234] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-0c4ed53e-d274-4427-995e-e2d8d9d3b2cc tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] [instance: 676f18b3-3da1-4a53-b96d-fe64ee8f9101] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:66:40:96', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '82ca17df-257e-40e6-9ec9-310ed6f05ccb', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'd3a921a7-f1fa-4352-ae53-9f2996576fef', 'vif_model': 'vmxnet3'}] {{(pid=61806) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 932.709647] env[61806]: DEBUG oslo.service.loopingcall [None req-0c4ed53e-d274-4427-995e-e2d8d9d3b2cc tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61806) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 932.712920] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 676f18b3-3da1-4a53-b96d-fe64ee8f9101] Creating VM on the ESX host {{(pid=61806) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 932.713658] env[61806]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-a5beeee7-7add-439c-8b0a-40524b044b4b {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 932.734534] env[61806]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 932.734534] env[61806]: value = "task-1294926" [ 932.734534] env[61806]: _type = "Task" [ 932.734534] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 932.743127] env[61806]: DEBUG oslo_vmware.api [-] Task: {'id': task-1294926, 'name': CreateVM_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 932.816812] env[61806]: DEBUG nova.compute.resource_tracker [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=61806) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 932.817060] env[61806]: DEBUG oslo_concurrency.lockutils [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 2.232s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 932.927511] env[61806]: DEBUG nova.network.neutron [req-302fce67-d7cd-4027-bf8e-66ad1c968b7e req-7cbde3ec-592b-43b2-82ba-0b41bb5791d8 service nova] [instance: 676f18b3-3da1-4a53-b96d-fe64ee8f9101] Updated VIF entry in instance network info cache for port d3a921a7-f1fa-4352-ae53-9f2996576fef. {{(pid=61806) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 932.927879] env[61806]: DEBUG nova.network.neutron [req-302fce67-d7cd-4027-bf8e-66ad1c968b7e req-7cbde3ec-592b-43b2-82ba-0b41bb5791d8 service nova] [instance: 676f18b3-3da1-4a53-b96d-fe64ee8f9101] Updating instance_info_cache with network_info: [{"id": "d3a921a7-f1fa-4352-ae53-9f2996576fef", "address": "fa:16:3e:66:40:96", "network": {"id": "46515a4e-610f-42fb-a424-205a03df0640", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-1133623547-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4b9d94c3d684474b99e305fb6cd9b95f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "82ca17df-257e-40e6-9ec9-310ed6f05ccb", "external-id": "nsx-vlan-transportzone-778", "segmentation_id": 778, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd3a921a7-f1", "ovs_interfaceid": "d3a921a7-f1fa-4352-ae53-9f2996576fef", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61806) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 933.006821] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6faeec2e-a445-438e-9ffc-5b135b9366f1 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 933.009534] env[61806]: DEBUG nova.network.neutron [-] [instance: c486c00d-61fc-4795-9299-e6548b686ee8] Updating instance_info_cache with network_info: [] {{(pid=61806) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 933.034022] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eeee4daf-d3b7-4b63-b62d-dc4c9f78bf29 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 933.038562] env[61806]: DEBUG oslo_vmware.api [None req-bdb89e04-5c89-42d4-937f-829d8aee12b6 tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] Task: {'id': task-1294924, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.677167} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 933.038838] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [None req-bdb89e04-5c89-42d4-937f-829d8aee12b6 tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa.vmdk to [datastore2] fd9e520a-2e7b-43d4-b84d-8933c32cc6a6/fd9e520a-2e7b-43d4-b84d-8933c32cc6a6.vmdk {{(pid=61806) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 933.039060] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-bdb89e04-5c89-42d4-937f-829d8aee12b6 tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] [instance: fd9e520a-2e7b-43d4-b84d-8933c32cc6a6] Extending root virtual disk to 1048576 {{(pid=61806) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 933.039893] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-58f7a255-5a9d-4b29-a638-98bdd9c0d59b {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 933.046300] env[61806]: DEBUG oslo_vmware.api [None req-d76979a9-290f-44f9-bf19-39a08fc72d03 tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Task: {'id': task-1294925, 'name': Rename_Task, 'duration_secs': 0.164073} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 933.048879] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [None req-d76979a9-290f-44f9-bf19-39a08fc72d03 tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] [instance: 380aa63f-46ba-437d-bc29-0fc141ad1cf3] Powering on the VM {{(pid=61806) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 933.049315] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-effb72ed-b1e5-4e96-9798-f22bc0ab4a24 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 933.052822] env[61806]: DEBUG oslo_vmware.api [None req-bdb89e04-5c89-42d4-937f-829d8aee12b6 tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] Waiting for the task: (returnval){ [ 933.052822] env[61806]: value = "task-1294927" [ 933.052822] env[61806]: _type = "Task" [ 933.052822] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 933.056617] env[61806]: DEBUG oslo_vmware.api [None req-d76979a9-290f-44f9-bf19-39a08fc72d03 tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Waiting for the task: (returnval){ [ 933.056617] env[61806]: value = "task-1294928" [ 933.056617] env[61806]: _type = "Task" [ 933.056617] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 933.062960] env[61806]: DEBUG oslo_vmware.api [None req-bdb89e04-5c89-42d4-937f-829d8aee12b6 tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] Task: {'id': task-1294927, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 933.068088] env[61806]: DEBUG oslo_vmware.api [None req-d76979a9-290f-44f9-bf19-39a08fc72d03 tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Task: {'id': task-1294928, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 933.160060] env[61806]: DEBUG oslo_concurrency.lockutils [None req-d547963d-0e28-4458-8b96-eaec3d80a721 tempest-ServersAaction247Test-1786617390 tempest-ServersAaction247Test-1786617390-project-member] Releasing lock "refresh_cache-3f4fa1ba-a6ad-4e0f-aebc-973816ee09d5" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 933.160597] env[61806]: DEBUG nova.compute.manager [None req-d547963d-0e28-4458-8b96-eaec3d80a721 tempest-ServersAaction247Test-1786617390 tempest-ServersAaction247Test-1786617390-project-member] [instance: 3f4fa1ba-a6ad-4e0f-aebc-973816ee09d5] Start destroying the instance on the hypervisor. {{(pid=61806) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 933.160847] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-d547963d-0e28-4458-8b96-eaec3d80a721 tempest-ServersAaction247Test-1786617390 tempest-ServersAaction247Test-1786617390-project-member] [instance: 3f4fa1ba-a6ad-4e0f-aebc-973816ee09d5] Destroying instance {{(pid=61806) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 933.161834] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6cb88e25-3cc8-431f-9733-b3b45dedaf6a {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 933.170245] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [None req-d547963d-0e28-4458-8b96-eaec3d80a721 tempest-ServersAaction247Test-1786617390 tempest-ServersAaction247Test-1786617390-project-member] [instance: 3f4fa1ba-a6ad-4e0f-aebc-973816ee09d5] Powering off the VM {{(pid=61806) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 933.170548] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-be17410f-48c7-4c53-be9a-8a09fa9b042d {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 933.176715] env[61806]: DEBUG oslo_vmware.api [None req-d547963d-0e28-4458-8b96-eaec3d80a721 tempest-ServersAaction247Test-1786617390 tempest-ServersAaction247Test-1786617390-project-member] Waiting for the task: (returnval){ [ 933.176715] env[61806]: value = "task-1294929" [ 933.176715] env[61806]: _type = "Task" [ 933.176715] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 933.184217] env[61806]: DEBUG oslo_vmware.api [None req-d547963d-0e28-4458-8b96-eaec3d80a721 tempest-ServersAaction247Test-1786617390 tempest-ServersAaction247Test-1786617390-project-member] Task: {'id': task-1294929, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 933.244904] env[61806]: DEBUG oslo_vmware.api [-] Task: {'id': task-1294926, 'name': CreateVM_Task} progress is 99%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 933.430595] env[61806]: DEBUG oslo_concurrency.lockutils [req-302fce67-d7cd-4027-bf8e-66ad1c968b7e req-7cbde3ec-592b-43b2-82ba-0b41bb5791d8 service nova] Releasing lock "refresh_cache-676f18b3-3da1-4a53-b96d-fe64ee8f9101" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 933.430904] env[61806]: DEBUG nova.compute.manager [req-302fce67-d7cd-4027-bf8e-66ad1c968b7e req-7cbde3ec-592b-43b2-82ba-0b41bb5791d8 service nova] [instance: c486c00d-61fc-4795-9299-e6548b686ee8] Received event network-vif-deleted-df6e9523-3657-4405-8d44-5b35cfc6fa9f {{(pid=61806) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 933.431110] env[61806]: INFO nova.compute.manager [req-302fce67-d7cd-4027-bf8e-66ad1c968b7e req-7cbde3ec-592b-43b2-82ba-0b41bb5791d8 service nova] [instance: c486c00d-61fc-4795-9299-e6548b686ee8] Neutron deleted interface df6e9523-3657-4405-8d44-5b35cfc6fa9f; detaching it from the instance and deleting it from the info cache [ 933.431294] env[61806]: DEBUG nova.network.neutron [req-302fce67-d7cd-4027-bf8e-66ad1c968b7e req-7cbde3ec-592b-43b2-82ba-0b41bb5791d8 service nova] [instance: c486c00d-61fc-4795-9299-e6548b686ee8] Updating instance_info_cache with network_info: [] {{(pid=61806) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 933.515683] env[61806]: INFO nova.compute.manager [-] [instance: c486c00d-61fc-4795-9299-e6548b686ee8] Took 1.26 seconds to deallocate network for instance. [ 933.549854] env[61806]: DEBUG nova.compute.manager [None req-e9736ad2-6387-42e9-9138-c30403fdc0ce tempest-ServersAaction247Test-1786617390 tempest-ServersAaction247Test-1786617390-project-member] [instance: 3f4fa1ba-a6ad-4e0f-aebc-973816ee09d5] Instance disappeared during snapshot {{(pid=61806) _snapshot_instance /opt/stack/nova/nova/compute/manager.py:4494}} [ 933.562991] env[61806]: DEBUG oslo_vmware.api [None req-bdb89e04-5c89-42d4-937f-829d8aee12b6 tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] Task: {'id': task-1294927, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.067263} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 933.563588] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-bdb89e04-5c89-42d4-937f-829d8aee12b6 tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] [instance: fd9e520a-2e7b-43d4-b84d-8933c32cc6a6] Extended root virtual disk {{(pid=61806) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 933.564405] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8f18a87a-9951-4b99-bbe7-111365ed7061 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 933.569690] env[61806]: DEBUG oslo_vmware.api [None req-d76979a9-290f-44f9-bf19-39a08fc72d03 tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Task: {'id': task-1294928, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 933.590901] env[61806]: DEBUG nova.virt.vmwareapi.volumeops [None req-bdb89e04-5c89-42d4-937f-829d8aee12b6 tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] [instance: fd9e520a-2e7b-43d4-b84d-8933c32cc6a6] Reconfiguring VM instance instance-00000057 to attach disk [datastore2] fd9e520a-2e7b-43d4-b84d-8933c32cc6a6/fd9e520a-2e7b-43d4-b84d-8933c32cc6a6.vmdk or device None with type sparse {{(pid=61806) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 933.594575] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-e7e06205-328b-4b72-8766-0017be2da5dd {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 933.614226] env[61806]: DEBUG oslo_vmware.api [None req-bdb89e04-5c89-42d4-937f-829d8aee12b6 tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] Waiting for the task: (returnval){ [ 933.614226] env[61806]: value = "task-1294930" [ 933.614226] env[61806]: _type = "Task" [ 933.614226] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 933.621911] env[61806]: DEBUG oslo_vmware.api [None req-bdb89e04-5c89-42d4-937f-829d8aee12b6 tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] Task: {'id': task-1294930, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 933.689057] env[61806]: DEBUG oslo_vmware.api [None req-d547963d-0e28-4458-8b96-eaec3d80a721 tempest-ServersAaction247Test-1786617390 tempest-ServersAaction247Test-1786617390-project-member] Task: {'id': task-1294929, 'name': PowerOffVM_Task, 'duration_secs': 0.112186} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 933.689355] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [None req-d547963d-0e28-4458-8b96-eaec3d80a721 tempest-ServersAaction247Test-1786617390 tempest-ServersAaction247Test-1786617390-project-member] [instance: 3f4fa1ba-a6ad-4e0f-aebc-973816ee09d5] Powered off the VM {{(pid=61806) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 933.689530] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-d547963d-0e28-4458-8b96-eaec3d80a721 tempest-ServersAaction247Test-1786617390 tempest-ServersAaction247Test-1786617390-project-member] [instance: 3f4fa1ba-a6ad-4e0f-aebc-973816ee09d5] Unregistering the VM {{(pid=61806) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 933.689786] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-043b4700-2452-4a70-bb83-eb7c5a329989 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 933.713864] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-d547963d-0e28-4458-8b96-eaec3d80a721 tempest-ServersAaction247Test-1786617390 tempest-ServersAaction247Test-1786617390-project-member] [instance: 3f4fa1ba-a6ad-4e0f-aebc-973816ee09d5] Unregistered the VM {{(pid=61806) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 933.714211] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-d547963d-0e28-4458-8b96-eaec3d80a721 tempest-ServersAaction247Test-1786617390 tempest-ServersAaction247Test-1786617390-project-member] [instance: 3f4fa1ba-a6ad-4e0f-aebc-973816ee09d5] Deleting contents of the VM from datastore datastore1 {{(pid=61806) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 933.714421] env[61806]: DEBUG nova.virt.vmwareapi.ds_util [None req-d547963d-0e28-4458-8b96-eaec3d80a721 tempest-ServersAaction247Test-1786617390 tempest-ServersAaction247Test-1786617390-project-member] Deleting the datastore file [datastore1] 3f4fa1ba-a6ad-4e0f-aebc-973816ee09d5 {{(pid=61806) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 933.714702] env[61806]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-0e9cb05c-1e9d-4f55-91e0-e9f3046a76b4 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 933.719203] env[61806]: DEBUG nova.compute.manager [None req-e9736ad2-6387-42e9-9138-c30403fdc0ce tempest-ServersAaction247Test-1786617390 tempest-ServersAaction247Test-1786617390-project-member] [instance: 3f4fa1ba-a6ad-4e0f-aebc-973816ee09d5] Found 0 images (rotation: 2) {{(pid=61806) _rotate_backups /opt/stack/nova/nova/compute/manager.py:4554}} [ 933.721342] env[61806]: DEBUG oslo_vmware.api [None req-d547963d-0e28-4458-8b96-eaec3d80a721 tempest-ServersAaction247Test-1786617390 tempest-ServersAaction247Test-1786617390-project-member] Waiting for the task: (returnval){ [ 933.721342] env[61806]: value = "task-1294932" [ 933.721342] env[61806]: _type = "Task" [ 933.721342] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 933.730037] env[61806]: DEBUG oslo_vmware.api [None req-d547963d-0e28-4458-8b96-eaec3d80a721 tempest-ServersAaction247Test-1786617390 tempest-ServersAaction247Test-1786617390-project-member] Task: {'id': task-1294932, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 933.746421] env[61806]: DEBUG oslo_vmware.api [-] Task: {'id': task-1294926, 'name': CreateVM_Task} progress is 99%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 933.934723] env[61806]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-e6667a56-2c60-4c5d-a9fd-2f866c26cec9 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 933.943954] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-40c3fd16-30ab-4b91-9bf2-acc76f93a674 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 933.970850] env[61806]: DEBUG nova.compute.manager [req-302fce67-d7cd-4027-bf8e-66ad1c968b7e req-7cbde3ec-592b-43b2-82ba-0b41bb5791d8 service nova] [instance: c486c00d-61fc-4795-9299-e6548b686ee8] Detach interface failed, port_id=df6e9523-3657-4405-8d44-5b35cfc6fa9f, reason: Instance c486c00d-61fc-4795-9299-e6548b686ee8 could not be found. {{(pid=61806) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10963}} [ 934.021937] env[61806]: DEBUG oslo_concurrency.lockutils [None req-9dda5c0d-d86c-4217-825a-c2d7c73e302d tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 934.022283] env[61806]: DEBUG oslo_concurrency.lockutils [None req-9dda5c0d-d86c-4217-825a-c2d7c73e302d tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.001s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 934.022547] env[61806]: DEBUG nova.objects.instance [None req-9dda5c0d-d86c-4217-825a-c2d7c73e302d tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] Lazy-loading 'resources' on Instance uuid c486c00d-61fc-4795-9299-e6548b686ee8 {{(pid=61806) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 934.067805] env[61806]: DEBUG oslo_vmware.api [None req-d76979a9-290f-44f9-bf19-39a08fc72d03 tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Task: {'id': task-1294928, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 934.124201] env[61806]: DEBUG oslo_vmware.api [None req-bdb89e04-5c89-42d4-937f-829d8aee12b6 tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] Task: {'id': task-1294930, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 934.232165] env[61806]: DEBUG oslo_vmware.api [None req-d547963d-0e28-4458-8b96-eaec3d80a721 tempest-ServersAaction247Test-1786617390 tempest-ServersAaction247Test-1786617390-project-member] Task: {'id': task-1294932, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.117639} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 934.232430] env[61806]: DEBUG nova.virt.vmwareapi.ds_util [None req-d547963d-0e28-4458-8b96-eaec3d80a721 tempest-ServersAaction247Test-1786617390 tempest-ServersAaction247Test-1786617390-project-member] Deleted the datastore file {{(pid=61806) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 934.232621] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-d547963d-0e28-4458-8b96-eaec3d80a721 tempest-ServersAaction247Test-1786617390 tempest-ServersAaction247Test-1786617390-project-member] [instance: 3f4fa1ba-a6ad-4e0f-aebc-973816ee09d5] Deleted contents of the VM from datastore datastore1 {{(pid=61806) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 934.232802] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-d547963d-0e28-4458-8b96-eaec3d80a721 tempest-ServersAaction247Test-1786617390 tempest-ServersAaction247Test-1786617390-project-member] [instance: 3f4fa1ba-a6ad-4e0f-aebc-973816ee09d5] Instance destroyed {{(pid=61806) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 934.232985] env[61806]: INFO nova.compute.manager [None req-d547963d-0e28-4458-8b96-eaec3d80a721 tempest-ServersAaction247Test-1786617390 tempest-ServersAaction247Test-1786617390-project-member] [instance: 3f4fa1ba-a6ad-4e0f-aebc-973816ee09d5] Took 1.07 seconds to destroy the instance on the hypervisor. [ 934.233255] env[61806]: DEBUG oslo.service.loopingcall [None req-d547963d-0e28-4458-8b96-eaec3d80a721 tempest-ServersAaction247Test-1786617390 tempest-ServersAaction247Test-1786617390-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61806) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 934.233459] env[61806]: DEBUG nova.compute.manager [-] [instance: 3f4fa1ba-a6ad-4e0f-aebc-973816ee09d5] Deallocating network for instance {{(pid=61806) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 934.233553] env[61806]: DEBUG nova.network.neutron [-] [instance: 3f4fa1ba-a6ad-4e0f-aebc-973816ee09d5] deallocate_for_instance() {{(pid=61806) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 934.244167] env[61806]: DEBUG oslo_vmware.api [-] Task: {'id': task-1294926, 'name': CreateVM_Task, 'duration_secs': 1.365589} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 934.244321] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 676f18b3-3da1-4a53-b96d-fe64ee8f9101] Created VM on the ESX host {{(pid=61806) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 934.244954] env[61806]: DEBUG oslo_concurrency.lockutils [None req-0c4ed53e-d274-4427-995e-e2d8d9d3b2cc tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 934.245228] env[61806]: DEBUG oslo_concurrency.lockutils [None req-0c4ed53e-d274-4427-995e-e2d8d9d3b2cc tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Acquired lock "[datastore2] devstack-image-cache_base/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 934.245457] env[61806]: DEBUG oslo_concurrency.lockutils [None req-0c4ed53e-d274-4427-995e-e2d8d9d3b2cc tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 934.245704] env[61806]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-51d152e5-e650-4f39-a4e1-4c4e5d0dd4fb {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 934.250094] env[61806]: DEBUG oslo_vmware.api [None req-0c4ed53e-d274-4427-995e-e2d8d9d3b2cc tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Waiting for the task: (returnval){ [ 934.250094] env[61806]: value = "session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]5288db90-433b-a09f-886a-c4031358fc81" [ 934.250094] env[61806]: _type = "Task" [ 934.250094] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 934.253550] env[61806]: DEBUG nova.network.neutron [-] [instance: 3f4fa1ba-a6ad-4e0f-aebc-973816ee09d5] Instance cache missing network info. {{(pid=61806) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 934.260254] env[61806]: DEBUG oslo_vmware.api [None req-0c4ed53e-d274-4427-995e-e2d8d9d3b2cc tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Task: {'id': session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]5288db90-433b-a09f-886a-c4031358fc81, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 934.569779] env[61806]: DEBUG oslo_vmware.api [None req-d76979a9-290f-44f9-bf19-39a08fc72d03 tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Task: {'id': task-1294928, 'name': PowerOnVM_Task} progress is 88%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 934.625584] env[61806]: DEBUG oslo_vmware.api [None req-bdb89e04-5c89-42d4-937f-829d8aee12b6 tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] Task: {'id': task-1294930, 'name': ReconfigVM_Task, 'duration_secs': 0.768423} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 934.625912] env[61806]: DEBUG nova.virt.vmwareapi.volumeops [None req-bdb89e04-5c89-42d4-937f-829d8aee12b6 tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] [instance: fd9e520a-2e7b-43d4-b84d-8933c32cc6a6] Reconfigured VM instance instance-00000057 to attach disk [datastore2] fd9e520a-2e7b-43d4-b84d-8933c32cc6a6/fd9e520a-2e7b-43d4-b84d-8933c32cc6a6.vmdk or device None with type sparse {{(pid=61806) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 934.626550] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-bbe2f8c0-40a1-4d88-87e7-2c08c6a83fe2 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 934.634566] env[61806]: DEBUG oslo_vmware.api [None req-bdb89e04-5c89-42d4-937f-829d8aee12b6 tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] Waiting for the task: (returnval){ [ 934.634566] env[61806]: value = "task-1294933" [ 934.634566] env[61806]: _type = "Task" [ 934.634566] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 934.646053] env[61806]: DEBUG oslo_vmware.api [None req-bdb89e04-5c89-42d4-937f-829d8aee12b6 tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] Task: {'id': task-1294933, 'name': Rename_Task} progress is 5%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 934.672790] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ca39c6d5-3b93-4de7-a706-0a42c1d6d345 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 934.679786] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ca0bee77-0874-4cd7-b408-a753ec671bac {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 934.710321] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bece6323-119c-4bff-92d6-f9ab33b8855f {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 934.718076] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2c4f665c-f558-4cc4-92da-0ba75a5e9ef9 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 934.731150] env[61806]: DEBUG nova.compute.provider_tree [None req-9dda5c0d-d86c-4217-825a-c2d7c73e302d tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] Inventory has not changed in ProviderTree for provider: a2858be1-fd22-4e08-979e-87ad25293407 {{(pid=61806) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 934.756397] env[61806]: DEBUG nova.network.neutron [-] [instance: 3f4fa1ba-a6ad-4e0f-aebc-973816ee09d5] Updating instance_info_cache with network_info: [] {{(pid=61806) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 934.761888] env[61806]: DEBUG oslo_vmware.api [None req-0c4ed53e-d274-4427-995e-e2d8d9d3b2cc tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Task: {'id': session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]5288db90-433b-a09f-886a-c4031358fc81, 'name': SearchDatastore_Task, 'duration_secs': 0.009559} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 934.762434] env[61806]: DEBUG oslo_concurrency.lockutils [None req-0c4ed53e-d274-4427-995e-e2d8d9d3b2cc tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Releasing lock "[datastore2] devstack-image-cache_base/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 934.762673] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-0c4ed53e-d274-4427-995e-e2d8d9d3b2cc tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] [instance: 676f18b3-3da1-4a53-b96d-fe64ee8f9101] Processing image ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa {{(pid=61806) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 934.762924] env[61806]: DEBUG oslo_concurrency.lockutils [None req-0c4ed53e-d274-4427-995e-e2d8d9d3b2cc tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa.vmdk" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 934.763176] env[61806]: DEBUG oslo_concurrency.lockutils [None req-0c4ed53e-d274-4427-995e-e2d8d9d3b2cc tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Acquired lock "[datastore2] devstack-image-cache_base/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa.vmdk" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 934.763401] env[61806]: DEBUG nova.virt.vmwareapi.ds_util [None req-0c4ed53e-d274-4427-995e-e2d8d9d3b2cc tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61806) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 934.763683] env[61806]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-8700071d-da1b-42ea-afbd-c86d04e61070 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 934.773914] env[61806]: DEBUG nova.virt.vmwareapi.ds_util [None req-0c4ed53e-d274-4427-995e-e2d8d9d3b2cc tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61806) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 934.774153] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-0c4ed53e-d274-4427-995e-e2d8d9d3b2cc tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61806) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 934.775124] env[61806]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ff5cc8fb-d37f-409b-86f3-d13d693a20dc {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 934.780434] env[61806]: DEBUG oslo_vmware.api [None req-0c4ed53e-d274-4427-995e-e2d8d9d3b2cc tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Waiting for the task: (returnval){ [ 934.780434] env[61806]: value = "session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]52845e69-46c0-ccf1-eec3-4563fb22bd82" [ 934.780434] env[61806]: _type = "Task" [ 934.780434] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 934.787944] env[61806]: DEBUG oslo_vmware.api [None req-0c4ed53e-d274-4427-995e-e2d8d9d3b2cc tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Task: {'id': session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]52845e69-46c0-ccf1-eec3-4563fb22bd82, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 935.068602] env[61806]: DEBUG oslo_vmware.api [None req-d76979a9-290f-44f9-bf19-39a08fc72d03 tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Task: {'id': task-1294928, 'name': PowerOnVM_Task, 'duration_secs': 1.686331} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 935.068877] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [None req-d76979a9-290f-44f9-bf19-39a08fc72d03 tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] [instance: 380aa63f-46ba-437d-bc29-0fc141ad1cf3] Powered on the VM {{(pid=61806) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 935.069098] env[61806]: INFO nova.compute.manager [None req-d76979a9-290f-44f9-bf19-39a08fc72d03 tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] [instance: 380aa63f-46ba-437d-bc29-0fc141ad1cf3] Took 10.65 seconds to spawn the instance on the hypervisor. [ 935.069288] env[61806]: DEBUG nova.compute.manager [None req-d76979a9-290f-44f9-bf19-39a08fc72d03 tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] [instance: 380aa63f-46ba-437d-bc29-0fc141ad1cf3] Checking state {{(pid=61806) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 935.070140] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2958c143-a62c-4be3-96e9-f9c923a210bb {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 935.146053] env[61806]: DEBUG oslo_vmware.api [None req-bdb89e04-5c89-42d4-937f-829d8aee12b6 tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] Task: {'id': task-1294933, 'name': Rename_Task, 'duration_secs': 0.137291} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 935.146374] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [None req-bdb89e04-5c89-42d4-937f-829d8aee12b6 tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] [instance: fd9e520a-2e7b-43d4-b84d-8933c32cc6a6] Powering on the VM {{(pid=61806) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 935.146626] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-9fef1998-5248-4a41-9025-5d096d635a47 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 935.154484] env[61806]: DEBUG oslo_vmware.api [None req-bdb89e04-5c89-42d4-937f-829d8aee12b6 tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] Waiting for the task: (returnval){ [ 935.154484] env[61806]: value = "task-1294934" [ 935.154484] env[61806]: _type = "Task" [ 935.154484] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 935.161711] env[61806]: DEBUG oslo_vmware.api [None req-bdb89e04-5c89-42d4-937f-829d8aee12b6 tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] Task: {'id': task-1294934, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 935.234393] env[61806]: DEBUG nova.scheduler.client.report [None req-9dda5c0d-d86c-4217-825a-c2d7c73e302d tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] Inventory has not changed for provider a2858be1-fd22-4e08-979e-87ad25293407 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 138, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61806) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 935.262638] env[61806]: INFO nova.compute.manager [-] [instance: 3f4fa1ba-a6ad-4e0f-aebc-973816ee09d5] Took 1.03 seconds to deallocate network for instance. [ 935.290746] env[61806]: DEBUG oslo_vmware.api [None req-0c4ed53e-d274-4427-995e-e2d8d9d3b2cc tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Task: {'id': session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]52845e69-46c0-ccf1-eec3-4563fb22bd82, 'name': SearchDatastore_Task, 'duration_secs': 0.009084} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 935.291513] env[61806]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-309fba7b-2808-4871-a914-83b9fc092e10 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 935.296955] env[61806]: DEBUG oslo_vmware.api [None req-0c4ed53e-d274-4427-995e-e2d8d9d3b2cc tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Waiting for the task: (returnval){ [ 935.296955] env[61806]: value = "session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]52ba6daf-fa5e-2348-9d0d-edb284f3acbc" [ 935.296955] env[61806]: _type = "Task" [ 935.296955] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 935.304593] env[61806]: DEBUG oslo_vmware.api [None req-0c4ed53e-d274-4427-995e-e2d8d9d3b2cc tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Task: {'id': session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]52ba6daf-fa5e-2348-9d0d-edb284f3acbc, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 935.593272] env[61806]: INFO nova.compute.manager [None req-d76979a9-290f-44f9-bf19-39a08fc72d03 tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] [instance: 380aa63f-46ba-437d-bc29-0fc141ad1cf3] Took 18.48 seconds to build instance. [ 935.665964] env[61806]: DEBUG oslo_vmware.api [None req-bdb89e04-5c89-42d4-937f-829d8aee12b6 tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] Task: {'id': task-1294934, 'name': PowerOnVM_Task, 'duration_secs': 0.439476} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 935.666314] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [None req-bdb89e04-5c89-42d4-937f-829d8aee12b6 tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] [instance: fd9e520a-2e7b-43d4-b84d-8933c32cc6a6] Powered on the VM {{(pid=61806) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 935.666527] env[61806]: INFO nova.compute.manager [None req-bdb89e04-5c89-42d4-937f-829d8aee12b6 tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] [instance: fd9e520a-2e7b-43d4-b84d-8933c32cc6a6] Took 7.39 seconds to spawn the instance on the hypervisor. [ 935.666724] env[61806]: DEBUG nova.compute.manager [None req-bdb89e04-5c89-42d4-937f-829d8aee12b6 tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] [instance: fd9e520a-2e7b-43d4-b84d-8933c32cc6a6] Checking state {{(pid=61806) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 935.667562] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-df97acdd-1c32-41b5-8c42-15acba749915 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 935.739316] env[61806]: DEBUG oslo_concurrency.lockutils [None req-9dda5c0d-d86c-4217-825a-c2d7c73e302d tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.717s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 935.757381] env[61806]: INFO nova.scheduler.client.report [None req-9dda5c0d-d86c-4217-825a-c2d7c73e302d tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] Deleted allocations for instance c486c00d-61fc-4795-9299-e6548b686ee8 [ 935.769361] env[61806]: DEBUG oslo_concurrency.lockutils [None req-d547963d-0e28-4458-8b96-eaec3d80a721 tempest-ServersAaction247Test-1786617390 tempest-ServersAaction247Test-1786617390-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 935.769361] env[61806]: DEBUG oslo_concurrency.lockutils [None req-d547963d-0e28-4458-8b96-eaec3d80a721 tempest-ServersAaction247Test-1786617390 tempest-ServersAaction247Test-1786617390-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.000s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 935.769361] env[61806]: DEBUG nova.objects.instance [None req-d547963d-0e28-4458-8b96-eaec3d80a721 tempest-ServersAaction247Test-1786617390 tempest-ServersAaction247Test-1786617390-project-member] Lazy-loading 'resources' on Instance uuid 3f4fa1ba-a6ad-4e0f-aebc-973816ee09d5 {{(pid=61806) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 935.806969] env[61806]: DEBUG oslo_vmware.api [None req-0c4ed53e-d274-4427-995e-e2d8d9d3b2cc tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Task: {'id': session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]52ba6daf-fa5e-2348-9d0d-edb284f3acbc, 'name': SearchDatastore_Task, 'duration_secs': 0.008437} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 935.807843] env[61806]: DEBUG oslo_concurrency.lockutils [None req-0c4ed53e-d274-4427-995e-e2d8d9d3b2cc tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Releasing lock "[datastore2] devstack-image-cache_base/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa.vmdk" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 935.808124] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [None req-0c4ed53e-d274-4427-995e-e2d8d9d3b2cc tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa.vmdk to [datastore2] 676f18b3-3da1-4a53-b96d-fe64ee8f9101/676f18b3-3da1-4a53-b96d-fe64ee8f9101.vmdk {{(pid=61806) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 935.808396] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-76ec4565-5265-4fec-98f5-6d64d7aeaf77 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 935.815432] env[61806]: DEBUG oslo_vmware.api [None req-0c4ed53e-d274-4427-995e-e2d8d9d3b2cc tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Waiting for the task: (returnval){ [ 935.815432] env[61806]: value = "task-1294935" [ 935.815432] env[61806]: _type = "Task" [ 935.815432] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 935.823479] env[61806]: DEBUG oslo_vmware.api [None req-0c4ed53e-d274-4427-995e-e2d8d9d3b2cc tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Task: {'id': task-1294935, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 936.095609] env[61806]: DEBUG oslo_concurrency.lockutils [None req-d76979a9-290f-44f9-bf19-39a08fc72d03 tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Lock "380aa63f-46ba-437d-bc29-0fc141ad1cf3" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 19.990s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 936.190031] env[61806]: INFO nova.compute.manager [None req-bdb89e04-5c89-42d4-937f-829d8aee12b6 tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] [instance: fd9e520a-2e7b-43d4-b84d-8933c32cc6a6] Took 12.23 seconds to build instance. [ 936.264544] env[61806]: DEBUG oslo_concurrency.lockutils [None req-9dda5c0d-d86c-4217-825a-c2d7c73e302d tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] Lock "c486c00d-61fc-4795-9299-e6548b686ee8" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.124s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 936.324871] env[61806]: DEBUG oslo_vmware.api [None req-0c4ed53e-d274-4427-995e-e2d8d9d3b2cc tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Task: {'id': task-1294935, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.458504} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 936.327135] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [None req-0c4ed53e-d274-4427-995e-e2d8d9d3b2cc tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa.vmdk to [datastore2] 676f18b3-3da1-4a53-b96d-fe64ee8f9101/676f18b3-3da1-4a53-b96d-fe64ee8f9101.vmdk {{(pid=61806) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 936.327367] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-0c4ed53e-d274-4427-995e-e2d8d9d3b2cc tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] [instance: 676f18b3-3da1-4a53-b96d-fe64ee8f9101] Extending root virtual disk to 1048576 {{(pid=61806) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 936.327798] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-0c7a69b6-e057-485b-98e6-17da6c6c14ae {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 936.334333] env[61806]: DEBUG oslo_vmware.api [None req-0c4ed53e-d274-4427-995e-e2d8d9d3b2cc tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Waiting for the task: (returnval){ [ 936.334333] env[61806]: value = "task-1294936" [ 936.334333] env[61806]: _type = "Task" [ 936.334333] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 936.344384] env[61806]: DEBUG oslo_vmware.api [None req-0c4ed53e-d274-4427-995e-e2d8d9d3b2cc tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Task: {'id': task-1294936, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 936.406970] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e89df660-af0e-4096-985a-e2817819fa8d {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 936.414245] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5cf2a964-6fd7-44cd-a097-4bd38b332df2 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 936.446492] env[61806]: DEBUG oslo_concurrency.lockutils [None req-03f49469-0c97-4bf3-b87e-03f023c054e5 tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Acquiring lock "380aa63f-46ba-437d-bc29-0fc141ad1cf3" by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 936.448972] env[61806]: DEBUG oslo_concurrency.lockutils [None req-03f49469-0c97-4bf3-b87e-03f023c054e5 tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Lock "380aa63f-46ba-437d-bc29-0fc141ad1cf3" acquired by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" :: waited 0.000s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 936.448972] env[61806]: INFO nova.compute.manager [None req-03f49469-0c97-4bf3-b87e-03f023c054e5 tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] [instance: 380aa63f-46ba-437d-bc29-0fc141ad1cf3] Shelving [ 936.449773] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1e7dd652-ebd3-429b-aaea-c4edee181ef1 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 936.458897] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b58ce0c8-d659-4e36-aa4e-f68c7f822fed {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 936.473891] env[61806]: DEBUG nova.compute.provider_tree [None req-d547963d-0e28-4458-8b96-eaec3d80a721 tempest-ServersAaction247Test-1786617390 tempest-ServersAaction247Test-1786617390-project-member] Inventory has not changed in ProviderTree for provider: a2858be1-fd22-4e08-979e-87ad25293407 {{(pid=61806) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 936.692633] env[61806]: DEBUG oslo_concurrency.lockutils [None req-bdb89e04-5c89-42d4-937f-829d8aee12b6 tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] Lock "fd9e520a-2e7b-43d4-b84d-8933c32cc6a6" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 13.741s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 936.846357] env[61806]: DEBUG oslo_vmware.api [None req-0c4ed53e-d274-4427-995e-e2d8d9d3b2cc tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Task: {'id': task-1294936, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.066246} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 936.846678] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-0c4ed53e-d274-4427-995e-e2d8d9d3b2cc tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] [instance: 676f18b3-3da1-4a53-b96d-fe64ee8f9101] Extended root virtual disk {{(pid=61806) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 936.847501] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-de575d72-f97a-446b-af7b-6690354172dc {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 936.870295] env[61806]: DEBUG nova.virt.vmwareapi.volumeops [None req-0c4ed53e-d274-4427-995e-e2d8d9d3b2cc tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] [instance: 676f18b3-3da1-4a53-b96d-fe64ee8f9101] Reconfiguring VM instance instance-00000058 to attach disk [datastore2] 676f18b3-3da1-4a53-b96d-fe64ee8f9101/676f18b3-3da1-4a53-b96d-fe64ee8f9101.vmdk or device None with type sparse {{(pid=61806) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 936.870660] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-ae0dc250-2bc6-4f5a-9a3b-72a7f1c97591 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 936.891924] env[61806]: DEBUG oslo_vmware.api [None req-0c4ed53e-d274-4427-995e-e2d8d9d3b2cc tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Waiting for the task: (returnval){ [ 936.891924] env[61806]: value = "task-1294937" [ 936.891924] env[61806]: _type = "Task" [ 936.891924] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 936.900059] env[61806]: DEBUG oslo_vmware.api [None req-0c4ed53e-d274-4427-995e-e2d8d9d3b2cc tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Task: {'id': task-1294937, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 936.956783] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [None req-03f49469-0c97-4bf3-b87e-03f023c054e5 tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] [instance: 380aa63f-46ba-437d-bc29-0fc141ad1cf3] Powering off the VM {{(pid=61806) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 936.957099] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-a2822507-3fe5-429f-b79e-97c02252f955 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 936.963896] env[61806]: DEBUG oslo_vmware.api [None req-03f49469-0c97-4bf3-b87e-03f023c054e5 tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Waiting for the task: (returnval){ [ 936.963896] env[61806]: value = "task-1294938" [ 936.963896] env[61806]: _type = "Task" [ 936.963896] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 936.972352] env[61806]: DEBUG oslo_vmware.api [None req-03f49469-0c97-4bf3-b87e-03f023c054e5 tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Task: {'id': task-1294938, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 936.976326] env[61806]: DEBUG nova.scheduler.client.report [None req-d547963d-0e28-4458-8b96-eaec3d80a721 tempest-ServersAaction247Test-1786617390 tempest-ServersAaction247Test-1786617390-project-member] Inventory has not changed for provider a2858be1-fd22-4e08-979e-87ad25293407 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 138, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61806) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 937.163564] env[61806]: DEBUG oslo_concurrency.lockutils [None req-9cd2108a-a1e5-403d-bbda-81099a62d59d tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] Acquiring lock "38e96e22-4200-4175-9085-31977357da2d" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 937.163846] env[61806]: DEBUG oslo_concurrency.lockutils [None req-9cd2108a-a1e5-403d-bbda-81099a62d59d tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] Lock "38e96e22-4200-4175-9085-31977357da2d" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 937.323621] env[61806]: DEBUG nova.compute.manager [req-ca83317e-d75a-42f7-b8ad-714b38e470ef req-2c5d6ad2-fd20-45c0-a508-c02f7d8c6708 service nova] [instance: fd9e520a-2e7b-43d4-b84d-8933c32cc6a6] Received event network-changed-670f75e3-58f9-493b-9148-416e2e9cea5d {{(pid=61806) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 937.323911] env[61806]: DEBUG nova.compute.manager [req-ca83317e-d75a-42f7-b8ad-714b38e470ef req-2c5d6ad2-fd20-45c0-a508-c02f7d8c6708 service nova] [instance: fd9e520a-2e7b-43d4-b84d-8933c32cc6a6] Refreshing instance network info cache due to event network-changed-670f75e3-58f9-493b-9148-416e2e9cea5d. {{(pid=61806) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 937.324085] env[61806]: DEBUG oslo_concurrency.lockutils [req-ca83317e-d75a-42f7-b8ad-714b38e470ef req-2c5d6ad2-fd20-45c0-a508-c02f7d8c6708 service nova] Acquiring lock "refresh_cache-fd9e520a-2e7b-43d4-b84d-8933c32cc6a6" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 937.324241] env[61806]: DEBUG oslo_concurrency.lockutils [req-ca83317e-d75a-42f7-b8ad-714b38e470ef req-2c5d6ad2-fd20-45c0-a508-c02f7d8c6708 service nova] Acquired lock "refresh_cache-fd9e520a-2e7b-43d4-b84d-8933c32cc6a6" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 937.324403] env[61806]: DEBUG nova.network.neutron [req-ca83317e-d75a-42f7-b8ad-714b38e470ef req-2c5d6ad2-fd20-45c0-a508-c02f7d8c6708 service nova] [instance: fd9e520a-2e7b-43d4-b84d-8933c32cc6a6] Refreshing network info cache for port 670f75e3-58f9-493b-9148-416e2e9cea5d {{(pid=61806) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 937.401886] env[61806]: DEBUG oslo_vmware.api [None req-0c4ed53e-d274-4427-995e-e2d8d9d3b2cc tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Task: {'id': task-1294937, 'name': ReconfigVM_Task, 'duration_secs': 0.402329} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 937.402217] env[61806]: DEBUG nova.virt.vmwareapi.volumeops [None req-0c4ed53e-d274-4427-995e-e2d8d9d3b2cc tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] [instance: 676f18b3-3da1-4a53-b96d-fe64ee8f9101] Reconfigured VM instance instance-00000058 to attach disk [datastore2] 676f18b3-3da1-4a53-b96d-fe64ee8f9101/676f18b3-3da1-4a53-b96d-fe64ee8f9101.vmdk or device None with type sparse {{(pid=61806) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 937.402826] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-7e8dd372-5a7f-4218-a163-fc958cae26dd {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 937.408829] env[61806]: DEBUG oslo_vmware.api [None req-0c4ed53e-d274-4427-995e-e2d8d9d3b2cc tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Waiting for the task: (returnval){ [ 937.408829] env[61806]: value = "task-1294939" [ 937.408829] env[61806]: _type = "Task" [ 937.408829] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 937.415965] env[61806]: DEBUG oslo_vmware.api [None req-0c4ed53e-d274-4427-995e-e2d8d9d3b2cc tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Task: {'id': task-1294939, 'name': Rename_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 937.473763] env[61806]: DEBUG oslo_vmware.api [None req-03f49469-0c97-4bf3-b87e-03f023c054e5 tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Task: {'id': task-1294938, 'name': PowerOffVM_Task, 'duration_secs': 0.204681} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 937.474051] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [None req-03f49469-0c97-4bf3-b87e-03f023c054e5 tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] [instance: 380aa63f-46ba-437d-bc29-0fc141ad1cf3] Powered off the VM {{(pid=61806) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 937.475079] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b4e7d36a-e783-4ef2-ad4f-cd34df09ab8d {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 937.493139] env[61806]: DEBUG oslo_concurrency.lockutils [None req-d547963d-0e28-4458-8b96-eaec3d80a721 tempest-ServersAaction247Test-1786617390 tempest-ServersAaction247Test-1786617390-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.725s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 937.496655] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9e3b2a91-7aee-4cf4-9873-1b534fee6f30 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 937.516611] env[61806]: INFO nova.scheduler.client.report [None req-d547963d-0e28-4458-8b96-eaec3d80a721 tempest-ServersAaction247Test-1786617390 tempest-ServersAaction247Test-1786617390-project-member] Deleted allocations for instance 3f4fa1ba-a6ad-4e0f-aebc-973816ee09d5 [ 937.666376] env[61806]: DEBUG nova.compute.manager [None req-9cd2108a-a1e5-403d-bbda-81099a62d59d tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] [instance: 38e96e22-4200-4175-9085-31977357da2d] Starting instance... {{(pid=61806) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 937.918971] env[61806]: DEBUG oslo_vmware.api [None req-0c4ed53e-d274-4427-995e-e2d8d9d3b2cc tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Task: {'id': task-1294939, 'name': Rename_Task, 'duration_secs': 0.140715} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 937.919296] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [None req-0c4ed53e-d274-4427-995e-e2d8d9d3b2cc tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] [instance: 676f18b3-3da1-4a53-b96d-fe64ee8f9101] Powering on the VM {{(pid=61806) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 937.919725] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-9d5aa526-e662-4f2d-87eb-16a298548ac0 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 937.927643] env[61806]: DEBUG oslo_vmware.api [None req-0c4ed53e-d274-4427-995e-e2d8d9d3b2cc tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Waiting for the task: (returnval){ [ 937.927643] env[61806]: value = "task-1294940" [ 937.927643] env[61806]: _type = "Task" [ 937.927643] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 937.935317] env[61806]: DEBUG oslo_vmware.api [None req-0c4ed53e-d274-4427-995e-e2d8d9d3b2cc tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Task: {'id': task-1294940, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 938.006281] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-03f49469-0c97-4bf3-b87e-03f023c054e5 tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] [instance: 380aa63f-46ba-437d-bc29-0fc141ad1cf3] Creating Snapshot of the VM instance {{(pid=61806) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 938.006592] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-6de5b072-4d02-45b2-9360-8756d2139db1 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 938.013507] env[61806]: DEBUG oslo_vmware.api [None req-03f49469-0c97-4bf3-b87e-03f023c054e5 tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Waiting for the task: (returnval){ [ 938.013507] env[61806]: value = "task-1294941" [ 938.013507] env[61806]: _type = "Task" [ 938.013507] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 938.023748] env[61806]: DEBUG oslo_vmware.api [None req-03f49469-0c97-4bf3-b87e-03f023c054e5 tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Task: {'id': task-1294941, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 938.024282] env[61806]: DEBUG oslo_concurrency.lockutils [None req-d547963d-0e28-4458-8b96-eaec3d80a721 tempest-ServersAaction247Test-1786617390 tempest-ServersAaction247Test-1786617390-project-member] Lock "3f4fa1ba-a6ad-4e0f-aebc-973816ee09d5" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.996s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 938.036598] env[61806]: DEBUG nova.network.neutron [req-ca83317e-d75a-42f7-b8ad-714b38e470ef req-2c5d6ad2-fd20-45c0-a508-c02f7d8c6708 service nova] [instance: fd9e520a-2e7b-43d4-b84d-8933c32cc6a6] Updated VIF entry in instance network info cache for port 670f75e3-58f9-493b-9148-416e2e9cea5d. {{(pid=61806) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 938.036960] env[61806]: DEBUG nova.network.neutron [req-ca83317e-d75a-42f7-b8ad-714b38e470ef req-2c5d6ad2-fd20-45c0-a508-c02f7d8c6708 service nova] [instance: fd9e520a-2e7b-43d4-b84d-8933c32cc6a6] Updating instance_info_cache with network_info: [{"id": "670f75e3-58f9-493b-9148-416e2e9cea5d", "address": "fa:16:3e:76:be:e1", "network": {"id": "a9d993bc-0817-4ee6-ab6e-1ef9432a110f", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-939246736-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.226", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "bbbb09a49203421d807f733616c854df", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ad4c015b-4a8b-46ca-9556-74bad8db9fb3", "external-id": "nsx-vlan-transportzone-246", "segmentation_id": 246, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap670f75e3-58", "ovs_interfaceid": "670f75e3-58f9-493b-9148-416e2e9cea5d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61806) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 938.189878] env[61806]: DEBUG oslo_concurrency.lockutils [None req-9cd2108a-a1e5-403d-bbda-81099a62d59d tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 938.190266] env[61806]: DEBUG oslo_concurrency.lockutils [None req-9cd2108a-a1e5-403d-bbda-81099a62d59d tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.001s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 938.192082] env[61806]: INFO nova.compute.claims [None req-9cd2108a-a1e5-403d-bbda-81099a62d59d tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] [instance: 38e96e22-4200-4175-9085-31977357da2d] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 938.209639] env[61806]: DEBUG oslo_concurrency.lockutils [None req-232a749a-b989-4ce4-a547-d7aa6acd6397 tempest-AttachVolumeNegativeTest-1683033704 tempest-AttachVolumeNegativeTest-1683033704-project-member] Acquiring lock "daff849e-eb4f-411e-92f8-9362e3bd91e3" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 938.209955] env[61806]: DEBUG oslo_concurrency.lockutils [None req-232a749a-b989-4ce4-a547-d7aa6acd6397 tempest-AttachVolumeNegativeTest-1683033704 tempest-AttachVolumeNegativeTest-1683033704-project-member] Lock "daff849e-eb4f-411e-92f8-9362e3bd91e3" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 938.210201] env[61806]: DEBUG oslo_concurrency.lockutils [None req-232a749a-b989-4ce4-a547-d7aa6acd6397 tempest-AttachVolumeNegativeTest-1683033704 tempest-AttachVolumeNegativeTest-1683033704-project-member] Acquiring lock "daff849e-eb4f-411e-92f8-9362e3bd91e3-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 938.210417] env[61806]: DEBUG oslo_concurrency.lockutils [None req-232a749a-b989-4ce4-a547-d7aa6acd6397 tempest-AttachVolumeNegativeTest-1683033704 tempest-AttachVolumeNegativeTest-1683033704-project-member] Lock "daff849e-eb4f-411e-92f8-9362e3bd91e3-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 938.210605] env[61806]: DEBUG oslo_concurrency.lockutils [None req-232a749a-b989-4ce4-a547-d7aa6acd6397 tempest-AttachVolumeNegativeTest-1683033704 tempest-AttachVolumeNegativeTest-1683033704-project-member] Lock "daff849e-eb4f-411e-92f8-9362e3bd91e3-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 938.212909] env[61806]: INFO nova.compute.manager [None req-232a749a-b989-4ce4-a547-d7aa6acd6397 tempest-AttachVolumeNegativeTest-1683033704 tempest-AttachVolumeNegativeTest-1683033704-project-member] [instance: daff849e-eb4f-411e-92f8-9362e3bd91e3] Terminating instance [ 938.214922] env[61806]: DEBUG nova.compute.manager [None req-232a749a-b989-4ce4-a547-d7aa6acd6397 tempest-AttachVolumeNegativeTest-1683033704 tempest-AttachVolumeNegativeTest-1683033704-project-member] [instance: daff849e-eb4f-411e-92f8-9362e3bd91e3] Start destroying the instance on the hypervisor. {{(pid=61806) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 938.215099] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-232a749a-b989-4ce4-a547-d7aa6acd6397 tempest-AttachVolumeNegativeTest-1683033704 tempest-AttachVolumeNegativeTest-1683033704-project-member] [instance: daff849e-eb4f-411e-92f8-9362e3bd91e3] Destroying instance {{(pid=61806) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 938.216155] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-421a88c8-1938-48c7-a2bb-38d8f0a2346e {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 938.224044] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [None req-232a749a-b989-4ce4-a547-d7aa6acd6397 tempest-AttachVolumeNegativeTest-1683033704 tempest-AttachVolumeNegativeTest-1683033704-project-member] [instance: daff849e-eb4f-411e-92f8-9362e3bd91e3] Powering off the VM {{(pid=61806) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 938.224332] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-e23d4ab3-ee8d-40cc-a53d-f2ee44b83757 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 938.231471] env[61806]: DEBUG oslo_vmware.api [None req-232a749a-b989-4ce4-a547-d7aa6acd6397 tempest-AttachVolumeNegativeTest-1683033704 tempest-AttachVolumeNegativeTest-1683033704-project-member] Waiting for the task: (returnval){ [ 938.231471] env[61806]: value = "task-1294942" [ 938.231471] env[61806]: _type = "Task" [ 938.231471] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 938.240123] env[61806]: DEBUG oslo_vmware.api [None req-232a749a-b989-4ce4-a547-d7aa6acd6397 tempest-AttachVolumeNegativeTest-1683033704 tempest-AttachVolumeNegativeTest-1683033704-project-member] Task: {'id': task-1294942, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 938.443034] env[61806]: DEBUG oslo_vmware.api [None req-0c4ed53e-d274-4427-995e-e2d8d9d3b2cc tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Task: {'id': task-1294940, 'name': PowerOnVM_Task, 'duration_secs': 0.495647} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 938.443034] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [None req-0c4ed53e-d274-4427-995e-e2d8d9d3b2cc tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] [instance: 676f18b3-3da1-4a53-b96d-fe64ee8f9101] Powered on the VM {{(pid=61806) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 938.443034] env[61806]: INFO nova.compute.manager [None req-0c4ed53e-d274-4427-995e-e2d8d9d3b2cc tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] [instance: 676f18b3-3da1-4a53-b96d-fe64ee8f9101] Took 7.86 seconds to spawn the instance on the hypervisor. [ 938.443034] env[61806]: DEBUG nova.compute.manager [None req-0c4ed53e-d274-4427-995e-e2d8d9d3b2cc tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] [instance: 676f18b3-3da1-4a53-b96d-fe64ee8f9101] Checking state {{(pid=61806) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 938.443034] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8f88b4c1-29bc-4433-aa36-ff926f9859ae {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 938.525267] env[61806]: DEBUG oslo_vmware.api [None req-03f49469-0c97-4bf3-b87e-03f023c054e5 tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Task: {'id': task-1294941, 'name': CreateSnapshot_Task, 'duration_secs': 0.489055} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 938.525595] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-03f49469-0c97-4bf3-b87e-03f023c054e5 tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] [instance: 380aa63f-46ba-437d-bc29-0fc141ad1cf3] Created Snapshot of the VM instance {{(pid=61806) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 938.526635] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1602c4d4-c8b9-49cd-9995-4af9de4681a2 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 938.540160] env[61806]: DEBUG oslo_concurrency.lockutils [req-ca83317e-d75a-42f7-b8ad-714b38e470ef req-2c5d6ad2-fd20-45c0-a508-c02f7d8c6708 service nova] Releasing lock "refresh_cache-fd9e520a-2e7b-43d4-b84d-8933c32cc6a6" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 938.741459] env[61806]: DEBUG oslo_vmware.api [None req-232a749a-b989-4ce4-a547-d7aa6acd6397 tempest-AttachVolumeNegativeTest-1683033704 tempest-AttachVolumeNegativeTest-1683033704-project-member] Task: {'id': task-1294942, 'name': PowerOffVM_Task, 'duration_secs': 0.252107} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 938.741775] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [None req-232a749a-b989-4ce4-a547-d7aa6acd6397 tempest-AttachVolumeNegativeTest-1683033704 tempest-AttachVolumeNegativeTest-1683033704-project-member] [instance: daff849e-eb4f-411e-92f8-9362e3bd91e3] Powered off the VM {{(pid=61806) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 938.741937] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-232a749a-b989-4ce4-a547-d7aa6acd6397 tempest-AttachVolumeNegativeTest-1683033704 tempest-AttachVolumeNegativeTest-1683033704-project-member] [instance: daff849e-eb4f-411e-92f8-9362e3bd91e3] Unregistering the VM {{(pid=61806) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 938.742214] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-d715f717-f146-493e-8292-39e448ad8d8e {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 938.840884] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-232a749a-b989-4ce4-a547-d7aa6acd6397 tempest-AttachVolumeNegativeTest-1683033704 tempest-AttachVolumeNegativeTest-1683033704-project-member] [instance: daff849e-eb4f-411e-92f8-9362e3bd91e3] Unregistered the VM {{(pid=61806) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 938.841209] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-232a749a-b989-4ce4-a547-d7aa6acd6397 tempest-AttachVolumeNegativeTest-1683033704 tempest-AttachVolumeNegativeTest-1683033704-project-member] [instance: daff849e-eb4f-411e-92f8-9362e3bd91e3] Deleting contents of the VM from datastore datastore2 {{(pid=61806) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 938.841410] env[61806]: DEBUG nova.virt.vmwareapi.ds_util [None req-232a749a-b989-4ce4-a547-d7aa6acd6397 tempest-AttachVolumeNegativeTest-1683033704 tempest-AttachVolumeNegativeTest-1683033704-project-member] Deleting the datastore file [datastore2] daff849e-eb4f-411e-92f8-9362e3bd91e3 {{(pid=61806) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 938.841717] env[61806]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-7b8f0007-25f9-4510-b17d-6f35e37bb70e {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 938.847758] env[61806]: DEBUG oslo_vmware.api [None req-232a749a-b989-4ce4-a547-d7aa6acd6397 tempest-AttachVolumeNegativeTest-1683033704 tempest-AttachVolumeNegativeTest-1683033704-project-member] Waiting for the task: (returnval){ [ 938.847758] env[61806]: value = "task-1294944" [ 938.847758] env[61806]: _type = "Task" [ 938.847758] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 938.855729] env[61806]: DEBUG oslo_vmware.api [None req-232a749a-b989-4ce4-a547-d7aa6acd6397 tempest-AttachVolumeNegativeTest-1683033704 tempest-AttachVolumeNegativeTest-1683033704-project-member] Task: {'id': task-1294944, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 938.959088] env[61806]: INFO nova.compute.manager [None req-0c4ed53e-d274-4427-995e-e2d8d9d3b2cc tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] [instance: 676f18b3-3da1-4a53-b96d-fe64ee8f9101] Took 14.66 seconds to build instance. [ 939.044933] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-03f49469-0c97-4bf3-b87e-03f023c054e5 tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] [instance: 380aa63f-46ba-437d-bc29-0fc141ad1cf3] Creating linked-clone VM from snapshot {{(pid=61806) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 939.045642] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-97498c00-1f36-48c8-8c74-63b9c1038e55 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 939.055263] env[61806]: DEBUG oslo_vmware.api [None req-03f49469-0c97-4bf3-b87e-03f023c054e5 tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Waiting for the task: (returnval){ [ 939.055263] env[61806]: value = "task-1294945" [ 939.055263] env[61806]: _type = "Task" [ 939.055263] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 939.063813] env[61806]: DEBUG oslo_vmware.api [None req-03f49469-0c97-4bf3-b87e-03f023c054e5 tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Task: {'id': task-1294945, 'name': CloneVM_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 939.341441] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c1638daf-7db5-4a1f-a65c-b328c93feaec {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 939.352179] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eabf77dc-c7fe-48ef-badd-147f2297953b {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 939.360146] env[61806]: DEBUG oslo_vmware.api [None req-232a749a-b989-4ce4-a547-d7aa6acd6397 tempest-AttachVolumeNegativeTest-1683033704 tempest-AttachVolumeNegativeTest-1683033704-project-member] Task: {'id': task-1294944, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.144417} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 939.384901] env[61806]: DEBUG nova.virt.vmwareapi.ds_util [None req-232a749a-b989-4ce4-a547-d7aa6acd6397 tempest-AttachVolumeNegativeTest-1683033704 tempest-AttachVolumeNegativeTest-1683033704-project-member] Deleted the datastore file {{(pid=61806) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 939.385145] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-232a749a-b989-4ce4-a547-d7aa6acd6397 tempest-AttachVolumeNegativeTest-1683033704 tempest-AttachVolumeNegativeTest-1683033704-project-member] [instance: daff849e-eb4f-411e-92f8-9362e3bd91e3] Deleted contents of the VM from datastore datastore2 {{(pid=61806) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 939.385349] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-232a749a-b989-4ce4-a547-d7aa6acd6397 tempest-AttachVolumeNegativeTest-1683033704 tempest-AttachVolumeNegativeTest-1683033704-project-member] [instance: daff849e-eb4f-411e-92f8-9362e3bd91e3] Instance destroyed {{(pid=61806) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 939.385541] env[61806]: INFO nova.compute.manager [None req-232a749a-b989-4ce4-a547-d7aa6acd6397 tempest-AttachVolumeNegativeTest-1683033704 tempest-AttachVolumeNegativeTest-1683033704-project-member] [instance: daff849e-eb4f-411e-92f8-9362e3bd91e3] Took 1.17 seconds to destroy the instance on the hypervisor. [ 939.385797] env[61806]: DEBUG oslo.service.loopingcall [None req-232a749a-b989-4ce4-a547-d7aa6acd6397 tempest-AttachVolumeNegativeTest-1683033704 tempest-AttachVolumeNegativeTest-1683033704-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61806) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 939.386955] env[61806]: DEBUG nova.compute.manager [-] [instance: daff849e-eb4f-411e-92f8-9362e3bd91e3] Deallocating network for instance {{(pid=61806) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 939.387086] env[61806]: DEBUG nova.network.neutron [-] [instance: daff849e-eb4f-411e-92f8-9362e3bd91e3] deallocate_for_instance() {{(pid=61806) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 939.389242] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6cf6e28b-25c1-4aec-9d1a-963abebaca4a {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 939.397327] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4e690684-5878-4012-9015-988ed4142107 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 939.410945] env[61806]: DEBUG nova.compute.provider_tree [None req-9cd2108a-a1e5-403d-bbda-81099a62d59d tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] Inventory has not changed in ProviderTree for provider: a2858be1-fd22-4e08-979e-87ad25293407 {{(pid=61806) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 939.461878] env[61806]: DEBUG oslo_concurrency.lockutils [None req-0c4ed53e-d274-4427-995e-e2d8d9d3b2cc tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Lock "676f18b3-3da1-4a53-b96d-fe64ee8f9101" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 16.174s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 939.566565] env[61806]: DEBUG oslo_vmware.api [None req-03f49469-0c97-4bf3-b87e-03f023c054e5 tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Task: {'id': task-1294945, 'name': CloneVM_Task} progress is 94%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 939.603345] env[61806]: DEBUG oslo_concurrency.lockutils [None req-ccfe4ccf-a36d-4d36-adb7-3f5c04a0f60f tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] Acquiring lock "interface-6f6ba57a-e2d6-4749-a53a-e263861cb1c0-7e031e88-6928-472d-a65e-b818d89e6fc5" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 939.604037] env[61806]: DEBUG oslo_concurrency.lockutils [None req-ccfe4ccf-a36d-4d36-adb7-3f5c04a0f60f tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] Lock "interface-6f6ba57a-e2d6-4749-a53a-e263861cb1c0-7e031e88-6928-472d-a65e-b818d89e6fc5" acquired by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: waited 0.001s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 939.604508] env[61806]: DEBUG nova.objects.instance [None req-ccfe4ccf-a36d-4d36-adb7-3f5c04a0f60f tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] Lazy-loading 'flavor' on Instance uuid 6f6ba57a-e2d6-4749-a53a-e263861cb1c0 {{(pid=61806) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 939.914373] env[61806]: DEBUG nova.scheduler.client.report [None req-9cd2108a-a1e5-403d-bbda-81099a62d59d tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] Inventory has not changed for provider a2858be1-fd22-4e08-979e-87ad25293407 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 138, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61806) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 939.952048] env[61806]: DEBUG nova.compute.manager [req-cdc45bbd-7ca6-4d10-a1aa-b41eb605eff3 req-7295a49f-be0e-4da2-a9c0-2ed50e3245fb service nova] [instance: daff849e-eb4f-411e-92f8-9362e3bd91e3] Received event network-vif-deleted-9fdf6139-082b-49e0-8ce8-7870749b2c7a {{(pid=61806) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 939.952275] env[61806]: INFO nova.compute.manager [req-cdc45bbd-7ca6-4d10-a1aa-b41eb605eff3 req-7295a49f-be0e-4da2-a9c0-2ed50e3245fb service nova] [instance: daff849e-eb4f-411e-92f8-9362e3bd91e3] Neutron deleted interface 9fdf6139-082b-49e0-8ce8-7870749b2c7a; detaching it from the instance and deleting it from the info cache [ 939.952473] env[61806]: DEBUG nova.network.neutron [req-cdc45bbd-7ca6-4d10-a1aa-b41eb605eff3 req-7295a49f-be0e-4da2-a9c0-2ed50e3245fb service nova] [instance: daff849e-eb4f-411e-92f8-9362e3bd91e3] Updating instance_info_cache with network_info: [] {{(pid=61806) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 940.071834] env[61806]: DEBUG oslo_vmware.api [None req-03f49469-0c97-4bf3-b87e-03f023c054e5 tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Task: {'id': task-1294945, 'name': CloneVM_Task} progress is 94%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 940.202222] env[61806]: DEBUG nova.objects.instance [None req-ccfe4ccf-a36d-4d36-adb7-3f5c04a0f60f tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] Lazy-loading 'pci_requests' on Instance uuid 6f6ba57a-e2d6-4749-a53a-e263861cb1c0 {{(pid=61806) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 940.341562] env[61806]: DEBUG nova.network.neutron [-] [instance: daff849e-eb4f-411e-92f8-9362e3bd91e3] Updating instance_info_cache with network_info: [] {{(pid=61806) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 940.421447] env[61806]: DEBUG oslo_concurrency.lockutils [None req-9cd2108a-a1e5-403d-bbda-81099a62d59d tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.231s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 940.422303] env[61806]: DEBUG nova.compute.manager [None req-9cd2108a-a1e5-403d-bbda-81099a62d59d tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] [instance: 38e96e22-4200-4175-9085-31977357da2d] Start building networks asynchronously for instance. {{(pid=61806) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 940.457501] env[61806]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-564e06c4-b2f8-4643-b1be-8c48a3b9f2ae {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 940.468924] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-96d5a573-9bec-4673-8dd3-3046a3785cba {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 940.519657] env[61806]: DEBUG nova.compute.manager [req-cdc45bbd-7ca6-4d10-a1aa-b41eb605eff3 req-7295a49f-be0e-4da2-a9c0-2ed50e3245fb service nova] [instance: daff849e-eb4f-411e-92f8-9362e3bd91e3] Detach interface failed, port_id=9fdf6139-082b-49e0-8ce8-7870749b2c7a, reason: Instance daff849e-eb4f-411e-92f8-9362e3bd91e3 could not be found. {{(pid=61806) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10963}} [ 940.567757] env[61806]: DEBUG oslo_vmware.api [None req-03f49469-0c97-4bf3-b87e-03f023c054e5 tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Task: {'id': task-1294945, 'name': CloneVM_Task} progress is 95%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 940.704590] env[61806]: DEBUG nova.objects.base [None req-ccfe4ccf-a36d-4d36-adb7-3f5c04a0f60f tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] Object Instance<6f6ba57a-e2d6-4749-a53a-e263861cb1c0> lazy-loaded attributes: flavor,pci_requests {{(pid=61806) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 940.704848] env[61806]: DEBUG nova.network.neutron [None req-ccfe4ccf-a36d-4d36-adb7-3f5c04a0f60f tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] [instance: 6f6ba57a-e2d6-4749-a53a-e263861cb1c0] allocate_for_instance() {{(pid=61806) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 940.783844] env[61806]: DEBUG nova.policy [None req-ccfe4ccf-a36d-4d36-adb7-3f5c04a0f60f tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'b1eac9072ccb4b3bbd23f240e941a76e', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '6e391e1267014fafbb8f5a3211299819', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61806) authorize /opt/stack/nova/nova/policy.py:201}} [ 940.849018] env[61806]: INFO nova.compute.manager [-] [instance: daff849e-eb4f-411e-92f8-9362e3bd91e3] Took 1.46 seconds to deallocate network for instance. [ 940.931368] env[61806]: DEBUG nova.compute.utils [None req-9cd2108a-a1e5-403d-bbda-81099a62d59d tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] Using /dev/sd instead of None {{(pid=61806) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 940.934200] env[61806]: DEBUG nova.compute.manager [None req-9cd2108a-a1e5-403d-bbda-81099a62d59d tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] [instance: 38e96e22-4200-4175-9085-31977357da2d] Allocating IP information in the background. {{(pid=61806) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 940.938123] env[61806]: DEBUG nova.network.neutron [None req-9cd2108a-a1e5-403d-bbda-81099a62d59d tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] [instance: 38e96e22-4200-4175-9085-31977357da2d] allocate_for_instance() {{(pid=61806) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 941.068058] env[61806]: DEBUG oslo_vmware.api [None req-03f49469-0c97-4bf3-b87e-03f023c054e5 tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Task: {'id': task-1294945, 'name': CloneVM_Task, 'duration_secs': 1.817488} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 941.068537] env[61806]: INFO nova.virt.vmwareapi.vmops [None req-03f49469-0c97-4bf3-b87e-03f023c054e5 tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] [instance: 380aa63f-46ba-437d-bc29-0fc141ad1cf3] Created linked-clone VM from snapshot [ 941.069190] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e55ac4fc-3c97-45dc-95e2-9fb28846803b {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 941.079032] env[61806]: DEBUG nova.virt.vmwareapi.images [None req-03f49469-0c97-4bf3-b87e-03f023c054e5 tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] [instance: 380aa63f-46ba-437d-bc29-0fc141ad1cf3] Uploading image 3047e10f-a04a-4487-8696-28cac698ab9e {{(pid=61806) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:442}} [ 941.104148] env[61806]: DEBUG oslo_vmware.rw_handles [None req-03f49469-0c97-4bf3-b87e-03f023c054e5 tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Creating HttpNfcLease lease for exporting VM: (result){ [ 941.104148] env[61806]: value = "vm-277722" [ 941.104148] env[61806]: _type = "VirtualMachine" [ 941.104148] env[61806]: }. {{(pid=61806) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:478}} [ 941.104512] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ExportVm with opID=oslo.vmware-811da68d-0898-44ce-bf6d-28dc6546f855 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 941.112629] env[61806]: DEBUG oslo_vmware.rw_handles [None req-03f49469-0c97-4bf3-b87e-03f023c054e5 tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Lease: (returnval){ [ 941.112629] env[61806]: value = "session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]5236b516-f639-5310-6af1-8a6f6315ea7c" [ 941.112629] env[61806]: _type = "HttpNfcLease" [ 941.112629] env[61806]: } obtained for exporting VM: (result){ [ 941.112629] env[61806]: value = "vm-277722" [ 941.112629] env[61806]: _type = "VirtualMachine" [ 941.112629] env[61806]: }. {{(pid=61806) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:481}} [ 941.112629] env[61806]: DEBUG oslo_vmware.api [None req-03f49469-0c97-4bf3-b87e-03f023c054e5 tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Waiting for the lease: (returnval){ [ 941.112629] env[61806]: value = "session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]5236b516-f639-5310-6af1-8a6f6315ea7c" [ 941.112629] env[61806]: _type = "HttpNfcLease" [ 941.112629] env[61806]: } to be ready. {{(pid=61806) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 941.118429] env[61806]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 941.118429] env[61806]: value = "session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]5236b516-f639-5310-6af1-8a6f6315ea7c" [ 941.118429] env[61806]: _type = "HttpNfcLease" [ 941.118429] env[61806]: } is initializing. {{(pid=61806) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 941.183952] env[61806]: DEBUG nova.policy [None req-9cd2108a-a1e5-403d-bbda-81099a62d59d tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '4c5d4c6e1ad844a585c08c4644a60bbf', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '73ebcc4d79d248efb653a084a5e44302', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61806) authorize /opt/stack/nova/nova/policy.py:201}} [ 941.353978] env[61806]: DEBUG oslo_concurrency.lockutils [None req-232a749a-b989-4ce4-a547-d7aa6acd6397 tempest-AttachVolumeNegativeTest-1683033704 tempest-AttachVolumeNegativeTest-1683033704-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 941.354307] env[61806]: DEBUG oslo_concurrency.lockutils [None req-232a749a-b989-4ce4-a547-d7aa6acd6397 tempest-AttachVolumeNegativeTest-1683033704 tempest-AttachVolumeNegativeTest-1683033704-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.001s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 941.354540] env[61806]: DEBUG nova.objects.instance [None req-232a749a-b989-4ce4-a547-d7aa6acd6397 tempest-AttachVolumeNegativeTest-1683033704 tempest-AttachVolumeNegativeTest-1683033704-project-member] Lazy-loading 'resources' on Instance uuid daff849e-eb4f-411e-92f8-9362e3bd91e3 {{(pid=61806) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 941.441316] env[61806]: DEBUG nova.compute.manager [None req-9cd2108a-a1e5-403d-bbda-81099a62d59d tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] [instance: 38e96e22-4200-4175-9085-31977357da2d] Start building block device mappings for instance. {{(pid=61806) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 941.573607] env[61806]: DEBUG nova.network.neutron [None req-9cd2108a-a1e5-403d-bbda-81099a62d59d tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] [instance: 38e96e22-4200-4175-9085-31977357da2d] Successfully created port: 85d68e9b-a57d-4078-baf3-cd7c620f9366 {{(pid=61806) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 941.622304] env[61806]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 941.622304] env[61806]: value = "session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]5236b516-f639-5310-6af1-8a6f6315ea7c" [ 941.622304] env[61806]: _type = "HttpNfcLease" [ 941.622304] env[61806]: } is ready. {{(pid=61806) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 941.622304] env[61806]: DEBUG oslo_vmware.rw_handles [None req-03f49469-0c97-4bf3-b87e-03f023c054e5 tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 941.622304] env[61806]: value = "session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]5236b516-f639-5310-6af1-8a6f6315ea7c" [ 941.622304] env[61806]: _type = "HttpNfcLease" [ 941.622304] env[61806]: }. {{(pid=61806) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:486}} [ 941.622661] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-44de6547-71f0-47df-a4ae-5f07bdfbdb6f {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 941.630044] env[61806]: DEBUG oslo_vmware.rw_handles [None req-03f49469-0c97-4bf3-b87e-03f023c054e5 tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Found VMDK URL: https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52ba91b7-8fef-044e-b400-61ffb14f4c5e/disk-0.vmdk from lease info. {{(pid=61806) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 941.630242] env[61806]: DEBUG oslo_vmware.rw_handles [None req-03f49469-0c97-4bf3-b87e-03f023c054e5 tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Opening URL: https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52ba91b7-8fef-044e-b400-61ffb14f4c5e/disk-0.vmdk for reading. {{(pid=61806) _create_read_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:99}} [ 941.808208] env[61806]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-f640a359-e4f9-49c7-ba05-f6d74594cec0 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 941.968540] env[61806]: DEBUG nova.compute.manager [None req-cf5202db-26c0-4960-9218-bd95b8e81f62 tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] [instance: 676f18b3-3da1-4a53-b96d-fe64ee8f9101] Stashing vm_state: active {{(pid=61806) _prep_resize /opt/stack/nova/nova/compute/manager.py:5624}} [ 942.066362] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b21d018f-32e4-441a-b82f-4c839d60ff41 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 942.075049] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eeef7fde-21b0-4ca6-8e46-d78ac9782ab6 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 942.104759] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a30fd92e-72ac-4f36-9dc9-29311a2b66af {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 942.112641] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-24a9891d-ff56-4a92-a42f-f3d14b021a38 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 942.126059] env[61806]: DEBUG nova.compute.provider_tree [None req-232a749a-b989-4ce4-a547-d7aa6acd6397 tempest-AttachVolumeNegativeTest-1683033704 tempest-AttachVolumeNegativeTest-1683033704-project-member] Inventory has not changed in ProviderTree for provider: a2858be1-fd22-4e08-979e-87ad25293407 {{(pid=61806) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 942.250652] env[61806]: DEBUG oslo_concurrency.lockutils [None req-5d1f922c-8037-4fb1-9c2a-dcd72461a463 tempest-ServersNegativeTestJSON-891856253 tempest-ServersNegativeTestJSON-891856253-project-member] Acquiring lock "bdbaef31-7d37-40cb-ac3e-ad152b37e2d2" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 942.251109] env[61806]: DEBUG oslo_concurrency.lockutils [None req-5d1f922c-8037-4fb1-9c2a-dcd72461a463 tempest-ServersNegativeTestJSON-891856253 tempest-ServersNegativeTestJSON-891856253-project-member] Lock "bdbaef31-7d37-40cb-ac3e-ad152b37e2d2" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 942.325189] env[61806]: DEBUG nova.compute.manager [req-9b125768-62ab-47de-b08f-c36f3a0cb700 req-aa0818a9-f9c6-49af-b198-6e8821151f65 service nova] [instance: 6f6ba57a-e2d6-4749-a53a-e263861cb1c0] Received event network-vif-plugged-7e031e88-6928-472d-a65e-b818d89e6fc5 {{(pid=61806) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 942.325757] env[61806]: DEBUG oslo_concurrency.lockutils [req-9b125768-62ab-47de-b08f-c36f3a0cb700 req-aa0818a9-f9c6-49af-b198-6e8821151f65 service nova] Acquiring lock "6f6ba57a-e2d6-4749-a53a-e263861cb1c0-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 942.326227] env[61806]: DEBUG oslo_concurrency.lockutils [req-9b125768-62ab-47de-b08f-c36f3a0cb700 req-aa0818a9-f9c6-49af-b198-6e8821151f65 service nova] Lock "6f6ba57a-e2d6-4749-a53a-e263861cb1c0-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 942.326912] env[61806]: DEBUG oslo_concurrency.lockutils [req-9b125768-62ab-47de-b08f-c36f3a0cb700 req-aa0818a9-f9c6-49af-b198-6e8821151f65 service nova] Lock "6f6ba57a-e2d6-4749-a53a-e263861cb1c0-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 942.326912] env[61806]: DEBUG nova.compute.manager [req-9b125768-62ab-47de-b08f-c36f3a0cb700 req-aa0818a9-f9c6-49af-b198-6e8821151f65 service nova] [instance: 6f6ba57a-e2d6-4749-a53a-e263861cb1c0] No waiting events found dispatching network-vif-plugged-7e031e88-6928-472d-a65e-b818d89e6fc5 {{(pid=61806) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 942.326912] env[61806]: WARNING nova.compute.manager [req-9b125768-62ab-47de-b08f-c36f3a0cb700 req-aa0818a9-f9c6-49af-b198-6e8821151f65 service nova] [instance: 6f6ba57a-e2d6-4749-a53a-e263861cb1c0] Received unexpected event network-vif-plugged-7e031e88-6928-472d-a65e-b818d89e6fc5 for instance with vm_state active and task_state None. [ 942.380146] env[61806]: DEBUG nova.network.neutron [None req-ccfe4ccf-a36d-4d36-adb7-3f5c04a0f60f tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] [instance: 6f6ba57a-e2d6-4749-a53a-e263861cb1c0] Successfully updated port: 7e031e88-6928-472d-a65e-b818d89e6fc5 {{(pid=61806) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 942.453393] env[61806]: DEBUG nova.compute.manager [None req-9cd2108a-a1e5-403d-bbda-81099a62d59d tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] [instance: 38e96e22-4200-4175-9085-31977357da2d] Start spawning the instance on the hypervisor. {{(pid=61806) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 942.493438] env[61806]: DEBUG nova.virt.hardware [None req-9cd2108a-a1e5-403d-bbda-81099a62d59d tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-15T18:05:10Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-15T18:04:52Z,direct_url=,disk_format='vmdk',id=ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='84ba83fce2c349a6988173c5d6fc3b07',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-15T18:04:53Z,virtual_size=,visibility=), allow threads: False {{(pid=61806) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 942.493776] env[61806]: DEBUG nova.virt.hardware [None req-9cd2108a-a1e5-403d-bbda-81099a62d59d tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] Flavor limits 0:0:0 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 942.493987] env[61806]: DEBUG nova.virt.hardware [None req-9cd2108a-a1e5-403d-bbda-81099a62d59d tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] Image limits 0:0:0 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 942.494218] env[61806]: DEBUG nova.virt.hardware [None req-9cd2108a-a1e5-403d-bbda-81099a62d59d tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] Flavor pref 0:0:0 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 942.494395] env[61806]: DEBUG nova.virt.hardware [None req-9cd2108a-a1e5-403d-bbda-81099a62d59d tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] Image pref 0:0:0 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 942.494565] env[61806]: DEBUG nova.virt.hardware [None req-9cd2108a-a1e5-403d-bbda-81099a62d59d tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 942.494803] env[61806]: DEBUG nova.virt.hardware [None req-9cd2108a-a1e5-403d-bbda-81099a62d59d tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61806) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 942.494994] env[61806]: DEBUG nova.virt.hardware [None req-9cd2108a-a1e5-403d-bbda-81099a62d59d tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61806) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 942.495210] env[61806]: DEBUG nova.virt.hardware [None req-9cd2108a-a1e5-403d-bbda-81099a62d59d tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] Got 1 possible topologies {{(pid=61806) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 942.495401] env[61806]: DEBUG nova.virt.hardware [None req-9cd2108a-a1e5-403d-bbda-81099a62d59d tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61806) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 942.495588] env[61806]: DEBUG nova.virt.hardware [None req-9cd2108a-a1e5-403d-bbda-81099a62d59d tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61806) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 942.496534] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-76b64e30-f97f-4591-a59d-cde6d7731298 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 942.499918] env[61806]: DEBUG oslo_concurrency.lockutils [None req-cf5202db-26c0-4960-9218-bd95b8e81f62 tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 942.505772] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-644009f5-da20-4d28-bb18-11177baa9538 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 942.629615] env[61806]: DEBUG nova.scheduler.client.report [None req-232a749a-b989-4ce4-a547-d7aa6acd6397 tempest-AttachVolumeNegativeTest-1683033704 tempest-AttachVolumeNegativeTest-1683033704-project-member] Inventory has not changed for provider a2858be1-fd22-4e08-979e-87ad25293407 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 138, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61806) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 942.754113] env[61806]: DEBUG nova.compute.manager [None req-5d1f922c-8037-4fb1-9c2a-dcd72461a463 tempest-ServersNegativeTestJSON-891856253 tempest-ServersNegativeTestJSON-891856253-project-member] [instance: bdbaef31-7d37-40cb-ac3e-ad152b37e2d2] Starting instance... {{(pid=61806) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 942.884438] env[61806]: DEBUG oslo_concurrency.lockutils [None req-ccfe4ccf-a36d-4d36-adb7-3f5c04a0f60f tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] Acquiring lock "refresh_cache-6f6ba57a-e2d6-4749-a53a-e263861cb1c0" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 942.887243] env[61806]: DEBUG oslo_concurrency.lockutils [None req-ccfe4ccf-a36d-4d36-adb7-3f5c04a0f60f tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] Acquired lock "refresh_cache-6f6ba57a-e2d6-4749-a53a-e263861cb1c0" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 942.887243] env[61806]: DEBUG nova.network.neutron [None req-ccfe4ccf-a36d-4d36-adb7-3f5c04a0f60f tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] [instance: 6f6ba57a-e2d6-4749-a53a-e263861cb1c0] Building network info cache for instance {{(pid=61806) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 943.134922] env[61806]: DEBUG oslo_concurrency.lockutils [None req-232a749a-b989-4ce4-a547-d7aa6acd6397 tempest-AttachVolumeNegativeTest-1683033704 tempest-AttachVolumeNegativeTest-1683033704-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.780s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 943.137485] env[61806]: DEBUG oslo_concurrency.lockutils [None req-cf5202db-26c0-4960-9218-bd95b8e81f62 tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: waited 0.637s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 943.161018] env[61806]: INFO nova.scheduler.client.report [None req-232a749a-b989-4ce4-a547-d7aa6acd6397 tempest-AttachVolumeNegativeTest-1683033704 tempest-AttachVolumeNegativeTest-1683033704-project-member] Deleted allocations for instance daff849e-eb4f-411e-92f8-9362e3bd91e3 [ 943.237103] env[61806]: DEBUG nova.compute.manager [req-536fb375-c0d7-4a99-9b2e-ea55b0398c53 req-b26cf4f2-434b-4201-8efb-1a773408e048 service nova] [instance: 38e96e22-4200-4175-9085-31977357da2d] Received event network-vif-plugged-85d68e9b-a57d-4078-baf3-cd7c620f9366 {{(pid=61806) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 943.237460] env[61806]: DEBUG oslo_concurrency.lockutils [req-536fb375-c0d7-4a99-9b2e-ea55b0398c53 req-b26cf4f2-434b-4201-8efb-1a773408e048 service nova] Acquiring lock "38e96e22-4200-4175-9085-31977357da2d-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 943.237778] env[61806]: DEBUG oslo_concurrency.lockutils [req-536fb375-c0d7-4a99-9b2e-ea55b0398c53 req-b26cf4f2-434b-4201-8efb-1a773408e048 service nova] Lock "38e96e22-4200-4175-9085-31977357da2d-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 943.238175] env[61806]: DEBUG oslo_concurrency.lockutils [req-536fb375-c0d7-4a99-9b2e-ea55b0398c53 req-b26cf4f2-434b-4201-8efb-1a773408e048 service nova] Lock "38e96e22-4200-4175-9085-31977357da2d-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 943.238415] env[61806]: DEBUG nova.compute.manager [req-536fb375-c0d7-4a99-9b2e-ea55b0398c53 req-b26cf4f2-434b-4201-8efb-1a773408e048 service nova] [instance: 38e96e22-4200-4175-9085-31977357da2d] No waiting events found dispatching network-vif-plugged-85d68e9b-a57d-4078-baf3-cd7c620f9366 {{(pid=61806) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 943.238600] env[61806]: WARNING nova.compute.manager [req-536fb375-c0d7-4a99-9b2e-ea55b0398c53 req-b26cf4f2-434b-4201-8efb-1a773408e048 service nova] [instance: 38e96e22-4200-4175-9085-31977357da2d] Received unexpected event network-vif-plugged-85d68e9b-a57d-4078-baf3-cd7c620f9366 for instance with vm_state building and task_state spawning. [ 943.276468] env[61806]: DEBUG oslo_concurrency.lockutils [None req-5d1f922c-8037-4fb1-9c2a-dcd72461a463 tempest-ServersNegativeTestJSON-891856253 tempest-ServersNegativeTestJSON-891856253-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 943.330381] env[61806]: DEBUG nova.network.neutron [None req-9cd2108a-a1e5-403d-bbda-81099a62d59d tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] [instance: 38e96e22-4200-4175-9085-31977357da2d] Successfully updated port: 85d68e9b-a57d-4078-baf3-cd7c620f9366 {{(pid=61806) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 943.427707] env[61806]: WARNING nova.network.neutron [None req-ccfe4ccf-a36d-4d36-adb7-3f5c04a0f60f tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] [instance: 6f6ba57a-e2d6-4749-a53a-e263861cb1c0] a7ffda72-e57c-41fa-a7ed-d7b12084bc9d already exists in list: networks containing: ['a7ffda72-e57c-41fa-a7ed-d7b12084bc9d']. ignoring it [ 943.643124] env[61806]: INFO nova.compute.claims [None req-cf5202db-26c0-4960-9218-bd95b8e81f62 tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] [instance: 676f18b3-3da1-4a53-b96d-fe64ee8f9101] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 943.669021] env[61806]: DEBUG oslo_concurrency.lockutils [None req-232a749a-b989-4ce4-a547-d7aa6acd6397 tempest-AttachVolumeNegativeTest-1683033704 tempest-AttachVolumeNegativeTest-1683033704-project-member] Lock "daff849e-eb4f-411e-92f8-9362e3bd91e3" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.459s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 943.748958] env[61806]: DEBUG nova.network.neutron [None req-ccfe4ccf-a36d-4d36-adb7-3f5c04a0f60f tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] [instance: 6f6ba57a-e2d6-4749-a53a-e263861cb1c0] Updating instance_info_cache with network_info: [{"id": "013c9f0d-65b1-4279-a823-a7f1c14dfb91", "address": "fa:16:3e:15:56:4e", "network": {"id": "a7ffda72-e57c-41fa-a7ed-d7b12084bc9d", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-62794865-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.218", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "6e391e1267014fafbb8f5a3211299819", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8b3cfeb1-f262-4fd9-b506-8e9c0733e2d8", "external-id": "nsx-vlan-transportzone-119", "segmentation_id": 119, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap013c9f0d-65", "ovs_interfaceid": "013c9f0d-65b1-4279-a823-a7f1c14dfb91", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "7e031e88-6928-472d-a65e-b818d89e6fc5", "address": "fa:16:3e:1e:a6:b3", "network": {"id": "a7ffda72-e57c-41fa-a7ed-d7b12084bc9d", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-62794865-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "6e391e1267014fafbb8f5a3211299819", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8b3cfeb1-f262-4fd9-b506-8e9c0733e2d8", "external-id": "nsx-vlan-transportzone-119", "segmentation_id": 119, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7e031e88-69", "ovs_interfaceid": "7e031e88-6928-472d-a65e-b818d89e6fc5", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": true, "delegate_create": true, "meta": {}}] {{(pid=61806) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 943.834960] env[61806]: DEBUG oslo_concurrency.lockutils [None req-9cd2108a-a1e5-403d-bbda-81099a62d59d tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] Acquiring lock "refresh_cache-38e96e22-4200-4175-9085-31977357da2d" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 943.834960] env[61806]: DEBUG oslo_concurrency.lockutils [None req-9cd2108a-a1e5-403d-bbda-81099a62d59d tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] Acquired lock "refresh_cache-38e96e22-4200-4175-9085-31977357da2d" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 943.834960] env[61806]: DEBUG nova.network.neutron [None req-9cd2108a-a1e5-403d-bbda-81099a62d59d tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] [instance: 38e96e22-4200-4175-9085-31977357da2d] Building network info cache for instance {{(pid=61806) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 944.149570] env[61806]: INFO nova.compute.resource_tracker [None req-cf5202db-26c0-4960-9218-bd95b8e81f62 tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] [instance: 676f18b3-3da1-4a53-b96d-fe64ee8f9101] Updating resource usage from migration 07d765d6-8d01-4ca8-b027-cc87b221420f [ 944.252578] env[61806]: DEBUG oslo_concurrency.lockutils [None req-ccfe4ccf-a36d-4d36-adb7-3f5c04a0f60f tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] Releasing lock "refresh_cache-6f6ba57a-e2d6-4749-a53a-e263861cb1c0" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 944.253464] env[61806]: DEBUG oslo_concurrency.lockutils [None req-ccfe4ccf-a36d-4d36-adb7-3f5c04a0f60f tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] Acquiring lock "6f6ba57a-e2d6-4749-a53a-e263861cb1c0" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 944.253553] env[61806]: DEBUG oslo_concurrency.lockutils [None req-ccfe4ccf-a36d-4d36-adb7-3f5c04a0f60f tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] Acquired lock "6f6ba57a-e2d6-4749-a53a-e263861cb1c0" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 944.256994] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-863d126c-d4f4-4ac3-992e-2310069d61b9 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 944.277443] env[61806]: DEBUG nova.virt.hardware [None req-ccfe4ccf-a36d-4d36-adb7-3f5c04a0f60f tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-15T18:05:10Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format=,created_at=,direct_url=,disk_format=,id=,min_disk=,min_ram=,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=61806) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 944.277707] env[61806]: DEBUG nova.virt.hardware [None req-ccfe4ccf-a36d-4d36-adb7-3f5c04a0f60f tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] Flavor limits 0:0:0 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 944.277871] env[61806]: DEBUG nova.virt.hardware [None req-ccfe4ccf-a36d-4d36-adb7-3f5c04a0f60f tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] Image limits 0:0:0 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 944.278115] env[61806]: DEBUG nova.virt.hardware [None req-ccfe4ccf-a36d-4d36-adb7-3f5c04a0f60f tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] Flavor pref 0:0:0 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 944.278281] env[61806]: DEBUG nova.virt.hardware [None req-ccfe4ccf-a36d-4d36-adb7-3f5c04a0f60f tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] Image pref 0:0:0 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 944.278458] env[61806]: DEBUG nova.virt.hardware [None req-ccfe4ccf-a36d-4d36-adb7-3f5c04a0f60f tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 944.278704] env[61806]: DEBUG nova.virt.hardware [None req-ccfe4ccf-a36d-4d36-adb7-3f5c04a0f60f tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61806) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 944.278895] env[61806]: DEBUG nova.virt.hardware [None req-ccfe4ccf-a36d-4d36-adb7-3f5c04a0f60f tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61806) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 944.279090] env[61806]: DEBUG nova.virt.hardware [None req-ccfe4ccf-a36d-4d36-adb7-3f5c04a0f60f tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] Got 1 possible topologies {{(pid=61806) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 944.279273] env[61806]: DEBUG nova.virt.hardware [None req-ccfe4ccf-a36d-4d36-adb7-3f5c04a0f60f tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61806) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 944.279455] env[61806]: DEBUG nova.virt.hardware [None req-ccfe4ccf-a36d-4d36-adb7-3f5c04a0f60f tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61806) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 944.286135] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-ccfe4ccf-a36d-4d36-adb7-3f5c04a0f60f tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] [instance: 6f6ba57a-e2d6-4749-a53a-e263861cb1c0] Reconfiguring VM to attach interface {{(pid=61806) attach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1929}} [ 944.289915] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-2ae32a2e-2dd9-4ed6-bedd-33eca8e13fed {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 944.310875] env[61806]: DEBUG oslo_vmware.api [None req-ccfe4ccf-a36d-4d36-adb7-3f5c04a0f60f tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] Waiting for the task: (returnval){ [ 944.310875] env[61806]: value = "task-1294947" [ 944.310875] env[61806]: _type = "Task" [ 944.310875] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 944.321987] env[61806]: DEBUG oslo_vmware.api [None req-ccfe4ccf-a36d-4d36-adb7-3f5c04a0f60f tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] Task: {'id': task-1294947, 'name': ReconfigVM_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 944.343122] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-59583e2c-cdfa-4542-ae53-3f3b36de8d54 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 944.351470] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7d74b552-8ede-44d5-8f26-4b03b35a9423 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 944.383046] env[61806]: DEBUG nova.network.neutron [None req-9cd2108a-a1e5-403d-bbda-81099a62d59d tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] [instance: 38e96e22-4200-4175-9085-31977357da2d] Instance cache missing network info. {{(pid=61806) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 944.386057] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8aeb5156-a7ff-4311-9fa2-0384727b4299 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 944.394404] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2653cfea-44b1-4d9c-8722-d585431a8f51 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 944.410436] env[61806]: DEBUG nova.compute.provider_tree [None req-cf5202db-26c0-4960-9218-bd95b8e81f62 tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Inventory has not changed in ProviderTree for provider: a2858be1-fd22-4e08-979e-87ad25293407 {{(pid=61806) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 944.451364] env[61806]: DEBUG nova.compute.manager [req-c5920898-70fb-41bb-9d6e-0da8f3b245e4 req-82f8fd60-6e87-4063-82ff-ea33bf5f860d service nova] [instance: 6f6ba57a-e2d6-4749-a53a-e263861cb1c0] Received event network-changed-7e031e88-6928-472d-a65e-b818d89e6fc5 {{(pid=61806) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 944.451644] env[61806]: DEBUG nova.compute.manager [req-c5920898-70fb-41bb-9d6e-0da8f3b245e4 req-82f8fd60-6e87-4063-82ff-ea33bf5f860d service nova] [instance: 6f6ba57a-e2d6-4749-a53a-e263861cb1c0] Refreshing instance network info cache due to event network-changed-7e031e88-6928-472d-a65e-b818d89e6fc5. {{(pid=61806) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 944.451926] env[61806]: DEBUG oslo_concurrency.lockutils [req-c5920898-70fb-41bb-9d6e-0da8f3b245e4 req-82f8fd60-6e87-4063-82ff-ea33bf5f860d service nova] Acquiring lock "refresh_cache-6f6ba57a-e2d6-4749-a53a-e263861cb1c0" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 944.452145] env[61806]: DEBUG oslo_concurrency.lockutils [req-c5920898-70fb-41bb-9d6e-0da8f3b245e4 req-82f8fd60-6e87-4063-82ff-ea33bf5f860d service nova] Acquired lock "refresh_cache-6f6ba57a-e2d6-4749-a53a-e263861cb1c0" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 944.452665] env[61806]: DEBUG nova.network.neutron [req-c5920898-70fb-41bb-9d6e-0da8f3b245e4 req-82f8fd60-6e87-4063-82ff-ea33bf5f860d service nova] [instance: 6f6ba57a-e2d6-4749-a53a-e263861cb1c0] Refreshing network info cache for port 7e031e88-6928-472d-a65e-b818d89e6fc5 {{(pid=61806) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 944.594109] env[61806]: DEBUG nova.network.neutron [None req-9cd2108a-a1e5-403d-bbda-81099a62d59d tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] [instance: 38e96e22-4200-4175-9085-31977357da2d] Updating instance_info_cache with network_info: [{"id": "85d68e9b-a57d-4078-baf3-cd7c620f9366", "address": "fa:16:3e:d0:4d:44", "network": {"id": "f280a01b-7000-4e54-ab9a-20eab156a410", "bridge": "br-int", "label": "tempest-ServersTestJSON-2080718146-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "73ebcc4d79d248efb653a084a5e44302", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "43798f54-0c07-4417-a23f-58bb6b7e204b", "external-id": "nsx-vlan-transportzone-571", "segmentation_id": 571, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap85d68e9b-a5", "ovs_interfaceid": "85d68e9b-a57d-4078-baf3-cd7c620f9366", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61806) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 944.807064] env[61806]: DEBUG oslo_concurrency.lockutils [None req-31f82ebb-fc83-41eb-84eb-02db4bac455e tempest-AttachVolumeNegativeTest-1683033704 tempest-AttachVolumeNegativeTest-1683033704-project-member] Acquiring lock "43dcce0d-4fdb-4ee1-8309-92e92d2331a9" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 944.807411] env[61806]: DEBUG oslo_concurrency.lockutils [None req-31f82ebb-fc83-41eb-84eb-02db4bac455e tempest-AttachVolumeNegativeTest-1683033704 tempest-AttachVolumeNegativeTest-1683033704-project-member] Lock "43dcce0d-4fdb-4ee1-8309-92e92d2331a9" acquired by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: waited 0.001s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 944.821731] env[61806]: DEBUG oslo_vmware.api [None req-ccfe4ccf-a36d-4d36-adb7-3f5c04a0f60f tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] Task: {'id': task-1294947, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 944.915305] env[61806]: DEBUG nova.scheduler.client.report [None req-cf5202db-26c0-4960-9218-bd95b8e81f62 tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Inventory has not changed for provider a2858be1-fd22-4e08-979e-87ad25293407 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 138, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61806) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 945.097277] env[61806]: DEBUG oslo_concurrency.lockutils [None req-9cd2108a-a1e5-403d-bbda-81099a62d59d tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] Releasing lock "refresh_cache-38e96e22-4200-4175-9085-31977357da2d" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 945.097489] env[61806]: DEBUG nova.compute.manager [None req-9cd2108a-a1e5-403d-bbda-81099a62d59d tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] [instance: 38e96e22-4200-4175-9085-31977357da2d] Instance network_info: |[{"id": "85d68e9b-a57d-4078-baf3-cd7c620f9366", "address": "fa:16:3e:d0:4d:44", "network": {"id": "f280a01b-7000-4e54-ab9a-20eab156a410", "bridge": "br-int", "label": "tempest-ServersTestJSON-2080718146-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "73ebcc4d79d248efb653a084a5e44302", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "43798f54-0c07-4417-a23f-58bb6b7e204b", "external-id": "nsx-vlan-transportzone-571", "segmentation_id": 571, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap85d68e9b-a5", "ovs_interfaceid": "85d68e9b-a57d-4078-baf3-cd7c620f9366", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61806) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 945.097922] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-9cd2108a-a1e5-403d-bbda-81099a62d59d tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] [instance: 38e96e22-4200-4175-9085-31977357da2d] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:d0:4d:44', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '43798f54-0c07-4417-a23f-58bb6b7e204b', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '85d68e9b-a57d-4078-baf3-cd7c620f9366', 'vif_model': 'vmxnet3'}] {{(pid=61806) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 945.105535] env[61806]: DEBUG oslo.service.loopingcall [None req-9cd2108a-a1e5-403d-bbda-81099a62d59d tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61806) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 945.105769] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 38e96e22-4200-4175-9085-31977357da2d] Creating VM on the ESX host {{(pid=61806) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 945.105995] env[61806]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-5070b185-6228-44e6-9d4f-1c5e996da6cd {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 945.127086] env[61806]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 945.127086] env[61806]: value = "task-1294948" [ 945.127086] env[61806]: _type = "Task" [ 945.127086] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 945.137357] env[61806]: DEBUG oslo_vmware.api [-] Task: {'id': task-1294948, 'name': CreateVM_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 945.194087] env[61806]: DEBUG nova.network.neutron [req-c5920898-70fb-41bb-9d6e-0da8f3b245e4 req-82f8fd60-6e87-4063-82ff-ea33bf5f860d service nova] [instance: 6f6ba57a-e2d6-4749-a53a-e263861cb1c0] Updated VIF entry in instance network info cache for port 7e031e88-6928-472d-a65e-b818d89e6fc5. {{(pid=61806) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 945.195056] env[61806]: DEBUG nova.network.neutron [req-c5920898-70fb-41bb-9d6e-0da8f3b245e4 req-82f8fd60-6e87-4063-82ff-ea33bf5f860d service nova] [instance: 6f6ba57a-e2d6-4749-a53a-e263861cb1c0] Updating instance_info_cache with network_info: [{"id": "013c9f0d-65b1-4279-a823-a7f1c14dfb91", "address": "fa:16:3e:15:56:4e", "network": {"id": "a7ffda72-e57c-41fa-a7ed-d7b12084bc9d", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-62794865-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.218", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "6e391e1267014fafbb8f5a3211299819", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8b3cfeb1-f262-4fd9-b506-8e9c0733e2d8", "external-id": "nsx-vlan-transportzone-119", "segmentation_id": 119, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap013c9f0d-65", "ovs_interfaceid": "013c9f0d-65b1-4279-a823-a7f1c14dfb91", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "7e031e88-6928-472d-a65e-b818d89e6fc5", "address": "fa:16:3e:1e:a6:b3", "network": {"id": "a7ffda72-e57c-41fa-a7ed-d7b12084bc9d", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-62794865-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "6e391e1267014fafbb8f5a3211299819", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8b3cfeb1-f262-4fd9-b506-8e9c0733e2d8", "external-id": "nsx-vlan-transportzone-119", "segmentation_id": 119, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7e031e88-69", "ovs_interfaceid": "7e031e88-6928-472d-a65e-b818d89e6fc5", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": true, "delegate_create": true, "meta": {}}] {{(pid=61806) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 945.274176] env[61806]: DEBUG nova.compute.manager [req-4a804f52-46e9-499a-9707-b6db54a4e2d3 req-9866007a-d0ee-4c52-b4a5-492a1358fa88 service nova] [instance: 38e96e22-4200-4175-9085-31977357da2d] Received event network-changed-85d68e9b-a57d-4078-baf3-cd7c620f9366 {{(pid=61806) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 945.274387] env[61806]: DEBUG nova.compute.manager [req-4a804f52-46e9-499a-9707-b6db54a4e2d3 req-9866007a-d0ee-4c52-b4a5-492a1358fa88 service nova] [instance: 38e96e22-4200-4175-9085-31977357da2d] Refreshing instance network info cache due to event network-changed-85d68e9b-a57d-4078-baf3-cd7c620f9366. {{(pid=61806) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 945.274611] env[61806]: DEBUG oslo_concurrency.lockutils [req-4a804f52-46e9-499a-9707-b6db54a4e2d3 req-9866007a-d0ee-4c52-b4a5-492a1358fa88 service nova] Acquiring lock "refresh_cache-38e96e22-4200-4175-9085-31977357da2d" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 945.274857] env[61806]: DEBUG oslo_concurrency.lockutils [req-4a804f52-46e9-499a-9707-b6db54a4e2d3 req-9866007a-d0ee-4c52-b4a5-492a1358fa88 service nova] Acquired lock "refresh_cache-38e96e22-4200-4175-9085-31977357da2d" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 945.275109] env[61806]: DEBUG nova.network.neutron [req-4a804f52-46e9-499a-9707-b6db54a4e2d3 req-9866007a-d0ee-4c52-b4a5-492a1358fa88 service nova] [instance: 38e96e22-4200-4175-9085-31977357da2d] Refreshing network info cache for port 85d68e9b-a57d-4078-baf3-cd7c620f9366 {{(pid=61806) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 945.309909] env[61806]: INFO nova.compute.manager [None req-31f82ebb-fc83-41eb-84eb-02db4bac455e tempest-AttachVolumeNegativeTest-1683033704 tempest-AttachVolumeNegativeTest-1683033704-project-member] [instance: 43dcce0d-4fdb-4ee1-8309-92e92d2331a9] Detaching volume 21f9e064-7d34-44c9-9759-32f3d73dec59 [ 945.323729] env[61806]: DEBUG oslo_vmware.api [None req-ccfe4ccf-a36d-4d36-adb7-3f5c04a0f60f tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] Task: {'id': task-1294947, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 945.341955] env[61806]: INFO nova.virt.block_device [None req-31f82ebb-fc83-41eb-84eb-02db4bac455e tempest-AttachVolumeNegativeTest-1683033704 tempest-AttachVolumeNegativeTest-1683033704-project-member] [instance: 43dcce0d-4fdb-4ee1-8309-92e92d2331a9] Attempting to driver detach volume 21f9e064-7d34-44c9-9759-32f3d73dec59 from mountpoint /dev/sdb [ 945.342262] env[61806]: DEBUG nova.virt.vmwareapi.volumeops [None req-31f82ebb-fc83-41eb-84eb-02db4bac455e tempest-AttachVolumeNegativeTest-1683033704 tempest-AttachVolumeNegativeTest-1683033704-project-member] [instance: 43dcce0d-4fdb-4ee1-8309-92e92d2331a9] Volume detach. Driver type: vmdk {{(pid=61806) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 945.342459] env[61806]: DEBUG nova.virt.vmwareapi.volumeops [None req-31f82ebb-fc83-41eb-84eb-02db4bac455e tempest-AttachVolumeNegativeTest-1683033704 tempest-AttachVolumeNegativeTest-1683033704-project-member] [instance: 43dcce0d-4fdb-4ee1-8309-92e92d2331a9] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-277686', 'volume_id': '21f9e064-7d34-44c9-9759-32f3d73dec59', 'name': 'volume-21f9e064-7d34-44c9-9759-32f3d73dec59', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '43dcce0d-4fdb-4ee1-8309-92e92d2331a9', 'attached_at': '', 'detached_at': '', 'volume_id': '21f9e064-7d34-44c9-9759-32f3d73dec59', 'serial': '21f9e064-7d34-44c9-9759-32f3d73dec59'} {{(pid=61806) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 945.343396] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cb977d42-de90-4c66-a40d-a9dbe8fe15c5 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 945.367229] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dcaeb67a-c694-42eb-b709-0167f5b80bc9 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 945.374505] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2ddb8284-7c4e-4813-8dc5-15370e83816b {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 945.395401] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e44cc01a-92ef-4308-a2ba-3d7f592cc48f {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 945.411329] env[61806]: DEBUG nova.virt.vmwareapi.volumeops [None req-31f82ebb-fc83-41eb-84eb-02db4bac455e tempest-AttachVolumeNegativeTest-1683033704 tempest-AttachVolumeNegativeTest-1683033704-project-member] The volume has not been displaced from its original location: [datastore2] volume-21f9e064-7d34-44c9-9759-32f3d73dec59/volume-21f9e064-7d34-44c9-9759-32f3d73dec59.vmdk. No consolidation needed. {{(pid=61806) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 945.416963] env[61806]: DEBUG nova.virt.vmwareapi.volumeops [None req-31f82ebb-fc83-41eb-84eb-02db4bac455e tempest-AttachVolumeNegativeTest-1683033704 tempest-AttachVolumeNegativeTest-1683033704-project-member] [instance: 43dcce0d-4fdb-4ee1-8309-92e92d2331a9] Reconfiguring VM instance instance-0000003f to detach disk 2001 {{(pid=61806) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 945.417347] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-f6f91a93-4752-4c2f-a85b-0a4842d96dec {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 945.430962] env[61806]: DEBUG oslo_concurrency.lockutils [None req-cf5202db-26c0-4960-9218-bd95b8e81f62 tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: held 2.293s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 945.431203] env[61806]: INFO nova.compute.manager [None req-cf5202db-26c0-4960-9218-bd95b8e81f62 tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] [instance: 676f18b3-3da1-4a53-b96d-fe64ee8f9101] Migrating [ 945.431445] env[61806]: DEBUG oslo_concurrency.lockutils [None req-cf5202db-26c0-4960-9218-bd95b8e81f62 tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Acquiring lock "compute-rpcapi-router" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 945.431601] env[61806]: DEBUG oslo_concurrency.lockutils [None req-cf5202db-26c0-4960-9218-bd95b8e81f62 tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Acquired lock "compute-rpcapi-router" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 945.433246] env[61806]: DEBUG oslo_concurrency.lockutils [None req-5d1f922c-8037-4fb1-9c2a-dcd72461a463 tempest-ServersNegativeTestJSON-891856253 tempest-ServersNegativeTestJSON-891856253-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 2.157s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 945.434813] env[61806]: INFO nova.compute.claims [None req-5d1f922c-8037-4fb1-9c2a-dcd72461a463 tempest-ServersNegativeTestJSON-891856253 tempest-ServersNegativeTestJSON-891856253-project-member] [instance: bdbaef31-7d37-40cb-ac3e-ad152b37e2d2] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 945.443499] env[61806]: DEBUG oslo_vmware.api [None req-31f82ebb-fc83-41eb-84eb-02db4bac455e tempest-AttachVolumeNegativeTest-1683033704 tempest-AttachVolumeNegativeTest-1683033704-project-member] Waiting for the task: (returnval){ [ 945.443499] env[61806]: value = "task-1294949" [ 945.443499] env[61806]: _type = "Task" [ 945.443499] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 945.452250] env[61806]: DEBUG oslo_vmware.api [None req-31f82ebb-fc83-41eb-84eb-02db4bac455e tempest-AttachVolumeNegativeTest-1683033704 tempest-AttachVolumeNegativeTest-1683033704-project-member] Task: {'id': task-1294949, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 945.636787] env[61806]: DEBUG oslo_vmware.api [-] Task: {'id': task-1294948, 'name': CreateVM_Task} progress is 6%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 945.697692] env[61806]: DEBUG oslo_concurrency.lockutils [req-c5920898-70fb-41bb-9d6e-0da8f3b245e4 req-82f8fd60-6e87-4063-82ff-ea33bf5f860d service nova] Releasing lock "refresh_cache-6f6ba57a-e2d6-4749-a53a-e263861cb1c0" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 945.822025] env[61806]: DEBUG oslo_vmware.api [None req-ccfe4ccf-a36d-4d36-adb7-3f5c04a0f60f tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] Task: {'id': task-1294947, 'name': ReconfigVM_Task, 'duration_secs': 1.344552} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 945.822572] env[61806]: DEBUG oslo_concurrency.lockutils [None req-ccfe4ccf-a36d-4d36-adb7-3f5c04a0f60f tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] Releasing lock "6f6ba57a-e2d6-4749-a53a-e263861cb1c0" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 945.822795] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-ccfe4ccf-a36d-4d36-adb7-3f5c04a0f60f tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] [instance: 6f6ba57a-e2d6-4749-a53a-e263861cb1c0] Reconfigured VM to attach interface {{(pid=61806) attach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1943}} [ 945.939055] env[61806]: INFO nova.compute.rpcapi [None req-cf5202db-26c0-4960-9218-bd95b8e81f62 tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Automatically selected compute RPC version 6.3 from minimum service version 67 [ 945.939636] env[61806]: DEBUG oslo_concurrency.lockutils [None req-cf5202db-26c0-4960-9218-bd95b8e81f62 tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Releasing lock "compute-rpcapi-router" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 945.970778] env[61806]: DEBUG oslo_vmware.api [None req-31f82ebb-fc83-41eb-84eb-02db4bac455e tempest-AttachVolumeNegativeTest-1683033704 tempest-AttachVolumeNegativeTest-1683033704-project-member] Task: {'id': task-1294949, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 945.992928] env[61806]: DEBUG nova.network.neutron [req-4a804f52-46e9-499a-9707-b6db54a4e2d3 req-9866007a-d0ee-4c52-b4a5-492a1358fa88 service nova] [instance: 38e96e22-4200-4175-9085-31977357da2d] Updated VIF entry in instance network info cache for port 85d68e9b-a57d-4078-baf3-cd7c620f9366. {{(pid=61806) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 945.993313] env[61806]: DEBUG nova.network.neutron [req-4a804f52-46e9-499a-9707-b6db54a4e2d3 req-9866007a-d0ee-4c52-b4a5-492a1358fa88 service nova] [instance: 38e96e22-4200-4175-9085-31977357da2d] Updating instance_info_cache with network_info: [{"id": "85d68e9b-a57d-4078-baf3-cd7c620f9366", "address": "fa:16:3e:d0:4d:44", "network": {"id": "f280a01b-7000-4e54-ab9a-20eab156a410", "bridge": "br-int", "label": "tempest-ServersTestJSON-2080718146-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "73ebcc4d79d248efb653a084a5e44302", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "43798f54-0c07-4417-a23f-58bb6b7e204b", "external-id": "nsx-vlan-transportzone-571", "segmentation_id": 571, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap85d68e9b-a5", "ovs_interfaceid": "85d68e9b-a57d-4078-baf3-cd7c620f9366", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61806) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 946.112059] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bb93a114-db51-4978-b165-6b6824eee398 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 946.119409] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-baae77f8-6db3-4f3f-9a85-13860929e711 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 946.152670] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fceaff97-4a14-4d3f-b976-73f5816b19c5 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 946.159923] env[61806]: DEBUG oslo_vmware.api [-] Task: {'id': task-1294948, 'name': CreateVM_Task} progress is 6%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 946.162994] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-464eeb09-b3f6-4e91-838c-a4d07c72c879 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 946.176442] env[61806]: DEBUG nova.compute.provider_tree [None req-5d1f922c-8037-4fb1-9c2a-dcd72461a463 tempest-ServersNegativeTestJSON-891856253 tempest-ServersNegativeTestJSON-891856253-project-member] Inventory has not changed in ProviderTree for provider: a2858be1-fd22-4e08-979e-87ad25293407 {{(pid=61806) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 946.327457] env[61806]: DEBUG oslo_concurrency.lockutils [None req-ccfe4ccf-a36d-4d36-adb7-3f5c04a0f60f tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] Lock "interface-6f6ba57a-e2d6-4749-a53a-e263861cb1c0-7e031e88-6928-472d-a65e-b818d89e6fc5" "released" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: held 6.724s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 946.461169] env[61806]: DEBUG oslo_vmware.api [None req-31f82ebb-fc83-41eb-84eb-02db4bac455e tempest-AttachVolumeNegativeTest-1683033704 tempest-AttachVolumeNegativeTest-1683033704-project-member] Task: {'id': task-1294949, 'name': ReconfigVM_Task} progress is 99%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 946.464804] env[61806]: DEBUG oslo_concurrency.lockutils [None req-cf5202db-26c0-4960-9218-bd95b8e81f62 tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Acquiring lock "refresh_cache-676f18b3-3da1-4a53-b96d-fe64ee8f9101" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 946.465055] env[61806]: DEBUG oslo_concurrency.lockutils [None req-cf5202db-26c0-4960-9218-bd95b8e81f62 tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Acquired lock "refresh_cache-676f18b3-3da1-4a53-b96d-fe64ee8f9101" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 946.465265] env[61806]: DEBUG nova.network.neutron [None req-cf5202db-26c0-4960-9218-bd95b8e81f62 tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] [instance: 676f18b3-3da1-4a53-b96d-fe64ee8f9101] Building network info cache for instance {{(pid=61806) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 946.496018] env[61806]: DEBUG oslo_concurrency.lockutils [req-4a804f52-46e9-499a-9707-b6db54a4e2d3 req-9866007a-d0ee-4c52-b4a5-492a1358fa88 service nova] Releasing lock "refresh_cache-38e96e22-4200-4175-9085-31977357da2d" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 946.660243] env[61806]: DEBUG oslo_vmware.api [-] Task: {'id': task-1294948, 'name': CreateVM_Task} progress is 99%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 946.679529] env[61806]: DEBUG nova.scheduler.client.report [None req-5d1f922c-8037-4fb1-9c2a-dcd72461a463 tempest-ServersNegativeTestJSON-891856253 tempest-ServersNegativeTestJSON-891856253-project-member] Inventory has not changed for provider a2858be1-fd22-4e08-979e-87ad25293407 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 138, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61806) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 946.961938] env[61806]: DEBUG oslo_vmware.api [None req-31f82ebb-fc83-41eb-84eb-02db4bac455e tempest-AttachVolumeNegativeTest-1683033704 tempest-AttachVolumeNegativeTest-1683033704-project-member] Task: {'id': task-1294949, 'name': ReconfigVM_Task, 'duration_secs': 1.025512} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 946.962304] env[61806]: DEBUG nova.virt.vmwareapi.volumeops [None req-31f82ebb-fc83-41eb-84eb-02db4bac455e tempest-AttachVolumeNegativeTest-1683033704 tempest-AttachVolumeNegativeTest-1683033704-project-member] [instance: 43dcce0d-4fdb-4ee1-8309-92e92d2331a9] Reconfigured VM instance instance-0000003f to detach disk 2001 {{(pid=61806) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 946.966827] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-b6a8365c-b094-474b-b548-d78556b9cb80 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 946.983348] env[61806]: DEBUG oslo_vmware.api [None req-31f82ebb-fc83-41eb-84eb-02db4bac455e tempest-AttachVolumeNegativeTest-1683033704 tempest-AttachVolumeNegativeTest-1683033704-project-member] Waiting for the task: (returnval){ [ 946.983348] env[61806]: value = "task-1294950" [ 946.983348] env[61806]: _type = "Task" [ 946.983348] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 946.992285] env[61806]: DEBUG oslo_vmware.api [None req-31f82ebb-fc83-41eb-84eb-02db4bac455e tempest-AttachVolumeNegativeTest-1683033704 tempest-AttachVolumeNegativeTest-1683033704-project-member] Task: {'id': task-1294950, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 947.159580] env[61806]: DEBUG oslo_vmware.api [-] Task: {'id': task-1294948, 'name': CreateVM_Task} progress is 99%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 947.184599] env[61806]: DEBUG oslo_concurrency.lockutils [None req-5d1f922c-8037-4fb1-9c2a-dcd72461a463 tempest-ServersNegativeTestJSON-891856253 tempest-ServersNegativeTestJSON-891856253-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 1.751s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 947.185267] env[61806]: DEBUG nova.compute.manager [None req-5d1f922c-8037-4fb1-9c2a-dcd72461a463 tempest-ServersNegativeTestJSON-891856253 tempest-ServersNegativeTestJSON-891856253-project-member] [instance: bdbaef31-7d37-40cb-ac3e-ad152b37e2d2] Start building networks asynchronously for instance. {{(pid=61806) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 947.196597] env[61806]: DEBUG nova.network.neutron [None req-cf5202db-26c0-4960-9218-bd95b8e81f62 tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] [instance: 676f18b3-3da1-4a53-b96d-fe64ee8f9101] Updating instance_info_cache with network_info: [{"id": "d3a921a7-f1fa-4352-ae53-9f2996576fef", "address": "fa:16:3e:66:40:96", "network": {"id": "46515a4e-610f-42fb-a424-205a03df0640", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-1133623547-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4b9d94c3d684474b99e305fb6cd9b95f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "82ca17df-257e-40e6-9ec9-310ed6f05ccb", "external-id": "nsx-vlan-transportzone-778", "segmentation_id": 778, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd3a921a7-f1", "ovs_interfaceid": "d3a921a7-f1fa-4352-ae53-9f2996576fef", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61806) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 947.493754] env[61806]: DEBUG oslo_vmware.api [None req-31f82ebb-fc83-41eb-84eb-02db4bac455e tempest-AttachVolumeNegativeTest-1683033704 tempest-AttachVolumeNegativeTest-1683033704-project-member] Task: {'id': task-1294950, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 947.659556] env[61806]: DEBUG oslo_vmware.api [-] Task: {'id': task-1294948, 'name': CreateVM_Task, 'duration_secs': 2.480891} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 947.659740] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 38e96e22-4200-4175-9085-31977357da2d] Created VM on the ESX host {{(pid=61806) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 947.660534] env[61806]: DEBUG oslo_concurrency.lockutils [None req-9cd2108a-a1e5-403d-bbda-81099a62d59d tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 947.660713] env[61806]: DEBUG oslo_concurrency.lockutils [None req-9cd2108a-a1e5-403d-bbda-81099a62d59d tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] Acquired lock "[datastore1] devstack-image-cache_base/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 947.661315] env[61806]: DEBUG oslo_concurrency.lockutils [None req-9cd2108a-a1e5-403d-bbda-81099a62d59d tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 947.661576] env[61806]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-06248aea-30bb-4899-be14-1331d3048599 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 947.665992] env[61806]: DEBUG oslo_vmware.api [None req-9cd2108a-a1e5-403d-bbda-81099a62d59d tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] Waiting for the task: (returnval){ [ 947.665992] env[61806]: value = "session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]529d97e0-f4fb-0c56-e567-67ee1690b798" [ 947.665992] env[61806]: _type = "Task" [ 947.665992] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 947.673718] env[61806]: DEBUG oslo_vmware.api [None req-9cd2108a-a1e5-403d-bbda-81099a62d59d tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] Task: {'id': session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]529d97e0-f4fb-0c56-e567-67ee1690b798, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 947.690898] env[61806]: DEBUG nova.compute.utils [None req-5d1f922c-8037-4fb1-9c2a-dcd72461a463 tempest-ServersNegativeTestJSON-891856253 tempest-ServersNegativeTestJSON-891856253-project-member] Using /dev/sd instead of None {{(pid=61806) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 947.691849] env[61806]: DEBUG nova.compute.manager [None req-5d1f922c-8037-4fb1-9c2a-dcd72461a463 tempest-ServersNegativeTestJSON-891856253 tempest-ServersNegativeTestJSON-891856253-project-member] [instance: bdbaef31-7d37-40cb-ac3e-ad152b37e2d2] Allocating IP information in the background. {{(pid=61806) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 947.692552] env[61806]: DEBUG nova.network.neutron [None req-5d1f922c-8037-4fb1-9c2a-dcd72461a463 tempest-ServersNegativeTestJSON-891856253 tempest-ServersNegativeTestJSON-891856253-project-member] [instance: bdbaef31-7d37-40cb-ac3e-ad152b37e2d2] allocate_for_instance() {{(pid=61806) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 947.699170] env[61806]: DEBUG oslo_concurrency.lockutils [None req-cf5202db-26c0-4960-9218-bd95b8e81f62 tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Releasing lock "refresh_cache-676f18b3-3da1-4a53-b96d-fe64ee8f9101" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 947.731760] env[61806]: DEBUG nova.policy [None req-5d1f922c-8037-4fb1-9c2a-dcd72461a463 tempest-ServersNegativeTestJSON-891856253 tempest-ServersNegativeTestJSON-891856253-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'bfd1d620a5ea4b11ae028bd6c719a47a', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'd2e2b7eec47e4b10898215123408692a', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61806) authorize /opt/stack/nova/nova/policy.py:201}} [ 947.817138] env[61806]: DEBUG oslo_concurrency.lockutils [None req-61d33433-c3e5-4624-95f4-86df358dffc7 tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] Acquiring lock "interface-6f6ba57a-e2d6-4749-a53a-e263861cb1c0-7e031e88-6928-472d-a65e-b818d89e6fc5" by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 947.817432] env[61806]: DEBUG oslo_concurrency.lockutils [None req-61d33433-c3e5-4624-95f4-86df358dffc7 tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] Lock "interface-6f6ba57a-e2d6-4749-a53a-e263861cb1c0-7e031e88-6928-472d-a65e-b818d89e6fc5" acquired by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" :: waited 0.001s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 947.994300] env[61806]: DEBUG oslo_vmware.api [None req-31f82ebb-fc83-41eb-84eb-02db4bac455e tempest-AttachVolumeNegativeTest-1683033704 tempest-AttachVolumeNegativeTest-1683033704-project-member] Task: {'id': task-1294950, 'name': ReconfigVM_Task, 'duration_secs': 0.650126} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 947.994759] env[61806]: DEBUG nova.virt.vmwareapi.volumeops [None req-31f82ebb-fc83-41eb-84eb-02db4bac455e tempest-AttachVolumeNegativeTest-1683033704 tempest-AttachVolumeNegativeTest-1683033704-project-member] [instance: 43dcce0d-4fdb-4ee1-8309-92e92d2331a9] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-277686', 'volume_id': '21f9e064-7d34-44c9-9759-32f3d73dec59', 'name': 'volume-21f9e064-7d34-44c9-9759-32f3d73dec59', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '43dcce0d-4fdb-4ee1-8309-92e92d2331a9', 'attached_at': '', 'detached_at': '', 'volume_id': '21f9e064-7d34-44c9-9759-32f3d73dec59', 'serial': '21f9e064-7d34-44c9-9759-32f3d73dec59'} {{(pid=61806) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 948.051586] env[61806]: DEBUG nova.network.neutron [None req-5d1f922c-8037-4fb1-9c2a-dcd72461a463 tempest-ServersNegativeTestJSON-891856253 tempest-ServersNegativeTestJSON-891856253-project-member] [instance: bdbaef31-7d37-40cb-ac3e-ad152b37e2d2] Successfully created port: b8b714a4-b17c-4b3d-889e-407bec6c30ae {{(pid=61806) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 948.178031] env[61806]: DEBUG oslo_vmware.api [None req-9cd2108a-a1e5-403d-bbda-81099a62d59d tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] Task: {'id': session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]529d97e0-f4fb-0c56-e567-67ee1690b798, 'name': SearchDatastore_Task, 'duration_secs': 0.011931} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 948.178363] env[61806]: DEBUG oslo_concurrency.lockutils [None req-9cd2108a-a1e5-403d-bbda-81099a62d59d tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] Releasing lock "[datastore1] devstack-image-cache_base/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 948.178622] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-9cd2108a-a1e5-403d-bbda-81099a62d59d tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] [instance: 38e96e22-4200-4175-9085-31977357da2d] Processing image ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa {{(pid=61806) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 948.178879] env[61806]: DEBUG oslo_concurrency.lockutils [None req-9cd2108a-a1e5-403d-bbda-81099a62d59d tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa.vmdk" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 948.179067] env[61806]: DEBUG oslo_concurrency.lockutils [None req-9cd2108a-a1e5-403d-bbda-81099a62d59d tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] Acquired lock "[datastore1] devstack-image-cache_base/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa.vmdk" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 948.179748] env[61806]: DEBUG nova.virt.vmwareapi.ds_util [None req-9cd2108a-a1e5-403d-bbda-81099a62d59d tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61806) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 948.179748] env[61806]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-142434ca-6f1f-445b-a383-4b02123e33c2 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 948.188370] env[61806]: DEBUG nova.virt.vmwareapi.ds_util [None req-9cd2108a-a1e5-403d-bbda-81099a62d59d tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61806) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 948.188607] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-9cd2108a-a1e5-403d-bbda-81099a62d59d tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61806) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 948.189412] env[61806]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-6e5f4816-4a32-4865-b0bd-5ace70f48491 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 948.195188] env[61806]: DEBUG nova.compute.manager [None req-5d1f922c-8037-4fb1-9c2a-dcd72461a463 tempest-ServersNegativeTestJSON-891856253 tempest-ServersNegativeTestJSON-891856253-project-member] [instance: bdbaef31-7d37-40cb-ac3e-ad152b37e2d2] Start building block device mappings for instance. {{(pid=61806) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 948.198176] env[61806]: DEBUG oslo_vmware.api [None req-9cd2108a-a1e5-403d-bbda-81099a62d59d tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] Waiting for the task: (returnval){ [ 948.198176] env[61806]: value = "session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]52ba8a3d-b7df-71d8-10fb-daa8f9ccd60d" [ 948.198176] env[61806]: _type = "Task" [ 948.198176] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 948.220882] env[61806]: DEBUG oslo_vmware.api [None req-9cd2108a-a1e5-403d-bbda-81099a62d59d tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] Task: {'id': session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]52ba8a3d-b7df-71d8-10fb-daa8f9ccd60d, 'name': SearchDatastore_Task, 'duration_secs': 0.009524} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 948.220882] env[61806]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ce2c7a5e-30ac-44c6-8618-3b8c2173d7c9 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 948.232758] env[61806]: DEBUG oslo_vmware.api [None req-9cd2108a-a1e5-403d-bbda-81099a62d59d tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] Waiting for the task: (returnval){ [ 948.232758] env[61806]: value = "session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]52fdc29a-f64b-dbbc-8304-0649fa5d9e04" [ 948.232758] env[61806]: _type = "Task" [ 948.232758] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 948.239343] env[61806]: DEBUG oslo_vmware.api [None req-9cd2108a-a1e5-403d-bbda-81099a62d59d tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] Task: {'id': session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]52fdc29a-f64b-dbbc-8304-0649fa5d9e04, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 948.323016] env[61806]: DEBUG oslo_concurrency.lockutils [None req-61d33433-c3e5-4624-95f4-86df358dffc7 tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] Acquiring lock "6f6ba57a-e2d6-4749-a53a-e263861cb1c0" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 948.323016] env[61806]: DEBUG oslo_concurrency.lockutils [None req-61d33433-c3e5-4624-95f4-86df358dffc7 tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] Acquired lock "6f6ba57a-e2d6-4749-a53a-e263861cb1c0" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 948.323016] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8f03e548-fe88-4c09-8767-b0e68e445108 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 948.342111] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a19a8446-eeba-4602-9f89-9b670ca22a55 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 948.368031] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-61d33433-c3e5-4624-95f4-86df358dffc7 tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] [instance: 6f6ba57a-e2d6-4749-a53a-e263861cb1c0] Reconfiguring VM to detach interface {{(pid=61806) detach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1974}} [ 948.368635] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-06923336-5a77-493f-8341-14418a16caad {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 948.390024] env[61806]: DEBUG oslo_vmware.api [None req-61d33433-c3e5-4624-95f4-86df358dffc7 tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] Waiting for the task: (returnval){ [ 948.390024] env[61806]: value = "task-1294951" [ 948.390024] env[61806]: _type = "Task" [ 948.390024] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 948.397860] env[61806]: DEBUG oslo_vmware.api [None req-61d33433-c3e5-4624-95f4-86df358dffc7 tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] Task: {'id': task-1294951, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 948.541203] env[61806]: DEBUG nova.objects.instance [None req-31f82ebb-fc83-41eb-84eb-02db4bac455e tempest-AttachVolumeNegativeTest-1683033704 tempest-AttachVolumeNegativeTest-1683033704-project-member] Lazy-loading 'flavor' on Instance uuid 43dcce0d-4fdb-4ee1-8309-92e92d2331a9 {{(pid=61806) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 948.739936] env[61806]: DEBUG oslo_vmware.api [None req-9cd2108a-a1e5-403d-bbda-81099a62d59d tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] Task: {'id': session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]52fdc29a-f64b-dbbc-8304-0649fa5d9e04, 'name': SearchDatastore_Task, 'duration_secs': 0.010511} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 948.740894] env[61806]: DEBUG oslo_concurrency.lockutils [None req-9cd2108a-a1e5-403d-bbda-81099a62d59d tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] Releasing lock "[datastore1] devstack-image-cache_base/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa.vmdk" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 948.741191] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [None req-9cd2108a-a1e5-403d-bbda-81099a62d59d tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa.vmdk to [datastore1] 38e96e22-4200-4175-9085-31977357da2d/38e96e22-4200-4175-9085-31977357da2d.vmdk {{(pid=61806) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 948.741470] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-327168d4-df05-4640-a98d-825d3252812f {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 948.748178] env[61806]: DEBUG oslo_vmware.api [None req-9cd2108a-a1e5-403d-bbda-81099a62d59d tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] Waiting for the task: (returnval){ [ 948.748178] env[61806]: value = "task-1294952" [ 948.748178] env[61806]: _type = "Task" [ 948.748178] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 948.758285] env[61806]: DEBUG oslo_vmware.api [None req-9cd2108a-a1e5-403d-bbda-81099a62d59d tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] Task: {'id': task-1294952, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 948.898849] env[61806]: DEBUG oslo_vmware.api [None req-61d33433-c3e5-4624-95f4-86df358dffc7 tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] Task: {'id': task-1294951, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 949.206302] env[61806]: DEBUG nova.compute.manager [None req-5d1f922c-8037-4fb1-9c2a-dcd72461a463 tempest-ServersNegativeTestJSON-891856253 tempest-ServersNegativeTestJSON-891856253-project-member] [instance: bdbaef31-7d37-40cb-ac3e-ad152b37e2d2] Start spawning the instance on the hypervisor. {{(pid=61806) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 949.214177] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0ef080c3-d079-481a-8001-11cfbc9afebd {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 949.233312] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-cf5202db-26c0-4960-9218-bd95b8e81f62 tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] [instance: 676f18b3-3da1-4a53-b96d-fe64ee8f9101] Updating instance '676f18b3-3da1-4a53-b96d-fe64ee8f9101' progress to 0 {{(pid=61806) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 949.246727] env[61806]: DEBUG nova.virt.hardware [None req-5d1f922c-8037-4fb1-9c2a-dcd72461a463 tempest-ServersNegativeTestJSON-891856253 tempest-ServersNegativeTestJSON-891856253-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-15T18:05:10Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-15T18:04:52Z,direct_url=,disk_format='vmdk',id=ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='84ba83fce2c349a6988173c5d6fc3b07',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-15T18:04:53Z,virtual_size=,visibility=), allow threads: False {{(pid=61806) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 949.247279] env[61806]: DEBUG nova.virt.hardware [None req-5d1f922c-8037-4fb1-9c2a-dcd72461a463 tempest-ServersNegativeTestJSON-891856253 tempest-ServersNegativeTestJSON-891856253-project-member] Flavor limits 0:0:0 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 949.247279] env[61806]: DEBUG nova.virt.hardware [None req-5d1f922c-8037-4fb1-9c2a-dcd72461a463 tempest-ServersNegativeTestJSON-891856253 tempest-ServersNegativeTestJSON-891856253-project-member] Image limits 0:0:0 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 949.247360] env[61806]: DEBUG nova.virt.hardware [None req-5d1f922c-8037-4fb1-9c2a-dcd72461a463 tempest-ServersNegativeTestJSON-891856253 tempest-ServersNegativeTestJSON-891856253-project-member] Flavor pref 0:0:0 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 949.247504] env[61806]: DEBUG nova.virt.hardware [None req-5d1f922c-8037-4fb1-9c2a-dcd72461a463 tempest-ServersNegativeTestJSON-891856253 tempest-ServersNegativeTestJSON-891856253-project-member] Image pref 0:0:0 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 949.247659] env[61806]: DEBUG nova.virt.hardware [None req-5d1f922c-8037-4fb1-9c2a-dcd72461a463 tempest-ServersNegativeTestJSON-891856253 tempest-ServersNegativeTestJSON-891856253-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 949.247893] env[61806]: DEBUG nova.virt.hardware [None req-5d1f922c-8037-4fb1-9c2a-dcd72461a463 tempest-ServersNegativeTestJSON-891856253 tempest-ServersNegativeTestJSON-891856253-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61806) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 949.248112] env[61806]: DEBUG nova.virt.hardware [None req-5d1f922c-8037-4fb1-9c2a-dcd72461a463 tempest-ServersNegativeTestJSON-891856253 tempest-ServersNegativeTestJSON-891856253-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61806) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 949.249028] env[61806]: DEBUG nova.virt.hardware [None req-5d1f922c-8037-4fb1-9c2a-dcd72461a463 tempest-ServersNegativeTestJSON-891856253 tempest-ServersNegativeTestJSON-891856253-project-member] Got 1 possible topologies {{(pid=61806) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 949.249028] env[61806]: DEBUG nova.virt.hardware [None req-5d1f922c-8037-4fb1-9c2a-dcd72461a463 tempest-ServersNegativeTestJSON-891856253 tempest-ServersNegativeTestJSON-891856253-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61806) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 949.249028] env[61806]: DEBUG nova.virt.hardware [None req-5d1f922c-8037-4fb1-9c2a-dcd72461a463 tempest-ServersNegativeTestJSON-891856253 tempest-ServersNegativeTestJSON-891856253-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61806) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 949.249629] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b824c092-3cd5-41ec-b64f-455a1066a643 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 949.263856] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7340b02c-956d-4405-8212-83180ef34913 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 949.268748] env[61806]: DEBUG oslo_vmware.api [None req-9cd2108a-a1e5-403d-bbda-81099a62d59d tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] Task: {'id': task-1294952, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 949.400962] env[61806]: DEBUG oslo_vmware.api [None req-61d33433-c3e5-4624-95f4-86df358dffc7 tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] Task: {'id': task-1294951, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 949.456249] env[61806]: DEBUG nova.compute.manager [req-acf02e73-044c-462f-8b78-f4c23f24aea8 req-b4ee36dd-7282-4f84-9f14-5fa3551d9510 service nova] [instance: bdbaef31-7d37-40cb-ac3e-ad152b37e2d2] Received event network-vif-plugged-b8b714a4-b17c-4b3d-889e-407bec6c30ae {{(pid=61806) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 949.456249] env[61806]: DEBUG oslo_concurrency.lockutils [req-acf02e73-044c-462f-8b78-f4c23f24aea8 req-b4ee36dd-7282-4f84-9f14-5fa3551d9510 service nova] Acquiring lock "bdbaef31-7d37-40cb-ac3e-ad152b37e2d2-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 949.456249] env[61806]: DEBUG oslo_concurrency.lockutils [req-acf02e73-044c-462f-8b78-f4c23f24aea8 req-b4ee36dd-7282-4f84-9f14-5fa3551d9510 service nova] Lock "bdbaef31-7d37-40cb-ac3e-ad152b37e2d2-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 949.456249] env[61806]: DEBUG oslo_concurrency.lockutils [req-acf02e73-044c-462f-8b78-f4c23f24aea8 req-b4ee36dd-7282-4f84-9f14-5fa3551d9510 service nova] Lock "bdbaef31-7d37-40cb-ac3e-ad152b37e2d2-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 949.456516] env[61806]: DEBUG nova.compute.manager [req-acf02e73-044c-462f-8b78-f4c23f24aea8 req-b4ee36dd-7282-4f84-9f14-5fa3551d9510 service nova] [instance: bdbaef31-7d37-40cb-ac3e-ad152b37e2d2] No waiting events found dispatching network-vif-plugged-b8b714a4-b17c-4b3d-889e-407bec6c30ae {{(pid=61806) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 949.456892] env[61806]: WARNING nova.compute.manager [req-acf02e73-044c-462f-8b78-f4c23f24aea8 req-b4ee36dd-7282-4f84-9f14-5fa3551d9510 service nova] [instance: bdbaef31-7d37-40cb-ac3e-ad152b37e2d2] Received unexpected event network-vif-plugged-b8b714a4-b17c-4b3d-889e-407bec6c30ae for instance with vm_state building and task_state spawning. [ 949.549803] env[61806]: DEBUG oslo_concurrency.lockutils [None req-31f82ebb-fc83-41eb-84eb-02db4bac455e tempest-AttachVolumeNegativeTest-1683033704 tempest-AttachVolumeNegativeTest-1683033704-project-member] Lock "43dcce0d-4fdb-4ee1-8309-92e92d2331a9" "released" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: held 4.742s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 949.612767] env[61806]: DEBUG nova.network.neutron [None req-5d1f922c-8037-4fb1-9c2a-dcd72461a463 tempest-ServersNegativeTestJSON-891856253 tempest-ServersNegativeTestJSON-891856253-project-member] [instance: bdbaef31-7d37-40cb-ac3e-ad152b37e2d2] Successfully updated port: b8b714a4-b17c-4b3d-889e-407bec6c30ae {{(pid=61806) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 949.741637] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [None req-cf5202db-26c0-4960-9218-bd95b8e81f62 tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] [instance: 676f18b3-3da1-4a53-b96d-fe64ee8f9101] Powering off the VM {{(pid=61806) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 949.741982] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-454b848e-445e-4200-8ea1-208cd618c5ac {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 949.749056] env[61806]: DEBUG oslo_vmware.api [None req-cf5202db-26c0-4960-9218-bd95b8e81f62 tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Waiting for the task: (returnval){ [ 949.749056] env[61806]: value = "task-1294953" [ 949.749056] env[61806]: _type = "Task" [ 949.749056] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 949.761348] env[61806]: DEBUG oslo_vmware.api [None req-cf5202db-26c0-4960-9218-bd95b8e81f62 tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Task: {'id': task-1294953, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 949.764785] env[61806]: DEBUG oslo_vmware.api [None req-9cd2108a-a1e5-403d-bbda-81099a62d59d tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] Task: {'id': task-1294952, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.561643} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 949.765146] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [None req-9cd2108a-a1e5-403d-bbda-81099a62d59d tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa.vmdk to [datastore1] 38e96e22-4200-4175-9085-31977357da2d/38e96e22-4200-4175-9085-31977357da2d.vmdk {{(pid=61806) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 949.765346] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-9cd2108a-a1e5-403d-bbda-81099a62d59d tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] [instance: 38e96e22-4200-4175-9085-31977357da2d] Extending root virtual disk to 1048576 {{(pid=61806) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 949.765615] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-beebde80-d4ad-4717-8fed-d90e4f094e21 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 949.773072] env[61806]: DEBUG oslo_vmware.api [None req-9cd2108a-a1e5-403d-bbda-81099a62d59d tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] Waiting for the task: (returnval){ [ 949.773072] env[61806]: value = "task-1294954" [ 949.773072] env[61806]: _type = "Task" [ 949.773072] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 949.782258] env[61806]: DEBUG oslo_vmware.api [None req-9cd2108a-a1e5-403d-bbda-81099a62d59d tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] Task: {'id': task-1294954, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 949.901329] env[61806]: DEBUG oslo_vmware.api [None req-61d33433-c3e5-4624-95f4-86df358dffc7 tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] Task: {'id': task-1294951, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 950.116342] env[61806]: DEBUG oslo_concurrency.lockutils [None req-5d1f922c-8037-4fb1-9c2a-dcd72461a463 tempest-ServersNegativeTestJSON-891856253 tempest-ServersNegativeTestJSON-891856253-project-member] Acquiring lock "refresh_cache-bdbaef31-7d37-40cb-ac3e-ad152b37e2d2" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 950.116580] env[61806]: DEBUG oslo_concurrency.lockutils [None req-5d1f922c-8037-4fb1-9c2a-dcd72461a463 tempest-ServersNegativeTestJSON-891856253 tempest-ServersNegativeTestJSON-891856253-project-member] Acquired lock "refresh_cache-bdbaef31-7d37-40cb-ac3e-ad152b37e2d2" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 950.116745] env[61806]: DEBUG nova.network.neutron [None req-5d1f922c-8037-4fb1-9c2a-dcd72461a463 tempest-ServersNegativeTestJSON-891856253 tempest-ServersNegativeTestJSON-891856253-project-member] [instance: bdbaef31-7d37-40cb-ac3e-ad152b37e2d2] Building network info cache for instance {{(pid=61806) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 950.259118] env[61806]: DEBUG oslo_vmware.api [None req-cf5202db-26c0-4960-9218-bd95b8e81f62 tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Task: {'id': task-1294953, 'name': PowerOffVM_Task, 'duration_secs': 0.250768} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 950.259443] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [None req-cf5202db-26c0-4960-9218-bd95b8e81f62 tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] [instance: 676f18b3-3da1-4a53-b96d-fe64ee8f9101] Powered off the VM {{(pid=61806) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 950.259610] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-cf5202db-26c0-4960-9218-bd95b8e81f62 tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] [instance: 676f18b3-3da1-4a53-b96d-fe64ee8f9101] Updating instance '676f18b3-3da1-4a53-b96d-fe64ee8f9101' progress to 17 {{(pid=61806) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 950.282242] env[61806]: DEBUG oslo_vmware.api [None req-9cd2108a-a1e5-403d-bbda-81099a62d59d tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] Task: {'id': task-1294954, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.069538} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 950.282532] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-9cd2108a-a1e5-403d-bbda-81099a62d59d tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] [instance: 38e96e22-4200-4175-9085-31977357da2d] Extended root virtual disk {{(pid=61806) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 950.283973] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-25bdba22-29bf-4676-9c43-8350912a3434 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 950.306065] env[61806]: DEBUG nova.virt.vmwareapi.volumeops [None req-9cd2108a-a1e5-403d-bbda-81099a62d59d tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] [instance: 38e96e22-4200-4175-9085-31977357da2d] Reconfiguring VM instance instance-00000059 to attach disk [datastore1] 38e96e22-4200-4175-9085-31977357da2d/38e96e22-4200-4175-9085-31977357da2d.vmdk or device None with type sparse {{(pid=61806) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 950.307127] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-0932787c-aa2f-4451-93e3-297cd1f5e47c {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 950.326954] env[61806]: DEBUG oslo_vmware.api [None req-9cd2108a-a1e5-403d-bbda-81099a62d59d tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] Waiting for the task: (returnval){ [ 950.326954] env[61806]: value = "task-1294955" [ 950.326954] env[61806]: _type = "Task" [ 950.326954] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 950.335285] env[61806]: DEBUG oslo_vmware.api [None req-9cd2108a-a1e5-403d-bbda-81099a62d59d tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] Task: {'id': task-1294955, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 950.372033] env[61806]: DEBUG oslo_concurrency.lockutils [None req-0efd4a4f-9a56-4502-a76c-4d8b926237a6 tempest-AttachVolumeNegativeTest-1683033704 tempest-AttachVolumeNegativeTest-1683033704-project-member] Acquiring lock "43dcce0d-4fdb-4ee1-8309-92e92d2331a9" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 950.372259] env[61806]: DEBUG oslo_concurrency.lockutils [None req-0efd4a4f-9a56-4502-a76c-4d8b926237a6 tempest-AttachVolumeNegativeTest-1683033704 tempest-AttachVolumeNegativeTest-1683033704-project-member] Lock "43dcce0d-4fdb-4ee1-8309-92e92d2331a9" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 950.372484] env[61806]: DEBUG oslo_concurrency.lockutils [None req-0efd4a4f-9a56-4502-a76c-4d8b926237a6 tempest-AttachVolumeNegativeTest-1683033704 tempest-AttachVolumeNegativeTest-1683033704-project-member] Acquiring lock "43dcce0d-4fdb-4ee1-8309-92e92d2331a9-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 950.372671] env[61806]: DEBUG oslo_concurrency.lockutils [None req-0efd4a4f-9a56-4502-a76c-4d8b926237a6 tempest-AttachVolumeNegativeTest-1683033704 tempest-AttachVolumeNegativeTest-1683033704-project-member] Lock "43dcce0d-4fdb-4ee1-8309-92e92d2331a9-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 950.372846] env[61806]: DEBUG oslo_concurrency.lockutils [None req-0efd4a4f-9a56-4502-a76c-4d8b926237a6 tempest-AttachVolumeNegativeTest-1683033704 tempest-AttachVolumeNegativeTest-1683033704-project-member] Lock "43dcce0d-4fdb-4ee1-8309-92e92d2331a9-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 950.375247] env[61806]: INFO nova.compute.manager [None req-0efd4a4f-9a56-4502-a76c-4d8b926237a6 tempest-AttachVolumeNegativeTest-1683033704 tempest-AttachVolumeNegativeTest-1683033704-project-member] [instance: 43dcce0d-4fdb-4ee1-8309-92e92d2331a9] Terminating instance [ 950.377087] env[61806]: DEBUG nova.compute.manager [None req-0efd4a4f-9a56-4502-a76c-4d8b926237a6 tempest-AttachVolumeNegativeTest-1683033704 tempest-AttachVolumeNegativeTest-1683033704-project-member] [instance: 43dcce0d-4fdb-4ee1-8309-92e92d2331a9] Start destroying the instance on the hypervisor. {{(pid=61806) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 950.377291] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-0efd4a4f-9a56-4502-a76c-4d8b926237a6 tempest-AttachVolumeNegativeTest-1683033704 tempest-AttachVolumeNegativeTest-1683033704-project-member] [instance: 43dcce0d-4fdb-4ee1-8309-92e92d2331a9] Destroying instance {{(pid=61806) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 950.378126] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b381cc56-531b-444d-8fa8-90b4c12a2031 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 950.386141] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [None req-0efd4a4f-9a56-4502-a76c-4d8b926237a6 tempest-AttachVolumeNegativeTest-1683033704 tempest-AttachVolumeNegativeTest-1683033704-project-member] [instance: 43dcce0d-4fdb-4ee1-8309-92e92d2331a9] Powering off the VM {{(pid=61806) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 950.386405] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-fffff6d1-f43f-41cf-ab94-5566379294e9 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 950.392325] env[61806]: DEBUG oslo_vmware.api [None req-0efd4a4f-9a56-4502-a76c-4d8b926237a6 tempest-AttachVolumeNegativeTest-1683033704 tempest-AttachVolumeNegativeTest-1683033704-project-member] Waiting for the task: (returnval){ [ 950.392325] env[61806]: value = "task-1294956" [ 950.392325] env[61806]: _type = "Task" [ 950.392325] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 950.403191] env[61806]: DEBUG oslo_vmware.api [None req-61d33433-c3e5-4624-95f4-86df358dffc7 tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] Task: {'id': task-1294951, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 950.406342] env[61806]: DEBUG oslo_vmware.api [None req-0efd4a4f-9a56-4502-a76c-4d8b926237a6 tempest-AttachVolumeNegativeTest-1683033704 tempest-AttachVolumeNegativeTest-1683033704-project-member] Task: {'id': task-1294956, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 950.651593] env[61806]: DEBUG nova.network.neutron [None req-5d1f922c-8037-4fb1-9c2a-dcd72461a463 tempest-ServersNegativeTestJSON-891856253 tempest-ServersNegativeTestJSON-891856253-project-member] [instance: bdbaef31-7d37-40cb-ac3e-ad152b37e2d2] Instance cache missing network info. {{(pid=61806) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 950.766331] env[61806]: DEBUG nova.virt.hardware [None req-cf5202db-26c0-4960-9218-bd95b8e81f62 tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-15T18:05:12Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='84',id=12,is_public=True,memory_mb=256,name='m1.micro',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=61806) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 950.766639] env[61806]: DEBUG nova.virt.hardware [None req-cf5202db-26c0-4960-9218-bd95b8e81f62 tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Flavor limits 0:0:0 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 950.766819] env[61806]: DEBUG nova.virt.hardware [None req-cf5202db-26c0-4960-9218-bd95b8e81f62 tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Image limits 0:0:0 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 950.767036] env[61806]: DEBUG nova.virt.hardware [None req-cf5202db-26c0-4960-9218-bd95b8e81f62 tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Flavor pref 0:0:0 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 950.767200] env[61806]: DEBUG nova.virt.hardware [None req-cf5202db-26c0-4960-9218-bd95b8e81f62 tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Image pref 0:0:0 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 950.767357] env[61806]: DEBUG nova.virt.hardware [None req-cf5202db-26c0-4960-9218-bd95b8e81f62 tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 950.767605] env[61806]: DEBUG nova.virt.hardware [None req-cf5202db-26c0-4960-9218-bd95b8e81f62 tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61806) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 950.767775] env[61806]: DEBUG nova.virt.hardware [None req-cf5202db-26c0-4960-9218-bd95b8e81f62 tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61806) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 950.767951] env[61806]: DEBUG nova.virt.hardware [None req-cf5202db-26c0-4960-9218-bd95b8e81f62 tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Got 1 possible topologies {{(pid=61806) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 950.768139] env[61806]: DEBUG nova.virt.hardware [None req-cf5202db-26c0-4960-9218-bd95b8e81f62 tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61806) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 950.768323] env[61806]: DEBUG nova.virt.hardware [None req-cf5202db-26c0-4960-9218-bd95b8e81f62 tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61806) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 950.773877] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-7e7afd33-d571-401b-bbb0-a81045fd2b12 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 950.791693] env[61806]: DEBUG oslo_vmware.api [None req-cf5202db-26c0-4960-9218-bd95b8e81f62 tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Waiting for the task: (returnval){ [ 950.791693] env[61806]: value = "task-1294957" [ 950.791693] env[61806]: _type = "Task" [ 950.791693] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 950.801803] env[61806]: DEBUG oslo_vmware.api [None req-cf5202db-26c0-4960-9218-bd95b8e81f62 tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Task: {'id': task-1294957, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 950.843556] env[61806]: DEBUG oslo_vmware.api [None req-9cd2108a-a1e5-403d-bbda-81099a62d59d tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] Task: {'id': task-1294955, 'name': ReconfigVM_Task, 'duration_secs': 0.360657} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 950.845381] env[61806]: DEBUG nova.network.neutron [None req-5d1f922c-8037-4fb1-9c2a-dcd72461a463 tempest-ServersNegativeTestJSON-891856253 tempest-ServersNegativeTestJSON-891856253-project-member] [instance: bdbaef31-7d37-40cb-ac3e-ad152b37e2d2] Updating instance_info_cache with network_info: [{"id": "b8b714a4-b17c-4b3d-889e-407bec6c30ae", "address": "fa:16:3e:33:74:bb", "network": {"id": "fd35ea93-8914-4b54-a4ea-f45fc750982b", "bridge": "br-int", "label": "tempest-ServersNegativeTestJSON-2117068882-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d2e2b7eec47e4b10898215123408692a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "2d94740a-bce8-4103-8ecf-230d02ec0a44", "external-id": "nsx-vlan-transportzone-149", "segmentation_id": 149, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb8b714a4-b1", "ovs_interfaceid": "b8b714a4-b17c-4b3d-889e-407bec6c30ae", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61806) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 950.847405] env[61806]: DEBUG nova.virt.vmwareapi.volumeops [None req-9cd2108a-a1e5-403d-bbda-81099a62d59d tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] [instance: 38e96e22-4200-4175-9085-31977357da2d] Reconfigured VM instance instance-00000059 to attach disk [datastore1] 38e96e22-4200-4175-9085-31977357da2d/38e96e22-4200-4175-9085-31977357da2d.vmdk or device None with type sparse {{(pid=61806) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 950.848947] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-8b6dc6f8-e1dc-4cb8-bd79-d7ea7990c2d5 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 950.856945] env[61806]: DEBUG oslo_vmware.api [None req-9cd2108a-a1e5-403d-bbda-81099a62d59d tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] Waiting for the task: (returnval){ [ 950.856945] env[61806]: value = "task-1294958" [ 950.856945] env[61806]: _type = "Task" [ 950.856945] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 950.869107] env[61806]: DEBUG oslo_vmware.api [None req-9cd2108a-a1e5-403d-bbda-81099a62d59d tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] Task: {'id': task-1294958, 'name': Rename_Task} progress is 5%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 950.908199] env[61806]: DEBUG oslo_vmware.api [None req-61d33433-c3e5-4624-95f4-86df358dffc7 tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] Task: {'id': task-1294951, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 950.911331] env[61806]: DEBUG oslo_vmware.api [None req-0efd4a4f-9a56-4502-a76c-4d8b926237a6 tempest-AttachVolumeNegativeTest-1683033704 tempest-AttachVolumeNegativeTest-1683033704-project-member] Task: {'id': task-1294956, 'name': PowerOffVM_Task, 'duration_secs': 0.189805} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 950.911595] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [None req-0efd4a4f-9a56-4502-a76c-4d8b926237a6 tempest-AttachVolumeNegativeTest-1683033704 tempest-AttachVolumeNegativeTest-1683033704-project-member] [instance: 43dcce0d-4fdb-4ee1-8309-92e92d2331a9] Powered off the VM {{(pid=61806) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 950.911783] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-0efd4a4f-9a56-4502-a76c-4d8b926237a6 tempest-AttachVolumeNegativeTest-1683033704 tempest-AttachVolumeNegativeTest-1683033704-project-member] [instance: 43dcce0d-4fdb-4ee1-8309-92e92d2331a9] Unregistering the VM {{(pid=61806) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 950.912111] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-e7422ea8-9ff7-44dd-9d19-a300b97352c2 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 950.983040] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-0efd4a4f-9a56-4502-a76c-4d8b926237a6 tempest-AttachVolumeNegativeTest-1683033704 tempest-AttachVolumeNegativeTest-1683033704-project-member] [instance: 43dcce0d-4fdb-4ee1-8309-92e92d2331a9] Unregistered the VM {{(pid=61806) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 950.983237] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-0efd4a4f-9a56-4502-a76c-4d8b926237a6 tempest-AttachVolumeNegativeTest-1683033704 tempest-AttachVolumeNegativeTest-1683033704-project-member] [instance: 43dcce0d-4fdb-4ee1-8309-92e92d2331a9] Deleting contents of the VM from datastore datastore1 {{(pid=61806) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 950.983455] env[61806]: DEBUG nova.virt.vmwareapi.ds_util [None req-0efd4a4f-9a56-4502-a76c-4d8b926237a6 tempest-AttachVolumeNegativeTest-1683033704 tempest-AttachVolumeNegativeTest-1683033704-project-member] Deleting the datastore file [datastore1] 43dcce0d-4fdb-4ee1-8309-92e92d2331a9 {{(pid=61806) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 950.983806] env[61806]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-39e2b398-7953-4453-8b81-a6e4fd8c02f3 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 950.990925] env[61806]: DEBUG oslo_vmware.api [None req-0efd4a4f-9a56-4502-a76c-4d8b926237a6 tempest-AttachVolumeNegativeTest-1683033704 tempest-AttachVolumeNegativeTest-1683033704-project-member] Waiting for the task: (returnval){ [ 950.990925] env[61806]: value = "task-1294960" [ 950.990925] env[61806]: _type = "Task" [ 950.990925] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 950.999107] env[61806]: DEBUG oslo_vmware.api [None req-0efd4a4f-9a56-4502-a76c-4d8b926237a6 tempest-AttachVolumeNegativeTest-1683033704 tempest-AttachVolumeNegativeTest-1683033704-project-member] Task: {'id': task-1294960, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 951.033129] env[61806]: DEBUG oslo_vmware.rw_handles [None req-03f49469-0c97-4bf3-b87e-03f023c054e5 tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Getting lease state for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52ba91b7-8fef-044e-b400-61ffb14f4c5e/disk-0.vmdk. {{(pid=61806) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 951.034241] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dd2a6dab-9fd2-45dc-8151-3220f4eba89d {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 951.041128] env[61806]: DEBUG oslo_vmware.rw_handles [None req-03f49469-0c97-4bf3-b87e-03f023c054e5 tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Lease for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52ba91b7-8fef-044e-b400-61ffb14f4c5e/disk-0.vmdk is in state: ready. {{(pid=61806) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 951.041322] env[61806]: ERROR oslo_vmware.rw_handles [None req-03f49469-0c97-4bf3-b87e-03f023c054e5 tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Aborting lease for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52ba91b7-8fef-044e-b400-61ffb14f4c5e/disk-0.vmdk due to incomplete transfer. [ 951.041568] env[61806]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseAbort with opID=oslo.vmware-7f065c0e-5adf-4d46-8d66-190e39d9b530 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 951.048953] env[61806]: DEBUG oslo_vmware.rw_handles [None req-03f49469-0c97-4bf3-b87e-03f023c054e5 tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Closed VMDK read handle for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52ba91b7-8fef-044e-b400-61ffb14f4c5e/disk-0.vmdk. {{(pid=61806) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:735}} [ 951.049180] env[61806]: DEBUG nova.virt.vmwareapi.images [None req-03f49469-0c97-4bf3-b87e-03f023c054e5 tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] [instance: 380aa63f-46ba-437d-bc29-0fc141ad1cf3] Uploaded image 3047e10f-a04a-4487-8696-28cac698ab9e to the Glance image server {{(pid=61806) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:473}} [ 951.051627] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [None req-03f49469-0c97-4bf3-b87e-03f023c054e5 tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] [instance: 380aa63f-46ba-437d-bc29-0fc141ad1cf3] Destroying the VM {{(pid=61806) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1358}} [ 951.051902] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-4294d406-3bee-4612-b5e3-4b0938ac1e82 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 951.057841] env[61806]: DEBUG oslo_vmware.api [None req-03f49469-0c97-4bf3-b87e-03f023c054e5 tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Waiting for the task: (returnval){ [ 951.057841] env[61806]: value = "task-1294961" [ 951.057841] env[61806]: _type = "Task" [ 951.057841] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 951.065780] env[61806]: DEBUG oslo_vmware.api [None req-03f49469-0c97-4bf3-b87e-03f023c054e5 tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Task: {'id': task-1294961, 'name': Destroy_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 951.303781] env[61806]: DEBUG oslo_vmware.api [None req-cf5202db-26c0-4960-9218-bd95b8e81f62 tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Task: {'id': task-1294957, 'name': ReconfigVM_Task, 'duration_secs': 0.28228} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 951.304167] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-cf5202db-26c0-4960-9218-bd95b8e81f62 tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] [instance: 676f18b3-3da1-4a53-b96d-fe64ee8f9101] Updating instance '676f18b3-3da1-4a53-b96d-fe64ee8f9101' progress to 33 {{(pid=61806) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 951.349469] env[61806]: DEBUG oslo_concurrency.lockutils [None req-5d1f922c-8037-4fb1-9c2a-dcd72461a463 tempest-ServersNegativeTestJSON-891856253 tempest-ServersNegativeTestJSON-891856253-project-member] Releasing lock "refresh_cache-bdbaef31-7d37-40cb-ac3e-ad152b37e2d2" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 951.349768] env[61806]: DEBUG nova.compute.manager [None req-5d1f922c-8037-4fb1-9c2a-dcd72461a463 tempest-ServersNegativeTestJSON-891856253 tempest-ServersNegativeTestJSON-891856253-project-member] [instance: bdbaef31-7d37-40cb-ac3e-ad152b37e2d2] Instance network_info: |[{"id": "b8b714a4-b17c-4b3d-889e-407bec6c30ae", "address": "fa:16:3e:33:74:bb", "network": {"id": "fd35ea93-8914-4b54-a4ea-f45fc750982b", "bridge": "br-int", "label": "tempest-ServersNegativeTestJSON-2117068882-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d2e2b7eec47e4b10898215123408692a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "2d94740a-bce8-4103-8ecf-230d02ec0a44", "external-id": "nsx-vlan-transportzone-149", "segmentation_id": 149, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb8b714a4-b1", "ovs_interfaceid": "b8b714a4-b17c-4b3d-889e-407bec6c30ae", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61806) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 951.350340] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-5d1f922c-8037-4fb1-9c2a-dcd72461a463 tempest-ServersNegativeTestJSON-891856253 tempest-ServersNegativeTestJSON-891856253-project-member] [instance: bdbaef31-7d37-40cb-ac3e-ad152b37e2d2] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:33:74:bb', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '2d94740a-bce8-4103-8ecf-230d02ec0a44', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'b8b714a4-b17c-4b3d-889e-407bec6c30ae', 'vif_model': 'vmxnet3'}] {{(pid=61806) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 951.358396] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [None req-5d1f922c-8037-4fb1-9c2a-dcd72461a463 tempest-ServersNegativeTestJSON-891856253 tempest-ServersNegativeTestJSON-891856253-project-member] Creating folder: Project (d2e2b7eec47e4b10898215123408692a). Parent ref: group-v277609. {{(pid=61806) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 951.358950] env[61806]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-6de30727-7792-4c08-8912-84d6b05b53c8 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 951.369835] env[61806]: DEBUG oslo_vmware.api [None req-9cd2108a-a1e5-403d-bbda-81099a62d59d tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] Task: {'id': task-1294958, 'name': Rename_Task, 'duration_secs': 0.17917} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 951.370108] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [None req-9cd2108a-a1e5-403d-bbda-81099a62d59d tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] [instance: 38e96e22-4200-4175-9085-31977357da2d] Powering on the VM {{(pid=61806) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 951.371400] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-2e59a4d8-0b31-4272-a598-9cba2541e8dd {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 951.372923] env[61806]: INFO nova.virt.vmwareapi.vm_util [None req-5d1f922c-8037-4fb1-9c2a-dcd72461a463 tempest-ServersNegativeTestJSON-891856253 tempest-ServersNegativeTestJSON-891856253-project-member] Created folder: Project (d2e2b7eec47e4b10898215123408692a) in parent group-v277609. [ 951.373149] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [None req-5d1f922c-8037-4fb1-9c2a-dcd72461a463 tempest-ServersNegativeTestJSON-891856253 tempest-ServersNegativeTestJSON-891856253-project-member] Creating folder: Instances. Parent ref: group-v277724. {{(pid=61806) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 951.373377] env[61806]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-952f24fd-62b4-4330-9d95-0ae7bfee1f0f {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 951.379606] env[61806]: DEBUG oslo_vmware.api [None req-9cd2108a-a1e5-403d-bbda-81099a62d59d tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] Waiting for the task: (returnval){ [ 951.379606] env[61806]: value = "task-1294963" [ 951.379606] env[61806]: _type = "Task" [ 951.379606] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 951.387870] env[61806]: DEBUG oslo_vmware.api [None req-9cd2108a-a1e5-403d-bbda-81099a62d59d tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] Task: {'id': task-1294963, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 951.389042] env[61806]: INFO nova.virt.vmwareapi.vm_util [None req-5d1f922c-8037-4fb1-9c2a-dcd72461a463 tempest-ServersNegativeTestJSON-891856253 tempest-ServersNegativeTestJSON-891856253-project-member] Created folder: Instances in parent group-v277724. [ 951.389283] env[61806]: DEBUG oslo.service.loopingcall [None req-5d1f922c-8037-4fb1-9c2a-dcd72461a463 tempest-ServersNegativeTestJSON-891856253 tempest-ServersNegativeTestJSON-891856253-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61806) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 951.389468] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: bdbaef31-7d37-40cb-ac3e-ad152b37e2d2] Creating VM on the ESX host {{(pid=61806) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 951.389688] env[61806]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-401599fd-c0f3-4378-9518-c6202e2057c9 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 951.412153] env[61806]: DEBUG oslo_vmware.api [None req-61d33433-c3e5-4624-95f4-86df358dffc7 tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] Task: {'id': task-1294951, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 951.413396] env[61806]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 951.413396] env[61806]: value = "task-1294965" [ 951.413396] env[61806]: _type = "Task" [ 951.413396] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 951.421363] env[61806]: DEBUG oslo_vmware.api [-] Task: {'id': task-1294965, 'name': CreateVM_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 951.484771] env[61806]: DEBUG nova.compute.manager [req-5f91f4b9-4bd2-4135-abdc-d666a8f033ef req-eac5152f-0447-4a35-8775-899cdf2cec75 service nova] [instance: bdbaef31-7d37-40cb-ac3e-ad152b37e2d2] Received event network-changed-b8b714a4-b17c-4b3d-889e-407bec6c30ae {{(pid=61806) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 951.484998] env[61806]: DEBUG nova.compute.manager [req-5f91f4b9-4bd2-4135-abdc-d666a8f033ef req-eac5152f-0447-4a35-8775-899cdf2cec75 service nova] [instance: bdbaef31-7d37-40cb-ac3e-ad152b37e2d2] Refreshing instance network info cache due to event network-changed-b8b714a4-b17c-4b3d-889e-407bec6c30ae. {{(pid=61806) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 951.485293] env[61806]: DEBUG oslo_concurrency.lockutils [req-5f91f4b9-4bd2-4135-abdc-d666a8f033ef req-eac5152f-0447-4a35-8775-899cdf2cec75 service nova] Acquiring lock "refresh_cache-bdbaef31-7d37-40cb-ac3e-ad152b37e2d2" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 951.485451] env[61806]: DEBUG oslo_concurrency.lockutils [req-5f91f4b9-4bd2-4135-abdc-d666a8f033ef req-eac5152f-0447-4a35-8775-899cdf2cec75 service nova] Acquired lock "refresh_cache-bdbaef31-7d37-40cb-ac3e-ad152b37e2d2" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 951.485641] env[61806]: DEBUG nova.network.neutron [req-5f91f4b9-4bd2-4135-abdc-d666a8f033ef req-eac5152f-0447-4a35-8775-899cdf2cec75 service nova] [instance: bdbaef31-7d37-40cb-ac3e-ad152b37e2d2] Refreshing network info cache for port b8b714a4-b17c-4b3d-889e-407bec6c30ae {{(pid=61806) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 951.502483] env[61806]: DEBUG oslo_vmware.api [None req-0efd4a4f-9a56-4502-a76c-4d8b926237a6 tempest-AttachVolumeNegativeTest-1683033704 tempest-AttachVolumeNegativeTest-1683033704-project-member] Task: {'id': task-1294960, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.1639} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 951.502755] env[61806]: DEBUG nova.virt.vmwareapi.ds_util [None req-0efd4a4f-9a56-4502-a76c-4d8b926237a6 tempest-AttachVolumeNegativeTest-1683033704 tempest-AttachVolumeNegativeTest-1683033704-project-member] Deleted the datastore file {{(pid=61806) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 951.502947] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-0efd4a4f-9a56-4502-a76c-4d8b926237a6 tempest-AttachVolumeNegativeTest-1683033704 tempest-AttachVolumeNegativeTest-1683033704-project-member] [instance: 43dcce0d-4fdb-4ee1-8309-92e92d2331a9] Deleted contents of the VM from datastore datastore1 {{(pid=61806) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 951.503150] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-0efd4a4f-9a56-4502-a76c-4d8b926237a6 tempest-AttachVolumeNegativeTest-1683033704 tempest-AttachVolumeNegativeTest-1683033704-project-member] [instance: 43dcce0d-4fdb-4ee1-8309-92e92d2331a9] Instance destroyed {{(pid=61806) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 951.503350] env[61806]: INFO nova.compute.manager [None req-0efd4a4f-9a56-4502-a76c-4d8b926237a6 tempest-AttachVolumeNegativeTest-1683033704 tempest-AttachVolumeNegativeTest-1683033704-project-member] [instance: 43dcce0d-4fdb-4ee1-8309-92e92d2331a9] Took 1.13 seconds to destroy the instance on the hypervisor. [ 951.503597] env[61806]: DEBUG oslo.service.loopingcall [None req-0efd4a4f-9a56-4502-a76c-4d8b926237a6 tempest-AttachVolumeNegativeTest-1683033704 tempest-AttachVolumeNegativeTest-1683033704-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61806) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 951.503789] env[61806]: DEBUG nova.compute.manager [-] [instance: 43dcce0d-4fdb-4ee1-8309-92e92d2331a9] Deallocating network for instance {{(pid=61806) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 951.503880] env[61806]: DEBUG nova.network.neutron [-] [instance: 43dcce0d-4fdb-4ee1-8309-92e92d2331a9] deallocate_for_instance() {{(pid=61806) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 951.567897] env[61806]: DEBUG oslo_vmware.api [None req-03f49469-0c97-4bf3-b87e-03f023c054e5 tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Task: {'id': task-1294961, 'name': Destroy_Task} progress is 33%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 951.811043] env[61806]: DEBUG nova.virt.hardware [None req-cf5202db-26c0-4960-9218-bd95b8e81f62 tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-15T18:05:10Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=61806) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 951.811399] env[61806]: DEBUG nova.virt.hardware [None req-cf5202db-26c0-4960-9218-bd95b8e81f62 tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Flavor limits 0:0:0 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 951.811653] env[61806]: DEBUG nova.virt.hardware [None req-cf5202db-26c0-4960-9218-bd95b8e81f62 tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Image limits 0:0:0 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 951.811927] env[61806]: DEBUG nova.virt.hardware [None req-cf5202db-26c0-4960-9218-bd95b8e81f62 tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Flavor pref 0:0:0 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 951.812187] env[61806]: DEBUG nova.virt.hardware [None req-cf5202db-26c0-4960-9218-bd95b8e81f62 tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Image pref 0:0:0 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 951.812362] env[61806]: DEBUG nova.virt.hardware [None req-cf5202db-26c0-4960-9218-bd95b8e81f62 tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 951.812580] env[61806]: DEBUG nova.virt.hardware [None req-cf5202db-26c0-4960-9218-bd95b8e81f62 tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61806) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 951.812959] env[61806]: DEBUG nova.virt.hardware [None req-cf5202db-26c0-4960-9218-bd95b8e81f62 tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61806) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 951.813128] env[61806]: DEBUG nova.virt.hardware [None req-cf5202db-26c0-4960-9218-bd95b8e81f62 tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Got 1 possible topologies {{(pid=61806) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 951.813196] env[61806]: DEBUG nova.virt.hardware [None req-cf5202db-26c0-4960-9218-bd95b8e81f62 tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61806) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 951.813384] env[61806]: DEBUG nova.virt.hardware [None req-cf5202db-26c0-4960-9218-bd95b8e81f62 tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61806) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 951.819289] env[61806]: DEBUG nova.virt.vmwareapi.volumeops [None req-cf5202db-26c0-4960-9218-bd95b8e81f62 tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] [instance: 676f18b3-3da1-4a53-b96d-fe64ee8f9101] Reconfiguring VM instance instance-00000058 to detach disk 2000 {{(pid=61806) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 951.819586] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-e4e88fa8-6460-469f-8230-9661fbf40b6b {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 951.839397] env[61806]: DEBUG oslo_vmware.api [None req-cf5202db-26c0-4960-9218-bd95b8e81f62 tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Waiting for the task: (returnval){ [ 951.839397] env[61806]: value = "task-1294966" [ 951.839397] env[61806]: _type = "Task" [ 951.839397] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 951.848388] env[61806]: DEBUG oslo_vmware.api [None req-cf5202db-26c0-4960-9218-bd95b8e81f62 tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Task: {'id': task-1294966, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 951.889903] env[61806]: DEBUG oslo_vmware.api [None req-9cd2108a-a1e5-403d-bbda-81099a62d59d tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] Task: {'id': task-1294963, 'name': PowerOnVM_Task} progress is 87%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 951.914072] env[61806]: DEBUG oslo_vmware.api [None req-61d33433-c3e5-4624-95f4-86df358dffc7 tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] Task: {'id': task-1294951, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 951.924489] env[61806]: DEBUG oslo_vmware.api [-] Task: {'id': task-1294965, 'name': CreateVM_Task} progress is 99%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 952.069689] env[61806]: DEBUG oslo_vmware.api [None req-03f49469-0c97-4bf3-b87e-03f023c054e5 tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Task: {'id': task-1294961, 'name': Destroy_Task, 'duration_secs': 0.570057} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 952.070419] env[61806]: INFO nova.virt.vmwareapi.vm_util [None req-03f49469-0c97-4bf3-b87e-03f023c054e5 tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] [instance: 380aa63f-46ba-437d-bc29-0fc141ad1cf3] Destroyed the VM [ 952.070419] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-03f49469-0c97-4bf3-b87e-03f023c054e5 tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] [instance: 380aa63f-46ba-437d-bc29-0fc141ad1cf3] Deleting Snapshot of the VM instance {{(pid=61806) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 952.070640] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-2457d522-1d1e-45a5-9c51-ec5c23f1657a {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 952.081407] env[61806]: DEBUG oslo_vmware.api [None req-03f49469-0c97-4bf3-b87e-03f023c054e5 tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Waiting for the task: (returnval){ [ 952.081407] env[61806]: value = "task-1294967" [ 952.081407] env[61806]: _type = "Task" [ 952.081407] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 952.089624] env[61806]: DEBUG oslo_vmware.api [None req-03f49469-0c97-4bf3-b87e-03f023c054e5 tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Task: {'id': task-1294967, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 952.260853] env[61806]: DEBUG nova.network.neutron [req-5f91f4b9-4bd2-4135-abdc-d666a8f033ef req-eac5152f-0447-4a35-8775-899cdf2cec75 service nova] [instance: bdbaef31-7d37-40cb-ac3e-ad152b37e2d2] Updated VIF entry in instance network info cache for port b8b714a4-b17c-4b3d-889e-407bec6c30ae. {{(pid=61806) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 952.261784] env[61806]: DEBUG nova.network.neutron [req-5f91f4b9-4bd2-4135-abdc-d666a8f033ef req-eac5152f-0447-4a35-8775-899cdf2cec75 service nova] [instance: bdbaef31-7d37-40cb-ac3e-ad152b37e2d2] Updating instance_info_cache with network_info: [{"id": "b8b714a4-b17c-4b3d-889e-407bec6c30ae", "address": "fa:16:3e:33:74:bb", "network": {"id": "fd35ea93-8914-4b54-a4ea-f45fc750982b", "bridge": "br-int", "label": "tempest-ServersNegativeTestJSON-2117068882-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d2e2b7eec47e4b10898215123408692a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "2d94740a-bce8-4103-8ecf-230d02ec0a44", "external-id": "nsx-vlan-transportzone-149", "segmentation_id": 149, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb8b714a4-b1", "ovs_interfaceid": "b8b714a4-b17c-4b3d-889e-407bec6c30ae", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61806) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 952.349982] env[61806]: DEBUG oslo_vmware.api [None req-cf5202db-26c0-4960-9218-bd95b8e81f62 tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Task: {'id': task-1294966, 'name': ReconfigVM_Task, 'duration_secs': 0.176163} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 952.350316] env[61806]: DEBUG nova.virt.vmwareapi.volumeops [None req-cf5202db-26c0-4960-9218-bd95b8e81f62 tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] [instance: 676f18b3-3da1-4a53-b96d-fe64ee8f9101] Reconfigured VM instance instance-00000058 to detach disk 2000 {{(pid=61806) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 952.351126] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b9f6c026-d2c2-4952-aa77-ef6adbf3c3b3 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 952.373892] env[61806]: DEBUG nova.virt.vmwareapi.volumeops [None req-cf5202db-26c0-4960-9218-bd95b8e81f62 tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] [instance: 676f18b3-3da1-4a53-b96d-fe64ee8f9101] Reconfiguring VM instance instance-00000058 to attach disk [datastore2] 676f18b3-3da1-4a53-b96d-fe64ee8f9101/676f18b3-3da1-4a53-b96d-fe64ee8f9101.vmdk or device None with type thin {{(pid=61806) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 952.374224] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-eb01b725-a162-4e78-a606-f14c084925db {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 952.396308] env[61806]: DEBUG oslo_vmware.api [None req-9cd2108a-a1e5-403d-bbda-81099a62d59d tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] Task: {'id': task-1294963, 'name': PowerOnVM_Task, 'duration_secs': 0.854936} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 952.397646] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [None req-9cd2108a-a1e5-403d-bbda-81099a62d59d tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] [instance: 38e96e22-4200-4175-9085-31977357da2d] Powered on the VM {{(pid=61806) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 952.397863] env[61806]: INFO nova.compute.manager [None req-9cd2108a-a1e5-403d-bbda-81099a62d59d tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] [instance: 38e96e22-4200-4175-9085-31977357da2d] Took 9.94 seconds to spawn the instance on the hypervisor. [ 952.398126] env[61806]: DEBUG nova.compute.manager [None req-9cd2108a-a1e5-403d-bbda-81099a62d59d tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] [instance: 38e96e22-4200-4175-9085-31977357da2d] Checking state {{(pid=61806) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 952.398518] env[61806]: DEBUG oslo_vmware.api [None req-cf5202db-26c0-4960-9218-bd95b8e81f62 tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Waiting for the task: (returnval){ [ 952.398518] env[61806]: value = "task-1294968" [ 952.398518] env[61806]: _type = "Task" [ 952.398518] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 952.399260] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e2c3ce74-65af-48d9-b5b6-2614de5d004f {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 952.417519] env[61806]: DEBUG oslo_vmware.api [None req-cf5202db-26c0-4960-9218-bd95b8e81f62 tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Task: {'id': task-1294968, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 952.425953] env[61806]: DEBUG oslo_vmware.api [None req-61d33433-c3e5-4624-95f4-86df358dffc7 tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] Task: {'id': task-1294951, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 952.431797] env[61806]: DEBUG oslo_vmware.api [-] Task: {'id': task-1294965, 'name': CreateVM_Task, 'duration_secs': 0.618778} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 952.431873] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: bdbaef31-7d37-40cb-ac3e-ad152b37e2d2] Created VM on the ESX host {{(pid=61806) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 952.432595] env[61806]: DEBUG oslo_concurrency.lockutils [None req-5d1f922c-8037-4fb1-9c2a-dcd72461a463 tempest-ServersNegativeTestJSON-891856253 tempest-ServersNegativeTestJSON-891856253-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 952.432769] env[61806]: DEBUG oslo_concurrency.lockutils [None req-5d1f922c-8037-4fb1-9c2a-dcd72461a463 tempest-ServersNegativeTestJSON-891856253 tempest-ServersNegativeTestJSON-891856253-project-member] Acquired lock "[datastore1] devstack-image-cache_base/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 952.433112] env[61806]: DEBUG oslo_concurrency.lockutils [None req-5d1f922c-8037-4fb1-9c2a-dcd72461a463 tempest-ServersNegativeTestJSON-891856253 tempest-ServersNegativeTestJSON-891856253-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 952.433410] env[61806]: DEBUG nova.network.neutron [-] [instance: 43dcce0d-4fdb-4ee1-8309-92e92d2331a9] Updating instance_info_cache with network_info: [] {{(pid=61806) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 952.434786] env[61806]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-dde77ea8-3307-4f9c-8a29-1facb4474c93 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 952.442403] env[61806]: DEBUG oslo_vmware.api [None req-5d1f922c-8037-4fb1-9c2a-dcd72461a463 tempest-ServersNegativeTestJSON-891856253 tempest-ServersNegativeTestJSON-891856253-project-member] Waiting for the task: (returnval){ [ 952.442403] env[61806]: value = "session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]522f72b6-78e6-28a3-be1f-846dcbefa188" [ 952.442403] env[61806]: _type = "Task" [ 952.442403] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 952.452357] env[61806]: DEBUG oslo_vmware.api [None req-5d1f922c-8037-4fb1-9c2a-dcd72461a463 tempest-ServersNegativeTestJSON-891856253 tempest-ServersNegativeTestJSON-891856253-project-member] Task: {'id': session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]522f72b6-78e6-28a3-be1f-846dcbefa188, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 952.594461] env[61806]: DEBUG oslo_vmware.api [None req-03f49469-0c97-4bf3-b87e-03f023c054e5 tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Task: {'id': task-1294967, 'name': RemoveSnapshot_Task, 'duration_secs': 0.364289} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 952.594461] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-03f49469-0c97-4bf3-b87e-03f023c054e5 tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] [instance: 380aa63f-46ba-437d-bc29-0fc141ad1cf3] Deleted Snapshot of the VM instance {{(pid=61806) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 952.594461] env[61806]: DEBUG nova.compute.manager [None req-03f49469-0c97-4bf3-b87e-03f023c054e5 tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] [instance: 380aa63f-46ba-437d-bc29-0fc141ad1cf3] Checking state {{(pid=61806) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 952.595399] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1b3cecf9-b174-4242-859f-a64fe3b42328 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 952.763658] env[61806]: DEBUG oslo_concurrency.lockutils [req-5f91f4b9-4bd2-4135-abdc-d666a8f033ef req-eac5152f-0447-4a35-8775-899cdf2cec75 service nova] Releasing lock "refresh_cache-bdbaef31-7d37-40cb-ac3e-ad152b37e2d2" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 952.914769] env[61806]: DEBUG oslo_vmware.api [None req-61d33433-c3e5-4624-95f4-86df358dffc7 tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] Task: {'id': task-1294951, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 952.917724] env[61806]: DEBUG oslo_vmware.api [None req-cf5202db-26c0-4960-9218-bd95b8e81f62 tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Task: {'id': task-1294968, 'name': ReconfigVM_Task, 'duration_secs': 0.276266} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 952.918040] env[61806]: DEBUG nova.virt.vmwareapi.volumeops [None req-cf5202db-26c0-4960-9218-bd95b8e81f62 tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] [instance: 676f18b3-3da1-4a53-b96d-fe64ee8f9101] Reconfigured VM instance instance-00000058 to attach disk [datastore2] 676f18b3-3da1-4a53-b96d-fe64ee8f9101/676f18b3-3da1-4a53-b96d-fe64ee8f9101.vmdk or device None with type thin {{(pid=61806) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 952.918337] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-cf5202db-26c0-4960-9218-bd95b8e81f62 tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] [instance: 676f18b3-3da1-4a53-b96d-fe64ee8f9101] Updating instance '676f18b3-3da1-4a53-b96d-fe64ee8f9101' progress to 50 {{(pid=61806) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 952.927642] env[61806]: INFO nova.compute.manager [None req-9cd2108a-a1e5-403d-bbda-81099a62d59d tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] [instance: 38e96e22-4200-4175-9085-31977357da2d] Took 14.75 seconds to build instance. [ 952.937924] env[61806]: INFO nova.compute.manager [-] [instance: 43dcce0d-4fdb-4ee1-8309-92e92d2331a9] Took 1.43 seconds to deallocate network for instance. [ 952.954958] env[61806]: DEBUG oslo_vmware.api [None req-5d1f922c-8037-4fb1-9c2a-dcd72461a463 tempest-ServersNegativeTestJSON-891856253 tempest-ServersNegativeTestJSON-891856253-project-member] Task: {'id': session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]522f72b6-78e6-28a3-be1f-846dcbefa188, 'name': SearchDatastore_Task, 'duration_secs': 0.010793} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 952.955651] env[61806]: DEBUG oslo_concurrency.lockutils [None req-5d1f922c-8037-4fb1-9c2a-dcd72461a463 tempest-ServersNegativeTestJSON-891856253 tempest-ServersNegativeTestJSON-891856253-project-member] Releasing lock "[datastore1] devstack-image-cache_base/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 952.955651] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-5d1f922c-8037-4fb1-9c2a-dcd72461a463 tempest-ServersNegativeTestJSON-891856253 tempest-ServersNegativeTestJSON-891856253-project-member] [instance: bdbaef31-7d37-40cb-ac3e-ad152b37e2d2] Processing image ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa {{(pid=61806) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 952.955868] env[61806]: DEBUG oslo_concurrency.lockutils [None req-5d1f922c-8037-4fb1-9c2a-dcd72461a463 tempest-ServersNegativeTestJSON-891856253 tempest-ServersNegativeTestJSON-891856253-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa.vmdk" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 952.956037] env[61806]: DEBUG oslo_concurrency.lockutils [None req-5d1f922c-8037-4fb1-9c2a-dcd72461a463 tempest-ServersNegativeTestJSON-891856253 tempest-ServersNegativeTestJSON-891856253-project-member] Acquired lock "[datastore1] devstack-image-cache_base/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa.vmdk" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 952.956232] env[61806]: DEBUG nova.virt.vmwareapi.ds_util [None req-5d1f922c-8037-4fb1-9c2a-dcd72461a463 tempest-ServersNegativeTestJSON-891856253 tempest-ServersNegativeTestJSON-891856253-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61806) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 952.956497] env[61806]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-a69e0891-f6fc-4010-a733-597440cd4cb1 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 952.968109] env[61806]: DEBUG nova.virt.vmwareapi.ds_util [None req-5d1f922c-8037-4fb1-9c2a-dcd72461a463 tempest-ServersNegativeTestJSON-891856253 tempest-ServersNegativeTestJSON-891856253-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61806) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 952.968337] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-5d1f922c-8037-4fb1-9c2a-dcd72461a463 tempest-ServersNegativeTestJSON-891856253 tempest-ServersNegativeTestJSON-891856253-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61806) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 952.969313] env[61806]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-bfff275c-f474-4f6c-94ef-e07b952bcbbb {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 952.975433] env[61806]: DEBUG oslo_vmware.api [None req-5d1f922c-8037-4fb1-9c2a-dcd72461a463 tempest-ServersNegativeTestJSON-891856253 tempest-ServersNegativeTestJSON-891856253-project-member] Waiting for the task: (returnval){ [ 952.975433] env[61806]: value = "session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]5248bc99-5077-18d7-ca57-208abc6e4721" [ 952.975433] env[61806]: _type = "Task" [ 952.975433] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 952.984027] env[61806]: DEBUG oslo_vmware.api [None req-5d1f922c-8037-4fb1-9c2a-dcd72461a463 tempest-ServersNegativeTestJSON-891856253 tempest-ServersNegativeTestJSON-891856253-project-member] Task: {'id': session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]5248bc99-5077-18d7-ca57-208abc6e4721, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 953.110242] env[61806]: INFO nova.compute.manager [None req-03f49469-0c97-4bf3-b87e-03f023c054e5 tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] [instance: 380aa63f-46ba-437d-bc29-0fc141ad1cf3] Shelve offloading [ 953.112490] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [None req-03f49469-0c97-4bf3-b87e-03f023c054e5 tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] [instance: 380aa63f-46ba-437d-bc29-0fc141ad1cf3] Powering off the VM {{(pid=61806) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 953.112789] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-af87862f-81c4-4fbf-9304-1e4c7c6612c6 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 953.120711] env[61806]: DEBUG oslo_vmware.api [None req-03f49469-0c97-4bf3-b87e-03f023c054e5 tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Waiting for the task: (returnval){ [ 953.120711] env[61806]: value = "task-1294969" [ 953.120711] env[61806]: _type = "Task" [ 953.120711] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 953.128889] env[61806]: DEBUG oslo_vmware.api [None req-03f49469-0c97-4bf3-b87e-03f023c054e5 tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Task: {'id': task-1294969, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 953.416164] env[61806]: DEBUG oslo_vmware.api [None req-61d33433-c3e5-4624-95f4-86df358dffc7 tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] Task: {'id': task-1294951, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 953.427772] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f49b9755-cbb3-4191-9647-11c7bd856701 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 953.430613] env[61806]: DEBUG oslo_concurrency.lockutils [None req-9cd2108a-a1e5-403d-bbda-81099a62d59d tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] Lock "38e96e22-4200-4175-9085-31977357da2d" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 16.267s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 953.448244] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d0835dbb-a770-4e27-8745-04c7d6de7ae0 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 953.451544] env[61806]: DEBUG oslo_concurrency.lockutils [None req-0efd4a4f-9a56-4502-a76c-4d8b926237a6 tempest-AttachVolumeNegativeTest-1683033704 tempest-AttachVolumeNegativeTest-1683033704-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 953.451800] env[61806]: DEBUG oslo_concurrency.lockutils [None req-0efd4a4f-9a56-4502-a76c-4d8b926237a6 tempest-AttachVolumeNegativeTest-1683033704 tempest-AttachVolumeNegativeTest-1683033704-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.000s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 953.452042] env[61806]: DEBUG nova.objects.instance [None req-0efd4a4f-9a56-4502-a76c-4d8b926237a6 tempest-AttachVolumeNegativeTest-1683033704 tempest-AttachVolumeNegativeTest-1683033704-project-member] Lazy-loading 'resources' on Instance uuid 43dcce0d-4fdb-4ee1-8309-92e92d2331a9 {{(pid=61806) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 953.471523] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-cf5202db-26c0-4960-9218-bd95b8e81f62 tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] [instance: 676f18b3-3da1-4a53-b96d-fe64ee8f9101] Updating instance '676f18b3-3da1-4a53-b96d-fe64ee8f9101' progress to 67 {{(pid=61806) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 953.487200] env[61806]: DEBUG oslo_vmware.api [None req-5d1f922c-8037-4fb1-9c2a-dcd72461a463 tempest-ServersNegativeTestJSON-891856253 tempest-ServersNegativeTestJSON-891856253-project-member] Task: {'id': session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]5248bc99-5077-18d7-ca57-208abc6e4721, 'name': SearchDatastore_Task, 'duration_secs': 0.043215} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 953.488041] env[61806]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-de1ab134-17c4-4f80-b038-fbdf67a33668 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 953.494987] env[61806]: DEBUG oslo_vmware.api [None req-5d1f922c-8037-4fb1-9c2a-dcd72461a463 tempest-ServersNegativeTestJSON-891856253 tempest-ServersNegativeTestJSON-891856253-project-member] Waiting for the task: (returnval){ [ 953.494987] env[61806]: value = "session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]526f4cb4-7700-38d4-dc3f-97219ae48f94" [ 953.494987] env[61806]: _type = "Task" [ 953.494987] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 953.505069] env[61806]: DEBUG oslo_vmware.api [None req-5d1f922c-8037-4fb1-9c2a-dcd72461a463 tempest-ServersNegativeTestJSON-891856253 tempest-ServersNegativeTestJSON-891856253-project-member] Task: {'id': session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]526f4cb4-7700-38d4-dc3f-97219ae48f94, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 953.551232] env[61806]: DEBUG nova.compute.manager [req-48d3cc39-bb0c-4742-b6aa-5bad69baea93 req-ca470bad-f570-4c72-9a0a-9b31ccd8f935 service nova] [instance: 43dcce0d-4fdb-4ee1-8309-92e92d2331a9] Received event network-vif-deleted-6361e800-fbf2-48e0-a11d-b00d9a9955b4 {{(pid=61806) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 953.631476] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [None req-03f49469-0c97-4bf3-b87e-03f023c054e5 tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] [instance: 380aa63f-46ba-437d-bc29-0fc141ad1cf3] VM already powered off {{(pid=61806) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1509}} [ 953.631701] env[61806]: DEBUG nova.compute.manager [None req-03f49469-0c97-4bf3-b87e-03f023c054e5 tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] [instance: 380aa63f-46ba-437d-bc29-0fc141ad1cf3] Checking state {{(pid=61806) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 953.632482] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-811bdb2c-87f4-48af-a619-62bc445ed83e {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 953.639113] env[61806]: DEBUG oslo_concurrency.lockutils [None req-03f49469-0c97-4bf3-b87e-03f023c054e5 tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Acquiring lock "refresh_cache-380aa63f-46ba-437d-bc29-0fc141ad1cf3" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 953.639303] env[61806]: DEBUG oslo_concurrency.lockutils [None req-03f49469-0c97-4bf3-b87e-03f023c054e5 tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Acquired lock "refresh_cache-380aa63f-46ba-437d-bc29-0fc141ad1cf3" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 953.639484] env[61806]: DEBUG nova.network.neutron [None req-03f49469-0c97-4bf3-b87e-03f023c054e5 tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] [instance: 380aa63f-46ba-437d-bc29-0fc141ad1cf3] Building network info cache for instance {{(pid=61806) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 953.918313] env[61806]: DEBUG oslo_vmware.api [None req-61d33433-c3e5-4624-95f4-86df358dffc7 tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] Task: {'id': task-1294951, 'name': ReconfigVM_Task} progress is 18%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 954.010634] env[61806]: DEBUG oslo_vmware.api [None req-5d1f922c-8037-4fb1-9c2a-dcd72461a463 tempest-ServersNegativeTestJSON-891856253 tempest-ServersNegativeTestJSON-891856253-project-member] Task: {'id': session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]526f4cb4-7700-38d4-dc3f-97219ae48f94, 'name': SearchDatastore_Task, 'duration_secs': 0.012262} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 954.011488] env[61806]: DEBUG oslo_concurrency.lockutils [None req-5d1f922c-8037-4fb1-9c2a-dcd72461a463 tempest-ServersNegativeTestJSON-891856253 tempest-ServersNegativeTestJSON-891856253-project-member] Releasing lock "[datastore1] devstack-image-cache_base/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa.vmdk" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 954.011678] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [None req-5d1f922c-8037-4fb1-9c2a-dcd72461a463 tempest-ServersNegativeTestJSON-891856253 tempest-ServersNegativeTestJSON-891856253-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa.vmdk to [datastore1] bdbaef31-7d37-40cb-ac3e-ad152b37e2d2/bdbaef31-7d37-40cb-ac3e-ad152b37e2d2.vmdk {{(pid=61806) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 954.012558] env[61806]: DEBUG nova.network.neutron [None req-cf5202db-26c0-4960-9218-bd95b8e81f62 tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] [instance: 676f18b3-3da1-4a53-b96d-fe64ee8f9101] Port d3a921a7-f1fa-4352-ae53-9f2996576fef binding to destination host cpu-1 is already ACTIVE {{(pid=61806) migrate_instance_start /opt/stack/nova/nova/network/neutron.py:3171}} [ 954.013973] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-f8c18be0-a84c-4204-b089-5826a38dbe73 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 954.026050] env[61806]: DEBUG oslo_vmware.api [None req-5d1f922c-8037-4fb1-9c2a-dcd72461a463 tempest-ServersNegativeTestJSON-891856253 tempest-ServersNegativeTestJSON-891856253-project-member] Waiting for the task: (returnval){ [ 954.026050] env[61806]: value = "task-1294970" [ 954.026050] env[61806]: _type = "Task" [ 954.026050] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 954.039146] env[61806]: DEBUG oslo_vmware.api [None req-5d1f922c-8037-4fb1-9c2a-dcd72461a463 tempest-ServersNegativeTestJSON-891856253 tempest-ServersNegativeTestJSON-891856253-project-member] Task: {'id': task-1294970, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 954.144916] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bc4894a5-3e24-4225-8aeb-e853630b647e {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 954.153360] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f8e9a018-12b5-453d-9e1d-424364a79610 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 954.185031] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4e9cb601-85c9-42af-92cf-32fe9ad01a40 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 954.195617] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bec49710-4c1d-4f80-bd63-41374b84739f {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 954.209526] env[61806]: DEBUG nova.compute.provider_tree [None req-0efd4a4f-9a56-4502-a76c-4d8b926237a6 tempest-AttachVolumeNegativeTest-1683033704 tempest-AttachVolumeNegativeTest-1683033704-project-member] Inventory has not changed in ProviderTree for provider: a2858be1-fd22-4e08-979e-87ad25293407 {{(pid=61806) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 954.255127] env[61806]: DEBUG oslo_concurrency.lockutils [None req-60f88c72-8f2b-4f98-b3cc-7787383864bd tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Acquiring lock "461f98f2-80ba-4088-9e7f-1ba2a9b0c1d9" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 954.255399] env[61806]: DEBUG oslo_concurrency.lockutils [None req-60f88c72-8f2b-4f98-b3cc-7787383864bd tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Lock "461f98f2-80ba-4088-9e7f-1ba2a9b0c1d9" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 954.400679] env[61806]: DEBUG nova.network.neutron [None req-03f49469-0c97-4bf3-b87e-03f023c054e5 tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] [instance: 380aa63f-46ba-437d-bc29-0fc141ad1cf3] Updating instance_info_cache with network_info: [{"id": "2c4c1295-84da-4bdf-8e6d-742b6358064c", "address": "fa:16:3e:14:c5:ea", "network": {"id": "b7e78559-d885-4b59-9c7f-9d966b424d04", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-1456721354-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7ab95fb72b5d46c3b6c7bebbccf897cd", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d891a662-7da0-4be5-8d0d-01b4ff70552a", "external-id": "nsx-vlan-transportzone-815", "segmentation_id": 815, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2c4c1295-84", "ovs_interfaceid": "2c4c1295-84da-4bdf-8e6d-742b6358064c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61806) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 954.419274] env[61806]: DEBUG oslo_vmware.api [None req-61d33433-c3e5-4624-95f4-86df358dffc7 tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] Task: {'id': task-1294951, 'name': ReconfigVM_Task, 'duration_secs': 5.745886} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 954.419752] env[61806]: DEBUG oslo_concurrency.lockutils [None req-61d33433-c3e5-4624-95f4-86df358dffc7 tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] Releasing lock "6f6ba57a-e2d6-4749-a53a-e263861cb1c0" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 954.419824] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-61d33433-c3e5-4624-95f4-86df358dffc7 tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] [instance: 6f6ba57a-e2d6-4749-a53a-e263861cb1c0] Reconfigured VM to detach interface {{(pid=61806) detach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1984}} [ 954.441835] env[61806]: DEBUG oslo_concurrency.lockutils [None req-c8cf9c01-54a9-4c86-b57f-b2e6a2fbc11a tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] Acquiring lock "0bac3048-759a-47d9-a6ef-a0386a49a974" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 954.442183] env[61806]: DEBUG oslo_concurrency.lockutils [None req-c8cf9c01-54a9-4c86-b57f-b2e6a2fbc11a tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] Lock "0bac3048-759a-47d9-a6ef-a0386a49a974" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 954.536318] env[61806]: DEBUG oslo_vmware.api [None req-5d1f922c-8037-4fb1-9c2a-dcd72461a463 tempest-ServersNegativeTestJSON-891856253 tempest-ServersNegativeTestJSON-891856253-project-member] Task: {'id': task-1294970, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 954.712963] env[61806]: DEBUG nova.scheduler.client.report [None req-0efd4a4f-9a56-4502-a76c-4d8b926237a6 tempest-AttachVolumeNegativeTest-1683033704 tempest-AttachVolumeNegativeTest-1683033704-project-member] Inventory has not changed for provider a2858be1-fd22-4e08-979e-87ad25293407 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 138, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61806) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 954.757764] env[61806]: DEBUG nova.compute.manager [None req-60f88c72-8f2b-4f98-b3cc-7787383864bd tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] [instance: 461f98f2-80ba-4088-9e7f-1ba2a9b0c1d9] Starting instance... {{(pid=61806) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 954.903485] env[61806]: DEBUG oslo_concurrency.lockutils [None req-03f49469-0c97-4bf3-b87e-03f023c054e5 tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Releasing lock "refresh_cache-380aa63f-46ba-437d-bc29-0fc141ad1cf3" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 954.944400] env[61806]: DEBUG nova.compute.manager [None req-c8cf9c01-54a9-4c86-b57f-b2e6a2fbc11a tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] [instance: 0bac3048-759a-47d9-a6ef-a0386a49a974] Starting instance... {{(pid=61806) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 955.047868] env[61806]: DEBUG oslo_concurrency.lockutils [None req-cf5202db-26c0-4960-9218-bd95b8e81f62 tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Acquiring lock "676f18b3-3da1-4a53-b96d-fe64ee8f9101-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 955.048243] env[61806]: DEBUG oslo_concurrency.lockutils [None req-cf5202db-26c0-4960-9218-bd95b8e81f62 tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Lock "676f18b3-3da1-4a53-b96d-fe64ee8f9101-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 955.048539] env[61806]: DEBUG oslo_concurrency.lockutils [None req-cf5202db-26c0-4960-9218-bd95b8e81f62 tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Lock "676f18b3-3da1-4a53-b96d-fe64ee8f9101-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 955.056592] env[61806]: DEBUG oslo_vmware.api [None req-5d1f922c-8037-4fb1-9c2a-dcd72461a463 tempest-ServersNegativeTestJSON-891856253 tempest-ServersNegativeTestJSON-891856253-project-member] Task: {'id': task-1294970, 'name': CopyVirtualDisk_Task} progress is 100%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 955.167692] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-03f49469-0c97-4bf3-b87e-03f023c054e5 tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] [instance: 380aa63f-46ba-437d-bc29-0fc141ad1cf3] Destroying instance {{(pid=61806) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 955.168690] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6d4d6cde-9332-4a1a-a87a-832a6d3c76eb {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 955.176480] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-03f49469-0c97-4bf3-b87e-03f023c054e5 tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] [instance: 380aa63f-46ba-437d-bc29-0fc141ad1cf3] Unregistering the VM {{(pid=61806) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 955.176755] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-9a68693f-1591-44c6-ae20-b25423b3761d {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 955.219959] env[61806]: DEBUG oslo_concurrency.lockutils [None req-0efd4a4f-9a56-4502-a76c-4d8b926237a6 tempest-AttachVolumeNegativeTest-1683033704 tempest-AttachVolumeNegativeTest-1683033704-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.768s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 955.243780] env[61806]: INFO nova.scheduler.client.report [None req-0efd4a4f-9a56-4502-a76c-4d8b926237a6 tempest-AttachVolumeNegativeTest-1683033704 tempest-AttachVolumeNegativeTest-1683033704-project-member] Deleted allocations for instance 43dcce0d-4fdb-4ee1-8309-92e92d2331a9 [ 955.279415] env[61806]: DEBUG oslo_concurrency.lockutils [None req-60f88c72-8f2b-4f98-b3cc-7787383864bd tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 955.279634] env[61806]: DEBUG oslo_concurrency.lockutils [None req-60f88c72-8f2b-4f98-b3cc-7787383864bd tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.000s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 955.281540] env[61806]: INFO nova.compute.claims [None req-60f88c72-8f2b-4f98-b3cc-7787383864bd tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] [instance: 461f98f2-80ba-4088-9e7f-1ba2a9b0c1d9] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 955.304395] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-03f49469-0c97-4bf3-b87e-03f023c054e5 tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] [instance: 380aa63f-46ba-437d-bc29-0fc141ad1cf3] Unregistered the VM {{(pid=61806) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 955.304699] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-03f49469-0c97-4bf3-b87e-03f023c054e5 tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] [instance: 380aa63f-46ba-437d-bc29-0fc141ad1cf3] Deleting contents of the VM from datastore datastore1 {{(pid=61806) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 955.304807] env[61806]: DEBUG nova.virt.vmwareapi.ds_util [None req-03f49469-0c97-4bf3-b87e-03f023c054e5 tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Deleting the datastore file [datastore1] 380aa63f-46ba-437d-bc29-0fc141ad1cf3 {{(pid=61806) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 955.305381] env[61806]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-add54be5-07dd-4493-a22b-790b7e1d491f {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 955.314873] env[61806]: DEBUG oslo_vmware.api [None req-03f49469-0c97-4bf3-b87e-03f023c054e5 tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Waiting for the task: (returnval){ [ 955.314873] env[61806]: value = "task-1294972" [ 955.314873] env[61806]: _type = "Task" [ 955.314873] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 955.323486] env[61806]: DEBUG oslo_vmware.api [None req-03f49469-0c97-4bf3-b87e-03f023c054e5 tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Task: {'id': task-1294972, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 955.465480] env[61806]: DEBUG oslo_concurrency.lockutils [None req-c8cf9c01-54a9-4c86-b57f-b2e6a2fbc11a tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 955.544414] env[61806]: DEBUG oslo_vmware.api [None req-5d1f922c-8037-4fb1-9c2a-dcd72461a463 tempest-ServersNegativeTestJSON-891856253 tempest-ServersNegativeTestJSON-891856253-project-member] Task: {'id': task-1294970, 'name': CopyVirtualDisk_Task} progress is 100%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 955.579210] env[61806]: DEBUG nova.compute.manager [req-db07c76c-07f5-41e8-87d5-fd66dbd7a61d req-33cb9627-8e10-4f83-93ac-a8f37c649283 service nova] [instance: 380aa63f-46ba-437d-bc29-0fc141ad1cf3] Received event network-vif-unplugged-2c4c1295-84da-4bdf-8e6d-742b6358064c {{(pid=61806) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 955.579489] env[61806]: DEBUG oslo_concurrency.lockutils [req-db07c76c-07f5-41e8-87d5-fd66dbd7a61d req-33cb9627-8e10-4f83-93ac-a8f37c649283 service nova] Acquiring lock "380aa63f-46ba-437d-bc29-0fc141ad1cf3-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 955.579747] env[61806]: DEBUG oslo_concurrency.lockutils [req-db07c76c-07f5-41e8-87d5-fd66dbd7a61d req-33cb9627-8e10-4f83-93ac-a8f37c649283 service nova] Lock "380aa63f-46ba-437d-bc29-0fc141ad1cf3-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.001s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 955.579931] env[61806]: DEBUG oslo_concurrency.lockutils [req-db07c76c-07f5-41e8-87d5-fd66dbd7a61d req-33cb9627-8e10-4f83-93ac-a8f37c649283 service nova] Lock "380aa63f-46ba-437d-bc29-0fc141ad1cf3-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 955.580159] env[61806]: DEBUG nova.compute.manager [req-db07c76c-07f5-41e8-87d5-fd66dbd7a61d req-33cb9627-8e10-4f83-93ac-a8f37c649283 service nova] [instance: 380aa63f-46ba-437d-bc29-0fc141ad1cf3] No waiting events found dispatching network-vif-unplugged-2c4c1295-84da-4bdf-8e6d-742b6358064c {{(pid=61806) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 955.580305] env[61806]: WARNING nova.compute.manager [req-db07c76c-07f5-41e8-87d5-fd66dbd7a61d req-33cb9627-8e10-4f83-93ac-a8f37c649283 service nova] [instance: 380aa63f-46ba-437d-bc29-0fc141ad1cf3] Received unexpected event network-vif-unplugged-2c4c1295-84da-4bdf-8e6d-742b6358064c for instance with vm_state shelved and task_state shelving_offloading. [ 955.580474] env[61806]: DEBUG nova.compute.manager [req-db07c76c-07f5-41e8-87d5-fd66dbd7a61d req-33cb9627-8e10-4f83-93ac-a8f37c649283 service nova] [instance: 380aa63f-46ba-437d-bc29-0fc141ad1cf3] Received event network-changed-2c4c1295-84da-4bdf-8e6d-742b6358064c {{(pid=61806) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 955.580636] env[61806]: DEBUG nova.compute.manager [req-db07c76c-07f5-41e8-87d5-fd66dbd7a61d req-33cb9627-8e10-4f83-93ac-a8f37c649283 service nova] [instance: 380aa63f-46ba-437d-bc29-0fc141ad1cf3] Refreshing instance network info cache due to event network-changed-2c4c1295-84da-4bdf-8e6d-742b6358064c. {{(pid=61806) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 955.580827] env[61806]: DEBUG oslo_concurrency.lockutils [req-db07c76c-07f5-41e8-87d5-fd66dbd7a61d req-33cb9627-8e10-4f83-93ac-a8f37c649283 service nova] Acquiring lock "refresh_cache-380aa63f-46ba-437d-bc29-0fc141ad1cf3" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 955.581077] env[61806]: DEBUG oslo_concurrency.lockutils [req-db07c76c-07f5-41e8-87d5-fd66dbd7a61d req-33cb9627-8e10-4f83-93ac-a8f37c649283 service nova] Acquired lock "refresh_cache-380aa63f-46ba-437d-bc29-0fc141ad1cf3" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 955.581143] env[61806]: DEBUG nova.network.neutron [req-db07c76c-07f5-41e8-87d5-fd66dbd7a61d req-33cb9627-8e10-4f83-93ac-a8f37c649283 service nova] [instance: 380aa63f-46ba-437d-bc29-0fc141ad1cf3] Refreshing network info cache for port 2c4c1295-84da-4bdf-8e6d-742b6358064c {{(pid=61806) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 955.726561] env[61806]: DEBUG oslo_concurrency.lockutils [None req-61d33433-c3e5-4624-95f4-86df358dffc7 tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] Acquiring lock "refresh_cache-6f6ba57a-e2d6-4749-a53a-e263861cb1c0" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 955.726561] env[61806]: DEBUG oslo_concurrency.lockutils [None req-61d33433-c3e5-4624-95f4-86df358dffc7 tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] Acquired lock "refresh_cache-6f6ba57a-e2d6-4749-a53a-e263861cb1c0" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 955.726561] env[61806]: DEBUG nova.network.neutron [None req-61d33433-c3e5-4624-95f4-86df358dffc7 tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] [instance: 6f6ba57a-e2d6-4749-a53a-e263861cb1c0] Building network info cache for instance {{(pid=61806) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 955.751292] env[61806]: DEBUG oslo_concurrency.lockutils [None req-0efd4a4f-9a56-4502-a76c-4d8b926237a6 tempest-AttachVolumeNegativeTest-1683033704 tempest-AttachVolumeNegativeTest-1683033704-project-member] Lock "43dcce0d-4fdb-4ee1-8309-92e92d2331a9" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.378s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 955.826528] env[61806]: DEBUG oslo_vmware.api [None req-03f49469-0c97-4bf3-b87e-03f023c054e5 tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Task: {'id': task-1294972, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 956.046186] env[61806]: DEBUG oslo_vmware.api [None req-5d1f922c-8037-4fb1-9c2a-dcd72461a463 tempest-ServersNegativeTestJSON-891856253 tempest-ServersNegativeTestJSON-891856253-project-member] Task: {'id': task-1294970, 'name': CopyVirtualDisk_Task, 'duration_secs': 1.56158} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 956.046186] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [None req-5d1f922c-8037-4fb1-9c2a-dcd72461a463 tempest-ServersNegativeTestJSON-891856253 tempest-ServersNegativeTestJSON-891856253-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa.vmdk to [datastore1] bdbaef31-7d37-40cb-ac3e-ad152b37e2d2/bdbaef31-7d37-40cb-ac3e-ad152b37e2d2.vmdk {{(pid=61806) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 956.046186] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-5d1f922c-8037-4fb1-9c2a-dcd72461a463 tempest-ServersNegativeTestJSON-891856253 tempest-ServersNegativeTestJSON-891856253-project-member] [instance: bdbaef31-7d37-40cb-ac3e-ad152b37e2d2] Extending root virtual disk to 1048576 {{(pid=61806) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 956.047162] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-5a828f63-7a09-4508-9cc7-d137f68ecb06 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 956.053074] env[61806]: DEBUG oslo_vmware.api [None req-5d1f922c-8037-4fb1-9c2a-dcd72461a463 tempest-ServersNegativeTestJSON-891856253 tempest-ServersNegativeTestJSON-891856253-project-member] Waiting for the task: (returnval){ [ 956.053074] env[61806]: value = "task-1294973" [ 956.053074] env[61806]: _type = "Task" [ 956.053074] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 956.064847] env[61806]: DEBUG oslo_vmware.api [None req-5d1f922c-8037-4fb1-9c2a-dcd72461a463 tempest-ServersNegativeTestJSON-891856253 tempest-ServersNegativeTestJSON-891856253-project-member] Task: {'id': task-1294973, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 956.115312] env[61806]: DEBUG oslo_concurrency.lockutils [None req-cf5202db-26c0-4960-9218-bd95b8e81f62 tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Acquiring lock "refresh_cache-676f18b3-3da1-4a53-b96d-fe64ee8f9101" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 956.115477] env[61806]: DEBUG oslo_concurrency.lockutils [None req-cf5202db-26c0-4960-9218-bd95b8e81f62 tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Acquired lock "refresh_cache-676f18b3-3da1-4a53-b96d-fe64ee8f9101" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 956.115655] env[61806]: DEBUG nova.network.neutron [None req-cf5202db-26c0-4960-9218-bd95b8e81f62 tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] [instance: 676f18b3-3da1-4a53-b96d-fe64ee8f9101] Building network info cache for instance {{(pid=61806) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 956.327114] env[61806]: DEBUG oslo_vmware.api [None req-03f49469-0c97-4bf3-b87e-03f023c054e5 tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Task: {'id': task-1294972, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.881221} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 956.329896] env[61806]: DEBUG nova.virt.vmwareapi.ds_util [None req-03f49469-0c97-4bf3-b87e-03f023c054e5 tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Deleted the datastore file {{(pid=61806) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 956.330140] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-03f49469-0c97-4bf3-b87e-03f023c054e5 tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] [instance: 380aa63f-46ba-437d-bc29-0fc141ad1cf3] Deleted contents of the VM from datastore datastore1 {{(pid=61806) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 956.330332] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-03f49469-0c97-4bf3-b87e-03f023c054e5 tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] [instance: 380aa63f-46ba-437d-bc29-0fc141ad1cf3] Instance destroyed {{(pid=61806) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 956.355680] env[61806]: INFO nova.scheduler.client.report [None req-03f49469-0c97-4bf3-b87e-03f023c054e5 tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Deleted allocations for instance 380aa63f-46ba-437d-bc29-0fc141ad1cf3 [ 956.475342] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-942b9b3d-a403-4388-85a0-d210db0df2f6 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 956.484174] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cf9646b0-d6ac-4790-abf4-ad112e5feee4 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 956.516103] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7fef5f84-0cc3-45f1-9846-f8b127d83364 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 956.524807] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-14a89bee-1ddb-4471-abac-ac7b6f8a5bcb {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 956.541715] env[61806]: DEBUG nova.compute.provider_tree [None req-60f88c72-8f2b-4f98-b3cc-7787383864bd tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Inventory has not changed in ProviderTree for provider: a2858be1-fd22-4e08-979e-87ad25293407 {{(pid=61806) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 956.543830] env[61806]: DEBUG nova.network.neutron [req-db07c76c-07f5-41e8-87d5-fd66dbd7a61d req-33cb9627-8e10-4f83-93ac-a8f37c649283 service nova] [instance: 380aa63f-46ba-437d-bc29-0fc141ad1cf3] Updated VIF entry in instance network info cache for port 2c4c1295-84da-4bdf-8e6d-742b6358064c. {{(pid=61806) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 956.543982] env[61806]: DEBUG nova.network.neutron [req-db07c76c-07f5-41e8-87d5-fd66dbd7a61d req-33cb9627-8e10-4f83-93ac-a8f37c649283 service nova] [instance: 380aa63f-46ba-437d-bc29-0fc141ad1cf3] Updating instance_info_cache with network_info: [{"id": "2c4c1295-84da-4bdf-8e6d-742b6358064c", "address": "fa:16:3e:14:c5:ea", "network": {"id": "b7e78559-d885-4b59-9c7f-9d966b424d04", "bridge": null, "label": "tempest-DeleteServersTestJSON-1456721354-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7ab95fb72b5d46c3b6c7bebbccf897cd", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "unbound", "details": {}, "devname": "tap2c4c1295-84", "ovs_interfaceid": null, "qbh_params": null, "qbg_params": null, "active": false, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61806) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 956.564186] env[61806]: DEBUG oslo_vmware.api [None req-5d1f922c-8037-4fb1-9c2a-dcd72461a463 tempest-ServersNegativeTestJSON-891856253 tempest-ServersNegativeTestJSON-891856253-project-member] Task: {'id': task-1294973, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.235835} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 956.566665] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-5d1f922c-8037-4fb1-9c2a-dcd72461a463 tempest-ServersNegativeTestJSON-891856253 tempest-ServersNegativeTestJSON-891856253-project-member] [instance: bdbaef31-7d37-40cb-ac3e-ad152b37e2d2] Extended root virtual disk {{(pid=61806) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 956.568095] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e4e2ebf5-4e51-475e-b6f9-9ee94ba20f06 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 956.595467] env[61806]: DEBUG nova.virt.vmwareapi.volumeops [None req-5d1f922c-8037-4fb1-9c2a-dcd72461a463 tempest-ServersNegativeTestJSON-891856253 tempest-ServersNegativeTestJSON-891856253-project-member] [instance: bdbaef31-7d37-40cb-ac3e-ad152b37e2d2] Reconfiguring VM instance instance-0000005a to attach disk [datastore1] bdbaef31-7d37-40cb-ac3e-ad152b37e2d2/bdbaef31-7d37-40cb-ac3e-ad152b37e2d2.vmdk or device None with type sparse {{(pid=61806) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 956.600662] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-1059452f-2dfd-4d96-bb5b-8ac032a71912 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 956.622934] env[61806]: DEBUG oslo_vmware.api [None req-5d1f922c-8037-4fb1-9c2a-dcd72461a463 tempest-ServersNegativeTestJSON-891856253 tempest-ServersNegativeTestJSON-891856253-project-member] Waiting for the task: (returnval){ [ 956.622934] env[61806]: value = "task-1294974" [ 956.622934] env[61806]: _type = "Task" [ 956.622934] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 956.632329] env[61806]: DEBUG oslo_vmware.api [None req-5d1f922c-8037-4fb1-9c2a-dcd72461a463 tempest-ServersNegativeTestJSON-891856253 tempest-ServersNegativeTestJSON-891856253-project-member] Task: {'id': task-1294974, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 956.711696] env[61806]: INFO nova.network.neutron [None req-61d33433-c3e5-4624-95f4-86df358dffc7 tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] [instance: 6f6ba57a-e2d6-4749-a53a-e263861cb1c0] Port 7e031e88-6928-472d-a65e-b818d89e6fc5 from network info_cache is no longer associated with instance in Neutron. Removing from network info_cache. [ 956.712446] env[61806]: DEBUG nova.network.neutron [None req-61d33433-c3e5-4624-95f4-86df358dffc7 tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] [instance: 6f6ba57a-e2d6-4749-a53a-e263861cb1c0] Updating instance_info_cache with network_info: [{"id": "013c9f0d-65b1-4279-a823-a7f1c14dfb91", "address": "fa:16:3e:15:56:4e", "network": {"id": "a7ffda72-e57c-41fa-a7ed-d7b12084bc9d", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-62794865-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "6e391e1267014fafbb8f5a3211299819", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8b3cfeb1-f262-4fd9-b506-8e9c0733e2d8", "external-id": "nsx-vlan-transportzone-119", "segmentation_id": 119, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap013c9f0d-65", "ovs_interfaceid": "013c9f0d-65b1-4279-a823-a7f1c14dfb91", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61806) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 956.860422] env[61806]: DEBUG oslo_concurrency.lockutils [None req-03f49469-0c97-4bf3-b87e-03f023c054e5 tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 957.047349] env[61806]: DEBUG nova.scheduler.client.report [None req-60f88c72-8f2b-4f98-b3cc-7787383864bd tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Inventory has not changed for provider a2858be1-fd22-4e08-979e-87ad25293407 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 138, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61806) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 957.050987] env[61806]: DEBUG oslo_concurrency.lockutils [req-db07c76c-07f5-41e8-87d5-fd66dbd7a61d req-33cb9627-8e10-4f83-93ac-a8f37c649283 service nova] Releasing lock "refresh_cache-380aa63f-46ba-437d-bc29-0fc141ad1cf3" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 957.134238] env[61806]: DEBUG oslo_vmware.api [None req-5d1f922c-8037-4fb1-9c2a-dcd72461a463 tempest-ServersNegativeTestJSON-891856253 tempest-ServersNegativeTestJSON-891856253-project-member] Task: {'id': task-1294974, 'name': ReconfigVM_Task, 'duration_secs': 0.363087} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 957.134520] env[61806]: DEBUG nova.virt.vmwareapi.volumeops [None req-5d1f922c-8037-4fb1-9c2a-dcd72461a463 tempest-ServersNegativeTestJSON-891856253 tempest-ServersNegativeTestJSON-891856253-project-member] [instance: bdbaef31-7d37-40cb-ac3e-ad152b37e2d2] Reconfigured VM instance instance-0000005a to attach disk [datastore1] bdbaef31-7d37-40cb-ac3e-ad152b37e2d2/bdbaef31-7d37-40cb-ac3e-ad152b37e2d2.vmdk or device None with type sparse {{(pid=61806) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 957.135203] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-ca995e80-8b46-4388-bdc6-9f1d399f69cd {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 957.142239] env[61806]: DEBUG oslo_vmware.api [None req-5d1f922c-8037-4fb1-9c2a-dcd72461a463 tempest-ServersNegativeTestJSON-891856253 tempest-ServersNegativeTestJSON-891856253-project-member] Waiting for the task: (returnval){ [ 957.142239] env[61806]: value = "task-1294975" [ 957.142239] env[61806]: _type = "Task" [ 957.142239] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 957.151321] env[61806]: DEBUG oslo_vmware.api [None req-5d1f922c-8037-4fb1-9c2a-dcd72461a463 tempest-ServersNegativeTestJSON-891856253 tempest-ServersNegativeTestJSON-891856253-project-member] Task: {'id': task-1294975, 'name': Rename_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 957.169093] env[61806]: DEBUG nova.network.neutron [None req-cf5202db-26c0-4960-9218-bd95b8e81f62 tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] [instance: 676f18b3-3da1-4a53-b96d-fe64ee8f9101] Updating instance_info_cache with network_info: [{"id": "d3a921a7-f1fa-4352-ae53-9f2996576fef", "address": "fa:16:3e:66:40:96", "network": {"id": "46515a4e-610f-42fb-a424-205a03df0640", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-1133623547-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4b9d94c3d684474b99e305fb6cd9b95f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "82ca17df-257e-40e6-9ec9-310ed6f05ccb", "external-id": "nsx-vlan-transportzone-778", "segmentation_id": 778, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd3a921a7-f1", "ovs_interfaceid": "d3a921a7-f1fa-4352-ae53-9f2996576fef", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61806) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 957.214711] env[61806]: DEBUG oslo_concurrency.lockutils [None req-61d33433-c3e5-4624-95f4-86df358dffc7 tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] Releasing lock "refresh_cache-6f6ba57a-e2d6-4749-a53a-e263861cb1c0" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 957.268448] env[61806]: DEBUG nova.compute.manager [req-8de5a2aa-dc74-44aa-966b-389a1eed275a req-89a36e07-cfbc-4f77-87fb-052d46577c71 service nova] [instance: 6f6ba57a-e2d6-4749-a53a-e263861cb1c0] Received event network-changed-013c9f0d-65b1-4279-a823-a7f1c14dfb91 {{(pid=61806) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 957.268760] env[61806]: DEBUG nova.compute.manager [req-8de5a2aa-dc74-44aa-966b-389a1eed275a req-89a36e07-cfbc-4f77-87fb-052d46577c71 service nova] [instance: 6f6ba57a-e2d6-4749-a53a-e263861cb1c0] Refreshing instance network info cache due to event network-changed-013c9f0d-65b1-4279-a823-a7f1c14dfb91. {{(pid=61806) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 957.269009] env[61806]: DEBUG oslo_concurrency.lockutils [req-8de5a2aa-dc74-44aa-966b-389a1eed275a req-89a36e07-cfbc-4f77-87fb-052d46577c71 service nova] Acquiring lock "refresh_cache-6f6ba57a-e2d6-4749-a53a-e263861cb1c0" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 957.269194] env[61806]: DEBUG oslo_concurrency.lockutils [req-8de5a2aa-dc74-44aa-966b-389a1eed275a req-89a36e07-cfbc-4f77-87fb-052d46577c71 service nova] Acquired lock "refresh_cache-6f6ba57a-e2d6-4749-a53a-e263861cb1c0" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 957.269365] env[61806]: DEBUG nova.network.neutron [req-8de5a2aa-dc74-44aa-966b-389a1eed275a req-89a36e07-cfbc-4f77-87fb-052d46577c71 service nova] [instance: 6f6ba57a-e2d6-4749-a53a-e263861cb1c0] Refreshing network info cache for port 013c9f0d-65b1-4279-a823-a7f1c14dfb91 {{(pid=61806) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 957.272077] env[61806]: DEBUG oslo_concurrency.lockutils [None req-d1459cd5-a70e-4111-a41f-72b97e54aa17 tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Acquiring lock "380aa63f-46ba-437d-bc29-0fc141ad1cf3" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 957.552839] env[61806]: DEBUG oslo_concurrency.lockutils [None req-60f88c72-8f2b-4f98-b3cc-7787383864bd tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.273s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 957.553349] env[61806]: DEBUG nova.compute.manager [None req-60f88c72-8f2b-4f98-b3cc-7787383864bd tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] [instance: 461f98f2-80ba-4088-9e7f-1ba2a9b0c1d9] Start building networks asynchronously for instance. {{(pid=61806) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 957.555938] env[61806]: DEBUG oslo_concurrency.lockutils [None req-c8cf9c01-54a9-4c86-b57f-b2e6a2fbc11a tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 2.091s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 957.557307] env[61806]: INFO nova.compute.claims [None req-c8cf9c01-54a9-4c86-b57f-b2e6a2fbc11a tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] [instance: 0bac3048-759a-47d9-a6ef-a0386a49a974] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 957.655588] env[61806]: DEBUG oslo_vmware.api [None req-5d1f922c-8037-4fb1-9c2a-dcd72461a463 tempest-ServersNegativeTestJSON-891856253 tempest-ServersNegativeTestJSON-891856253-project-member] Task: {'id': task-1294975, 'name': Rename_Task, 'duration_secs': 0.167183} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 957.655977] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [None req-5d1f922c-8037-4fb1-9c2a-dcd72461a463 tempest-ServersNegativeTestJSON-891856253 tempest-ServersNegativeTestJSON-891856253-project-member] [instance: bdbaef31-7d37-40cb-ac3e-ad152b37e2d2] Powering on the VM {{(pid=61806) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 957.656398] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-72ace49b-38b0-457d-b777-07d4b70c3a71 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 957.665016] env[61806]: DEBUG oslo_vmware.api [None req-5d1f922c-8037-4fb1-9c2a-dcd72461a463 tempest-ServersNegativeTestJSON-891856253 tempest-ServersNegativeTestJSON-891856253-project-member] Waiting for the task: (returnval){ [ 957.665016] env[61806]: value = "task-1294976" [ 957.665016] env[61806]: _type = "Task" [ 957.665016] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 957.674083] env[61806]: DEBUG oslo_concurrency.lockutils [None req-cf5202db-26c0-4960-9218-bd95b8e81f62 tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Releasing lock "refresh_cache-676f18b3-3da1-4a53-b96d-fe64ee8f9101" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 957.677575] env[61806]: DEBUG oslo_vmware.api [None req-5d1f922c-8037-4fb1-9c2a-dcd72461a463 tempest-ServersNegativeTestJSON-891856253 tempest-ServersNegativeTestJSON-891856253-project-member] Task: {'id': task-1294976, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 957.718852] env[61806]: DEBUG oslo_concurrency.lockutils [None req-61d33433-c3e5-4624-95f4-86df358dffc7 tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] Lock "interface-6f6ba57a-e2d6-4749-a53a-e263861cb1c0-7e031e88-6928-472d-a65e-b818d89e6fc5" "released" by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" :: held 9.901s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 958.045273] env[61806]: DEBUG oslo_concurrency.lockutils [None req-98c10467-8c15-41cc-ac16-68156711acc7 tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] Acquiring lock "interface-d9caee4a-a00c-4e50-b01c-99a8f796a09f-7e031e88-6928-472d-a65e-b818d89e6fc5" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 958.045273] env[61806]: DEBUG oslo_concurrency.lockutils [None req-98c10467-8c15-41cc-ac16-68156711acc7 tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] Lock "interface-d9caee4a-a00c-4e50-b01c-99a8f796a09f-7e031e88-6928-472d-a65e-b818d89e6fc5" acquired by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: waited 0.002s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 958.045273] env[61806]: DEBUG nova.objects.instance [None req-98c10467-8c15-41cc-ac16-68156711acc7 tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] Lazy-loading 'flavor' on Instance uuid d9caee4a-a00c-4e50-b01c-99a8f796a09f {{(pid=61806) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 958.045273] env[61806]: DEBUG nova.network.neutron [req-8de5a2aa-dc74-44aa-966b-389a1eed275a req-89a36e07-cfbc-4f77-87fb-052d46577c71 service nova] [instance: 6f6ba57a-e2d6-4749-a53a-e263861cb1c0] Updated VIF entry in instance network info cache for port 013c9f0d-65b1-4279-a823-a7f1c14dfb91. {{(pid=61806) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 958.045273] env[61806]: DEBUG nova.network.neutron [req-8de5a2aa-dc74-44aa-966b-389a1eed275a req-89a36e07-cfbc-4f77-87fb-052d46577c71 service nova] [instance: 6f6ba57a-e2d6-4749-a53a-e263861cb1c0] Updating instance_info_cache with network_info: [{"id": "013c9f0d-65b1-4279-a823-a7f1c14dfb91", "address": "fa:16:3e:15:56:4e", "network": {"id": "a7ffda72-e57c-41fa-a7ed-d7b12084bc9d", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-62794865-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "6e391e1267014fafbb8f5a3211299819", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8b3cfeb1-f262-4fd9-b506-8e9c0733e2d8", "external-id": "nsx-vlan-transportzone-119", "segmentation_id": 119, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap013c9f0d-65", "ovs_interfaceid": "013c9f0d-65b1-4279-a823-a7f1c14dfb91", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61806) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 958.063898] env[61806]: DEBUG nova.compute.utils [None req-60f88c72-8f2b-4f98-b3cc-7787383864bd tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Using /dev/sd instead of None {{(pid=61806) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 958.066031] env[61806]: DEBUG nova.compute.manager [None req-60f88c72-8f2b-4f98-b3cc-7787383864bd tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] [instance: 461f98f2-80ba-4088-9e7f-1ba2a9b0c1d9] Allocating IP information in the background. {{(pid=61806) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 958.066031] env[61806]: DEBUG nova.network.neutron [None req-60f88c72-8f2b-4f98-b3cc-7787383864bd tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] [instance: 461f98f2-80ba-4088-9e7f-1ba2a9b0c1d9] allocate_for_instance() {{(pid=61806) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 958.103816] env[61806]: DEBUG nova.policy [None req-60f88c72-8f2b-4f98-b3cc-7787383864bd tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '921aeb4e90e64f89888a9c65f873d66b', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'c1f2d161c50f4b70987b90e6364c3779', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61806) authorize /opt/stack/nova/nova/policy.py:201}} [ 958.175787] env[61806]: DEBUG oslo_vmware.api [None req-5d1f922c-8037-4fb1-9c2a-dcd72461a463 tempest-ServersNegativeTestJSON-891856253 tempest-ServersNegativeTestJSON-891856253-project-member] Task: {'id': task-1294976, 'name': PowerOnVM_Task} progress is 89%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 958.201378] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-940116ee-beb2-432e-b5aa-5906b35ffebe {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 958.223195] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2276c0b3-f6c8-46ef-bc67-4919a3d78330 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 958.231691] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-cf5202db-26c0-4960-9218-bd95b8e81f62 tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] [instance: 676f18b3-3da1-4a53-b96d-fe64ee8f9101] Updating instance '676f18b3-3da1-4a53-b96d-fe64ee8f9101' progress to 83 {{(pid=61806) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 958.274667] env[61806]: DEBUG oslo_concurrency.lockutils [None req-2c70fcbb-efa7-459d-81c0-a2a8a184893c tempest-AttachVolumeNegativeTest-1683033704 tempest-AttachVolumeNegativeTest-1683033704-project-member] Acquiring lock "e4638faa-7be1-4909-a595-c437837fa314" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 958.274922] env[61806]: DEBUG oslo_concurrency.lockutils [None req-2c70fcbb-efa7-459d-81c0-a2a8a184893c tempest-AttachVolumeNegativeTest-1683033704 tempest-AttachVolumeNegativeTest-1683033704-project-member] Lock "e4638faa-7be1-4909-a595-c437837fa314" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 958.369164] env[61806]: DEBUG nova.network.neutron [None req-60f88c72-8f2b-4f98-b3cc-7787383864bd tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] [instance: 461f98f2-80ba-4088-9e7f-1ba2a9b0c1d9] Successfully created port: 91000a2c-19c5-404d-90b3-367698085aec {{(pid=61806) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 958.509505] env[61806]: DEBUG oslo_concurrency.lockutils [req-8de5a2aa-dc74-44aa-966b-389a1eed275a req-89a36e07-cfbc-4f77-87fb-052d46577c71 service nova] Releasing lock "refresh_cache-6f6ba57a-e2d6-4749-a53a-e263861cb1c0" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 958.553333] env[61806]: DEBUG nova.objects.instance [None req-98c10467-8c15-41cc-ac16-68156711acc7 tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] Lazy-loading 'pci_requests' on Instance uuid d9caee4a-a00c-4e50-b01c-99a8f796a09f {{(pid=61806) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 958.566723] env[61806]: DEBUG nova.compute.manager [None req-60f88c72-8f2b-4f98-b3cc-7787383864bd tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] [instance: 461f98f2-80ba-4088-9e7f-1ba2a9b0c1d9] Start building block device mappings for instance. {{(pid=61806) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 958.675567] env[61806]: DEBUG oslo_vmware.api [None req-5d1f922c-8037-4fb1-9c2a-dcd72461a463 tempest-ServersNegativeTestJSON-891856253 tempest-ServersNegativeTestJSON-891856253-project-member] Task: {'id': task-1294976, 'name': PowerOnVM_Task, 'duration_secs': 0.634879} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 958.678237] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [None req-5d1f922c-8037-4fb1-9c2a-dcd72461a463 tempest-ServersNegativeTestJSON-891856253 tempest-ServersNegativeTestJSON-891856253-project-member] [instance: bdbaef31-7d37-40cb-ac3e-ad152b37e2d2] Powered on the VM {{(pid=61806) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 958.678455] env[61806]: INFO nova.compute.manager [None req-5d1f922c-8037-4fb1-9c2a-dcd72461a463 tempest-ServersNegativeTestJSON-891856253 tempest-ServersNegativeTestJSON-891856253-project-member] [instance: bdbaef31-7d37-40cb-ac3e-ad152b37e2d2] Took 9.47 seconds to spawn the instance on the hypervisor. [ 958.678640] env[61806]: DEBUG nova.compute.manager [None req-5d1f922c-8037-4fb1-9c2a-dcd72461a463 tempest-ServersNegativeTestJSON-891856253 tempest-ServersNegativeTestJSON-891856253-project-member] [instance: bdbaef31-7d37-40cb-ac3e-ad152b37e2d2] Checking state {{(pid=61806) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 958.680058] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a6cbe5c1-a9c4-4f95-b39b-75069cba8b9f {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 958.738460] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [None req-cf5202db-26c0-4960-9218-bd95b8e81f62 tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] [instance: 676f18b3-3da1-4a53-b96d-fe64ee8f9101] Powering on the VM {{(pid=61806) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 958.738776] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-8acd1872-3ace-4162-945c-50bb38b7f7ff {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 958.742902] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-11b3e46f-7801-4bec-8de1-54c00353cf75 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 958.747633] env[61806]: DEBUG oslo_vmware.api [None req-cf5202db-26c0-4960-9218-bd95b8e81f62 tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Waiting for the task: (returnval){ [ 958.747633] env[61806]: value = "task-1294977" [ 958.747633] env[61806]: _type = "Task" [ 958.747633] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 958.753189] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d89bc8c8-8237-42e2-ae8f-1743d97f8675 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 958.764029] env[61806]: DEBUG oslo_vmware.api [None req-cf5202db-26c0-4960-9218-bd95b8e81f62 tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Task: {'id': task-1294977, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 958.786805] env[61806]: DEBUG nova.compute.manager [None req-2c70fcbb-efa7-459d-81c0-a2a8a184893c tempest-AttachVolumeNegativeTest-1683033704 tempest-AttachVolumeNegativeTest-1683033704-project-member] [instance: e4638faa-7be1-4909-a595-c437837fa314] Starting instance... {{(pid=61806) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 958.791493] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-add03579-91ec-4ac7-bbc4-c9000108b688 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 958.800292] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b71dcf83-40e1-4beb-9e0a-df1340573d9a {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 958.816437] env[61806]: DEBUG nova.compute.provider_tree [None req-c8cf9c01-54a9-4c86-b57f-b2e6a2fbc11a tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] Inventory has not changed in ProviderTree for provider: a2858be1-fd22-4e08-979e-87ad25293407 {{(pid=61806) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 959.056304] env[61806]: DEBUG nova.objects.base [None req-98c10467-8c15-41cc-ac16-68156711acc7 tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] Object Instance lazy-loaded attributes: flavor,pci_requests {{(pid=61806) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 959.056591] env[61806]: DEBUG nova.network.neutron [None req-98c10467-8c15-41cc-ac16-68156711acc7 tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] [instance: d9caee4a-a00c-4e50-b01c-99a8f796a09f] allocate_for_instance() {{(pid=61806) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 959.165689] env[61806]: DEBUG nova.policy [None req-98c10467-8c15-41cc-ac16-68156711acc7 tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'b1eac9072ccb4b3bbd23f240e941a76e', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '6e391e1267014fafbb8f5a3211299819', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61806) authorize /opt/stack/nova/nova/policy.py:201}} [ 959.196603] env[61806]: INFO nova.compute.manager [None req-5d1f922c-8037-4fb1-9c2a-dcd72461a463 tempest-ServersNegativeTestJSON-891856253 tempest-ServersNegativeTestJSON-891856253-project-member] [instance: bdbaef31-7d37-40cb-ac3e-ad152b37e2d2] Took 15.94 seconds to build instance. [ 959.265814] env[61806]: DEBUG oslo_vmware.api [None req-cf5202db-26c0-4960-9218-bd95b8e81f62 tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Task: {'id': task-1294977, 'name': PowerOnVM_Task, 'duration_secs': 0.40587} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 959.269743] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [None req-cf5202db-26c0-4960-9218-bd95b8e81f62 tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] [instance: 676f18b3-3da1-4a53-b96d-fe64ee8f9101] Powered on the VM {{(pid=61806) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 959.269743] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-cf5202db-26c0-4960-9218-bd95b8e81f62 tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] [instance: 676f18b3-3da1-4a53-b96d-fe64ee8f9101] Updating instance '676f18b3-3da1-4a53-b96d-fe64ee8f9101' progress to 100 {{(pid=61806) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 959.311593] env[61806]: DEBUG oslo_concurrency.lockutils [None req-2c70fcbb-efa7-459d-81c0-a2a8a184893c tempest-AttachVolumeNegativeTest-1683033704 tempest-AttachVolumeNegativeTest-1683033704-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 959.319741] env[61806]: DEBUG nova.scheduler.client.report [None req-c8cf9c01-54a9-4c86-b57f-b2e6a2fbc11a tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] Inventory has not changed for provider a2858be1-fd22-4e08-979e-87ad25293407 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 138, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61806) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 959.394752] env[61806]: DEBUG oslo_service.periodic_task [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=61806) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 959.395066] env[61806]: DEBUG oslo_service.periodic_task [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Running periodic task ComputeManager._sync_scheduler_instance_info {{(pid=61806) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 959.578666] env[61806]: DEBUG nova.compute.manager [None req-60f88c72-8f2b-4f98-b3cc-7787383864bd tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] [instance: 461f98f2-80ba-4088-9e7f-1ba2a9b0c1d9] Start spawning the instance on the hypervisor. {{(pid=61806) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 959.604511] env[61806]: DEBUG nova.virt.hardware [None req-60f88c72-8f2b-4f98-b3cc-7787383864bd tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-15T18:05:10Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-15T18:04:52Z,direct_url=,disk_format='vmdk',id=ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='84ba83fce2c349a6988173c5d6fc3b07',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-15T18:04:53Z,virtual_size=,visibility=), allow threads: False {{(pid=61806) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 959.604784] env[61806]: DEBUG nova.virt.hardware [None req-60f88c72-8f2b-4f98-b3cc-7787383864bd tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Flavor limits 0:0:0 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 959.604947] env[61806]: DEBUG nova.virt.hardware [None req-60f88c72-8f2b-4f98-b3cc-7787383864bd tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Image limits 0:0:0 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 959.605172] env[61806]: DEBUG nova.virt.hardware [None req-60f88c72-8f2b-4f98-b3cc-7787383864bd tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Flavor pref 0:0:0 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 959.605344] env[61806]: DEBUG nova.virt.hardware [None req-60f88c72-8f2b-4f98-b3cc-7787383864bd tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Image pref 0:0:0 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 959.605502] env[61806]: DEBUG nova.virt.hardware [None req-60f88c72-8f2b-4f98-b3cc-7787383864bd tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 959.605721] env[61806]: DEBUG nova.virt.hardware [None req-60f88c72-8f2b-4f98-b3cc-7787383864bd tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61806) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 959.606080] env[61806]: DEBUG nova.virt.hardware [None req-60f88c72-8f2b-4f98-b3cc-7787383864bd tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61806) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 959.606289] env[61806]: DEBUG nova.virt.hardware [None req-60f88c72-8f2b-4f98-b3cc-7787383864bd tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Got 1 possible topologies {{(pid=61806) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 959.606465] env[61806]: DEBUG nova.virt.hardware [None req-60f88c72-8f2b-4f98-b3cc-7787383864bd tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61806) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 959.606644] env[61806]: DEBUG nova.virt.hardware [None req-60f88c72-8f2b-4f98-b3cc-7787383864bd tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61806) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 959.607536] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-591e1d1c-b078-4fde-9eb5-354e3d36f17d {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 959.616634] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-05861d59-1c38-4048-92a1-551fa0f28e50 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 959.698904] env[61806]: DEBUG oslo_concurrency.lockutils [None req-5d1f922c-8037-4fb1-9c2a-dcd72461a463 tempest-ServersNegativeTestJSON-891856253 tempest-ServersNegativeTestJSON-891856253-project-member] Lock "bdbaef31-7d37-40cb-ac3e-ad152b37e2d2" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 17.448s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 959.742284] env[61806]: DEBUG nova.compute.manager [req-8e83f681-fd89-4353-9e35-b5795b86bf49 req-9030c0ae-9b69-49c2-ab90-697249160cc1 service nova] [instance: d9caee4a-a00c-4e50-b01c-99a8f796a09f] Received event network-changed-c725ee28-d7a8-4b31-92bc-f63174564f5e {{(pid=61806) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 959.742613] env[61806]: DEBUG nova.compute.manager [req-8e83f681-fd89-4353-9e35-b5795b86bf49 req-9030c0ae-9b69-49c2-ab90-697249160cc1 service nova] [instance: d9caee4a-a00c-4e50-b01c-99a8f796a09f] Refreshing instance network info cache due to event network-changed-c725ee28-d7a8-4b31-92bc-f63174564f5e. {{(pid=61806) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 959.742926] env[61806]: DEBUG oslo_concurrency.lockutils [req-8e83f681-fd89-4353-9e35-b5795b86bf49 req-9030c0ae-9b69-49c2-ab90-697249160cc1 service nova] Acquiring lock "refresh_cache-d9caee4a-a00c-4e50-b01c-99a8f796a09f" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 959.743171] env[61806]: DEBUG oslo_concurrency.lockutils [req-8e83f681-fd89-4353-9e35-b5795b86bf49 req-9030c0ae-9b69-49c2-ab90-697249160cc1 service nova] Acquired lock "refresh_cache-d9caee4a-a00c-4e50-b01c-99a8f796a09f" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 959.743426] env[61806]: DEBUG nova.network.neutron [req-8e83f681-fd89-4353-9e35-b5795b86bf49 req-9030c0ae-9b69-49c2-ab90-697249160cc1 service nova] [instance: d9caee4a-a00c-4e50-b01c-99a8f796a09f] Refreshing network info cache for port c725ee28-d7a8-4b31-92bc-f63174564f5e {{(pid=61806) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 959.824721] env[61806]: DEBUG oslo_concurrency.lockutils [None req-c8cf9c01-54a9-4c86-b57f-b2e6a2fbc11a tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.269s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 959.825332] env[61806]: DEBUG nova.compute.manager [None req-c8cf9c01-54a9-4c86-b57f-b2e6a2fbc11a tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] [instance: 0bac3048-759a-47d9-a6ef-a0386a49a974] Start building networks asynchronously for instance. {{(pid=61806) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 959.829177] env[61806]: DEBUG oslo_concurrency.lockutils [None req-03f49469-0c97-4bf3-b87e-03f023c054e5 tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 2.969s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 959.829411] env[61806]: DEBUG nova.objects.instance [None req-03f49469-0c97-4bf3-b87e-03f023c054e5 tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Lazy-loading 'resources' on Instance uuid 380aa63f-46ba-437d-bc29-0fc141ad1cf3 {{(pid=61806) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 959.832758] env[61806]: DEBUG oslo_concurrency.lockutils [None req-28d96f96-d23f-4c45-9d59-0355de5e0c4f tempest-ServersNegativeTestJSON-891856253 tempest-ServersNegativeTestJSON-891856253-project-member] Acquiring lock "e146f7e5-bbf8-4adf-a2e9-b7e753438976" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 959.833073] env[61806]: DEBUG oslo_concurrency.lockutils [None req-28d96f96-d23f-4c45-9d59-0355de5e0c4f tempest-ServersNegativeTestJSON-891856253 tempest-ServersNegativeTestJSON-891856253-project-member] Lock "e146f7e5-bbf8-4adf-a2e9-b7e753438976" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 959.900357] env[61806]: DEBUG oslo_service.periodic_task [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=61806) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 959.900600] env[61806]: DEBUG nova.compute.manager [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Starting heal instance info cache {{(pid=61806) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9926}} [ 960.080871] env[61806]: DEBUG nova.network.neutron [None req-60f88c72-8f2b-4f98-b3cc-7787383864bd tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] [instance: 461f98f2-80ba-4088-9e7f-1ba2a9b0c1d9] Successfully updated port: 91000a2c-19c5-404d-90b3-367698085aec {{(pid=61806) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 960.334639] env[61806]: DEBUG nova.compute.utils [None req-c8cf9c01-54a9-4c86-b57f-b2e6a2fbc11a tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] Using /dev/sd instead of None {{(pid=61806) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 960.336105] env[61806]: DEBUG nova.objects.instance [None req-03f49469-0c97-4bf3-b87e-03f023c054e5 tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Lazy-loading 'numa_topology' on Instance uuid 380aa63f-46ba-437d-bc29-0fc141ad1cf3 {{(pid=61806) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 960.336980] env[61806]: DEBUG nova.compute.manager [None req-28d96f96-d23f-4c45-9d59-0355de5e0c4f tempest-ServersNegativeTestJSON-891856253 tempest-ServersNegativeTestJSON-891856253-project-member] [instance: e146f7e5-bbf8-4adf-a2e9-b7e753438976] Starting instance... {{(pid=61806) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 960.339219] env[61806]: DEBUG nova.compute.manager [None req-c8cf9c01-54a9-4c86-b57f-b2e6a2fbc11a tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] [instance: 0bac3048-759a-47d9-a6ef-a0386a49a974] Allocating IP information in the background. {{(pid=61806) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 960.339389] env[61806]: DEBUG nova.network.neutron [None req-c8cf9c01-54a9-4c86-b57f-b2e6a2fbc11a tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] [instance: 0bac3048-759a-47d9-a6ef-a0386a49a974] allocate_for_instance() {{(pid=61806) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 960.380503] env[61806]: DEBUG nova.policy [None req-c8cf9c01-54a9-4c86-b57f-b2e6a2fbc11a tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '4c5d4c6e1ad844a585c08c4644a60bbf', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '73ebcc4d79d248efb653a084a5e44302', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61806) authorize /opt/stack/nova/nova/policy.py:201}} [ 960.453797] env[61806]: DEBUG oslo_concurrency.lockutils [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Acquiring lock "refresh_cache-67d1bdd8-cf8f-4fb5-bd3c-42008e4d60e3" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 960.453950] env[61806]: DEBUG oslo_concurrency.lockutils [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Acquired lock "refresh_cache-67d1bdd8-cf8f-4fb5-bd3c-42008e4d60e3" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 960.454170] env[61806]: DEBUG nova.network.neutron [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] [instance: 67d1bdd8-cf8f-4fb5-bd3c-42008e4d60e3] Forcefully refreshing network info cache for instance {{(pid=61806) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2004}} [ 960.490063] env[61806]: DEBUG nova.network.neutron [req-8e83f681-fd89-4353-9e35-b5795b86bf49 req-9030c0ae-9b69-49c2-ab90-697249160cc1 service nova] [instance: d9caee4a-a00c-4e50-b01c-99a8f796a09f] Updated VIF entry in instance network info cache for port c725ee28-d7a8-4b31-92bc-f63174564f5e. {{(pid=61806) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 960.490468] env[61806]: DEBUG nova.network.neutron [req-8e83f681-fd89-4353-9e35-b5795b86bf49 req-9030c0ae-9b69-49c2-ab90-697249160cc1 service nova] [instance: d9caee4a-a00c-4e50-b01c-99a8f796a09f] Updating instance_info_cache with network_info: [{"id": "c725ee28-d7a8-4b31-92bc-f63174564f5e", "address": "fa:16:3e:3a:18:8e", "network": {"id": "a7ffda72-e57c-41fa-a7ed-d7b12084bc9d", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-62794865-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.218", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "6e391e1267014fafbb8f5a3211299819", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8b3cfeb1-f262-4fd9-b506-8e9c0733e2d8", "external-id": "nsx-vlan-transportzone-119", "segmentation_id": 119, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc725ee28-d7", "ovs_interfaceid": "c725ee28-d7a8-4b31-92bc-f63174564f5e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61806) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 960.588664] env[61806]: DEBUG oslo_concurrency.lockutils [None req-60f88c72-8f2b-4f98-b3cc-7787383864bd tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Acquiring lock "refresh_cache-461f98f2-80ba-4088-9e7f-1ba2a9b0c1d9" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 960.588664] env[61806]: DEBUG oslo_concurrency.lockutils [None req-60f88c72-8f2b-4f98-b3cc-7787383864bd tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Acquired lock "refresh_cache-461f98f2-80ba-4088-9e7f-1ba2a9b0c1d9" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 960.588664] env[61806]: DEBUG nova.network.neutron [None req-60f88c72-8f2b-4f98-b3cc-7787383864bd tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] [instance: 461f98f2-80ba-4088-9e7f-1ba2a9b0c1d9] Building network info cache for instance {{(pid=61806) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 960.845825] env[61806]: DEBUG nova.objects.base [None req-03f49469-0c97-4bf3-b87e-03f023c054e5 tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Object Instance<380aa63f-46ba-437d-bc29-0fc141ad1cf3> lazy-loaded attributes: resources,numa_topology {{(pid=61806) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 960.848440] env[61806]: DEBUG nova.compute.manager [None req-c8cf9c01-54a9-4c86-b57f-b2e6a2fbc11a tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] [instance: 0bac3048-759a-47d9-a6ef-a0386a49a974] Start building block device mappings for instance. {{(pid=61806) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 960.862595] env[61806]: DEBUG nova.network.neutron [None req-c8cf9c01-54a9-4c86-b57f-b2e6a2fbc11a tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] [instance: 0bac3048-759a-47d9-a6ef-a0386a49a974] Successfully created port: 69c51913-810b-4b60-8249-2960056a8035 {{(pid=61806) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 960.873234] env[61806]: DEBUG oslo_concurrency.lockutils [None req-28d96f96-d23f-4c45-9d59-0355de5e0c4f tempest-ServersNegativeTestJSON-891856253 tempest-ServersNegativeTestJSON-891856253-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 960.994059] env[61806]: DEBUG oslo_concurrency.lockutils [req-8e83f681-fd89-4353-9e35-b5795b86bf49 req-9030c0ae-9b69-49c2-ab90-697249160cc1 service nova] Releasing lock "refresh_cache-d9caee4a-a00c-4e50-b01c-99a8f796a09f" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 961.041850] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5c524610-7933-4a31-aa8c-e736b47a09f6 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 961.051412] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-767323fb-1302-49c2-9a99-e01ebe323090 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 961.083215] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9bedbe46-89f4-42b8-9313-44643f437584 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 961.093352] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7ef05572-fc2e-49a2-87bd-0ac7826a4af4 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 961.108382] env[61806]: DEBUG nova.compute.provider_tree [None req-03f49469-0c97-4bf3-b87e-03f023c054e5 tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Inventory has not changed in ProviderTree for provider: a2858be1-fd22-4e08-979e-87ad25293407 {{(pid=61806) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 961.125451] env[61806]: DEBUG nova.network.neutron [None req-60f88c72-8f2b-4f98-b3cc-7787383864bd tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] [instance: 461f98f2-80ba-4088-9e7f-1ba2a9b0c1d9] Instance cache missing network info. {{(pid=61806) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 961.151685] env[61806]: DEBUG nova.network.neutron [None req-98c10467-8c15-41cc-ac16-68156711acc7 tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] [instance: d9caee4a-a00c-4e50-b01c-99a8f796a09f] Successfully updated port: 7e031e88-6928-472d-a65e-b818d89e6fc5 {{(pid=61806) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 961.266866] env[61806]: DEBUG oslo_concurrency.lockutils [None req-0bbe3134-73ea-4d1b-8035-8d9ceb5bbb4c tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Acquiring lock "676f18b3-3da1-4a53-b96d-fe64ee8f9101" by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 961.267292] env[61806]: DEBUG oslo_concurrency.lockutils [None req-0bbe3134-73ea-4d1b-8035-8d9ceb5bbb4c tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Lock "676f18b3-3da1-4a53-b96d-fe64ee8f9101" acquired by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" :: waited 0.001s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 961.267535] env[61806]: DEBUG nova.compute.manager [None req-0bbe3134-73ea-4d1b-8035-8d9ceb5bbb4c tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] [instance: 676f18b3-3da1-4a53-b96d-fe64ee8f9101] Going to confirm migration 1 {{(pid=61806) do_confirm_resize /opt/stack/nova/nova/compute/manager.py:4783}} [ 961.278746] env[61806]: DEBUG nova.network.neutron [None req-60f88c72-8f2b-4f98-b3cc-7787383864bd tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] [instance: 461f98f2-80ba-4088-9e7f-1ba2a9b0c1d9] Updating instance_info_cache with network_info: [{"id": "91000a2c-19c5-404d-90b3-367698085aec", "address": "fa:16:3e:48:c6:32", "network": {"id": "746a03b6-c1a6-4d27-98b9-0bee6376634f", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-1971382587-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c1f2d161c50f4b70987b90e6364c3779", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6a31b224-77d7-48a4-af87-312758611b8e", "external-id": "nsx-vlan-transportzone-761", "segmentation_id": 761, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap91000a2c-19", "ovs_interfaceid": "91000a2c-19c5-404d-90b3-367698085aec", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61806) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 961.367996] env[61806]: DEBUG nova.compute.manager [req-a94df28d-57df-4c2f-968c-32da8ade6429 req-f5e14f86-3fef-4306-9fa1-e8946dcbea88 service nova] [instance: d9caee4a-a00c-4e50-b01c-99a8f796a09f] Received event network-vif-plugged-7e031e88-6928-472d-a65e-b818d89e6fc5 {{(pid=61806) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 961.368274] env[61806]: DEBUG oslo_concurrency.lockutils [req-a94df28d-57df-4c2f-968c-32da8ade6429 req-f5e14f86-3fef-4306-9fa1-e8946dcbea88 service nova] Acquiring lock "d9caee4a-a00c-4e50-b01c-99a8f796a09f-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 961.368507] env[61806]: DEBUG oslo_concurrency.lockutils [req-a94df28d-57df-4c2f-968c-32da8ade6429 req-f5e14f86-3fef-4306-9fa1-e8946dcbea88 service nova] Lock "d9caee4a-a00c-4e50-b01c-99a8f796a09f-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 961.368651] env[61806]: DEBUG oslo_concurrency.lockutils [req-a94df28d-57df-4c2f-968c-32da8ade6429 req-f5e14f86-3fef-4306-9fa1-e8946dcbea88 service nova] Lock "d9caee4a-a00c-4e50-b01c-99a8f796a09f-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 961.368862] env[61806]: DEBUG nova.compute.manager [req-a94df28d-57df-4c2f-968c-32da8ade6429 req-f5e14f86-3fef-4306-9fa1-e8946dcbea88 service nova] [instance: d9caee4a-a00c-4e50-b01c-99a8f796a09f] No waiting events found dispatching network-vif-plugged-7e031e88-6928-472d-a65e-b818d89e6fc5 {{(pid=61806) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 961.368993] env[61806]: WARNING nova.compute.manager [req-a94df28d-57df-4c2f-968c-32da8ade6429 req-f5e14f86-3fef-4306-9fa1-e8946dcbea88 service nova] [instance: d9caee4a-a00c-4e50-b01c-99a8f796a09f] Received unexpected event network-vif-plugged-7e031e88-6928-472d-a65e-b818d89e6fc5 for instance with vm_state active and task_state None. [ 961.612020] env[61806]: DEBUG nova.scheduler.client.report [None req-03f49469-0c97-4bf3-b87e-03f023c054e5 tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Inventory has not changed for provider a2858be1-fd22-4e08-979e-87ad25293407 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 138, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61806) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 961.653597] env[61806]: DEBUG oslo_concurrency.lockutils [None req-98c10467-8c15-41cc-ac16-68156711acc7 tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] Acquiring lock "refresh_cache-d9caee4a-a00c-4e50-b01c-99a8f796a09f" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 961.653800] env[61806]: DEBUG oslo_concurrency.lockutils [None req-98c10467-8c15-41cc-ac16-68156711acc7 tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] Acquired lock "refresh_cache-d9caee4a-a00c-4e50-b01c-99a8f796a09f" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 961.653983] env[61806]: DEBUG nova.network.neutron [None req-98c10467-8c15-41cc-ac16-68156711acc7 tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] [instance: d9caee4a-a00c-4e50-b01c-99a8f796a09f] Building network info cache for instance {{(pid=61806) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 961.687517] env[61806]: DEBUG nova.network.neutron [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] [instance: 67d1bdd8-cf8f-4fb5-bd3c-42008e4d60e3] Updating instance_info_cache with network_info: [{"id": "859fb723-4a1c-4ea8-bd7c-0ed308f3a9e3", "address": "fa:16:3e:3c:40:d6", "network": {"id": "746a03b6-c1a6-4d27-98b9-0bee6376634f", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-1971382587-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.245", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c1f2d161c50f4b70987b90e6364c3779", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6a31b224-77d7-48a4-af87-312758611b8e", "external-id": "nsx-vlan-transportzone-761", "segmentation_id": 761, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap859fb723-4a", "ovs_interfaceid": "859fb723-4a1c-4ea8-bd7c-0ed308f3a9e3", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61806) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 961.780925] env[61806]: DEBUG oslo_concurrency.lockutils [None req-60f88c72-8f2b-4f98-b3cc-7787383864bd tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Releasing lock "refresh_cache-461f98f2-80ba-4088-9e7f-1ba2a9b0c1d9" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 961.781406] env[61806]: DEBUG nova.compute.manager [None req-60f88c72-8f2b-4f98-b3cc-7787383864bd tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] [instance: 461f98f2-80ba-4088-9e7f-1ba2a9b0c1d9] Instance network_info: |[{"id": "91000a2c-19c5-404d-90b3-367698085aec", "address": "fa:16:3e:48:c6:32", "network": {"id": "746a03b6-c1a6-4d27-98b9-0bee6376634f", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-1971382587-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c1f2d161c50f4b70987b90e6364c3779", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6a31b224-77d7-48a4-af87-312758611b8e", "external-id": "nsx-vlan-transportzone-761", "segmentation_id": 761, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap91000a2c-19", "ovs_interfaceid": "91000a2c-19c5-404d-90b3-367698085aec", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61806) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 961.781773] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-60f88c72-8f2b-4f98-b3cc-7787383864bd tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] [instance: 461f98f2-80ba-4088-9e7f-1ba2a9b0c1d9] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:48:c6:32', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '6a31b224-77d7-48a4-af87-312758611b8e', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '91000a2c-19c5-404d-90b3-367698085aec', 'vif_model': 'vmxnet3'}] {{(pid=61806) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 961.789371] env[61806]: DEBUG oslo.service.loopingcall [None req-60f88c72-8f2b-4f98-b3cc-7787383864bd tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61806) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 961.789595] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 461f98f2-80ba-4088-9e7f-1ba2a9b0c1d9] Creating VM on the ESX host {{(pid=61806) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 961.789836] env[61806]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-3221af73-35ed-432a-92b1-c2901e78d748 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 961.811819] env[61806]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 961.811819] env[61806]: value = "task-1294978" [ 961.811819] env[61806]: _type = "Task" [ 961.811819] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 961.820747] env[61806]: DEBUG oslo_vmware.api [-] Task: {'id': task-1294978, 'name': CreateVM_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 961.843358] env[61806]: DEBUG nova.compute.manager [req-1341140b-d7b1-400f-a3bd-674eab7e3728 req-ce73f33a-6adf-4f46-9417-5ed97249320c service nova] [instance: 461f98f2-80ba-4088-9e7f-1ba2a9b0c1d9] Received event network-vif-plugged-91000a2c-19c5-404d-90b3-367698085aec {{(pid=61806) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 961.843581] env[61806]: DEBUG oslo_concurrency.lockutils [req-1341140b-d7b1-400f-a3bd-674eab7e3728 req-ce73f33a-6adf-4f46-9417-5ed97249320c service nova] Acquiring lock "461f98f2-80ba-4088-9e7f-1ba2a9b0c1d9-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 961.843839] env[61806]: DEBUG oslo_concurrency.lockutils [req-1341140b-d7b1-400f-a3bd-674eab7e3728 req-ce73f33a-6adf-4f46-9417-5ed97249320c service nova] Lock "461f98f2-80ba-4088-9e7f-1ba2a9b0c1d9-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 961.843983] env[61806]: DEBUG oslo_concurrency.lockutils [req-1341140b-d7b1-400f-a3bd-674eab7e3728 req-ce73f33a-6adf-4f46-9417-5ed97249320c service nova] Lock "461f98f2-80ba-4088-9e7f-1ba2a9b0c1d9-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 961.844477] env[61806]: DEBUG nova.compute.manager [req-1341140b-d7b1-400f-a3bd-674eab7e3728 req-ce73f33a-6adf-4f46-9417-5ed97249320c service nova] [instance: 461f98f2-80ba-4088-9e7f-1ba2a9b0c1d9] No waiting events found dispatching network-vif-plugged-91000a2c-19c5-404d-90b3-367698085aec {{(pid=61806) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 961.844713] env[61806]: WARNING nova.compute.manager [req-1341140b-d7b1-400f-a3bd-674eab7e3728 req-ce73f33a-6adf-4f46-9417-5ed97249320c service nova] [instance: 461f98f2-80ba-4088-9e7f-1ba2a9b0c1d9] Received unexpected event network-vif-plugged-91000a2c-19c5-404d-90b3-367698085aec for instance with vm_state building and task_state spawning. [ 961.844896] env[61806]: DEBUG nova.compute.manager [req-1341140b-d7b1-400f-a3bd-674eab7e3728 req-ce73f33a-6adf-4f46-9417-5ed97249320c service nova] [instance: 461f98f2-80ba-4088-9e7f-1ba2a9b0c1d9] Received event network-changed-91000a2c-19c5-404d-90b3-367698085aec {{(pid=61806) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 961.845102] env[61806]: DEBUG nova.compute.manager [req-1341140b-d7b1-400f-a3bd-674eab7e3728 req-ce73f33a-6adf-4f46-9417-5ed97249320c service nova] [instance: 461f98f2-80ba-4088-9e7f-1ba2a9b0c1d9] Refreshing instance network info cache due to event network-changed-91000a2c-19c5-404d-90b3-367698085aec. {{(pid=61806) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 961.845326] env[61806]: DEBUG oslo_concurrency.lockutils [req-1341140b-d7b1-400f-a3bd-674eab7e3728 req-ce73f33a-6adf-4f46-9417-5ed97249320c service nova] Acquiring lock "refresh_cache-461f98f2-80ba-4088-9e7f-1ba2a9b0c1d9" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 961.845476] env[61806]: DEBUG oslo_concurrency.lockutils [req-1341140b-d7b1-400f-a3bd-674eab7e3728 req-ce73f33a-6adf-4f46-9417-5ed97249320c service nova] Acquired lock "refresh_cache-461f98f2-80ba-4088-9e7f-1ba2a9b0c1d9" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 961.845643] env[61806]: DEBUG nova.network.neutron [req-1341140b-d7b1-400f-a3bd-674eab7e3728 req-ce73f33a-6adf-4f46-9417-5ed97249320c service nova] [instance: 461f98f2-80ba-4088-9e7f-1ba2a9b0c1d9] Refreshing network info cache for port 91000a2c-19c5-404d-90b3-367698085aec {{(pid=61806) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 961.858889] env[61806]: DEBUG oslo_concurrency.lockutils [None req-0bbe3134-73ea-4d1b-8035-8d9ceb5bbb4c tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Acquiring lock "refresh_cache-676f18b3-3da1-4a53-b96d-fe64ee8f9101" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 961.859101] env[61806]: DEBUG oslo_concurrency.lockutils [None req-0bbe3134-73ea-4d1b-8035-8d9ceb5bbb4c tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Acquired lock "refresh_cache-676f18b3-3da1-4a53-b96d-fe64ee8f9101" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 961.859306] env[61806]: DEBUG nova.network.neutron [None req-0bbe3134-73ea-4d1b-8035-8d9ceb5bbb4c tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] [instance: 676f18b3-3da1-4a53-b96d-fe64ee8f9101] Building network info cache for instance {{(pid=61806) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 961.859495] env[61806]: DEBUG nova.objects.instance [None req-0bbe3134-73ea-4d1b-8035-8d9ceb5bbb4c tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Lazy-loading 'info_cache' on Instance uuid 676f18b3-3da1-4a53-b96d-fe64ee8f9101 {{(pid=61806) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 961.861734] env[61806]: DEBUG nova.compute.manager [None req-c8cf9c01-54a9-4c86-b57f-b2e6a2fbc11a tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] [instance: 0bac3048-759a-47d9-a6ef-a0386a49a974] Start spawning the instance on the hypervisor. {{(pid=61806) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 961.889684] env[61806]: DEBUG nova.virt.hardware [None req-c8cf9c01-54a9-4c86-b57f-b2e6a2fbc11a tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-15T18:05:10Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-15T18:04:52Z,direct_url=,disk_format='vmdk',id=ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='84ba83fce2c349a6988173c5d6fc3b07',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-15T18:04:53Z,virtual_size=,visibility=), allow threads: False {{(pid=61806) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 961.889961] env[61806]: DEBUG nova.virt.hardware [None req-c8cf9c01-54a9-4c86-b57f-b2e6a2fbc11a tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] Flavor limits 0:0:0 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 961.890137] env[61806]: DEBUG nova.virt.hardware [None req-c8cf9c01-54a9-4c86-b57f-b2e6a2fbc11a tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] Image limits 0:0:0 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 961.890331] env[61806]: DEBUG nova.virt.hardware [None req-c8cf9c01-54a9-4c86-b57f-b2e6a2fbc11a tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] Flavor pref 0:0:0 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 961.890528] env[61806]: DEBUG nova.virt.hardware [None req-c8cf9c01-54a9-4c86-b57f-b2e6a2fbc11a tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] Image pref 0:0:0 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 961.890688] env[61806]: DEBUG nova.virt.hardware [None req-c8cf9c01-54a9-4c86-b57f-b2e6a2fbc11a tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 961.890915] env[61806]: DEBUG nova.virt.hardware [None req-c8cf9c01-54a9-4c86-b57f-b2e6a2fbc11a tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61806) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 961.891095] env[61806]: DEBUG nova.virt.hardware [None req-c8cf9c01-54a9-4c86-b57f-b2e6a2fbc11a tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61806) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 961.891314] env[61806]: DEBUG nova.virt.hardware [None req-c8cf9c01-54a9-4c86-b57f-b2e6a2fbc11a tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] Got 1 possible topologies {{(pid=61806) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 961.891495] env[61806]: DEBUG nova.virt.hardware [None req-c8cf9c01-54a9-4c86-b57f-b2e6a2fbc11a tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61806) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 961.891677] env[61806]: DEBUG nova.virt.hardware [None req-c8cf9c01-54a9-4c86-b57f-b2e6a2fbc11a tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61806) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 961.892624] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f2af7a36-dd7f-484d-afb4-25970fb564df {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 961.901305] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-63b1739e-71d3-487c-b626-659678e20340 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 962.116558] env[61806]: DEBUG oslo_concurrency.lockutils [None req-03f49469-0c97-4bf3-b87e-03f023c054e5 tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.287s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 962.119790] env[61806]: DEBUG oslo_concurrency.lockutils [None req-2c70fcbb-efa7-459d-81c0-a2a8a184893c tempest-AttachVolumeNegativeTest-1683033704 tempest-AttachVolumeNegativeTest-1683033704-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 2.808s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 962.121495] env[61806]: INFO nova.compute.claims [None req-2c70fcbb-efa7-459d-81c0-a2a8a184893c tempest-AttachVolumeNegativeTest-1683033704 tempest-AttachVolumeNegativeTest-1683033704-project-member] [instance: e4638faa-7be1-4909-a595-c437837fa314] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 962.189927] env[61806]: DEBUG oslo_concurrency.lockutils [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Releasing lock "refresh_cache-67d1bdd8-cf8f-4fb5-bd3c-42008e4d60e3" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 962.190517] env[61806]: DEBUG nova.compute.manager [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] [instance: 67d1bdd8-cf8f-4fb5-bd3c-42008e4d60e3] Updated the network info_cache for instance {{(pid=61806) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9997}} [ 962.191638] env[61806]: WARNING nova.network.neutron [None req-98c10467-8c15-41cc-ac16-68156711acc7 tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] [instance: d9caee4a-a00c-4e50-b01c-99a8f796a09f] a7ffda72-e57c-41fa-a7ed-d7b12084bc9d already exists in list: networks containing: ['a7ffda72-e57c-41fa-a7ed-d7b12084bc9d']. ignoring it [ 962.193748] env[61806]: DEBUG oslo_service.periodic_task [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=61806) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 962.194016] env[61806]: DEBUG oslo_service.periodic_task [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=61806) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 962.194692] env[61806]: DEBUG oslo_service.periodic_task [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=61806) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 962.194926] env[61806]: DEBUG oslo_service.periodic_task [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=61806) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 962.195199] env[61806]: DEBUG oslo_service.periodic_task [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=61806) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 962.195467] env[61806]: DEBUG oslo_service.periodic_task [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=61806) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 962.195634] env[61806]: DEBUG nova.compute.manager [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=61806) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10545}} [ 962.195796] env[61806]: DEBUG oslo_service.periodic_task [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Running periodic task ComputeManager.update_available_resource {{(pid=61806) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 962.324017] env[61806]: DEBUG oslo_vmware.api [-] Task: {'id': task-1294978, 'name': CreateVM_Task, 'duration_secs': 0.407869} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 962.324017] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 461f98f2-80ba-4088-9e7f-1ba2a9b0c1d9] Created VM on the ESX host {{(pid=61806) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 962.324017] env[61806]: DEBUG oslo_concurrency.lockutils [None req-60f88c72-8f2b-4f98-b3cc-7787383864bd tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 962.324292] env[61806]: DEBUG oslo_concurrency.lockutils [None req-60f88c72-8f2b-4f98-b3cc-7787383864bd tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Acquired lock "[datastore2] devstack-image-cache_base/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 962.324584] env[61806]: DEBUG oslo_concurrency.lockutils [None req-60f88c72-8f2b-4f98-b3cc-7787383864bd tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 962.324886] env[61806]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-7dc46440-bbbe-4b91-a676-f1d38008b007 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 962.329833] env[61806]: DEBUG oslo_vmware.api [None req-60f88c72-8f2b-4f98-b3cc-7787383864bd tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Waiting for the task: (returnval){ [ 962.329833] env[61806]: value = "session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]52fe1bc9-1fb0-4fe4-54ad-e57d54f4dc1b" [ 962.329833] env[61806]: _type = "Task" [ 962.329833] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 962.338270] env[61806]: DEBUG oslo_vmware.api [None req-60f88c72-8f2b-4f98-b3cc-7787383864bd tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Task: {'id': session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]52fe1bc9-1fb0-4fe4-54ad-e57d54f4dc1b, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 962.470872] env[61806]: DEBUG nova.network.neutron [None req-98c10467-8c15-41cc-ac16-68156711acc7 tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] [instance: d9caee4a-a00c-4e50-b01c-99a8f796a09f] Updating instance_info_cache with network_info: [{"id": "c725ee28-d7a8-4b31-92bc-f63174564f5e", "address": "fa:16:3e:3a:18:8e", "network": {"id": "a7ffda72-e57c-41fa-a7ed-d7b12084bc9d", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-62794865-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.218", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "6e391e1267014fafbb8f5a3211299819", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8b3cfeb1-f262-4fd9-b506-8e9c0733e2d8", "external-id": "nsx-vlan-transportzone-119", "segmentation_id": 119, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc725ee28-d7", "ovs_interfaceid": "c725ee28-d7a8-4b31-92bc-f63174564f5e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "7e031e88-6928-472d-a65e-b818d89e6fc5", "address": "fa:16:3e:1e:a6:b3", "network": {"id": "a7ffda72-e57c-41fa-a7ed-d7b12084bc9d", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-62794865-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "6e391e1267014fafbb8f5a3211299819", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8b3cfeb1-f262-4fd9-b506-8e9c0733e2d8", "external-id": "nsx-vlan-transportzone-119", "segmentation_id": 119, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7e031e88-69", "ovs_interfaceid": "7e031e88-6928-472d-a65e-b818d89e6fc5", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": true, "delegate_create": true, "meta": {}}] {{(pid=61806) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 962.633913] env[61806]: DEBUG oslo_concurrency.lockutils [None req-03f49469-0c97-4bf3-b87e-03f023c054e5 tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Lock "380aa63f-46ba-437d-bc29-0fc141ad1cf3" "released" by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" :: held 26.187s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 962.634683] env[61806]: DEBUG oslo_concurrency.lockutils [None req-d1459cd5-a70e-4111-a41f-72b97e54aa17 tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Lock "380aa63f-46ba-437d-bc29-0fc141ad1cf3" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 5.364s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 962.634902] env[61806]: DEBUG oslo_concurrency.lockutils [None req-d1459cd5-a70e-4111-a41f-72b97e54aa17 tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Acquiring lock "380aa63f-46ba-437d-bc29-0fc141ad1cf3-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 962.635137] env[61806]: DEBUG oslo_concurrency.lockutils [None req-d1459cd5-a70e-4111-a41f-72b97e54aa17 tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Lock "380aa63f-46ba-437d-bc29-0fc141ad1cf3-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 962.635333] env[61806]: DEBUG oslo_concurrency.lockutils [None req-d1459cd5-a70e-4111-a41f-72b97e54aa17 tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Lock "380aa63f-46ba-437d-bc29-0fc141ad1cf3-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 962.636818] env[61806]: INFO nova.compute.manager [None req-d1459cd5-a70e-4111-a41f-72b97e54aa17 tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] [instance: 380aa63f-46ba-437d-bc29-0fc141ad1cf3] Terminating instance [ 962.640786] env[61806]: DEBUG nova.compute.manager [None req-d1459cd5-a70e-4111-a41f-72b97e54aa17 tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] [instance: 380aa63f-46ba-437d-bc29-0fc141ad1cf3] Start destroying the instance on the hypervisor. {{(pid=61806) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 962.641023] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-d1459cd5-a70e-4111-a41f-72b97e54aa17 tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] [instance: 380aa63f-46ba-437d-bc29-0fc141ad1cf3] Destroying instance {{(pid=61806) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 962.641347] env[61806]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-9098ad1a-1bd5-4414-9283-6d9500d85a5c {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 962.652451] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-46f24c2b-b40d-4fee-9da9-7a0d57e26f9b {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 962.681587] env[61806]: WARNING nova.virt.vmwareapi.vmops [None req-d1459cd5-a70e-4111-a41f-72b97e54aa17 tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] [instance: 380aa63f-46ba-437d-bc29-0fc141ad1cf3] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 380aa63f-46ba-437d-bc29-0fc141ad1cf3 could not be found. [ 962.681807] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-d1459cd5-a70e-4111-a41f-72b97e54aa17 tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] [instance: 380aa63f-46ba-437d-bc29-0fc141ad1cf3] Instance destroyed {{(pid=61806) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 962.681984] env[61806]: INFO nova.compute.manager [None req-d1459cd5-a70e-4111-a41f-72b97e54aa17 tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] [instance: 380aa63f-46ba-437d-bc29-0fc141ad1cf3] Took 0.04 seconds to destroy the instance on the hypervisor. [ 962.682246] env[61806]: DEBUG oslo.service.loopingcall [None req-d1459cd5-a70e-4111-a41f-72b97e54aa17 tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61806) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 962.682549] env[61806]: DEBUG nova.compute.manager [-] [instance: 380aa63f-46ba-437d-bc29-0fc141ad1cf3] Deallocating network for instance {{(pid=61806) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 962.682647] env[61806]: DEBUG nova.network.neutron [-] [instance: 380aa63f-46ba-437d-bc29-0fc141ad1cf3] deallocate_for_instance() {{(pid=61806) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 962.698611] env[61806]: DEBUG oslo_concurrency.lockutils [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 962.727700] env[61806]: DEBUG nova.network.neutron [req-1341140b-d7b1-400f-a3bd-674eab7e3728 req-ce73f33a-6adf-4f46-9417-5ed97249320c service nova] [instance: 461f98f2-80ba-4088-9e7f-1ba2a9b0c1d9] Updated VIF entry in instance network info cache for port 91000a2c-19c5-404d-90b3-367698085aec. {{(pid=61806) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 962.728036] env[61806]: DEBUG nova.network.neutron [req-1341140b-d7b1-400f-a3bd-674eab7e3728 req-ce73f33a-6adf-4f46-9417-5ed97249320c service nova] [instance: 461f98f2-80ba-4088-9e7f-1ba2a9b0c1d9] Updating instance_info_cache with network_info: [{"id": "91000a2c-19c5-404d-90b3-367698085aec", "address": "fa:16:3e:48:c6:32", "network": {"id": "746a03b6-c1a6-4d27-98b9-0bee6376634f", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-1971382587-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c1f2d161c50f4b70987b90e6364c3779", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6a31b224-77d7-48a4-af87-312758611b8e", "external-id": "nsx-vlan-transportzone-761", "segmentation_id": 761, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap91000a2c-19", "ovs_interfaceid": "91000a2c-19c5-404d-90b3-367698085aec", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61806) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 962.802411] env[61806]: DEBUG nova.network.neutron [None req-c8cf9c01-54a9-4c86-b57f-b2e6a2fbc11a tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] [instance: 0bac3048-759a-47d9-a6ef-a0386a49a974] Successfully updated port: 69c51913-810b-4b60-8249-2960056a8035 {{(pid=61806) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 962.840620] env[61806]: DEBUG oslo_vmware.api [None req-60f88c72-8f2b-4f98-b3cc-7787383864bd tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Task: {'id': session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]52fe1bc9-1fb0-4fe4-54ad-e57d54f4dc1b, 'name': SearchDatastore_Task, 'duration_secs': 0.039543} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 962.840950] env[61806]: DEBUG oslo_concurrency.lockutils [None req-60f88c72-8f2b-4f98-b3cc-7787383864bd tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Releasing lock "[datastore2] devstack-image-cache_base/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 962.841289] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-60f88c72-8f2b-4f98-b3cc-7787383864bd tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] [instance: 461f98f2-80ba-4088-9e7f-1ba2a9b0c1d9] Processing image ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa {{(pid=61806) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 962.841549] env[61806]: DEBUG oslo_concurrency.lockutils [None req-60f88c72-8f2b-4f98-b3cc-7787383864bd tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa.vmdk" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 962.841703] env[61806]: DEBUG oslo_concurrency.lockutils [None req-60f88c72-8f2b-4f98-b3cc-7787383864bd tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Acquired lock "[datastore2] devstack-image-cache_base/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa.vmdk" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 962.841886] env[61806]: DEBUG nova.virt.vmwareapi.ds_util [None req-60f88c72-8f2b-4f98-b3cc-7787383864bd tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61806) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 962.842172] env[61806]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-c2498922-c70e-4e18-a8d8-236eaeda3ed8 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 962.851664] env[61806]: DEBUG nova.virt.vmwareapi.ds_util [None req-60f88c72-8f2b-4f98-b3cc-7787383864bd tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61806) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 962.851873] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-60f88c72-8f2b-4f98-b3cc-7787383864bd tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61806) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 962.852674] env[61806]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-48d8e83f-e99e-48e6-91f8-e8c08dd190af {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 962.859212] env[61806]: DEBUG oslo_vmware.api [None req-60f88c72-8f2b-4f98-b3cc-7787383864bd tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Waiting for the task: (returnval){ [ 962.859212] env[61806]: value = "session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]52553882-227c-2bda-29b8-2dfdd02107b9" [ 962.859212] env[61806]: _type = "Task" [ 962.859212] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 962.867855] env[61806]: DEBUG oslo_vmware.api [None req-60f88c72-8f2b-4f98-b3cc-7787383864bd tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Task: {'id': session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]52553882-227c-2bda-29b8-2dfdd02107b9, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 962.974024] env[61806]: DEBUG oslo_concurrency.lockutils [None req-98c10467-8c15-41cc-ac16-68156711acc7 tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] Releasing lock "refresh_cache-d9caee4a-a00c-4e50-b01c-99a8f796a09f" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 962.974777] env[61806]: DEBUG oslo_concurrency.lockutils [None req-98c10467-8c15-41cc-ac16-68156711acc7 tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] Acquiring lock "d9caee4a-a00c-4e50-b01c-99a8f796a09f" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 962.974945] env[61806]: DEBUG oslo_concurrency.lockutils [None req-98c10467-8c15-41cc-ac16-68156711acc7 tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] Acquired lock "d9caee4a-a00c-4e50-b01c-99a8f796a09f" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 962.975823] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a7fb8710-7ae9-4bae-afcc-7709f54aa836 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 962.996153] env[61806]: DEBUG nova.virt.hardware [None req-98c10467-8c15-41cc-ac16-68156711acc7 tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-15T18:05:10Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format=,created_at=,direct_url=,disk_format=,id=,min_disk=,min_ram=,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=61806) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 962.996481] env[61806]: DEBUG nova.virt.hardware [None req-98c10467-8c15-41cc-ac16-68156711acc7 tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] Flavor limits 0:0:0 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 962.996632] env[61806]: DEBUG nova.virt.hardware [None req-98c10467-8c15-41cc-ac16-68156711acc7 tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] Image limits 0:0:0 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 962.996822] env[61806]: DEBUG nova.virt.hardware [None req-98c10467-8c15-41cc-ac16-68156711acc7 tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] Flavor pref 0:0:0 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 962.996973] env[61806]: DEBUG nova.virt.hardware [None req-98c10467-8c15-41cc-ac16-68156711acc7 tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] Image pref 0:0:0 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 962.997142] env[61806]: DEBUG nova.virt.hardware [None req-98c10467-8c15-41cc-ac16-68156711acc7 tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 962.997358] env[61806]: DEBUG nova.virt.hardware [None req-98c10467-8c15-41cc-ac16-68156711acc7 tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61806) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 962.997523] env[61806]: DEBUG nova.virt.hardware [None req-98c10467-8c15-41cc-ac16-68156711acc7 tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61806) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 962.997696] env[61806]: DEBUG nova.virt.hardware [None req-98c10467-8c15-41cc-ac16-68156711acc7 tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] Got 1 possible topologies {{(pid=61806) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 962.997865] env[61806]: DEBUG nova.virt.hardware [None req-98c10467-8c15-41cc-ac16-68156711acc7 tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61806) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 962.998060] env[61806]: DEBUG nova.virt.hardware [None req-98c10467-8c15-41cc-ac16-68156711acc7 tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61806) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 963.004367] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-98c10467-8c15-41cc-ac16-68156711acc7 tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] [instance: d9caee4a-a00c-4e50-b01c-99a8f796a09f] Reconfiguring VM to attach interface {{(pid=61806) attach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1929}} [ 963.006721] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-f32ebf89-06fd-4814-a359-e5424ad7f1ee {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 963.025688] env[61806]: DEBUG oslo_vmware.api [None req-98c10467-8c15-41cc-ac16-68156711acc7 tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] Waiting for the task: (returnval){ [ 963.025688] env[61806]: value = "task-1294979" [ 963.025688] env[61806]: _type = "Task" [ 963.025688] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 963.033672] env[61806]: DEBUG oslo_vmware.api [None req-98c10467-8c15-41cc-ac16-68156711acc7 tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] Task: {'id': task-1294979, 'name': ReconfigVM_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 963.185443] env[61806]: DEBUG nova.network.neutron [None req-0bbe3134-73ea-4d1b-8035-8d9ceb5bbb4c tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] [instance: 676f18b3-3da1-4a53-b96d-fe64ee8f9101] Updating instance_info_cache with network_info: [{"id": "d3a921a7-f1fa-4352-ae53-9f2996576fef", "address": "fa:16:3e:66:40:96", "network": {"id": "46515a4e-610f-42fb-a424-205a03df0640", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-1133623547-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4b9d94c3d684474b99e305fb6cd9b95f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "82ca17df-257e-40e6-9ec9-310ed6f05ccb", "external-id": "nsx-vlan-transportzone-778", "segmentation_id": 778, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd3a921a7-f1", "ovs_interfaceid": "d3a921a7-f1fa-4352-ae53-9f2996576fef", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61806) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 963.230446] env[61806]: DEBUG oslo_concurrency.lockutils [req-1341140b-d7b1-400f-a3bd-674eab7e3728 req-ce73f33a-6adf-4f46-9417-5ed97249320c service nova] Releasing lock "refresh_cache-461f98f2-80ba-4088-9e7f-1ba2a9b0c1d9" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 963.304833] env[61806]: DEBUG oslo_concurrency.lockutils [None req-c8cf9c01-54a9-4c86-b57f-b2e6a2fbc11a tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] Acquiring lock "refresh_cache-0bac3048-759a-47d9-a6ef-a0386a49a974" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 963.304833] env[61806]: DEBUG oslo_concurrency.lockutils [None req-c8cf9c01-54a9-4c86-b57f-b2e6a2fbc11a tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] Acquired lock "refresh_cache-0bac3048-759a-47d9-a6ef-a0386a49a974" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 963.304833] env[61806]: DEBUG nova.network.neutron [None req-c8cf9c01-54a9-4c86-b57f-b2e6a2fbc11a tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] [instance: 0bac3048-759a-47d9-a6ef-a0386a49a974] Building network info cache for instance {{(pid=61806) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 963.314679] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c36925e2-b717-4f09-9a7a-f5016355879e {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 963.325804] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9452fb17-ccad-493b-a1ce-ec10928972c9 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 963.356751] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-efa38d0d-9d89-4782-953f-09f7b034d6c4 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 963.365784] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9a7c47ff-d01a-4e76-9c29-1a641671602c {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 963.373072] env[61806]: DEBUG oslo_vmware.api [None req-60f88c72-8f2b-4f98-b3cc-7787383864bd tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Task: {'id': session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]52553882-227c-2bda-29b8-2dfdd02107b9, 'name': SearchDatastore_Task, 'duration_secs': 0.0323} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 963.374161] env[61806]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-6437e259-474d-449a-b20f-9823c10c3e06 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 963.383842] env[61806]: DEBUG nova.compute.provider_tree [None req-2c70fcbb-efa7-459d-81c0-a2a8a184893c tempest-AttachVolumeNegativeTest-1683033704 tempest-AttachVolumeNegativeTest-1683033704-project-member] Inventory has not changed in ProviderTree for provider: a2858be1-fd22-4e08-979e-87ad25293407 {{(pid=61806) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 963.388284] env[61806]: DEBUG oslo_vmware.api [None req-60f88c72-8f2b-4f98-b3cc-7787383864bd tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Waiting for the task: (returnval){ [ 963.388284] env[61806]: value = "session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]5227e201-4947-6826-f2ff-54748702bbee" [ 963.388284] env[61806]: _type = "Task" [ 963.388284] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 963.397311] env[61806]: DEBUG oslo_vmware.api [None req-60f88c72-8f2b-4f98-b3cc-7787383864bd tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Task: {'id': session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]5227e201-4947-6826-f2ff-54748702bbee, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 963.413012] env[61806]: DEBUG nova.network.neutron [-] [instance: 380aa63f-46ba-437d-bc29-0fc141ad1cf3] Updating instance_info_cache with network_info: [] {{(pid=61806) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 963.474043] env[61806]: DEBUG nova.compute.manager [req-2ecdccfe-9d5e-42f9-a822-f7cab000e416 req-5f40af83-cd56-4043-a272-2f3fcc5ea3ec service nova] [instance: d9caee4a-a00c-4e50-b01c-99a8f796a09f] Received event network-changed-7e031e88-6928-472d-a65e-b818d89e6fc5 {{(pid=61806) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 963.474339] env[61806]: DEBUG nova.compute.manager [req-2ecdccfe-9d5e-42f9-a822-f7cab000e416 req-5f40af83-cd56-4043-a272-2f3fcc5ea3ec service nova] [instance: d9caee4a-a00c-4e50-b01c-99a8f796a09f] Refreshing instance network info cache due to event network-changed-7e031e88-6928-472d-a65e-b818d89e6fc5. {{(pid=61806) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 963.474613] env[61806]: DEBUG oslo_concurrency.lockutils [req-2ecdccfe-9d5e-42f9-a822-f7cab000e416 req-5f40af83-cd56-4043-a272-2f3fcc5ea3ec service nova] Acquiring lock "refresh_cache-d9caee4a-a00c-4e50-b01c-99a8f796a09f" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 963.474832] env[61806]: DEBUG oslo_concurrency.lockutils [req-2ecdccfe-9d5e-42f9-a822-f7cab000e416 req-5f40af83-cd56-4043-a272-2f3fcc5ea3ec service nova] Acquired lock "refresh_cache-d9caee4a-a00c-4e50-b01c-99a8f796a09f" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 963.475062] env[61806]: DEBUG nova.network.neutron [req-2ecdccfe-9d5e-42f9-a822-f7cab000e416 req-5f40af83-cd56-4043-a272-2f3fcc5ea3ec service nova] [instance: d9caee4a-a00c-4e50-b01c-99a8f796a09f] Refreshing network info cache for port 7e031e88-6928-472d-a65e-b818d89e6fc5 {{(pid=61806) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 963.535761] env[61806]: DEBUG oslo_vmware.api [None req-98c10467-8c15-41cc-ac16-68156711acc7 tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] Task: {'id': task-1294979, 'name': ReconfigVM_Task} progress is 99%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 963.690781] env[61806]: DEBUG oslo_concurrency.lockutils [None req-0bbe3134-73ea-4d1b-8035-8d9ceb5bbb4c tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Releasing lock "refresh_cache-676f18b3-3da1-4a53-b96d-fe64ee8f9101" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 963.691142] env[61806]: DEBUG nova.objects.instance [None req-0bbe3134-73ea-4d1b-8035-8d9ceb5bbb4c tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Lazy-loading 'migration_context' on Instance uuid 676f18b3-3da1-4a53-b96d-fe64ee8f9101 {{(pid=61806) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 963.834871] env[61806]: DEBUG nova.network.neutron [None req-c8cf9c01-54a9-4c86-b57f-b2e6a2fbc11a tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] [instance: 0bac3048-759a-47d9-a6ef-a0386a49a974] Instance cache missing network info. {{(pid=61806) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 963.889684] env[61806]: DEBUG nova.scheduler.client.report [None req-2c70fcbb-efa7-459d-81c0-a2a8a184893c tempest-AttachVolumeNegativeTest-1683033704 tempest-AttachVolumeNegativeTest-1683033704-project-member] Inventory has not changed for provider a2858be1-fd22-4e08-979e-87ad25293407 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 138, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61806) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 963.905554] env[61806]: DEBUG oslo_vmware.api [None req-60f88c72-8f2b-4f98-b3cc-7787383864bd tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Task: {'id': session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]5227e201-4947-6826-f2ff-54748702bbee, 'name': SearchDatastore_Task, 'duration_secs': 0.010783} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 963.905831] env[61806]: DEBUG oslo_concurrency.lockutils [None req-60f88c72-8f2b-4f98-b3cc-7787383864bd tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Releasing lock "[datastore2] devstack-image-cache_base/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa.vmdk" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 963.906102] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [None req-60f88c72-8f2b-4f98-b3cc-7787383864bd tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa.vmdk to [datastore2] 461f98f2-80ba-4088-9e7f-1ba2a9b0c1d9/461f98f2-80ba-4088-9e7f-1ba2a9b0c1d9.vmdk {{(pid=61806) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 963.906363] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-3dc0448a-8225-4982-9ebe-137e080f1900 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 963.915039] env[61806]: INFO nova.compute.manager [-] [instance: 380aa63f-46ba-437d-bc29-0fc141ad1cf3] Took 1.23 seconds to deallocate network for instance. [ 963.915404] env[61806]: DEBUG oslo_vmware.api [None req-60f88c72-8f2b-4f98-b3cc-7787383864bd tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Waiting for the task: (returnval){ [ 963.915404] env[61806]: value = "task-1294980" [ 963.915404] env[61806]: _type = "Task" [ 963.915404] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 963.928953] env[61806]: DEBUG oslo_vmware.api [None req-60f88c72-8f2b-4f98-b3cc-7787383864bd tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Task: {'id': task-1294980, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 963.969069] env[61806]: DEBUG nova.network.neutron [None req-c8cf9c01-54a9-4c86-b57f-b2e6a2fbc11a tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] [instance: 0bac3048-759a-47d9-a6ef-a0386a49a974] Updating instance_info_cache with network_info: [{"id": "69c51913-810b-4b60-8249-2960056a8035", "address": "fa:16:3e:a9:9e:06", "network": {"id": "f280a01b-7000-4e54-ab9a-20eab156a410", "bridge": "br-int", "label": "tempest-ServersTestJSON-2080718146-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "73ebcc4d79d248efb653a084a5e44302", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "43798f54-0c07-4417-a23f-58bb6b7e204b", "external-id": "nsx-vlan-transportzone-571", "segmentation_id": 571, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap69c51913-81", "ovs_interfaceid": "69c51913-810b-4b60-8249-2960056a8035", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61806) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 964.035898] env[61806]: DEBUG oslo_vmware.api [None req-98c10467-8c15-41cc-ac16-68156711acc7 tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] Task: {'id': task-1294979, 'name': ReconfigVM_Task, 'duration_secs': 0.566961} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 964.036465] env[61806]: DEBUG oslo_concurrency.lockutils [None req-98c10467-8c15-41cc-ac16-68156711acc7 tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] Releasing lock "d9caee4a-a00c-4e50-b01c-99a8f796a09f" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 964.036655] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-98c10467-8c15-41cc-ac16-68156711acc7 tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] [instance: d9caee4a-a00c-4e50-b01c-99a8f796a09f] Reconfigured VM to attach interface {{(pid=61806) attach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1943}} [ 964.194190] env[61806]: DEBUG nova.objects.base [None req-0bbe3134-73ea-4d1b-8035-8d9ceb5bbb4c tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Object Instance<676f18b3-3da1-4a53-b96d-fe64ee8f9101> lazy-loaded attributes: info_cache,migration_context {{(pid=61806) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 964.195217] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b813ddc5-0216-4734-a274-6c0c5e6992ec {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 964.202712] env[61806]: DEBUG nova.network.neutron [req-2ecdccfe-9d5e-42f9-a822-f7cab000e416 req-5f40af83-cd56-4043-a272-2f3fcc5ea3ec service nova] [instance: d9caee4a-a00c-4e50-b01c-99a8f796a09f] Updated VIF entry in instance network info cache for port 7e031e88-6928-472d-a65e-b818d89e6fc5. {{(pid=61806) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 964.203165] env[61806]: DEBUG nova.network.neutron [req-2ecdccfe-9d5e-42f9-a822-f7cab000e416 req-5f40af83-cd56-4043-a272-2f3fcc5ea3ec service nova] [instance: d9caee4a-a00c-4e50-b01c-99a8f796a09f] Updating instance_info_cache with network_info: [{"id": "c725ee28-d7a8-4b31-92bc-f63174564f5e", "address": "fa:16:3e:3a:18:8e", "network": {"id": "a7ffda72-e57c-41fa-a7ed-d7b12084bc9d", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-62794865-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.218", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "6e391e1267014fafbb8f5a3211299819", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8b3cfeb1-f262-4fd9-b506-8e9c0733e2d8", "external-id": "nsx-vlan-transportzone-119", "segmentation_id": 119, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc725ee28-d7", "ovs_interfaceid": "c725ee28-d7a8-4b31-92bc-f63174564f5e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "7e031e88-6928-472d-a65e-b818d89e6fc5", "address": "fa:16:3e:1e:a6:b3", "network": {"id": "a7ffda72-e57c-41fa-a7ed-d7b12084bc9d", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-62794865-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "6e391e1267014fafbb8f5a3211299819", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8b3cfeb1-f262-4fd9-b506-8e9c0733e2d8", "external-id": "nsx-vlan-transportzone-119", "segmentation_id": 119, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7e031e88-69", "ovs_interfaceid": "7e031e88-6928-472d-a65e-b818d89e6fc5", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": true, "delegate_create": true, "meta": {}}] {{(pid=61806) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 964.219262] env[61806]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-4771bf40-da13-430e-8a46-41d33a75988a {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 964.227227] env[61806]: DEBUG oslo_vmware.api [None req-0bbe3134-73ea-4d1b-8035-8d9ceb5bbb4c tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Waiting for the task: (returnval){ [ 964.227227] env[61806]: value = "session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]5280fac8-bc4c-1f16-d0c4-e7ae315fdf5d" [ 964.227227] env[61806]: _type = "Task" [ 964.227227] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 964.239798] env[61806]: DEBUG oslo_vmware.api [None req-0bbe3134-73ea-4d1b-8035-8d9ceb5bbb4c tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Task: {'id': session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]5280fac8-bc4c-1f16-d0c4-e7ae315fdf5d, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 964.397491] env[61806]: DEBUG oslo_concurrency.lockutils [None req-2c70fcbb-efa7-459d-81c0-a2a8a184893c tempest-AttachVolumeNegativeTest-1683033704 tempest-AttachVolumeNegativeTest-1683033704-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.278s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 964.398061] env[61806]: DEBUG nova.compute.manager [None req-2c70fcbb-efa7-459d-81c0-a2a8a184893c tempest-AttachVolumeNegativeTest-1683033704 tempest-AttachVolumeNegativeTest-1683033704-project-member] [instance: e4638faa-7be1-4909-a595-c437837fa314] Start building networks asynchronously for instance. {{(pid=61806) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 964.400994] env[61806]: DEBUG oslo_concurrency.lockutils [None req-28d96f96-d23f-4c45-9d59-0355de5e0c4f tempest-ServersNegativeTestJSON-891856253 tempest-ServersNegativeTestJSON-891856253-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 3.528s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 964.402460] env[61806]: INFO nova.compute.claims [None req-28d96f96-d23f-4c45-9d59-0355de5e0c4f tempest-ServersNegativeTestJSON-891856253 tempest-ServersNegativeTestJSON-891856253-project-member] [instance: e146f7e5-bbf8-4adf-a2e9-b7e753438976] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 964.429519] env[61806]: DEBUG oslo_vmware.api [None req-60f88c72-8f2b-4f98-b3cc-7787383864bd tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Task: {'id': task-1294980, 'name': CopyVirtualDisk_Task} progress is 89%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 964.471948] env[61806]: DEBUG oslo_concurrency.lockutils [None req-c8cf9c01-54a9-4c86-b57f-b2e6a2fbc11a tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] Releasing lock "refresh_cache-0bac3048-759a-47d9-a6ef-a0386a49a974" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 964.472335] env[61806]: DEBUG nova.compute.manager [None req-c8cf9c01-54a9-4c86-b57f-b2e6a2fbc11a tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] [instance: 0bac3048-759a-47d9-a6ef-a0386a49a974] Instance network_info: |[{"id": "69c51913-810b-4b60-8249-2960056a8035", "address": "fa:16:3e:a9:9e:06", "network": {"id": "f280a01b-7000-4e54-ab9a-20eab156a410", "bridge": "br-int", "label": "tempest-ServersTestJSON-2080718146-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "73ebcc4d79d248efb653a084a5e44302", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "43798f54-0c07-4417-a23f-58bb6b7e204b", "external-id": "nsx-vlan-transportzone-571", "segmentation_id": 571, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap69c51913-81", "ovs_interfaceid": "69c51913-810b-4b60-8249-2960056a8035", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61806) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 964.473204] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-c8cf9c01-54a9-4c86-b57f-b2e6a2fbc11a tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] [instance: 0bac3048-759a-47d9-a6ef-a0386a49a974] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:a9:9e:06', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '43798f54-0c07-4417-a23f-58bb6b7e204b', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '69c51913-810b-4b60-8249-2960056a8035', 'vif_model': 'vmxnet3'}] {{(pid=61806) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 964.480933] env[61806]: DEBUG oslo.service.loopingcall [None req-c8cf9c01-54a9-4c86-b57f-b2e6a2fbc11a tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61806) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 964.481212] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 0bac3048-759a-47d9-a6ef-a0386a49a974] Creating VM on the ESX host {{(pid=61806) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 964.481528] env[61806]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-ecef220b-1049-42f3-a2d5-662ad324e7e9 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 964.510031] env[61806]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 964.510031] env[61806]: value = "task-1294981" [ 964.510031] env[61806]: _type = "Task" [ 964.510031] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 964.519010] env[61806]: DEBUG oslo_vmware.api [-] Task: {'id': task-1294981, 'name': CreateVM_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 964.543028] env[61806]: DEBUG oslo_concurrency.lockutils [None req-98c10467-8c15-41cc-ac16-68156711acc7 tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] Lock "interface-d9caee4a-a00c-4e50-b01c-99a8f796a09f-7e031e88-6928-472d-a65e-b818d89e6fc5" "released" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: held 6.594s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 964.719441] env[61806]: DEBUG oslo_concurrency.lockutils [req-2ecdccfe-9d5e-42f9-a822-f7cab000e416 req-5f40af83-cd56-4043-a272-2f3fcc5ea3ec service nova] Releasing lock "refresh_cache-d9caee4a-a00c-4e50-b01c-99a8f796a09f" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 964.719810] env[61806]: DEBUG nova.compute.manager [req-2ecdccfe-9d5e-42f9-a822-f7cab000e416 req-5f40af83-cd56-4043-a272-2f3fcc5ea3ec service nova] [instance: 0bac3048-759a-47d9-a6ef-a0386a49a974] Received event network-vif-plugged-69c51913-810b-4b60-8249-2960056a8035 {{(pid=61806) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 964.719989] env[61806]: DEBUG oslo_concurrency.lockutils [req-2ecdccfe-9d5e-42f9-a822-f7cab000e416 req-5f40af83-cd56-4043-a272-2f3fcc5ea3ec service nova] Acquiring lock "0bac3048-759a-47d9-a6ef-a0386a49a974-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 964.720374] env[61806]: DEBUG oslo_concurrency.lockutils [req-2ecdccfe-9d5e-42f9-a822-f7cab000e416 req-5f40af83-cd56-4043-a272-2f3fcc5ea3ec service nova] Lock "0bac3048-759a-47d9-a6ef-a0386a49a974-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 964.720625] env[61806]: DEBUG oslo_concurrency.lockutils [req-2ecdccfe-9d5e-42f9-a822-f7cab000e416 req-5f40af83-cd56-4043-a272-2f3fcc5ea3ec service nova] Lock "0bac3048-759a-47d9-a6ef-a0386a49a974-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 964.720841] env[61806]: DEBUG nova.compute.manager [req-2ecdccfe-9d5e-42f9-a822-f7cab000e416 req-5f40af83-cd56-4043-a272-2f3fcc5ea3ec service nova] [instance: 0bac3048-759a-47d9-a6ef-a0386a49a974] No waiting events found dispatching network-vif-plugged-69c51913-810b-4b60-8249-2960056a8035 {{(pid=61806) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 964.721063] env[61806]: WARNING nova.compute.manager [req-2ecdccfe-9d5e-42f9-a822-f7cab000e416 req-5f40af83-cd56-4043-a272-2f3fcc5ea3ec service nova] [instance: 0bac3048-759a-47d9-a6ef-a0386a49a974] Received unexpected event network-vif-plugged-69c51913-810b-4b60-8249-2960056a8035 for instance with vm_state building and task_state spawning. [ 964.721281] env[61806]: DEBUG nova.compute.manager [req-2ecdccfe-9d5e-42f9-a822-f7cab000e416 req-5f40af83-cd56-4043-a272-2f3fcc5ea3ec service nova] [instance: 0bac3048-759a-47d9-a6ef-a0386a49a974] Received event network-changed-69c51913-810b-4b60-8249-2960056a8035 {{(pid=61806) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 964.721480] env[61806]: DEBUG nova.compute.manager [req-2ecdccfe-9d5e-42f9-a822-f7cab000e416 req-5f40af83-cd56-4043-a272-2f3fcc5ea3ec service nova] [instance: 0bac3048-759a-47d9-a6ef-a0386a49a974] Refreshing instance network info cache due to event network-changed-69c51913-810b-4b60-8249-2960056a8035. {{(pid=61806) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 964.721707] env[61806]: DEBUG oslo_concurrency.lockutils [req-2ecdccfe-9d5e-42f9-a822-f7cab000e416 req-5f40af83-cd56-4043-a272-2f3fcc5ea3ec service nova] Acquiring lock "refresh_cache-0bac3048-759a-47d9-a6ef-a0386a49a974" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 964.721862] env[61806]: DEBUG oslo_concurrency.lockutils [req-2ecdccfe-9d5e-42f9-a822-f7cab000e416 req-5f40af83-cd56-4043-a272-2f3fcc5ea3ec service nova] Acquired lock "refresh_cache-0bac3048-759a-47d9-a6ef-a0386a49a974" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 964.722089] env[61806]: DEBUG nova.network.neutron [req-2ecdccfe-9d5e-42f9-a822-f7cab000e416 req-5f40af83-cd56-4043-a272-2f3fcc5ea3ec service nova] [instance: 0bac3048-759a-47d9-a6ef-a0386a49a974] Refreshing network info cache for port 69c51913-810b-4b60-8249-2960056a8035 {{(pid=61806) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 964.739035] env[61806]: DEBUG oslo_vmware.api [None req-0bbe3134-73ea-4d1b-8035-8d9ceb5bbb4c tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Task: {'id': session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]5280fac8-bc4c-1f16-d0c4-e7ae315fdf5d, 'name': SearchDatastore_Task, 'duration_secs': 0.025854} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 964.739367] env[61806]: DEBUG oslo_concurrency.lockutils [None req-0bbe3134-73ea-4d1b-8035-8d9ceb5bbb4c tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 964.909977] env[61806]: DEBUG nova.compute.utils [None req-2c70fcbb-efa7-459d-81c0-a2a8a184893c tempest-AttachVolumeNegativeTest-1683033704 tempest-AttachVolumeNegativeTest-1683033704-project-member] Using /dev/sd instead of None {{(pid=61806) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 964.911455] env[61806]: DEBUG nova.compute.manager [None req-2c70fcbb-efa7-459d-81c0-a2a8a184893c tempest-AttachVolumeNegativeTest-1683033704 tempest-AttachVolumeNegativeTest-1683033704-project-member] [instance: e4638faa-7be1-4909-a595-c437837fa314] Allocating IP information in the background. {{(pid=61806) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 964.911621] env[61806]: DEBUG nova.network.neutron [None req-2c70fcbb-efa7-459d-81c0-a2a8a184893c tempest-AttachVolumeNegativeTest-1683033704 tempest-AttachVolumeNegativeTest-1683033704-project-member] [instance: e4638faa-7be1-4909-a595-c437837fa314] allocate_for_instance() {{(pid=61806) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 964.927555] env[61806]: DEBUG oslo_vmware.api [None req-60f88c72-8f2b-4f98-b3cc-7787383864bd tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Task: {'id': task-1294980, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.592739} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 964.927835] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [None req-60f88c72-8f2b-4f98-b3cc-7787383864bd tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa.vmdk to [datastore2] 461f98f2-80ba-4088-9e7f-1ba2a9b0c1d9/461f98f2-80ba-4088-9e7f-1ba2a9b0c1d9.vmdk {{(pid=61806) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 964.928064] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-60f88c72-8f2b-4f98-b3cc-7787383864bd tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] [instance: 461f98f2-80ba-4088-9e7f-1ba2a9b0c1d9] Extending root virtual disk to 1048576 {{(pid=61806) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 964.928326] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-bfeb8cd6-2487-49c4-881c-29d0c83a1561 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 964.937526] env[61806]: DEBUG oslo_vmware.api [None req-60f88c72-8f2b-4f98-b3cc-7787383864bd tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Waiting for the task: (returnval){ [ 964.937526] env[61806]: value = "task-1294982" [ 964.937526] env[61806]: _type = "Task" [ 964.937526] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 964.949895] env[61806]: DEBUG oslo_vmware.api [None req-60f88c72-8f2b-4f98-b3cc-7787383864bd tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Task: {'id': task-1294982, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 964.950501] env[61806]: DEBUG oslo_concurrency.lockutils [None req-d1459cd5-a70e-4111-a41f-72b97e54aa17 tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Lock "380aa63f-46ba-437d-bc29-0fc141ad1cf3" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 2.316s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 964.957529] env[61806]: DEBUG nova.policy [None req-2c70fcbb-efa7-459d-81c0-a2a8a184893c tempest-AttachVolumeNegativeTest-1683033704 tempest-AttachVolumeNegativeTest-1683033704-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '62b5e014cfeb4403a1edf89108c55e11', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '22e47882c2c44ce492dff6b1c5d782e2', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61806) authorize /opt/stack/nova/nova/policy.py:201}} [ 965.021302] env[61806]: DEBUG oslo_vmware.api [-] Task: {'id': task-1294981, 'name': CreateVM_Task} progress is 25%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 965.227335] env[61806]: DEBUG nova.network.neutron [None req-2c70fcbb-efa7-459d-81c0-a2a8a184893c tempest-AttachVolumeNegativeTest-1683033704 tempest-AttachVolumeNegativeTest-1683033704-project-member] [instance: e4638faa-7be1-4909-a595-c437837fa314] Successfully created port: bfb37474-8e7a-4820-9045-e1b388cfb87d {{(pid=61806) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 965.415931] env[61806]: DEBUG nova.compute.manager [None req-2c70fcbb-efa7-459d-81c0-a2a8a184893c tempest-AttachVolumeNegativeTest-1683033704 tempest-AttachVolumeNegativeTest-1683033704-project-member] [instance: e4638faa-7be1-4909-a595-c437837fa314] Start building block device mappings for instance. {{(pid=61806) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 965.445730] env[61806]: DEBUG nova.network.neutron [req-2ecdccfe-9d5e-42f9-a822-f7cab000e416 req-5f40af83-cd56-4043-a272-2f3fcc5ea3ec service nova] [instance: 0bac3048-759a-47d9-a6ef-a0386a49a974] Updated VIF entry in instance network info cache for port 69c51913-810b-4b60-8249-2960056a8035. {{(pid=61806) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 965.446742] env[61806]: DEBUG nova.network.neutron [req-2ecdccfe-9d5e-42f9-a822-f7cab000e416 req-5f40af83-cd56-4043-a272-2f3fcc5ea3ec service nova] [instance: 0bac3048-759a-47d9-a6ef-a0386a49a974] Updating instance_info_cache with network_info: [{"id": "69c51913-810b-4b60-8249-2960056a8035", "address": "fa:16:3e:a9:9e:06", "network": {"id": "f280a01b-7000-4e54-ab9a-20eab156a410", "bridge": "br-int", "label": "tempest-ServersTestJSON-2080718146-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "73ebcc4d79d248efb653a084a5e44302", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "43798f54-0c07-4417-a23f-58bb6b7e204b", "external-id": "nsx-vlan-transportzone-571", "segmentation_id": 571, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap69c51913-81", "ovs_interfaceid": "69c51913-810b-4b60-8249-2960056a8035", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61806) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 965.465477] env[61806]: DEBUG oslo_vmware.api [None req-60f88c72-8f2b-4f98-b3cc-7787383864bd tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Task: {'id': task-1294982, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.130052} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 965.469897] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-60f88c72-8f2b-4f98-b3cc-7787383864bd tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] [instance: 461f98f2-80ba-4088-9e7f-1ba2a9b0c1d9] Extended root virtual disk {{(pid=61806) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 965.472886] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-98fd4a0b-423d-426a-8716-2d19e0ad7242 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 965.500054] env[61806]: DEBUG nova.virt.vmwareapi.volumeops [None req-60f88c72-8f2b-4f98-b3cc-7787383864bd tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] [instance: 461f98f2-80ba-4088-9e7f-1ba2a9b0c1d9] Reconfiguring VM instance instance-0000005b to attach disk [datastore2] 461f98f2-80ba-4088-9e7f-1ba2a9b0c1d9/461f98f2-80ba-4088-9e7f-1ba2a9b0c1d9.vmdk or device None with type sparse {{(pid=61806) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 965.503770] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-df103582-6381-4dc5-836e-68f5a5816663 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 965.531098] env[61806]: DEBUG oslo_vmware.api [-] Task: {'id': task-1294981, 'name': CreateVM_Task, 'duration_secs': 0.761623} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 965.534971] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 0bac3048-759a-47d9-a6ef-a0386a49a974] Created VM on the ESX host {{(pid=61806) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 965.535448] env[61806]: DEBUG oslo_vmware.api [None req-60f88c72-8f2b-4f98-b3cc-7787383864bd tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Waiting for the task: (returnval){ [ 965.535448] env[61806]: value = "task-1294983" [ 965.535448] env[61806]: _type = "Task" [ 965.535448] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 965.536770] env[61806]: DEBUG oslo_concurrency.lockutils [None req-c8cf9c01-54a9-4c86-b57f-b2e6a2fbc11a tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 965.536839] env[61806]: DEBUG oslo_concurrency.lockutils [None req-c8cf9c01-54a9-4c86-b57f-b2e6a2fbc11a tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] Acquired lock "[datastore2] devstack-image-cache_base/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 965.537180] env[61806]: DEBUG oslo_concurrency.lockutils [None req-c8cf9c01-54a9-4c86-b57f-b2e6a2fbc11a tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 965.537499] env[61806]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-34b4c998-2533-49d4-a06a-c8d75c5a65e3 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 965.546692] env[61806]: DEBUG oslo_vmware.api [None req-c8cf9c01-54a9-4c86-b57f-b2e6a2fbc11a tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] Waiting for the task: (returnval){ [ 965.546692] env[61806]: value = "session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]5201941b-d79d-ce2b-f64b-84396a586818" [ 965.546692] env[61806]: _type = "Task" [ 965.546692] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 965.550037] env[61806]: DEBUG oslo_vmware.api [None req-60f88c72-8f2b-4f98-b3cc-7787383864bd tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Task: {'id': task-1294983, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 965.565622] env[61806]: DEBUG oslo_vmware.api [None req-c8cf9c01-54a9-4c86-b57f-b2e6a2fbc11a tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] Task: {'id': session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]5201941b-d79d-ce2b-f64b-84396a586818, 'name': SearchDatastore_Task, 'duration_secs': 0.011372} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 965.565985] env[61806]: DEBUG oslo_concurrency.lockutils [None req-c8cf9c01-54a9-4c86-b57f-b2e6a2fbc11a tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] Releasing lock "[datastore2] devstack-image-cache_base/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 965.566262] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-c8cf9c01-54a9-4c86-b57f-b2e6a2fbc11a tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] [instance: 0bac3048-759a-47d9-a6ef-a0386a49a974] Processing image ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa {{(pid=61806) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 965.566655] env[61806]: DEBUG oslo_concurrency.lockutils [None req-c8cf9c01-54a9-4c86-b57f-b2e6a2fbc11a tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa.vmdk" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 965.566741] env[61806]: DEBUG oslo_concurrency.lockutils [None req-c8cf9c01-54a9-4c86-b57f-b2e6a2fbc11a tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] Acquired lock "[datastore2] devstack-image-cache_base/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa.vmdk" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 965.566877] env[61806]: DEBUG nova.virt.vmwareapi.ds_util [None req-c8cf9c01-54a9-4c86-b57f-b2e6a2fbc11a tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61806) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 965.567208] env[61806]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-e5fa156f-0ea5-4c65-b047-bc6159c9c8a8 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 965.579163] env[61806]: DEBUG nova.virt.vmwareapi.ds_util [None req-c8cf9c01-54a9-4c86-b57f-b2e6a2fbc11a tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61806) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 965.579477] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-c8cf9c01-54a9-4c86-b57f-b2e6a2fbc11a tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61806) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 965.583020] env[61806]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-a4dc5bdd-d6de-41dc-a042-dbc85ae8bdb7 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 965.587701] env[61806]: DEBUG oslo_vmware.api [None req-c8cf9c01-54a9-4c86-b57f-b2e6a2fbc11a tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] Waiting for the task: (returnval){ [ 965.587701] env[61806]: value = "session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]524f8020-0c9a-67f0-1a33-42a136369650" [ 965.587701] env[61806]: _type = "Task" [ 965.587701] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 965.597968] env[61806]: DEBUG oslo_vmware.api [None req-c8cf9c01-54a9-4c86-b57f-b2e6a2fbc11a tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] Task: {'id': session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]524f8020-0c9a-67f0-1a33-42a136369650, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 965.684979] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-36a32935-c981-42c0-9f60-5d1eadde4c66 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 965.693433] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b30dacab-16d0-4911-b83c-53df548bcaf7 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 965.726666] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7138c894-e778-45a8-9536-3f6e6ffb9eb8 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 965.735358] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4b5f7974-c9d9-49a4-b204-7dce80b73a4a {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 965.750396] env[61806]: DEBUG nova.compute.provider_tree [None req-28d96f96-d23f-4c45-9d59-0355de5e0c4f tempest-ServersNegativeTestJSON-891856253 tempest-ServersNegativeTestJSON-891856253-project-member] Inventory has not changed in ProviderTree for provider: a2858be1-fd22-4e08-979e-87ad25293407 {{(pid=61806) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 965.954890] env[61806]: DEBUG oslo_concurrency.lockutils [req-2ecdccfe-9d5e-42f9-a822-f7cab000e416 req-5f40af83-cd56-4043-a272-2f3fcc5ea3ec service nova] Releasing lock "refresh_cache-0bac3048-759a-47d9-a6ef-a0386a49a974" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 966.049707] env[61806]: DEBUG oslo_vmware.api [None req-60f88c72-8f2b-4f98-b3cc-7787383864bd tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Task: {'id': task-1294983, 'name': ReconfigVM_Task, 'duration_secs': 0.294108} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 966.050017] env[61806]: DEBUG nova.virt.vmwareapi.volumeops [None req-60f88c72-8f2b-4f98-b3cc-7787383864bd tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] [instance: 461f98f2-80ba-4088-9e7f-1ba2a9b0c1d9] Reconfigured VM instance instance-0000005b to attach disk [datastore2] 461f98f2-80ba-4088-9e7f-1ba2a9b0c1d9/461f98f2-80ba-4088-9e7f-1ba2a9b0c1d9.vmdk or device None with type sparse {{(pid=61806) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 966.050788] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-ba94dfed-e911-45fe-b7db-7ff93ecf2869 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 966.059153] env[61806]: DEBUG oslo_vmware.api [None req-60f88c72-8f2b-4f98-b3cc-7787383864bd tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Waiting for the task: (returnval){ [ 966.059153] env[61806]: value = "task-1294984" [ 966.059153] env[61806]: _type = "Task" [ 966.059153] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 966.070904] env[61806]: DEBUG oslo_vmware.api [None req-60f88c72-8f2b-4f98-b3cc-7787383864bd tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Task: {'id': task-1294984, 'name': Rename_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 966.099972] env[61806]: DEBUG oslo_vmware.api [None req-c8cf9c01-54a9-4c86-b57f-b2e6a2fbc11a tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] Task: {'id': session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]524f8020-0c9a-67f0-1a33-42a136369650, 'name': SearchDatastore_Task, 'duration_secs': 0.01073} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 966.100690] env[61806]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-0af1f677-1992-498b-abce-58c6bf4f51c5 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 966.107523] env[61806]: DEBUG oslo_vmware.api [None req-c8cf9c01-54a9-4c86-b57f-b2e6a2fbc11a tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] Waiting for the task: (returnval){ [ 966.107523] env[61806]: value = "session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]52d1a14f-5fc5-c41f-4bac-fc473bb510e2" [ 966.107523] env[61806]: _type = "Task" [ 966.107523] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 966.116912] env[61806]: DEBUG oslo_vmware.api [None req-c8cf9c01-54a9-4c86-b57f-b2e6a2fbc11a tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] Task: {'id': session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]52d1a14f-5fc5-c41f-4bac-fc473bb510e2, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 966.164184] env[61806]: DEBUG oslo_concurrency.lockutils [None req-d01361ec-1270-4938-a4ab-1319441d7070 tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Acquiring lock "962d3d88-03ee-4ff2-9d79-15091671b330" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 966.164441] env[61806]: DEBUG oslo_concurrency.lockutils [None req-d01361ec-1270-4938-a4ab-1319441d7070 tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Lock "962d3d88-03ee-4ff2-9d79-15091671b330" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 966.253925] env[61806]: DEBUG nova.scheduler.client.report [None req-28d96f96-d23f-4c45-9d59-0355de5e0c4f tempest-ServersNegativeTestJSON-891856253 tempest-ServersNegativeTestJSON-891856253-project-member] Inventory has not changed for provider a2858be1-fd22-4e08-979e-87ad25293407 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 138, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61806) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 966.399646] env[61806]: DEBUG oslo_concurrency.lockutils [None req-49ec17b3-08f6-4e32-88fe-0409196f3352 tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] Acquiring lock "interface-d9caee4a-a00c-4e50-b01c-99a8f796a09f-7e031e88-6928-472d-a65e-b818d89e6fc5" by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 966.399880] env[61806]: DEBUG oslo_concurrency.lockutils [None req-49ec17b3-08f6-4e32-88fe-0409196f3352 tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] Lock "interface-d9caee4a-a00c-4e50-b01c-99a8f796a09f-7e031e88-6928-472d-a65e-b818d89e6fc5" acquired by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" :: waited 0.001s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 966.429527] env[61806]: DEBUG nova.compute.manager [None req-2c70fcbb-efa7-459d-81c0-a2a8a184893c tempest-AttachVolumeNegativeTest-1683033704 tempest-AttachVolumeNegativeTest-1683033704-project-member] [instance: e4638faa-7be1-4909-a595-c437837fa314] Start spawning the instance on the hypervisor. {{(pid=61806) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 966.456454] env[61806]: DEBUG nova.virt.hardware [None req-2c70fcbb-efa7-459d-81c0-a2a8a184893c tempest-AttachVolumeNegativeTest-1683033704 tempest-AttachVolumeNegativeTest-1683033704-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-15T18:05:10Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-15T18:04:52Z,direct_url=,disk_format='vmdk',id=ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='84ba83fce2c349a6988173c5d6fc3b07',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-15T18:04:53Z,virtual_size=,visibility=), allow threads: False {{(pid=61806) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 966.456755] env[61806]: DEBUG nova.virt.hardware [None req-2c70fcbb-efa7-459d-81c0-a2a8a184893c tempest-AttachVolumeNegativeTest-1683033704 tempest-AttachVolumeNegativeTest-1683033704-project-member] Flavor limits 0:0:0 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 966.456947] env[61806]: DEBUG nova.virt.hardware [None req-2c70fcbb-efa7-459d-81c0-a2a8a184893c tempest-AttachVolumeNegativeTest-1683033704 tempest-AttachVolumeNegativeTest-1683033704-project-member] Image limits 0:0:0 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 966.457177] env[61806]: DEBUG nova.virt.hardware [None req-2c70fcbb-efa7-459d-81c0-a2a8a184893c tempest-AttachVolumeNegativeTest-1683033704 tempest-AttachVolumeNegativeTest-1683033704-project-member] Flavor pref 0:0:0 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 966.457349] env[61806]: DEBUG nova.virt.hardware [None req-2c70fcbb-efa7-459d-81c0-a2a8a184893c tempest-AttachVolumeNegativeTest-1683033704 tempest-AttachVolumeNegativeTest-1683033704-project-member] Image pref 0:0:0 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 966.457527] env[61806]: DEBUG nova.virt.hardware [None req-2c70fcbb-efa7-459d-81c0-a2a8a184893c tempest-AttachVolumeNegativeTest-1683033704 tempest-AttachVolumeNegativeTest-1683033704-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 966.457764] env[61806]: DEBUG nova.virt.hardware [None req-2c70fcbb-efa7-459d-81c0-a2a8a184893c tempest-AttachVolumeNegativeTest-1683033704 tempest-AttachVolumeNegativeTest-1683033704-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61806) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 966.457951] env[61806]: DEBUG nova.virt.hardware [None req-2c70fcbb-efa7-459d-81c0-a2a8a184893c tempest-AttachVolumeNegativeTest-1683033704 tempest-AttachVolumeNegativeTest-1683033704-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61806) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 966.458172] env[61806]: DEBUG nova.virt.hardware [None req-2c70fcbb-efa7-459d-81c0-a2a8a184893c tempest-AttachVolumeNegativeTest-1683033704 tempest-AttachVolumeNegativeTest-1683033704-project-member] Got 1 possible topologies {{(pid=61806) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 966.458361] env[61806]: DEBUG nova.virt.hardware [None req-2c70fcbb-efa7-459d-81c0-a2a8a184893c tempest-AttachVolumeNegativeTest-1683033704 tempest-AttachVolumeNegativeTest-1683033704-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61806) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 966.458560] env[61806]: DEBUG nova.virt.hardware [None req-2c70fcbb-efa7-459d-81c0-a2a8a184893c tempest-AttachVolumeNegativeTest-1683033704 tempest-AttachVolumeNegativeTest-1683033704-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61806) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 966.459543] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cf325367-f1d0-43b5-91f4-d8f34f3563fe {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 966.469134] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ec066044-bb92-45bf-ab30-d4f476e86ac5 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 966.570418] env[61806]: DEBUG oslo_vmware.api [None req-60f88c72-8f2b-4f98-b3cc-7787383864bd tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Task: {'id': task-1294984, 'name': Rename_Task, 'duration_secs': 0.153536} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 966.571365] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [None req-60f88c72-8f2b-4f98-b3cc-7787383864bd tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] [instance: 461f98f2-80ba-4088-9e7f-1ba2a9b0c1d9] Powering on the VM {{(pid=61806) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 966.571730] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-f05b7d5b-3d9e-49f4-bf3a-40fc6a02511b {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 966.580391] env[61806]: DEBUG oslo_vmware.api [None req-60f88c72-8f2b-4f98-b3cc-7787383864bd tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Waiting for the task: (returnval){ [ 966.580391] env[61806]: value = "task-1294985" [ 966.580391] env[61806]: _type = "Task" [ 966.580391] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 966.600594] env[61806]: DEBUG oslo_vmware.api [None req-60f88c72-8f2b-4f98-b3cc-7787383864bd tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Task: {'id': task-1294985, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 966.619721] env[61806]: DEBUG oslo_vmware.api [None req-c8cf9c01-54a9-4c86-b57f-b2e6a2fbc11a tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] Task: {'id': session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]52d1a14f-5fc5-c41f-4bac-fc473bb510e2, 'name': SearchDatastore_Task, 'duration_secs': 0.0098} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 966.620085] env[61806]: DEBUG oslo_concurrency.lockutils [None req-c8cf9c01-54a9-4c86-b57f-b2e6a2fbc11a tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] Releasing lock "[datastore2] devstack-image-cache_base/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa.vmdk" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 966.620425] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [None req-c8cf9c01-54a9-4c86-b57f-b2e6a2fbc11a tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa.vmdk to [datastore2] 0bac3048-759a-47d9-a6ef-a0386a49a974/0bac3048-759a-47d9-a6ef-a0386a49a974.vmdk {{(pid=61806) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 966.620726] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-65008906-cb7e-46c2-bcf4-840dac6e1bfa {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 966.630109] env[61806]: DEBUG oslo_vmware.api [None req-c8cf9c01-54a9-4c86-b57f-b2e6a2fbc11a tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] Waiting for the task: (returnval){ [ 966.630109] env[61806]: value = "task-1294986" [ 966.630109] env[61806]: _type = "Task" [ 966.630109] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 966.640593] env[61806]: DEBUG oslo_vmware.api [None req-c8cf9c01-54a9-4c86-b57f-b2e6a2fbc11a tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] Task: {'id': task-1294986, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 966.667375] env[61806]: DEBUG nova.compute.manager [None req-d01361ec-1270-4938-a4ab-1319441d7070 tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] [instance: 962d3d88-03ee-4ff2-9d79-15091671b330] Starting instance... {{(pid=61806) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 966.759259] env[61806]: DEBUG oslo_concurrency.lockutils [None req-28d96f96-d23f-4c45-9d59-0355de5e0c4f tempest-ServersNegativeTestJSON-891856253 tempest-ServersNegativeTestJSON-891856253-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.358s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 966.759993] env[61806]: DEBUG nova.compute.manager [None req-28d96f96-d23f-4c45-9d59-0355de5e0c4f tempest-ServersNegativeTestJSON-891856253 tempest-ServersNegativeTestJSON-891856253-project-member] [instance: e146f7e5-bbf8-4adf-a2e9-b7e753438976] Start building networks asynchronously for instance. {{(pid=61806) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 966.765191] env[61806]: DEBUG oslo_concurrency.lockutils [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 4.067s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 966.765417] env[61806]: DEBUG oslo_concurrency.lockutils [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 966.765587] env[61806]: DEBUG nova.compute.resource_tracker [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=61806) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 966.765891] env[61806]: DEBUG oslo_concurrency.lockutils [None req-0bbe3134-73ea-4d1b-8035-8d9ceb5bbb4c tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" :: waited 2.027s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 966.767800] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1b208dd2-ed6d-4e19-b7c2-25022726ff56 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 966.778063] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-73547fb4-376c-48e7-9d52-52e2faa7bba3 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 966.797854] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b90ec3bb-8d1e-4396-a6a0-f3d2b4cf58ba {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 966.807659] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e4faf9b9-1fa1-44ce-b7c8-d91c5cbe605e {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 966.844487] env[61806]: DEBUG nova.compute.resource_tracker [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=180218MB free_disk=138GB free_vcpus=48 pci_devices=None {{(pid=61806) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 966.844704] env[61806]: DEBUG oslo_concurrency.lockutils [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 966.905043] env[61806]: DEBUG oslo_concurrency.lockutils [None req-49ec17b3-08f6-4e32-88fe-0409196f3352 tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] Acquiring lock "d9caee4a-a00c-4e50-b01c-99a8f796a09f" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 966.905043] env[61806]: DEBUG oslo_concurrency.lockutils [None req-49ec17b3-08f6-4e32-88fe-0409196f3352 tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] Acquired lock "d9caee4a-a00c-4e50-b01c-99a8f796a09f" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 966.905043] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3f5d375c-fe70-4f1c-8444-e9b02e5e7927 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 966.926489] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-086578c6-60ba-4bc6-9246-3db7c583a12f {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 966.957454] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-49ec17b3-08f6-4e32-88fe-0409196f3352 tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] [instance: d9caee4a-a00c-4e50-b01c-99a8f796a09f] Reconfiguring VM to detach interface {{(pid=61806) detach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1974}} [ 966.957826] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-4b44caf5-753c-411e-b5c2-2de736be928b {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 966.982988] env[61806]: DEBUG oslo_vmware.api [None req-49ec17b3-08f6-4e32-88fe-0409196f3352 tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] Waiting for the task: (returnval){ [ 966.982988] env[61806]: value = "task-1294987" [ 966.982988] env[61806]: _type = "Task" [ 966.982988] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 966.985222] env[61806]: DEBUG nova.compute.manager [req-ff722649-9093-4764-8602-de5502a4e0f6 req-fdeab1e6-ab01-4ea8-8e92-32adbf1ab280 service nova] [instance: e4638faa-7be1-4909-a595-c437837fa314] Received event network-vif-plugged-bfb37474-8e7a-4820-9045-e1b388cfb87d {{(pid=61806) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 966.985570] env[61806]: DEBUG oslo_concurrency.lockutils [req-ff722649-9093-4764-8602-de5502a4e0f6 req-fdeab1e6-ab01-4ea8-8e92-32adbf1ab280 service nova] Acquiring lock "e4638faa-7be1-4909-a595-c437837fa314-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 966.985641] env[61806]: DEBUG oslo_concurrency.lockutils [req-ff722649-9093-4764-8602-de5502a4e0f6 req-fdeab1e6-ab01-4ea8-8e92-32adbf1ab280 service nova] Lock "e4638faa-7be1-4909-a595-c437837fa314-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 966.985814] env[61806]: DEBUG oslo_concurrency.lockutils [req-ff722649-9093-4764-8602-de5502a4e0f6 req-fdeab1e6-ab01-4ea8-8e92-32adbf1ab280 service nova] Lock "e4638faa-7be1-4909-a595-c437837fa314-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 966.985989] env[61806]: DEBUG nova.compute.manager [req-ff722649-9093-4764-8602-de5502a4e0f6 req-fdeab1e6-ab01-4ea8-8e92-32adbf1ab280 service nova] [instance: e4638faa-7be1-4909-a595-c437837fa314] No waiting events found dispatching network-vif-plugged-bfb37474-8e7a-4820-9045-e1b388cfb87d {{(pid=61806) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 966.986250] env[61806]: WARNING nova.compute.manager [req-ff722649-9093-4764-8602-de5502a4e0f6 req-fdeab1e6-ab01-4ea8-8e92-32adbf1ab280 service nova] [instance: e4638faa-7be1-4909-a595-c437837fa314] Received unexpected event network-vif-plugged-bfb37474-8e7a-4820-9045-e1b388cfb87d for instance with vm_state building and task_state spawning. [ 967.001624] env[61806]: DEBUG oslo_vmware.api [None req-49ec17b3-08f6-4e32-88fe-0409196f3352 tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] Task: {'id': task-1294987, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 967.071818] env[61806]: DEBUG nova.network.neutron [None req-2c70fcbb-efa7-459d-81c0-a2a8a184893c tempest-AttachVolumeNegativeTest-1683033704 tempest-AttachVolumeNegativeTest-1683033704-project-member] [instance: e4638faa-7be1-4909-a595-c437837fa314] Successfully updated port: bfb37474-8e7a-4820-9045-e1b388cfb87d {{(pid=61806) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 967.094273] env[61806]: DEBUG oslo_vmware.api [None req-60f88c72-8f2b-4f98-b3cc-7787383864bd tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Task: {'id': task-1294985, 'name': PowerOnVM_Task} progress is 88%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 967.145911] env[61806]: DEBUG oslo_vmware.api [None req-c8cf9c01-54a9-4c86-b57f-b2e6a2fbc11a tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] Task: {'id': task-1294986, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 967.191744] env[61806]: DEBUG oslo_concurrency.lockutils [None req-d01361ec-1270-4938-a4ab-1319441d7070 tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 967.270070] env[61806]: DEBUG nova.compute.utils [None req-28d96f96-d23f-4c45-9d59-0355de5e0c4f tempest-ServersNegativeTestJSON-891856253 tempest-ServersNegativeTestJSON-891856253-project-member] Using /dev/sd instead of None {{(pid=61806) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 967.271946] env[61806]: DEBUG nova.compute.manager [None req-28d96f96-d23f-4c45-9d59-0355de5e0c4f tempest-ServersNegativeTestJSON-891856253 tempest-ServersNegativeTestJSON-891856253-project-member] [instance: e146f7e5-bbf8-4adf-a2e9-b7e753438976] Allocating IP information in the background. {{(pid=61806) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 967.272212] env[61806]: DEBUG nova.network.neutron [None req-28d96f96-d23f-4c45-9d59-0355de5e0c4f tempest-ServersNegativeTestJSON-891856253 tempest-ServersNegativeTestJSON-891856253-project-member] [instance: e146f7e5-bbf8-4adf-a2e9-b7e753438976] allocate_for_instance() {{(pid=61806) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 967.322953] env[61806]: DEBUG nova.policy [None req-28d96f96-d23f-4c45-9d59-0355de5e0c4f tempest-ServersNegativeTestJSON-891856253 tempest-ServersNegativeTestJSON-891856253-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'bfd1d620a5ea4b11ae028bd6c719a47a', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'd2e2b7eec47e4b10898215123408692a', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61806) authorize /opt/stack/nova/nova/policy.py:201}} [ 967.468972] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5d6c9e23-bf95-4a7d-8bb5-5adc0b35ef0e {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 967.478096] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0ba9bd9b-baa4-46dc-aae4-ea217bac34ce {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 967.510623] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5b6c780d-0372-42db-ac8c-0c2fae0f4cfb {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 967.522687] env[61806]: DEBUG oslo_vmware.api [None req-49ec17b3-08f6-4e32-88fe-0409196f3352 tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] Task: {'id': task-1294987, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 967.523947] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-103550a4-b2db-41ba-927d-aa530649b5f5 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 967.538796] env[61806]: DEBUG nova.compute.provider_tree [None req-0bbe3134-73ea-4d1b-8035-8d9ceb5bbb4c tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Inventory has not changed in ProviderTree for provider: a2858be1-fd22-4e08-979e-87ad25293407 {{(pid=61806) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 967.574885] env[61806]: DEBUG oslo_concurrency.lockutils [None req-2c70fcbb-efa7-459d-81c0-a2a8a184893c tempest-AttachVolumeNegativeTest-1683033704 tempest-AttachVolumeNegativeTest-1683033704-project-member] Acquiring lock "refresh_cache-e4638faa-7be1-4909-a595-c437837fa314" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 967.575037] env[61806]: DEBUG oslo_concurrency.lockutils [None req-2c70fcbb-efa7-459d-81c0-a2a8a184893c tempest-AttachVolumeNegativeTest-1683033704 tempest-AttachVolumeNegativeTest-1683033704-project-member] Acquired lock "refresh_cache-e4638faa-7be1-4909-a595-c437837fa314" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 967.575211] env[61806]: DEBUG nova.network.neutron [None req-2c70fcbb-efa7-459d-81c0-a2a8a184893c tempest-AttachVolumeNegativeTest-1683033704 tempest-AttachVolumeNegativeTest-1683033704-project-member] [instance: e4638faa-7be1-4909-a595-c437837fa314] Building network info cache for instance {{(pid=61806) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 967.585612] env[61806]: DEBUG nova.network.neutron [None req-28d96f96-d23f-4c45-9d59-0355de5e0c4f tempest-ServersNegativeTestJSON-891856253 tempest-ServersNegativeTestJSON-891856253-project-member] [instance: e146f7e5-bbf8-4adf-a2e9-b7e753438976] Successfully created port: b68f5483-a11f-4b35-aaa0-2dcd7f50e448 {{(pid=61806) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 967.593875] env[61806]: DEBUG oslo_vmware.api [None req-60f88c72-8f2b-4f98-b3cc-7787383864bd tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Task: {'id': task-1294985, 'name': PowerOnVM_Task, 'duration_secs': 0.818776} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 967.594167] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [None req-60f88c72-8f2b-4f98-b3cc-7787383864bd tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] [instance: 461f98f2-80ba-4088-9e7f-1ba2a9b0c1d9] Powered on the VM {{(pid=61806) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 967.594390] env[61806]: INFO nova.compute.manager [None req-60f88c72-8f2b-4f98-b3cc-7787383864bd tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] [instance: 461f98f2-80ba-4088-9e7f-1ba2a9b0c1d9] Took 8.02 seconds to spawn the instance on the hypervisor. [ 967.594577] env[61806]: DEBUG nova.compute.manager [None req-60f88c72-8f2b-4f98-b3cc-7787383864bd tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] [instance: 461f98f2-80ba-4088-9e7f-1ba2a9b0c1d9] Checking state {{(pid=61806) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 967.595390] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2dae585e-b119-43f3-96a0-df1e6ea555f3 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 967.642266] env[61806]: DEBUG oslo_vmware.api [None req-c8cf9c01-54a9-4c86-b57f-b2e6a2fbc11a tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] Task: {'id': task-1294986, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.604635} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 967.642592] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [None req-c8cf9c01-54a9-4c86-b57f-b2e6a2fbc11a tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa.vmdk to [datastore2] 0bac3048-759a-47d9-a6ef-a0386a49a974/0bac3048-759a-47d9-a6ef-a0386a49a974.vmdk {{(pid=61806) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 967.642827] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-c8cf9c01-54a9-4c86-b57f-b2e6a2fbc11a tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] [instance: 0bac3048-759a-47d9-a6ef-a0386a49a974] Extending root virtual disk to 1048576 {{(pid=61806) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 967.643113] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-232f8e10-1fdf-48ae-8f2e-ddf2ab0f43a9 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 967.654023] env[61806]: DEBUG oslo_vmware.api [None req-c8cf9c01-54a9-4c86-b57f-b2e6a2fbc11a tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] Waiting for the task: (returnval){ [ 967.654023] env[61806]: value = "task-1294988" [ 967.654023] env[61806]: _type = "Task" [ 967.654023] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 967.663483] env[61806]: DEBUG oslo_vmware.api [None req-c8cf9c01-54a9-4c86-b57f-b2e6a2fbc11a tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] Task: {'id': task-1294988, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 967.779746] env[61806]: DEBUG nova.compute.manager [None req-28d96f96-d23f-4c45-9d59-0355de5e0c4f tempest-ServersNegativeTestJSON-891856253 tempest-ServersNegativeTestJSON-891856253-project-member] [instance: e146f7e5-bbf8-4adf-a2e9-b7e753438976] Start building block device mappings for instance. {{(pid=61806) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 968.017594] env[61806]: DEBUG oslo_vmware.api [None req-49ec17b3-08f6-4e32-88fe-0409196f3352 tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] Task: {'id': task-1294987, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 968.042233] env[61806]: DEBUG nova.scheduler.client.report [None req-0bbe3134-73ea-4d1b-8035-8d9ceb5bbb4c tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Inventory has not changed for provider a2858be1-fd22-4e08-979e-87ad25293407 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 138, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61806) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 968.112324] env[61806]: DEBUG nova.network.neutron [None req-2c70fcbb-efa7-459d-81c0-a2a8a184893c tempest-AttachVolumeNegativeTest-1683033704 tempest-AttachVolumeNegativeTest-1683033704-project-member] [instance: e4638faa-7be1-4909-a595-c437837fa314] Instance cache missing network info. {{(pid=61806) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 968.115403] env[61806]: INFO nova.compute.manager [None req-60f88c72-8f2b-4f98-b3cc-7787383864bd tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] [instance: 461f98f2-80ba-4088-9e7f-1ba2a9b0c1d9] Took 12.85 seconds to build instance. [ 968.169464] env[61806]: DEBUG oslo_vmware.api [None req-c8cf9c01-54a9-4c86-b57f-b2e6a2fbc11a tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] Task: {'id': task-1294988, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.08356} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 968.169582] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-c8cf9c01-54a9-4c86-b57f-b2e6a2fbc11a tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] [instance: 0bac3048-759a-47d9-a6ef-a0386a49a974] Extended root virtual disk {{(pid=61806) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 968.170305] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6e29abc7-ca4f-4486-8399-01d54da5a673 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 968.194445] env[61806]: DEBUG nova.virt.vmwareapi.volumeops [None req-c8cf9c01-54a9-4c86-b57f-b2e6a2fbc11a tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] [instance: 0bac3048-759a-47d9-a6ef-a0386a49a974] Reconfiguring VM instance instance-0000005c to attach disk [datastore2] 0bac3048-759a-47d9-a6ef-a0386a49a974/0bac3048-759a-47d9-a6ef-a0386a49a974.vmdk or device None with type sparse {{(pid=61806) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 968.194895] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-e647e7ea-c6cc-4fda-9034-80acc4566d87 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 968.215703] env[61806]: DEBUG oslo_vmware.api [None req-c8cf9c01-54a9-4c86-b57f-b2e6a2fbc11a tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] Waiting for the task: (returnval){ [ 968.215703] env[61806]: value = "task-1294989" [ 968.215703] env[61806]: _type = "Task" [ 968.215703] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 968.224536] env[61806]: DEBUG oslo_vmware.api [None req-c8cf9c01-54a9-4c86-b57f-b2e6a2fbc11a tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] Task: {'id': task-1294989, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 968.313906] env[61806]: DEBUG nova.network.neutron [None req-2c70fcbb-efa7-459d-81c0-a2a8a184893c tempest-AttachVolumeNegativeTest-1683033704 tempest-AttachVolumeNegativeTest-1683033704-project-member] [instance: e4638faa-7be1-4909-a595-c437837fa314] Updating instance_info_cache with network_info: [{"id": "bfb37474-8e7a-4820-9045-e1b388cfb87d", "address": "fa:16:3e:b5:75:56", "network": {"id": "d636ca55-0737-48e4-8d9d-7fe0f797f49e", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-2090334349-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "22e47882c2c44ce492dff6b1c5d782e2", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6f493cd8-1cb4-42a1-8d56-bfa6ac7cf563", "external-id": "nsx-vlan-transportzone-931", "segmentation_id": 931, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapbfb37474-8e", "ovs_interfaceid": "bfb37474-8e7a-4820-9045-e1b388cfb87d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61806) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 968.518231] env[61806]: DEBUG oslo_vmware.api [None req-49ec17b3-08f6-4e32-88fe-0409196f3352 tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] Task: {'id': task-1294987, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 968.616354] env[61806]: DEBUG oslo_concurrency.lockutils [None req-60f88c72-8f2b-4f98-b3cc-7787383864bd tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Lock "461f98f2-80ba-4088-9e7f-1ba2a9b0c1d9" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 14.361s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 968.727220] env[61806]: DEBUG oslo_vmware.api [None req-c8cf9c01-54a9-4c86-b57f-b2e6a2fbc11a tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] Task: {'id': task-1294989, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 968.790724] env[61806]: DEBUG nova.compute.manager [None req-28d96f96-d23f-4c45-9d59-0355de5e0c4f tempest-ServersNegativeTestJSON-891856253 tempest-ServersNegativeTestJSON-891856253-project-member] [instance: e146f7e5-bbf8-4adf-a2e9-b7e753438976] Start spawning the instance on the hypervisor. {{(pid=61806) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 968.812712] env[61806]: DEBUG nova.virt.hardware [None req-28d96f96-d23f-4c45-9d59-0355de5e0c4f tempest-ServersNegativeTestJSON-891856253 tempest-ServersNegativeTestJSON-891856253-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-15T18:05:10Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-15T18:04:52Z,direct_url=,disk_format='vmdk',id=ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='84ba83fce2c349a6988173c5d6fc3b07',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-15T18:04:53Z,virtual_size=,visibility=), allow threads: False {{(pid=61806) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 968.812989] env[61806]: DEBUG nova.virt.hardware [None req-28d96f96-d23f-4c45-9d59-0355de5e0c4f tempest-ServersNegativeTestJSON-891856253 tempest-ServersNegativeTestJSON-891856253-project-member] Flavor limits 0:0:0 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 968.813171] env[61806]: DEBUG nova.virt.hardware [None req-28d96f96-d23f-4c45-9d59-0355de5e0c4f tempest-ServersNegativeTestJSON-891856253 tempest-ServersNegativeTestJSON-891856253-project-member] Image limits 0:0:0 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 968.813365] env[61806]: DEBUG nova.virt.hardware [None req-28d96f96-d23f-4c45-9d59-0355de5e0c4f tempest-ServersNegativeTestJSON-891856253 tempest-ServersNegativeTestJSON-891856253-project-member] Flavor pref 0:0:0 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 968.813562] env[61806]: DEBUG nova.virt.hardware [None req-28d96f96-d23f-4c45-9d59-0355de5e0c4f tempest-ServersNegativeTestJSON-891856253 tempest-ServersNegativeTestJSON-891856253-project-member] Image pref 0:0:0 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 968.813727] env[61806]: DEBUG nova.virt.hardware [None req-28d96f96-d23f-4c45-9d59-0355de5e0c4f tempest-ServersNegativeTestJSON-891856253 tempest-ServersNegativeTestJSON-891856253-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 968.813953] env[61806]: DEBUG nova.virt.hardware [None req-28d96f96-d23f-4c45-9d59-0355de5e0c4f tempest-ServersNegativeTestJSON-891856253 tempest-ServersNegativeTestJSON-891856253-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61806) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 968.814141] env[61806]: DEBUG nova.virt.hardware [None req-28d96f96-d23f-4c45-9d59-0355de5e0c4f tempest-ServersNegativeTestJSON-891856253 tempest-ServersNegativeTestJSON-891856253-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61806) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 968.814386] env[61806]: DEBUG nova.virt.hardware [None req-28d96f96-d23f-4c45-9d59-0355de5e0c4f tempest-ServersNegativeTestJSON-891856253 tempest-ServersNegativeTestJSON-891856253-project-member] Got 1 possible topologies {{(pid=61806) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 968.814563] env[61806]: DEBUG nova.virt.hardware [None req-28d96f96-d23f-4c45-9d59-0355de5e0c4f tempest-ServersNegativeTestJSON-891856253 tempest-ServersNegativeTestJSON-891856253-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61806) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 968.814742] env[61806]: DEBUG nova.virt.hardware [None req-28d96f96-d23f-4c45-9d59-0355de5e0c4f tempest-ServersNegativeTestJSON-891856253 tempest-ServersNegativeTestJSON-891856253-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61806) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 968.815648] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cc89f3f4-7bea-4f20-9a13-e2230850ffeb {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 968.818359] env[61806]: DEBUG oslo_concurrency.lockutils [None req-2c70fcbb-efa7-459d-81c0-a2a8a184893c tempest-AttachVolumeNegativeTest-1683033704 tempest-AttachVolumeNegativeTest-1683033704-project-member] Releasing lock "refresh_cache-e4638faa-7be1-4909-a595-c437837fa314" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 968.818637] env[61806]: DEBUG nova.compute.manager [None req-2c70fcbb-efa7-459d-81c0-a2a8a184893c tempest-AttachVolumeNegativeTest-1683033704 tempest-AttachVolumeNegativeTest-1683033704-project-member] [instance: e4638faa-7be1-4909-a595-c437837fa314] Instance network_info: |[{"id": "bfb37474-8e7a-4820-9045-e1b388cfb87d", "address": "fa:16:3e:b5:75:56", "network": {"id": "d636ca55-0737-48e4-8d9d-7fe0f797f49e", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-2090334349-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "22e47882c2c44ce492dff6b1c5d782e2", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6f493cd8-1cb4-42a1-8d56-bfa6ac7cf563", "external-id": "nsx-vlan-transportzone-931", "segmentation_id": 931, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapbfb37474-8e", "ovs_interfaceid": "bfb37474-8e7a-4820-9045-e1b388cfb87d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61806) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 968.819011] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-2c70fcbb-efa7-459d-81c0-a2a8a184893c tempest-AttachVolumeNegativeTest-1683033704 tempest-AttachVolumeNegativeTest-1683033704-project-member] [instance: e4638faa-7be1-4909-a595-c437837fa314] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:b5:75:56', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '6f493cd8-1cb4-42a1-8d56-bfa6ac7cf563', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'bfb37474-8e7a-4820-9045-e1b388cfb87d', 'vif_model': 'vmxnet3'}] {{(pid=61806) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 968.826445] env[61806]: DEBUG oslo.service.loopingcall [None req-2c70fcbb-efa7-459d-81c0-a2a8a184893c tempest-AttachVolumeNegativeTest-1683033704 tempest-AttachVolumeNegativeTest-1683033704-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61806) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 968.827459] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: e4638faa-7be1-4909-a595-c437837fa314] Creating VM on the ESX host {{(pid=61806) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 968.827684] env[61806]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-9fe52f6c-6130-4c3b-bdc5-f876355739df {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 968.845268] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a1d1d4f5-e91e-429c-ae89-4453e0e703b3 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 968.850060] env[61806]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 968.850060] env[61806]: value = "task-1294990" [ 968.850060] env[61806]: _type = "Task" [ 968.850060] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 968.858153] env[61806]: DEBUG oslo_vmware.api [-] Task: {'id': task-1294990, 'name': CreateVM_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 968.908020] env[61806]: INFO nova.compute.manager [None req-1e11ef2b-e661-483a-b1ad-9cb57e29c94f tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] [instance: 461f98f2-80ba-4088-9e7f-1ba2a9b0c1d9] Rebuilding instance [ 968.954395] env[61806]: DEBUG nova.compute.manager [None req-1e11ef2b-e661-483a-b1ad-9cb57e29c94f tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] [instance: 461f98f2-80ba-4088-9e7f-1ba2a9b0c1d9] Checking state {{(pid=61806) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 968.955288] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-258dd979-cff4-4873-953a-344a5b623800 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 969.018954] env[61806]: DEBUG oslo_vmware.api [None req-49ec17b3-08f6-4e32-88fe-0409196f3352 tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] Task: {'id': task-1294987, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 969.021596] env[61806]: DEBUG nova.compute.manager [req-63b8d971-7ca5-48b4-9161-8d5f7a29bed4 req-549aedc3-e4f0-482e-85de-f8f12921b3e9 service nova] [instance: e4638faa-7be1-4909-a595-c437837fa314] Received event network-changed-bfb37474-8e7a-4820-9045-e1b388cfb87d {{(pid=61806) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 969.021825] env[61806]: DEBUG nova.compute.manager [req-63b8d971-7ca5-48b4-9161-8d5f7a29bed4 req-549aedc3-e4f0-482e-85de-f8f12921b3e9 service nova] [instance: e4638faa-7be1-4909-a595-c437837fa314] Refreshing instance network info cache due to event network-changed-bfb37474-8e7a-4820-9045-e1b388cfb87d. {{(pid=61806) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 969.022614] env[61806]: DEBUG oslo_concurrency.lockutils [req-63b8d971-7ca5-48b4-9161-8d5f7a29bed4 req-549aedc3-e4f0-482e-85de-f8f12921b3e9 service nova] Acquiring lock "refresh_cache-e4638faa-7be1-4909-a595-c437837fa314" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 969.022614] env[61806]: DEBUG oslo_concurrency.lockutils [req-63b8d971-7ca5-48b4-9161-8d5f7a29bed4 req-549aedc3-e4f0-482e-85de-f8f12921b3e9 service nova] Acquired lock "refresh_cache-e4638faa-7be1-4909-a595-c437837fa314" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 969.022614] env[61806]: DEBUG nova.network.neutron [req-63b8d971-7ca5-48b4-9161-8d5f7a29bed4 req-549aedc3-e4f0-482e-85de-f8f12921b3e9 service nova] [instance: e4638faa-7be1-4909-a595-c437837fa314] Refreshing network info cache for port bfb37474-8e7a-4820-9045-e1b388cfb87d {{(pid=61806) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 969.052091] env[61806]: DEBUG oslo_concurrency.lockutils [None req-0bbe3134-73ea-4d1b-8035-8d9ceb5bbb4c tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" :: held 2.286s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 969.055164] env[61806]: DEBUG oslo_concurrency.lockutils [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 2.211s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 969.088159] env[61806]: DEBUG nova.network.neutron [None req-28d96f96-d23f-4c45-9d59-0355de5e0c4f tempest-ServersNegativeTestJSON-891856253 tempest-ServersNegativeTestJSON-891856253-project-member] [instance: e146f7e5-bbf8-4adf-a2e9-b7e753438976] Successfully updated port: b68f5483-a11f-4b35-aaa0-2dcd7f50e448 {{(pid=61806) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 969.226682] env[61806]: DEBUG oslo_vmware.api [None req-c8cf9c01-54a9-4c86-b57f-b2e6a2fbc11a tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] Task: {'id': task-1294989, 'name': ReconfigVM_Task, 'duration_secs': 0.568275} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 969.226984] env[61806]: DEBUG nova.virt.vmwareapi.volumeops [None req-c8cf9c01-54a9-4c86-b57f-b2e6a2fbc11a tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] [instance: 0bac3048-759a-47d9-a6ef-a0386a49a974] Reconfigured VM instance instance-0000005c to attach disk [datastore2] 0bac3048-759a-47d9-a6ef-a0386a49a974/0bac3048-759a-47d9-a6ef-a0386a49a974.vmdk or device None with type sparse {{(pid=61806) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 969.227766] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-2535b1e6-104c-41c2-9503-d130ad027211 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 969.234642] env[61806]: DEBUG oslo_vmware.api [None req-c8cf9c01-54a9-4c86-b57f-b2e6a2fbc11a tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] Waiting for the task: (returnval){ [ 969.234642] env[61806]: value = "task-1294991" [ 969.234642] env[61806]: _type = "Task" [ 969.234642] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 969.244144] env[61806]: DEBUG oslo_vmware.api [None req-c8cf9c01-54a9-4c86-b57f-b2e6a2fbc11a tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] Task: {'id': task-1294991, 'name': Rename_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 969.362227] env[61806]: DEBUG oslo_vmware.api [-] Task: {'id': task-1294990, 'name': CreateVM_Task, 'duration_secs': 0.46774} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 969.362450] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: e4638faa-7be1-4909-a595-c437837fa314] Created VM on the ESX host {{(pid=61806) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 969.363353] env[61806]: DEBUG oslo_concurrency.lockutils [None req-2c70fcbb-efa7-459d-81c0-a2a8a184893c tempest-AttachVolumeNegativeTest-1683033704 tempest-AttachVolumeNegativeTest-1683033704-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 969.363477] env[61806]: DEBUG oslo_concurrency.lockutils [None req-2c70fcbb-efa7-459d-81c0-a2a8a184893c tempest-AttachVolumeNegativeTest-1683033704 tempest-AttachVolumeNegativeTest-1683033704-project-member] Acquired lock "[datastore2] devstack-image-cache_base/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 969.363755] env[61806]: DEBUG oslo_concurrency.lockutils [None req-2c70fcbb-efa7-459d-81c0-a2a8a184893c tempest-AttachVolumeNegativeTest-1683033704 tempest-AttachVolumeNegativeTest-1683033704-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 969.364032] env[61806]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e1826ac6-2131-44e4-9878-ddd696a91e97 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 969.370737] env[61806]: DEBUG oslo_vmware.api [None req-2c70fcbb-efa7-459d-81c0-a2a8a184893c tempest-AttachVolumeNegativeTest-1683033704 tempest-AttachVolumeNegativeTest-1683033704-project-member] Waiting for the task: (returnval){ [ 969.370737] env[61806]: value = "session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]52e1ca2c-a1d7-23a9-c337-7da4d2c1fb34" [ 969.370737] env[61806]: _type = "Task" [ 969.370737] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 969.379774] env[61806]: DEBUG oslo_vmware.api [None req-2c70fcbb-efa7-459d-81c0-a2a8a184893c tempest-AttachVolumeNegativeTest-1683033704 tempest-AttachVolumeNegativeTest-1683033704-project-member] Task: {'id': session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]52e1ca2c-a1d7-23a9-c337-7da4d2c1fb34, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 969.467280] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [None req-1e11ef2b-e661-483a-b1ad-9cb57e29c94f tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] [instance: 461f98f2-80ba-4088-9e7f-1ba2a9b0c1d9] Powering off the VM {{(pid=61806) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 969.467615] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-71d43d9b-61c7-4011-97d5-ad7fd8bbbd55 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 969.475868] env[61806]: DEBUG oslo_vmware.api [None req-1e11ef2b-e661-483a-b1ad-9cb57e29c94f tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Waiting for the task: (returnval){ [ 969.475868] env[61806]: value = "task-1294992" [ 969.475868] env[61806]: _type = "Task" [ 969.475868] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 969.485414] env[61806]: DEBUG oslo_vmware.api [None req-1e11ef2b-e661-483a-b1ad-9cb57e29c94f tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Task: {'id': task-1294992, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 969.519033] env[61806]: DEBUG oslo_vmware.api [None req-49ec17b3-08f6-4e32-88fe-0409196f3352 tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] Task: {'id': task-1294987, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 969.591216] env[61806]: DEBUG oslo_concurrency.lockutils [None req-28d96f96-d23f-4c45-9d59-0355de5e0c4f tempest-ServersNegativeTestJSON-891856253 tempest-ServersNegativeTestJSON-891856253-project-member] Acquiring lock "refresh_cache-e146f7e5-bbf8-4adf-a2e9-b7e753438976" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 969.591372] env[61806]: DEBUG oslo_concurrency.lockutils [None req-28d96f96-d23f-4c45-9d59-0355de5e0c4f tempest-ServersNegativeTestJSON-891856253 tempest-ServersNegativeTestJSON-891856253-project-member] Acquired lock "refresh_cache-e146f7e5-bbf8-4adf-a2e9-b7e753438976" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 969.591525] env[61806]: DEBUG nova.network.neutron [None req-28d96f96-d23f-4c45-9d59-0355de5e0c4f tempest-ServersNegativeTestJSON-891856253 tempest-ServersNegativeTestJSON-891856253-project-member] [instance: e146f7e5-bbf8-4adf-a2e9-b7e753438976] Building network info cache for instance {{(pid=61806) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 969.635788] env[61806]: INFO nova.scheduler.client.report [None req-0bbe3134-73ea-4d1b-8035-8d9ceb5bbb4c tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Deleted allocation for migration 07d765d6-8d01-4ca8-b027-cc87b221420f [ 969.744587] env[61806]: DEBUG oslo_vmware.api [None req-c8cf9c01-54a9-4c86-b57f-b2e6a2fbc11a tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] Task: {'id': task-1294991, 'name': Rename_Task, 'duration_secs': 0.156879} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 969.744858] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [None req-c8cf9c01-54a9-4c86-b57f-b2e6a2fbc11a tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] [instance: 0bac3048-759a-47d9-a6ef-a0386a49a974] Powering on the VM {{(pid=61806) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 969.745238] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-a789dc6e-f00a-4227-8588-2a71695b7009 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 969.753192] env[61806]: DEBUG oslo_vmware.api [None req-c8cf9c01-54a9-4c86-b57f-b2e6a2fbc11a tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] Waiting for the task: (returnval){ [ 969.753192] env[61806]: value = "task-1294993" [ 969.753192] env[61806]: _type = "Task" [ 969.753192] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 969.763379] env[61806]: DEBUG oslo_vmware.api [None req-c8cf9c01-54a9-4c86-b57f-b2e6a2fbc11a tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] Task: {'id': task-1294993, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 969.801872] env[61806]: DEBUG nova.network.neutron [req-63b8d971-7ca5-48b4-9161-8d5f7a29bed4 req-549aedc3-e4f0-482e-85de-f8f12921b3e9 service nova] [instance: e4638faa-7be1-4909-a595-c437837fa314] Updated VIF entry in instance network info cache for port bfb37474-8e7a-4820-9045-e1b388cfb87d. {{(pid=61806) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 969.802449] env[61806]: DEBUG nova.network.neutron [req-63b8d971-7ca5-48b4-9161-8d5f7a29bed4 req-549aedc3-e4f0-482e-85de-f8f12921b3e9 service nova] [instance: e4638faa-7be1-4909-a595-c437837fa314] Updating instance_info_cache with network_info: [{"id": "bfb37474-8e7a-4820-9045-e1b388cfb87d", "address": "fa:16:3e:b5:75:56", "network": {"id": "d636ca55-0737-48e4-8d9d-7fe0f797f49e", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-2090334349-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "22e47882c2c44ce492dff6b1c5d782e2", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6f493cd8-1cb4-42a1-8d56-bfa6ac7cf563", "external-id": "nsx-vlan-transportzone-931", "segmentation_id": 931, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapbfb37474-8e", "ovs_interfaceid": "bfb37474-8e7a-4820-9045-e1b388cfb87d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61806) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 969.885418] env[61806]: DEBUG oslo_vmware.api [None req-2c70fcbb-efa7-459d-81c0-a2a8a184893c tempest-AttachVolumeNegativeTest-1683033704 tempest-AttachVolumeNegativeTest-1683033704-project-member] Task: {'id': session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]52e1ca2c-a1d7-23a9-c337-7da4d2c1fb34, 'name': SearchDatastore_Task, 'duration_secs': 0.011425} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 969.885769] env[61806]: DEBUG oslo_concurrency.lockutils [None req-2c70fcbb-efa7-459d-81c0-a2a8a184893c tempest-AttachVolumeNegativeTest-1683033704 tempest-AttachVolumeNegativeTest-1683033704-project-member] Releasing lock "[datastore2] devstack-image-cache_base/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 969.886015] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-2c70fcbb-efa7-459d-81c0-a2a8a184893c tempest-AttachVolumeNegativeTest-1683033704 tempest-AttachVolumeNegativeTest-1683033704-project-member] [instance: e4638faa-7be1-4909-a595-c437837fa314] Processing image ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa {{(pid=61806) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 969.886270] env[61806]: DEBUG oslo_concurrency.lockutils [None req-2c70fcbb-efa7-459d-81c0-a2a8a184893c tempest-AttachVolumeNegativeTest-1683033704 tempest-AttachVolumeNegativeTest-1683033704-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa.vmdk" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 969.886426] env[61806]: DEBUG oslo_concurrency.lockutils [None req-2c70fcbb-efa7-459d-81c0-a2a8a184893c tempest-AttachVolumeNegativeTest-1683033704 tempest-AttachVolumeNegativeTest-1683033704-project-member] Acquired lock "[datastore2] devstack-image-cache_base/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa.vmdk" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 969.886643] env[61806]: DEBUG nova.virt.vmwareapi.ds_util [None req-2c70fcbb-efa7-459d-81c0-a2a8a184893c tempest-AttachVolumeNegativeTest-1683033704 tempest-AttachVolumeNegativeTest-1683033704-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61806) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 969.886931] env[61806]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-b10cb037-70fa-47a7-84ee-6eb2fa698190 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 969.897872] env[61806]: DEBUG nova.virt.vmwareapi.ds_util [None req-2c70fcbb-efa7-459d-81c0-a2a8a184893c tempest-AttachVolumeNegativeTest-1683033704 tempest-AttachVolumeNegativeTest-1683033704-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61806) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 969.898102] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-2c70fcbb-efa7-459d-81c0-a2a8a184893c tempest-AttachVolumeNegativeTest-1683033704 tempest-AttachVolumeNegativeTest-1683033704-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61806) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 969.898938] env[61806]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-670ac388-e648-4623-b1ff-b49ba44fbf60 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 969.907218] env[61806]: DEBUG oslo_vmware.api [None req-2c70fcbb-efa7-459d-81c0-a2a8a184893c tempest-AttachVolumeNegativeTest-1683033704 tempest-AttachVolumeNegativeTest-1683033704-project-member] Waiting for the task: (returnval){ [ 969.907218] env[61806]: value = "session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]52960e53-1ceb-1044-00b0-bf08e27016c8" [ 969.907218] env[61806]: _type = "Task" [ 969.907218] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 969.917341] env[61806]: DEBUG oslo_vmware.api [None req-2c70fcbb-efa7-459d-81c0-a2a8a184893c tempest-AttachVolumeNegativeTest-1683033704 tempest-AttachVolumeNegativeTest-1683033704-project-member] Task: {'id': session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]52960e53-1ceb-1044-00b0-bf08e27016c8, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 969.986481] env[61806]: DEBUG oslo_vmware.api [None req-1e11ef2b-e661-483a-b1ad-9cb57e29c94f tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Task: {'id': task-1294992, 'name': PowerOffVM_Task, 'duration_secs': 0.244447} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 969.986761] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [None req-1e11ef2b-e661-483a-b1ad-9cb57e29c94f tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] [instance: 461f98f2-80ba-4088-9e7f-1ba2a9b0c1d9] Powered off the VM {{(pid=61806) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 969.986991] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-1e11ef2b-e661-483a-b1ad-9cb57e29c94f tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] [instance: 461f98f2-80ba-4088-9e7f-1ba2a9b0c1d9] Destroying instance {{(pid=61806) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 969.987817] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-11ee3dc3-91a1-459c-a95c-b52d227cf09c {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 969.996642] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-1e11ef2b-e661-483a-b1ad-9cb57e29c94f tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] [instance: 461f98f2-80ba-4088-9e7f-1ba2a9b0c1d9] Unregistering the VM {{(pid=61806) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 969.996936] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-00ba3018-19b0-4177-8746-c104f0ff7716 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 970.020203] env[61806]: DEBUG oslo_vmware.api [None req-49ec17b3-08f6-4e32-88fe-0409196f3352 tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] Task: {'id': task-1294987, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 970.077939] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-1e11ef2b-e661-483a-b1ad-9cb57e29c94f tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] [instance: 461f98f2-80ba-4088-9e7f-1ba2a9b0c1d9] Unregistered the VM {{(pid=61806) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 970.078207] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-1e11ef2b-e661-483a-b1ad-9cb57e29c94f tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] [instance: 461f98f2-80ba-4088-9e7f-1ba2a9b0c1d9] Deleting contents of the VM from datastore datastore2 {{(pid=61806) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 970.078435] env[61806]: DEBUG nova.virt.vmwareapi.ds_util [None req-1e11ef2b-e661-483a-b1ad-9cb57e29c94f tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Deleting the datastore file [datastore2] 461f98f2-80ba-4088-9e7f-1ba2a9b0c1d9 {{(pid=61806) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 970.078734] env[61806]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-119c16e8-239b-4d22-9f98-068bd444b5c6 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 970.087273] env[61806]: DEBUG oslo_vmware.api [None req-1e11ef2b-e661-483a-b1ad-9cb57e29c94f tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Waiting for the task: (returnval){ [ 970.087273] env[61806]: value = "task-1294995" [ 970.087273] env[61806]: _type = "Task" [ 970.087273] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 970.097729] env[61806]: DEBUG oslo_vmware.api [None req-1e11ef2b-e661-483a-b1ad-9cb57e29c94f tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Task: {'id': task-1294995, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 970.098870] env[61806]: DEBUG nova.compute.resource_tracker [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Instance 67d1bdd8-cf8f-4fb5-bd3c-42008e4d60e3 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61806) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 970.099022] env[61806]: DEBUG nova.compute.resource_tracker [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Instance d1802da4-a670-427d-84d7-0fcb4717e18f actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61806) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 970.099264] env[61806]: DEBUG nova.compute.resource_tracker [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Instance 6f6ba57a-e2d6-4749-a53a-e263861cb1c0 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61806) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 970.099333] env[61806]: DEBUG nova.compute.resource_tracker [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Instance d9caee4a-a00c-4e50-b01c-99a8f796a09f actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61806) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 970.099410] env[61806]: DEBUG nova.compute.resource_tracker [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Instance fd9e520a-2e7b-43d4-b84d-8933c32cc6a6 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61806) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 970.099525] env[61806]: DEBUG nova.compute.resource_tracker [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Instance 38e96e22-4200-4175-9085-31977357da2d actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61806) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 970.099640] env[61806]: DEBUG nova.compute.resource_tracker [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Instance 676f18b3-3da1-4a53-b96d-fe64ee8f9101 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 256, 'VCPU': 1}}. {{(pid=61806) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 970.099812] env[61806]: DEBUG nova.compute.resource_tracker [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Instance bdbaef31-7d37-40cb-ac3e-ad152b37e2d2 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61806) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 970.099854] env[61806]: DEBUG nova.compute.resource_tracker [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Instance 461f98f2-80ba-4088-9e7f-1ba2a9b0c1d9 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61806) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 970.099967] env[61806]: DEBUG nova.compute.resource_tracker [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Instance 0bac3048-759a-47d9-a6ef-a0386a49a974 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61806) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 970.100091] env[61806]: DEBUG nova.compute.resource_tracker [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Instance e4638faa-7be1-4909-a595-c437837fa314 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61806) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 970.100205] env[61806]: DEBUG nova.compute.resource_tracker [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Instance e146f7e5-bbf8-4adf-a2e9-b7e753438976 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61806) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 970.132306] env[61806]: DEBUG nova.network.neutron [None req-28d96f96-d23f-4c45-9d59-0355de5e0c4f tempest-ServersNegativeTestJSON-891856253 tempest-ServersNegativeTestJSON-891856253-project-member] [instance: e146f7e5-bbf8-4adf-a2e9-b7e753438976] Instance cache missing network info. {{(pid=61806) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 970.143617] env[61806]: DEBUG oslo_concurrency.lockutils [None req-0bbe3134-73ea-4d1b-8035-8d9ceb5bbb4c tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Lock "676f18b3-3da1-4a53-b96d-fe64ee8f9101" "released" by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" :: held 8.876s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 970.268050] env[61806]: DEBUG oslo_vmware.api [None req-c8cf9c01-54a9-4c86-b57f-b2e6a2fbc11a tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] Task: {'id': task-1294993, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 970.290294] env[61806]: DEBUG nova.network.neutron [None req-28d96f96-d23f-4c45-9d59-0355de5e0c4f tempest-ServersNegativeTestJSON-891856253 tempest-ServersNegativeTestJSON-891856253-project-member] [instance: e146f7e5-bbf8-4adf-a2e9-b7e753438976] Updating instance_info_cache with network_info: [{"id": "b68f5483-a11f-4b35-aaa0-2dcd7f50e448", "address": "fa:16:3e:eb:e5:76", "network": {"id": "fd35ea93-8914-4b54-a4ea-f45fc750982b", "bridge": "br-int", "label": "tempest-ServersNegativeTestJSON-2117068882-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d2e2b7eec47e4b10898215123408692a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "2d94740a-bce8-4103-8ecf-230d02ec0a44", "external-id": "nsx-vlan-transportzone-149", "segmentation_id": 149, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb68f5483-a1", "ovs_interfaceid": "b68f5483-a11f-4b35-aaa0-2dcd7f50e448", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61806) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 970.306105] env[61806]: DEBUG oslo_concurrency.lockutils [req-63b8d971-7ca5-48b4-9161-8d5f7a29bed4 req-549aedc3-e4f0-482e-85de-f8f12921b3e9 service nova] Releasing lock "refresh_cache-e4638faa-7be1-4909-a595-c437837fa314" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 970.306436] env[61806]: DEBUG nova.compute.manager [req-63b8d971-7ca5-48b4-9161-8d5f7a29bed4 req-549aedc3-e4f0-482e-85de-f8f12921b3e9 service nova] [instance: e146f7e5-bbf8-4adf-a2e9-b7e753438976] Received event network-vif-plugged-b68f5483-a11f-4b35-aaa0-2dcd7f50e448 {{(pid=61806) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 970.306711] env[61806]: DEBUG oslo_concurrency.lockutils [req-63b8d971-7ca5-48b4-9161-8d5f7a29bed4 req-549aedc3-e4f0-482e-85de-f8f12921b3e9 service nova] Acquiring lock "e146f7e5-bbf8-4adf-a2e9-b7e753438976-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 970.306941] env[61806]: DEBUG oslo_concurrency.lockutils [req-63b8d971-7ca5-48b4-9161-8d5f7a29bed4 req-549aedc3-e4f0-482e-85de-f8f12921b3e9 service nova] Lock "e146f7e5-bbf8-4adf-a2e9-b7e753438976-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 970.307169] env[61806]: DEBUG oslo_concurrency.lockutils [req-63b8d971-7ca5-48b4-9161-8d5f7a29bed4 req-549aedc3-e4f0-482e-85de-f8f12921b3e9 service nova] Lock "e146f7e5-bbf8-4adf-a2e9-b7e753438976-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 970.307360] env[61806]: DEBUG nova.compute.manager [req-63b8d971-7ca5-48b4-9161-8d5f7a29bed4 req-549aedc3-e4f0-482e-85de-f8f12921b3e9 service nova] [instance: e146f7e5-bbf8-4adf-a2e9-b7e753438976] No waiting events found dispatching network-vif-plugged-b68f5483-a11f-4b35-aaa0-2dcd7f50e448 {{(pid=61806) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 970.307542] env[61806]: WARNING nova.compute.manager [req-63b8d971-7ca5-48b4-9161-8d5f7a29bed4 req-549aedc3-e4f0-482e-85de-f8f12921b3e9 service nova] [instance: e146f7e5-bbf8-4adf-a2e9-b7e753438976] Received unexpected event network-vif-plugged-b68f5483-a11f-4b35-aaa0-2dcd7f50e448 for instance with vm_state building and task_state spawning. [ 970.419656] env[61806]: DEBUG oslo_vmware.api [None req-2c70fcbb-efa7-459d-81c0-a2a8a184893c tempest-AttachVolumeNegativeTest-1683033704 tempest-AttachVolumeNegativeTest-1683033704-project-member] Task: {'id': session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]52960e53-1ceb-1044-00b0-bf08e27016c8, 'name': SearchDatastore_Task, 'duration_secs': 0.017322} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 970.420601] env[61806]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ad2579af-0e46-4a88-b4e1-615873fbf9fe {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 970.427453] env[61806]: DEBUG oslo_vmware.api [None req-2c70fcbb-efa7-459d-81c0-a2a8a184893c tempest-AttachVolumeNegativeTest-1683033704 tempest-AttachVolumeNegativeTest-1683033704-project-member] Waiting for the task: (returnval){ [ 970.427453] env[61806]: value = "session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]5225ce23-3eff-b854-b9e2-5e8e68059eea" [ 970.427453] env[61806]: _type = "Task" [ 970.427453] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 970.442101] env[61806]: DEBUG oslo_vmware.api [None req-2c70fcbb-efa7-459d-81c0-a2a8a184893c tempest-AttachVolumeNegativeTest-1683033704 tempest-AttachVolumeNegativeTest-1683033704-project-member] Task: {'id': session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]5225ce23-3eff-b854-b9e2-5e8e68059eea, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 970.520723] env[61806]: DEBUG oslo_vmware.api [None req-49ec17b3-08f6-4e32-88fe-0409196f3352 tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] Task: {'id': task-1294987, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 970.600049] env[61806]: DEBUG oslo_vmware.api [None req-1e11ef2b-e661-483a-b1ad-9cb57e29c94f tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Task: {'id': task-1294995, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.321208} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 970.600523] env[61806]: DEBUG nova.virt.vmwareapi.ds_util [None req-1e11ef2b-e661-483a-b1ad-9cb57e29c94f tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Deleted the datastore file {{(pid=61806) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 970.600727] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-1e11ef2b-e661-483a-b1ad-9cb57e29c94f tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] [instance: 461f98f2-80ba-4088-9e7f-1ba2a9b0c1d9] Deleted contents of the VM from datastore datastore2 {{(pid=61806) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 970.601328] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-1e11ef2b-e661-483a-b1ad-9cb57e29c94f tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] [instance: 461f98f2-80ba-4088-9e7f-1ba2a9b0c1d9] Instance destroyed {{(pid=61806) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 970.605901] env[61806]: DEBUG nova.compute.resource_tracker [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Instance 962d3d88-03ee-4ff2-9d79-15091671b330 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61806) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 970.605901] env[61806]: DEBUG nova.compute.resource_tracker [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Total usable vcpus: 48, total allocated vcpus: 12 {{(pid=61806) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 970.605901] env[61806]: DEBUG nova.compute.resource_tracker [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=2880MB phys_disk=200GB used_disk=12GB total_vcpus=48 used_vcpus=12 pci_stats=[] {{(pid=61806) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 970.764675] env[61806]: DEBUG oslo_vmware.api [None req-c8cf9c01-54a9-4c86-b57f-b2e6a2fbc11a tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] Task: {'id': task-1294993, 'name': PowerOnVM_Task, 'duration_secs': 0.718456} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 970.767418] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [None req-c8cf9c01-54a9-4c86-b57f-b2e6a2fbc11a tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] [instance: 0bac3048-759a-47d9-a6ef-a0386a49a974] Powered on the VM {{(pid=61806) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 970.767637] env[61806]: INFO nova.compute.manager [None req-c8cf9c01-54a9-4c86-b57f-b2e6a2fbc11a tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] [instance: 0bac3048-759a-47d9-a6ef-a0386a49a974] Took 8.91 seconds to spawn the instance on the hypervisor. [ 970.767820] env[61806]: DEBUG nova.compute.manager [None req-c8cf9c01-54a9-4c86-b57f-b2e6a2fbc11a tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] [instance: 0bac3048-759a-47d9-a6ef-a0386a49a974] Checking state {{(pid=61806) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 970.768915] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6108fc45-af93-44ab-974a-ff60941ea4cb {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 970.793222] env[61806]: DEBUG oslo_concurrency.lockutils [None req-28d96f96-d23f-4c45-9d59-0355de5e0c4f tempest-ServersNegativeTestJSON-891856253 tempest-ServersNegativeTestJSON-891856253-project-member] Releasing lock "refresh_cache-e146f7e5-bbf8-4adf-a2e9-b7e753438976" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 970.793502] env[61806]: DEBUG nova.compute.manager [None req-28d96f96-d23f-4c45-9d59-0355de5e0c4f tempest-ServersNegativeTestJSON-891856253 tempest-ServersNegativeTestJSON-891856253-project-member] [instance: e146f7e5-bbf8-4adf-a2e9-b7e753438976] Instance network_info: |[{"id": "b68f5483-a11f-4b35-aaa0-2dcd7f50e448", "address": "fa:16:3e:eb:e5:76", "network": {"id": "fd35ea93-8914-4b54-a4ea-f45fc750982b", "bridge": "br-int", "label": "tempest-ServersNegativeTestJSON-2117068882-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d2e2b7eec47e4b10898215123408692a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "2d94740a-bce8-4103-8ecf-230d02ec0a44", "external-id": "nsx-vlan-transportzone-149", "segmentation_id": 149, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb68f5483-a1", "ovs_interfaceid": "b68f5483-a11f-4b35-aaa0-2dcd7f50e448", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61806) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 970.796452] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-28d96f96-d23f-4c45-9d59-0355de5e0c4f tempest-ServersNegativeTestJSON-891856253 tempest-ServersNegativeTestJSON-891856253-project-member] [instance: e146f7e5-bbf8-4adf-a2e9-b7e753438976] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:eb:e5:76', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '2d94740a-bce8-4103-8ecf-230d02ec0a44', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'b68f5483-a11f-4b35-aaa0-2dcd7f50e448', 'vif_model': 'vmxnet3'}] {{(pid=61806) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 970.804556] env[61806]: DEBUG oslo.service.loopingcall [None req-28d96f96-d23f-4c45-9d59-0355de5e0c4f tempest-ServersNegativeTestJSON-891856253 tempest-ServersNegativeTestJSON-891856253-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61806) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 970.807695] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: e146f7e5-bbf8-4adf-a2e9-b7e753438976] Creating VM on the ESX host {{(pid=61806) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 970.807695] env[61806]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-9ce13e4c-1897-467c-b1ee-a4936783b764 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 970.823991] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-084fd560-e507-459a-8392-3f17ab977bec {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 970.832759] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0e6f1ed6-7764-420c-9ac5-f46b06283174 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 970.838108] env[61806]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 970.838108] env[61806]: value = "task-1294996" [ 970.838108] env[61806]: _type = "Task" [ 970.838108] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 970.866205] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a670824d-9302-4bdc-903e-adc50fc77edf {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 970.871817] env[61806]: DEBUG oslo_vmware.api [-] Task: {'id': task-1294996, 'name': CreateVM_Task} progress is 10%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 970.878445] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-abe2b76a-05b2-44ae-b6e7-6153742bea46 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 970.894109] env[61806]: DEBUG nova.compute.provider_tree [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Inventory has not changed in ProviderTree for provider: a2858be1-fd22-4e08-979e-87ad25293407 {{(pid=61806) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 970.940811] env[61806]: DEBUG oslo_vmware.api [None req-2c70fcbb-efa7-459d-81c0-a2a8a184893c tempest-AttachVolumeNegativeTest-1683033704 tempest-AttachVolumeNegativeTest-1683033704-project-member] Task: {'id': session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]5225ce23-3eff-b854-b9e2-5e8e68059eea, 'name': SearchDatastore_Task, 'duration_secs': 0.014007} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 970.941264] env[61806]: DEBUG oslo_concurrency.lockutils [None req-2c70fcbb-efa7-459d-81c0-a2a8a184893c tempest-AttachVolumeNegativeTest-1683033704 tempest-AttachVolumeNegativeTest-1683033704-project-member] Releasing lock "[datastore2] devstack-image-cache_base/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa.vmdk" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 970.941652] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [None req-2c70fcbb-efa7-459d-81c0-a2a8a184893c tempest-AttachVolumeNegativeTest-1683033704 tempest-AttachVolumeNegativeTest-1683033704-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa.vmdk to [datastore2] e4638faa-7be1-4909-a595-c437837fa314/e4638faa-7be1-4909-a595-c437837fa314.vmdk {{(pid=61806) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 970.942142] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-2ace8664-b366-4edd-a1c0-18ce3fe23470 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 970.951670] env[61806]: DEBUG oslo_vmware.api [None req-2c70fcbb-efa7-459d-81c0-a2a8a184893c tempest-AttachVolumeNegativeTest-1683033704 tempest-AttachVolumeNegativeTest-1683033704-project-member] Waiting for the task: (returnval){ [ 970.951670] env[61806]: value = "task-1294997" [ 970.951670] env[61806]: _type = "Task" [ 970.951670] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 970.963791] env[61806]: DEBUG oslo_vmware.api [None req-2c70fcbb-efa7-459d-81c0-a2a8a184893c tempest-AttachVolumeNegativeTest-1683033704 tempest-AttachVolumeNegativeTest-1683033704-project-member] Task: {'id': task-1294997, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 971.023347] env[61806]: DEBUG oslo_vmware.api [None req-49ec17b3-08f6-4e32-88fe-0409196f3352 tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] Task: {'id': task-1294987, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 971.049642] env[61806]: DEBUG nova.compute.manager [req-93c00d77-6921-465d-8dc6-30c1613cbc22 req-16118fac-3f55-45b6-a4be-2ef7f83613eb service nova] [instance: e146f7e5-bbf8-4adf-a2e9-b7e753438976] Received event network-changed-b68f5483-a11f-4b35-aaa0-2dcd7f50e448 {{(pid=61806) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 971.050250] env[61806]: DEBUG nova.compute.manager [req-93c00d77-6921-465d-8dc6-30c1613cbc22 req-16118fac-3f55-45b6-a4be-2ef7f83613eb service nova] [instance: e146f7e5-bbf8-4adf-a2e9-b7e753438976] Refreshing instance network info cache due to event network-changed-b68f5483-a11f-4b35-aaa0-2dcd7f50e448. {{(pid=61806) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 971.050250] env[61806]: DEBUG oslo_concurrency.lockutils [req-93c00d77-6921-465d-8dc6-30c1613cbc22 req-16118fac-3f55-45b6-a4be-2ef7f83613eb service nova] Acquiring lock "refresh_cache-e146f7e5-bbf8-4adf-a2e9-b7e753438976" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 971.050404] env[61806]: DEBUG oslo_concurrency.lockutils [req-93c00d77-6921-465d-8dc6-30c1613cbc22 req-16118fac-3f55-45b6-a4be-2ef7f83613eb service nova] Acquired lock "refresh_cache-e146f7e5-bbf8-4adf-a2e9-b7e753438976" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 971.050612] env[61806]: DEBUG nova.network.neutron [req-93c00d77-6921-465d-8dc6-30c1613cbc22 req-16118fac-3f55-45b6-a4be-2ef7f83613eb service nova] [instance: e146f7e5-bbf8-4adf-a2e9-b7e753438976] Refreshing network info cache for port b68f5483-a11f-4b35-aaa0-2dcd7f50e448 {{(pid=61806) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 971.054919] env[61806]: DEBUG oslo_concurrency.lockutils [None req-bf67b140-cd2e-4476-9b1b-cb960c9fa13c tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Acquiring lock "676f18b3-3da1-4a53-b96d-fe64ee8f9101" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 971.055197] env[61806]: DEBUG oslo_concurrency.lockutils [None req-bf67b140-cd2e-4476-9b1b-cb960c9fa13c tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Lock "676f18b3-3da1-4a53-b96d-fe64ee8f9101" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 971.055542] env[61806]: DEBUG oslo_concurrency.lockutils [None req-bf67b140-cd2e-4476-9b1b-cb960c9fa13c tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Acquiring lock "676f18b3-3da1-4a53-b96d-fe64ee8f9101-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 971.055824] env[61806]: DEBUG oslo_concurrency.lockutils [None req-bf67b140-cd2e-4476-9b1b-cb960c9fa13c tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Lock "676f18b3-3da1-4a53-b96d-fe64ee8f9101-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 971.056885] env[61806]: DEBUG oslo_concurrency.lockutils [None req-bf67b140-cd2e-4476-9b1b-cb960c9fa13c tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Lock "676f18b3-3da1-4a53-b96d-fe64ee8f9101-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 971.058414] env[61806]: INFO nova.compute.manager [None req-bf67b140-cd2e-4476-9b1b-cb960c9fa13c tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] [instance: 676f18b3-3da1-4a53-b96d-fe64ee8f9101] Terminating instance [ 971.060794] env[61806]: DEBUG nova.compute.manager [None req-bf67b140-cd2e-4476-9b1b-cb960c9fa13c tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] [instance: 676f18b3-3da1-4a53-b96d-fe64ee8f9101] Start destroying the instance on the hypervisor. {{(pid=61806) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 971.061048] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-bf67b140-cd2e-4476-9b1b-cb960c9fa13c tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] [instance: 676f18b3-3da1-4a53-b96d-fe64ee8f9101] Destroying instance {{(pid=61806) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 971.061989] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bda60ec8-158c-459f-88af-6446d12c5a97 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 971.072063] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [None req-bf67b140-cd2e-4476-9b1b-cb960c9fa13c tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] [instance: 676f18b3-3da1-4a53-b96d-fe64ee8f9101] Powering off the VM {{(pid=61806) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 971.072618] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-25ee92fc-1867-4cd5-ac0d-8d4ceb1145a6 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 971.081483] env[61806]: DEBUG oslo_vmware.api [None req-bf67b140-cd2e-4476-9b1b-cb960c9fa13c tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Waiting for the task: (returnval){ [ 971.081483] env[61806]: value = "task-1294998" [ 971.081483] env[61806]: _type = "Task" [ 971.081483] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 971.092459] env[61806]: DEBUG oslo_vmware.api [None req-bf67b140-cd2e-4476-9b1b-cb960c9fa13c tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Task: {'id': task-1294998, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 971.291184] env[61806]: INFO nova.compute.manager [None req-c8cf9c01-54a9-4c86-b57f-b2e6a2fbc11a tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] [instance: 0bac3048-759a-47d9-a6ef-a0386a49a974] Took 15.84 seconds to build instance. [ 971.350048] env[61806]: DEBUG oslo_vmware.api [-] Task: {'id': task-1294996, 'name': CreateVM_Task, 'duration_secs': 0.428325} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 971.350048] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: e146f7e5-bbf8-4adf-a2e9-b7e753438976] Created VM on the ESX host {{(pid=61806) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 971.350800] env[61806]: DEBUG oslo_concurrency.lockutils [None req-28d96f96-d23f-4c45-9d59-0355de5e0c4f tempest-ServersNegativeTestJSON-891856253 tempest-ServersNegativeTestJSON-891856253-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 971.351056] env[61806]: DEBUG oslo_concurrency.lockutils [None req-28d96f96-d23f-4c45-9d59-0355de5e0c4f tempest-ServersNegativeTestJSON-891856253 tempest-ServersNegativeTestJSON-891856253-project-member] Acquired lock "[datastore1] devstack-image-cache_base/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 971.351489] env[61806]: DEBUG oslo_concurrency.lockutils [None req-28d96f96-d23f-4c45-9d59-0355de5e0c4f tempest-ServersNegativeTestJSON-891856253 tempest-ServersNegativeTestJSON-891856253-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 971.351767] env[61806]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b904b3cf-f6d6-4ee8-9378-243b0f0f7bf3 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 971.358895] env[61806]: DEBUG oslo_vmware.api [None req-28d96f96-d23f-4c45-9d59-0355de5e0c4f tempest-ServersNegativeTestJSON-891856253 tempest-ServersNegativeTestJSON-891856253-project-member] Waiting for the task: (returnval){ [ 971.358895] env[61806]: value = "session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]52ca4eba-705b-a50b-3d58-15babd304657" [ 971.358895] env[61806]: _type = "Task" [ 971.358895] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 971.370722] env[61806]: DEBUG oslo_vmware.api [None req-28d96f96-d23f-4c45-9d59-0355de5e0c4f tempest-ServersNegativeTestJSON-891856253 tempest-ServersNegativeTestJSON-891856253-project-member] Task: {'id': session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]52ca4eba-705b-a50b-3d58-15babd304657, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 971.397593] env[61806]: DEBUG nova.scheduler.client.report [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Inventory has not changed for provider a2858be1-fd22-4e08-979e-87ad25293407 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 138, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61806) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 971.464392] env[61806]: DEBUG oslo_vmware.api [None req-2c70fcbb-efa7-459d-81c0-a2a8a184893c tempest-AttachVolumeNegativeTest-1683033704 tempest-AttachVolumeNegativeTest-1683033704-project-member] Task: {'id': task-1294997, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 971.521417] env[61806]: DEBUG oslo_vmware.api [None req-49ec17b3-08f6-4e32-88fe-0409196f3352 tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] Task: {'id': task-1294987, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 971.586796] env[61806]: DEBUG oslo_concurrency.lockutils [None req-a06ee197-a4db-47dc-a9f0-aee2aa23de70 tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Acquiring lock "9357e61f-1628-43bd-ab46-de13c1529f51" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 971.587059] env[61806]: DEBUG oslo_concurrency.lockutils [None req-a06ee197-a4db-47dc-a9f0-aee2aa23de70 tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Lock "9357e61f-1628-43bd-ab46-de13c1529f51" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 971.597991] env[61806]: DEBUG oslo_vmware.api [None req-bf67b140-cd2e-4476-9b1b-cb960c9fa13c tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Task: {'id': task-1294998, 'name': PowerOffVM_Task, 'duration_secs': 0.2692} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 971.598702] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [None req-bf67b140-cd2e-4476-9b1b-cb960c9fa13c tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] [instance: 676f18b3-3da1-4a53-b96d-fe64ee8f9101] Powered off the VM {{(pid=61806) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 971.598702] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-bf67b140-cd2e-4476-9b1b-cb960c9fa13c tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] [instance: 676f18b3-3da1-4a53-b96d-fe64ee8f9101] Unregistering the VM {{(pid=61806) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 971.598702] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-4a5e0f71-9539-495c-859d-2b1a5578b27b {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 971.641199] env[61806]: DEBUG nova.virt.hardware [None req-1e11ef2b-e661-483a-b1ad-9cb57e29c94f tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-15T18:05:10Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-15T18:04:52Z,direct_url=,disk_format='vmdk',id=ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='84ba83fce2c349a6988173c5d6fc3b07',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-15T18:04:53Z,virtual_size=,visibility=), allow threads: False {{(pid=61806) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 971.641199] env[61806]: DEBUG nova.virt.hardware [None req-1e11ef2b-e661-483a-b1ad-9cb57e29c94f tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Flavor limits 0:0:0 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 971.641199] env[61806]: DEBUG nova.virt.hardware [None req-1e11ef2b-e661-483a-b1ad-9cb57e29c94f tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Image limits 0:0:0 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 971.641199] env[61806]: DEBUG nova.virt.hardware [None req-1e11ef2b-e661-483a-b1ad-9cb57e29c94f tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Flavor pref 0:0:0 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 971.641199] env[61806]: DEBUG nova.virt.hardware [None req-1e11ef2b-e661-483a-b1ad-9cb57e29c94f tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Image pref 0:0:0 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 971.641199] env[61806]: DEBUG nova.virt.hardware [None req-1e11ef2b-e661-483a-b1ad-9cb57e29c94f tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 971.641549] env[61806]: DEBUG nova.virt.hardware [None req-1e11ef2b-e661-483a-b1ad-9cb57e29c94f tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61806) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 971.641549] env[61806]: DEBUG nova.virt.hardware [None req-1e11ef2b-e661-483a-b1ad-9cb57e29c94f tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61806) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 971.641662] env[61806]: DEBUG nova.virt.hardware [None req-1e11ef2b-e661-483a-b1ad-9cb57e29c94f tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Got 1 possible topologies {{(pid=61806) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 971.641754] env[61806]: DEBUG nova.virt.hardware [None req-1e11ef2b-e661-483a-b1ad-9cb57e29c94f tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61806) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 971.641949] env[61806]: DEBUG nova.virt.hardware [None req-1e11ef2b-e661-483a-b1ad-9cb57e29c94f tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61806) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 971.643084] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4a2ab67d-64f2-4f2c-a905-b276d3ab8ab2 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 971.651442] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5529b2a8-7bfd-41d1-abae-79d70ae9eea1 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 971.668179] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-1e11ef2b-e661-483a-b1ad-9cb57e29c94f tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] [instance: 461f98f2-80ba-4088-9e7f-1ba2a9b0c1d9] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:48:c6:32', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '6a31b224-77d7-48a4-af87-312758611b8e', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '91000a2c-19c5-404d-90b3-367698085aec', 'vif_model': 'vmxnet3'}] {{(pid=61806) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 971.676974] env[61806]: DEBUG oslo.service.loopingcall [None req-1e11ef2b-e661-483a-b1ad-9cb57e29c94f tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61806) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 971.678945] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 461f98f2-80ba-4088-9e7f-1ba2a9b0c1d9] Creating VM on the ESX host {{(pid=61806) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 971.679269] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-bf67b140-cd2e-4476-9b1b-cb960c9fa13c tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] [instance: 676f18b3-3da1-4a53-b96d-fe64ee8f9101] Unregistered the VM {{(pid=61806) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 971.679608] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-bf67b140-cd2e-4476-9b1b-cb960c9fa13c tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] [instance: 676f18b3-3da1-4a53-b96d-fe64ee8f9101] Deleting contents of the VM from datastore datastore2 {{(pid=61806) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 971.679786] env[61806]: DEBUG nova.virt.vmwareapi.ds_util [None req-bf67b140-cd2e-4476-9b1b-cb960c9fa13c tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Deleting the datastore file [datastore2] 676f18b3-3da1-4a53-b96d-fe64ee8f9101 {{(pid=61806) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 971.682424] env[61806]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-3b2d4618-3dab-4272-ac66-2eac44311cd0 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 971.696762] env[61806]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-72e430b3-c054-4310-b935-d1dca81c4ce6 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 971.706752] env[61806]: DEBUG oslo_vmware.api [None req-bf67b140-cd2e-4476-9b1b-cb960c9fa13c tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Waiting for the task: (returnval){ [ 971.706752] env[61806]: value = "task-1295000" [ 971.706752] env[61806]: _type = "Task" [ 971.706752] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 971.711026] env[61806]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 971.711026] env[61806]: value = "task-1295001" [ 971.711026] env[61806]: _type = "Task" [ 971.711026] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 971.718030] env[61806]: DEBUG oslo_vmware.api [None req-bf67b140-cd2e-4476-9b1b-cb960c9fa13c tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Task: {'id': task-1295000, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 971.723754] env[61806]: DEBUG oslo_vmware.api [-] Task: {'id': task-1295001, 'name': CreateVM_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 971.794797] env[61806]: DEBUG oslo_concurrency.lockutils [None req-c8cf9c01-54a9-4c86-b57f-b2e6a2fbc11a tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] Lock "0bac3048-759a-47d9-a6ef-a0386a49a974" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 17.351s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 971.870436] env[61806]: DEBUG oslo_vmware.api [None req-28d96f96-d23f-4c45-9d59-0355de5e0c4f tempest-ServersNegativeTestJSON-891856253 tempest-ServersNegativeTestJSON-891856253-project-member] Task: {'id': session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]52ca4eba-705b-a50b-3d58-15babd304657, 'name': SearchDatastore_Task, 'duration_secs': 0.017918} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 971.871352] env[61806]: DEBUG nova.network.neutron [req-93c00d77-6921-465d-8dc6-30c1613cbc22 req-16118fac-3f55-45b6-a4be-2ef7f83613eb service nova] [instance: e146f7e5-bbf8-4adf-a2e9-b7e753438976] Updated VIF entry in instance network info cache for port b68f5483-a11f-4b35-aaa0-2dcd7f50e448. {{(pid=61806) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 971.871731] env[61806]: DEBUG nova.network.neutron [req-93c00d77-6921-465d-8dc6-30c1613cbc22 req-16118fac-3f55-45b6-a4be-2ef7f83613eb service nova] [instance: e146f7e5-bbf8-4adf-a2e9-b7e753438976] Updating instance_info_cache with network_info: [{"id": "b68f5483-a11f-4b35-aaa0-2dcd7f50e448", "address": "fa:16:3e:eb:e5:76", "network": {"id": "fd35ea93-8914-4b54-a4ea-f45fc750982b", "bridge": "br-int", "label": "tempest-ServersNegativeTestJSON-2117068882-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d2e2b7eec47e4b10898215123408692a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "2d94740a-bce8-4103-8ecf-230d02ec0a44", "external-id": "nsx-vlan-transportzone-149", "segmentation_id": 149, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb68f5483-a1", "ovs_interfaceid": "b68f5483-a11f-4b35-aaa0-2dcd7f50e448", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61806) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 971.873033] env[61806]: DEBUG oslo_concurrency.lockutils [None req-28d96f96-d23f-4c45-9d59-0355de5e0c4f tempest-ServersNegativeTestJSON-891856253 tempest-ServersNegativeTestJSON-891856253-project-member] Releasing lock "[datastore1] devstack-image-cache_base/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 971.873308] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-28d96f96-d23f-4c45-9d59-0355de5e0c4f tempest-ServersNegativeTestJSON-891856253 tempest-ServersNegativeTestJSON-891856253-project-member] [instance: e146f7e5-bbf8-4adf-a2e9-b7e753438976] Processing image ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa {{(pid=61806) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 971.873576] env[61806]: DEBUG oslo_concurrency.lockutils [None req-28d96f96-d23f-4c45-9d59-0355de5e0c4f tempest-ServersNegativeTestJSON-891856253 tempest-ServersNegativeTestJSON-891856253-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa.vmdk" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 971.873748] env[61806]: DEBUG oslo_concurrency.lockutils [None req-28d96f96-d23f-4c45-9d59-0355de5e0c4f tempest-ServersNegativeTestJSON-891856253 tempest-ServersNegativeTestJSON-891856253-project-member] Acquired lock "[datastore1] devstack-image-cache_base/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa.vmdk" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 971.873970] env[61806]: DEBUG nova.virt.vmwareapi.ds_util [None req-28d96f96-d23f-4c45-9d59-0355de5e0c4f tempest-ServersNegativeTestJSON-891856253 tempest-ServersNegativeTestJSON-891856253-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61806) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 971.874535] env[61806]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-ccec5eb6-7521-4e21-be58-9eca20b3fb1f {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 971.886465] env[61806]: DEBUG nova.virt.vmwareapi.ds_util [None req-28d96f96-d23f-4c45-9d59-0355de5e0c4f tempest-ServersNegativeTestJSON-891856253 tempest-ServersNegativeTestJSON-891856253-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61806) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 971.886667] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-28d96f96-d23f-4c45-9d59-0355de5e0c4f tempest-ServersNegativeTestJSON-891856253 tempest-ServersNegativeTestJSON-891856253-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61806) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 971.887385] env[61806]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-84caaaec-72a4-4d01-ada3-722c5c6258ff {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 971.894161] env[61806]: DEBUG oslo_vmware.api [None req-28d96f96-d23f-4c45-9d59-0355de5e0c4f tempest-ServersNegativeTestJSON-891856253 tempest-ServersNegativeTestJSON-891856253-project-member] Waiting for the task: (returnval){ [ 971.894161] env[61806]: value = "session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]52cef57a-382b-0375-7b3c-ac257a8bfc1c" [ 971.894161] env[61806]: _type = "Task" [ 971.894161] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 971.902490] env[61806]: DEBUG nova.compute.resource_tracker [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=61806) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 971.902673] env[61806]: DEBUG oslo_concurrency.lockutils [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 2.848s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 971.902937] env[61806]: DEBUG oslo_vmware.api [None req-28d96f96-d23f-4c45-9d59-0355de5e0c4f tempest-ServersNegativeTestJSON-891856253 tempest-ServersNegativeTestJSON-891856253-project-member] Task: {'id': session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]52cef57a-382b-0375-7b3c-ac257a8bfc1c, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 971.903187] env[61806]: DEBUG oslo_concurrency.lockutils [None req-d01361ec-1270-4938-a4ab-1319441d7070 tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 4.712s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 971.904708] env[61806]: INFO nova.compute.claims [None req-d01361ec-1270-4938-a4ab-1319441d7070 tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] [instance: 962d3d88-03ee-4ff2-9d79-15091671b330] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 971.963103] env[61806]: DEBUG oslo_vmware.api [None req-2c70fcbb-efa7-459d-81c0-a2a8a184893c tempest-AttachVolumeNegativeTest-1683033704 tempest-AttachVolumeNegativeTest-1683033704-project-member] Task: {'id': task-1294997, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.599067} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 971.963418] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [None req-2c70fcbb-efa7-459d-81c0-a2a8a184893c tempest-AttachVolumeNegativeTest-1683033704 tempest-AttachVolumeNegativeTest-1683033704-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa.vmdk to [datastore2] e4638faa-7be1-4909-a595-c437837fa314/e4638faa-7be1-4909-a595-c437837fa314.vmdk {{(pid=61806) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 971.963643] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-2c70fcbb-efa7-459d-81c0-a2a8a184893c tempest-AttachVolumeNegativeTest-1683033704 tempest-AttachVolumeNegativeTest-1683033704-project-member] [instance: e4638faa-7be1-4909-a595-c437837fa314] Extending root virtual disk to 1048576 {{(pid=61806) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 971.963901] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-9d542170-a29d-4e5d-951c-d15bcebed4fc {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 971.971199] env[61806]: DEBUG oslo_vmware.api [None req-2c70fcbb-efa7-459d-81c0-a2a8a184893c tempest-AttachVolumeNegativeTest-1683033704 tempest-AttachVolumeNegativeTest-1683033704-project-member] Waiting for the task: (returnval){ [ 971.971199] env[61806]: value = "task-1295002" [ 971.971199] env[61806]: _type = "Task" [ 971.971199] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 971.979321] env[61806]: DEBUG oslo_vmware.api [None req-2c70fcbb-efa7-459d-81c0-a2a8a184893c tempest-AttachVolumeNegativeTest-1683033704 tempest-AttachVolumeNegativeTest-1683033704-project-member] Task: {'id': task-1295002, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 972.021373] env[61806]: DEBUG oslo_vmware.api [None req-49ec17b3-08f6-4e32-88fe-0409196f3352 tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] Task: {'id': task-1294987, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 972.091829] env[61806]: DEBUG nova.compute.manager [None req-a06ee197-a4db-47dc-a9f0-aee2aa23de70 tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] [instance: 9357e61f-1628-43bd-ab46-de13c1529f51] Starting instance... {{(pid=61806) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 972.221655] env[61806]: DEBUG oslo_vmware.api [-] Task: {'id': task-1295001, 'name': CreateVM_Task, 'duration_secs': 0.433224} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 972.224465] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 461f98f2-80ba-4088-9e7f-1ba2a9b0c1d9] Created VM on the ESX host {{(pid=61806) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 972.224764] env[61806]: DEBUG oslo_vmware.api [None req-bf67b140-cd2e-4476-9b1b-cb960c9fa13c tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Task: {'id': task-1295000, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.224211} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 972.225402] env[61806]: DEBUG oslo_concurrency.lockutils [None req-1e11ef2b-e661-483a-b1ad-9cb57e29c94f tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 972.225566] env[61806]: DEBUG oslo_concurrency.lockutils [None req-1e11ef2b-e661-483a-b1ad-9cb57e29c94f tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Acquired lock "[datastore1] devstack-image-cache_base/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 972.225882] env[61806]: DEBUG oslo_concurrency.lockutils [None req-1e11ef2b-e661-483a-b1ad-9cb57e29c94f tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 972.226153] env[61806]: DEBUG nova.virt.vmwareapi.ds_util [None req-bf67b140-cd2e-4476-9b1b-cb960c9fa13c tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Deleted the datastore file {{(pid=61806) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 972.226329] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-bf67b140-cd2e-4476-9b1b-cb960c9fa13c tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] [instance: 676f18b3-3da1-4a53-b96d-fe64ee8f9101] Deleted contents of the VM from datastore datastore2 {{(pid=61806) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 972.226507] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-bf67b140-cd2e-4476-9b1b-cb960c9fa13c tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] [instance: 676f18b3-3da1-4a53-b96d-fe64ee8f9101] Instance destroyed {{(pid=61806) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 972.226737] env[61806]: INFO nova.compute.manager [None req-bf67b140-cd2e-4476-9b1b-cb960c9fa13c tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] [instance: 676f18b3-3da1-4a53-b96d-fe64ee8f9101] Took 1.17 seconds to destroy the instance on the hypervisor. [ 972.226907] env[61806]: DEBUG oslo.service.loopingcall [None req-bf67b140-cd2e-4476-9b1b-cb960c9fa13c tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61806) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 972.227116] env[61806]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d90fe5b0-a57a-48b2-a495-af097d059393 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 972.228609] env[61806]: DEBUG nova.compute.manager [-] [instance: 676f18b3-3da1-4a53-b96d-fe64ee8f9101] Deallocating network for instance {{(pid=61806) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 972.228705] env[61806]: DEBUG nova.network.neutron [-] [instance: 676f18b3-3da1-4a53-b96d-fe64ee8f9101] deallocate_for_instance() {{(pid=61806) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 972.234148] env[61806]: DEBUG oslo_vmware.api [None req-1e11ef2b-e661-483a-b1ad-9cb57e29c94f tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Waiting for the task: (returnval){ [ 972.234148] env[61806]: value = "session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]528482be-a93c-52af-3c0b-773eaceb3caa" [ 972.234148] env[61806]: _type = "Task" [ 972.234148] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 972.242719] env[61806]: DEBUG oslo_vmware.api [None req-1e11ef2b-e661-483a-b1ad-9cb57e29c94f tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Task: {'id': session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]528482be-a93c-52af-3c0b-773eaceb3caa, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 972.248247] env[61806]: DEBUG oslo_concurrency.lockutils [None req-38f9dcbc-e34b-458d-8ba6-e6b3c0e6b471 tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] Acquiring lock "0bac3048-759a-47d9-a6ef-a0386a49a974" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 972.248445] env[61806]: DEBUG oslo_concurrency.lockutils [None req-38f9dcbc-e34b-458d-8ba6-e6b3c0e6b471 tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] Lock "0bac3048-759a-47d9-a6ef-a0386a49a974" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 972.248642] env[61806]: DEBUG oslo_concurrency.lockutils [None req-38f9dcbc-e34b-458d-8ba6-e6b3c0e6b471 tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] Acquiring lock "0bac3048-759a-47d9-a6ef-a0386a49a974-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 972.248824] env[61806]: DEBUG oslo_concurrency.lockutils [None req-38f9dcbc-e34b-458d-8ba6-e6b3c0e6b471 tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] Lock "0bac3048-759a-47d9-a6ef-a0386a49a974-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 972.248997] env[61806]: DEBUG oslo_concurrency.lockutils [None req-38f9dcbc-e34b-458d-8ba6-e6b3c0e6b471 tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] Lock "0bac3048-759a-47d9-a6ef-a0386a49a974-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 972.251068] env[61806]: INFO nova.compute.manager [None req-38f9dcbc-e34b-458d-8ba6-e6b3c0e6b471 tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] [instance: 0bac3048-759a-47d9-a6ef-a0386a49a974] Terminating instance [ 972.252777] env[61806]: DEBUG nova.compute.manager [None req-38f9dcbc-e34b-458d-8ba6-e6b3c0e6b471 tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] [instance: 0bac3048-759a-47d9-a6ef-a0386a49a974] Start destroying the instance on the hypervisor. {{(pid=61806) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 972.253062] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-38f9dcbc-e34b-458d-8ba6-e6b3c0e6b471 tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] [instance: 0bac3048-759a-47d9-a6ef-a0386a49a974] Destroying instance {{(pid=61806) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 972.253855] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e3c29d6d-eaad-4dd7-a4c2-387ef8542559 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 972.263979] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [None req-38f9dcbc-e34b-458d-8ba6-e6b3c0e6b471 tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] [instance: 0bac3048-759a-47d9-a6ef-a0386a49a974] Powering off the VM {{(pid=61806) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 972.264270] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-0ea4d2d2-67e4-4789-9347-88e78f8d7389 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 972.273071] env[61806]: DEBUG oslo_vmware.api [None req-38f9dcbc-e34b-458d-8ba6-e6b3c0e6b471 tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] Waiting for the task: (returnval){ [ 972.273071] env[61806]: value = "task-1295003" [ 972.273071] env[61806]: _type = "Task" [ 972.273071] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 972.281641] env[61806]: DEBUG oslo_vmware.api [None req-38f9dcbc-e34b-458d-8ba6-e6b3c0e6b471 tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] Task: {'id': task-1295003, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 972.375748] env[61806]: DEBUG oslo_concurrency.lockutils [req-93c00d77-6921-465d-8dc6-30c1613cbc22 req-16118fac-3f55-45b6-a4be-2ef7f83613eb service nova] Releasing lock "refresh_cache-e146f7e5-bbf8-4adf-a2e9-b7e753438976" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 972.405664] env[61806]: DEBUG oslo_vmware.api [None req-28d96f96-d23f-4c45-9d59-0355de5e0c4f tempest-ServersNegativeTestJSON-891856253 tempest-ServersNegativeTestJSON-891856253-project-member] Task: {'id': session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]52cef57a-382b-0375-7b3c-ac257a8bfc1c, 'name': SearchDatastore_Task, 'duration_secs': 0.017677} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 972.406596] env[61806]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-60c09f33-4b14-48a0-88ac-2bb4343ad14e {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 972.415027] env[61806]: DEBUG oslo_vmware.api [None req-28d96f96-d23f-4c45-9d59-0355de5e0c4f tempest-ServersNegativeTestJSON-891856253 tempest-ServersNegativeTestJSON-891856253-project-member] Waiting for the task: (returnval){ [ 972.415027] env[61806]: value = "session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]52739599-2df4-f55f-5e95-da9c73049edf" [ 972.415027] env[61806]: _type = "Task" [ 972.415027] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 972.424549] env[61806]: DEBUG oslo_vmware.api [None req-28d96f96-d23f-4c45-9d59-0355de5e0c4f tempest-ServersNegativeTestJSON-891856253 tempest-ServersNegativeTestJSON-891856253-project-member] Task: {'id': session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]52739599-2df4-f55f-5e95-da9c73049edf, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 972.482067] env[61806]: DEBUG oslo_vmware.api [None req-2c70fcbb-efa7-459d-81c0-a2a8a184893c tempest-AttachVolumeNegativeTest-1683033704 tempest-AttachVolumeNegativeTest-1683033704-project-member] Task: {'id': task-1295002, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.107566} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 972.482623] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-2c70fcbb-efa7-459d-81c0-a2a8a184893c tempest-AttachVolumeNegativeTest-1683033704 tempest-AttachVolumeNegativeTest-1683033704-project-member] [instance: e4638faa-7be1-4909-a595-c437837fa314] Extended root virtual disk {{(pid=61806) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 972.483475] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b336fd70-dfce-4e2a-bc76-9c34793ec0b6 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 972.507604] env[61806]: DEBUG nova.virt.vmwareapi.volumeops [None req-2c70fcbb-efa7-459d-81c0-a2a8a184893c tempest-AttachVolumeNegativeTest-1683033704 tempest-AttachVolumeNegativeTest-1683033704-project-member] [instance: e4638faa-7be1-4909-a595-c437837fa314] Reconfiguring VM instance instance-0000005d to attach disk [datastore2] e4638faa-7be1-4909-a595-c437837fa314/e4638faa-7be1-4909-a595-c437837fa314.vmdk or device None with type sparse {{(pid=61806) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 972.507951] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-da89ad80-06b9-479e-b4b3-b6ef53a42e5e {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 972.532910] env[61806]: DEBUG oslo_vmware.api [None req-49ec17b3-08f6-4e32-88fe-0409196f3352 tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] Task: {'id': task-1294987, 'name': ReconfigVM_Task} progress is 18%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 972.534383] env[61806]: DEBUG oslo_vmware.api [None req-2c70fcbb-efa7-459d-81c0-a2a8a184893c tempest-AttachVolumeNegativeTest-1683033704 tempest-AttachVolumeNegativeTest-1683033704-project-member] Waiting for the task: (returnval){ [ 972.534383] env[61806]: value = "task-1295004" [ 972.534383] env[61806]: _type = "Task" [ 972.534383] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 972.543255] env[61806]: DEBUG oslo_vmware.api [None req-2c70fcbb-efa7-459d-81c0-a2a8a184893c tempest-AttachVolumeNegativeTest-1683033704 tempest-AttachVolumeNegativeTest-1683033704-project-member] Task: {'id': task-1295004, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 972.617143] env[61806]: DEBUG oslo_concurrency.lockutils [None req-a06ee197-a4db-47dc-a9f0-aee2aa23de70 tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 972.746108] env[61806]: DEBUG oslo_vmware.api [None req-1e11ef2b-e661-483a-b1ad-9cb57e29c94f tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Task: {'id': session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]528482be-a93c-52af-3c0b-773eaceb3caa, 'name': SearchDatastore_Task, 'duration_secs': 0.01043} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 972.746462] env[61806]: DEBUG oslo_concurrency.lockutils [None req-1e11ef2b-e661-483a-b1ad-9cb57e29c94f tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Releasing lock "[datastore1] devstack-image-cache_base/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 972.746642] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-1e11ef2b-e661-483a-b1ad-9cb57e29c94f tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] [instance: 461f98f2-80ba-4088-9e7f-1ba2a9b0c1d9] Processing image ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa {{(pid=61806) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 972.746860] env[61806]: DEBUG oslo_concurrency.lockutils [None req-1e11ef2b-e661-483a-b1ad-9cb57e29c94f tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa.vmdk" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 972.784169] env[61806]: DEBUG oslo_vmware.api [None req-38f9dcbc-e34b-458d-8ba6-e6b3c0e6b471 tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] Task: {'id': task-1295003, 'name': PowerOffVM_Task, 'duration_secs': 0.274396} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 972.784477] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [None req-38f9dcbc-e34b-458d-8ba6-e6b3c0e6b471 tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] [instance: 0bac3048-759a-47d9-a6ef-a0386a49a974] Powered off the VM {{(pid=61806) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 972.784692] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-38f9dcbc-e34b-458d-8ba6-e6b3c0e6b471 tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] [instance: 0bac3048-759a-47d9-a6ef-a0386a49a974] Unregistering the VM {{(pid=61806) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 972.784954] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-e2750492-e7e3-4b87-913d-cc1674f6a359 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 972.859428] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-38f9dcbc-e34b-458d-8ba6-e6b3c0e6b471 tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] [instance: 0bac3048-759a-47d9-a6ef-a0386a49a974] Unregistered the VM {{(pid=61806) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 972.859710] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-38f9dcbc-e34b-458d-8ba6-e6b3c0e6b471 tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] [instance: 0bac3048-759a-47d9-a6ef-a0386a49a974] Deleting contents of the VM from datastore datastore2 {{(pid=61806) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 972.859917] env[61806]: DEBUG nova.virt.vmwareapi.ds_util [None req-38f9dcbc-e34b-458d-8ba6-e6b3c0e6b471 tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] Deleting the datastore file [datastore2] 0bac3048-759a-47d9-a6ef-a0386a49a974 {{(pid=61806) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 972.860211] env[61806]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-7377def6-5912-4993-be6e-082d362df555 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 972.866823] env[61806]: DEBUG oslo_vmware.api [None req-38f9dcbc-e34b-458d-8ba6-e6b3c0e6b471 tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] Waiting for the task: (returnval){ [ 972.866823] env[61806]: value = "task-1295006" [ 972.866823] env[61806]: _type = "Task" [ 972.866823] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 972.876314] env[61806]: DEBUG oslo_vmware.api [None req-38f9dcbc-e34b-458d-8ba6-e6b3c0e6b471 tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] Task: {'id': task-1295006, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 972.925887] env[61806]: DEBUG oslo_vmware.api [None req-28d96f96-d23f-4c45-9d59-0355de5e0c4f tempest-ServersNegativeTestJSON-891856253 tempest-ServersNegativeTestJSON-891856253-project-member] Task: {'id': session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]52739599-2df4-f55f-5e95-da9c73049edf, 'name': SearchDatastore_Task, 'duration_secs': 0.010927} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 972.926325] env[61806]: DEBUG oslo_concurrency.lockutils [None req-28d96f96-d23f-4c45-9d59-0355de5e0c4f tempest-ServersNegativeTestJSON-891856253 tempest-ServersNegativeTestJSON-891856253-project-member] Releasing lock "[datastore1] devstack-image-cache_base/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa.vmdk" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 972.926662] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [None req-28d96f96-d23f-4c45-9d59-0355de5e0c4f tempest-ServersNegativeTestJSON-891856253 tempest-ServersNegativeTestJSON-891856253-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa.vmdk to [datastore1] e146f7e5-bbf8-4adf-a2e9-b7e753438976/e146f7e5-bbf8-4adf-a2e9-b7e753438976.vmdk {{(pid=61806) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 972.926978] env[61806]: DEBUG oslo_concurrency.lockutils [None req-1e11ef2b-e661-483a-b1ad-9cb57e29c94f tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Acquired lock "[datastore1] devstack-image-cache_base/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa.vmdk" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 972.927190] env[61806]: DEBUG nova.virt.vmwareapi.ds_util [None req-1e11ef2b-e661-483a-b1ad-9cb57e29c94f tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61806) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 972.927425] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-b5a11d76-0fe9-43ae-98ba-bb816348542a {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 972.929497] env[61806]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-ef0a03f6-71d1-45fa-8449-5768098b5ca3 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 972.938632] env[61806]: DEBUG oslo_vmware.api [None req-28d96f96-d23f-4c45-9d59-0355de5e0c4f tempest-ServersNegativeTestJSON-891856253 tempest-ServersNegativeTestJSON-891856253-project-member] Waiting for the task: (returnval){ [ 972.938632] env[61806]: value = "task-1295007" [ 972.938632] env[61806]: _type = "Task" [ 972.938632] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 972.948018] env[61806]: DEBUG oslo_vmware.api [None req-28d96f96-d23f-4c45-9d59-0355de5e0c4f tempest-ServersNegativeTestJSON-891856253 tempest-ServersNegativeTestJSON-891856253-project-member] Task: {'id': task-1295007, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 972.949354] env[61806]: DEBUG nova.virt.vmwareapi.ds_util [None req-1e11ef2b-e661-483a-b1ad-9cb57e29c94f tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61806) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 972.949585] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-1e11ef2b-e661-483a-b1ad-9cb57e29c94f tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61806) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 972.952925] env[61806]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-62ab393b-4035-4816-89b0-ca4c43a326e1 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 972.959070] env[61806]: DEBUG nova.network.neutron [-] [instance: 676f18b3-3da1-4a53-b96d-fe64ee8f9101] Updating instance_info_cache with network_info: [] {{(pid=61806) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 972.960466] env[61806]: DEBUG oslo_vmware.api [None req-1e11ef2b-e661-483a-b1ad-9cb57e29c94f tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Waiting for the task: (returnval){ [ 972.960466] env[61806]: value = "session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]52564958-1cad-8a6a-91ac-5ca7cd579cf9" [ 972.960466] env[61806]: _type = "Task" [ 972.960466] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 972.969985] env[61806]: DEBUG oslo_vmware.api [None req-1e11ef2b-e661-483a-b1ad-9cb57e29c94f tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Task: {'id': session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]52564958-1cad-8a6a-91ac-5ca7cd579cf9, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 973.032059] env[61806]: DEBUG oslo_vmware.api [None req-49ec17b3-08f6-4e32-88fe-0409196f3352 tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] Task: {'id': task-1294987, 'name': ReconfigVM_Task, 'duration_secs': 5.795779} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 973.034690] env[61806]: DEBUG oslo_concurrency.lockutils [None req-49ec17b3-08f6-4e32-88fe-0409196f3352 tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] Releasing lock "d9caee4a-a00c-4e50-b01c-99a8f796a09f" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 973.034912] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-49ec17b3-08f6-4e32-88fe-0409196f3352 tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] [instance: d9caee4a-a00c-4e50-b01c-99a8f796a09f] Reconfigured VM to detach interface {{(pid=61806) detach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1984}} [ 973.049463] env[61806]: DEBUG oslo_vmware.api [None req-2c70fcbb-efa7-459d-81c0-a2a8a184893c tempest-AttachVolumeNegativeTest-1683033704 tempest-AttachVolumeNegativeTest-1683033704-project-member] Task: {'id': task-1295004, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 973.082281] env[61806]: DEBUG nova.compute.manager [req-e098437f-8601-4859-b0f9-0e72eb33de4b req-b6956f4e-6025-45b0-8684-4b8f5967847e service nova] [instance: 676f18b3-3da1-4a53-b96d-fe64ee8f9101] Received event network-vif-deleted-d3a921a7-f1fa-4352-ae53-9f2996576fef {{(pid=61806) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 973.124187] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b8626eac-afac-408c-bd48-2c056040dd8f {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 973.132637] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-235bf3ed-e486-403d-aa8f-903a22eea0ad {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 973.167222] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e2977d00-1079-465e-a76b-9e7a378191fe {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 973.175768] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-505dd0c0-38e6-4ef2-914e-5d388aebde8a {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 973.191464] env[61806]: DEBUG nova.compute.provider_tree [None req-d01361ec-1270-4938-a4ab-1319441d7070 tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Inventory has not changed in ProviderTree for provider: a2858be1-fd22-4e08-979e-87ad25293407 {{(pid=61806) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 973.380778] env[61806]: DEBUG oslo_vmware.api [None req-38f9dcbc-e34b-458d-8ba6-e6b3c0e6b471 tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] Task: {'id': task-1295006, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.186766} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 973.380778] env[61806]: DEBUG nova.virt.vmwareapi.ds_util [None req-38f9dcbc-e34b-458d-8ba6-e6b3c0e6b471 tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] Deleted the datastore file {{(pid=61806) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 973.380778] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-38f9dcbc-e34b-458d-8ba6-e6b3c0e6b471 tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] [instance: 0bac3048-759a-47d9-a6ef-a0386a49a974] Deleted contents of the VM from datastore datastore2 {{(pid=61806) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 973.381138] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-38f9dcbc-e34b-458d-8ba6-e6b3c0e6b471 tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] [instance: 0bac3048-759a-47d9-a6ef-a0386a49a974] Instance destroyed {{(pid=61806) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 973.381138] env[61806]: INFO nova.compute.manager [None req-38f9dcbc-e34b-458d-8ba6-e6b3c0e6b471 tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] [instance: 0bac3048-759a-47d9-a6ef-a0386a49a974] Took 1.13 seconds to destroy the instance on the hypervisor. [ 973.381400] env[61806]: DEBUG oslo.service.loopingcall [None req-38f9dcbc-e34b-458d-8ba6-e6b3c0e6b471 tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61806) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 973.381646] env[61806]: DEBUG nova.compute.manager [-] [instance: 0bac3048-759a-47d9-a6ef-a0386a49a974] Deallocating network for instance {{(pid=61806) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 973.381743] env[61806]: DEBUG nova.network.neutron [-] [instance: 0bac3048-759a-47d9-a6ef-a0386a49a974] deallocate_for_instance() {{(pid=61806) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 973.450740] env[61806]: DEBUG oslo_vmware.api [None req-28d96f96-d23f-4c45-9d59-0355de5e0c4f tempest-ServersNegativeTestJSON-891856253 tempest-ServersNegativeTestJSON-891856253-project-member] Task: {'id': task-1295007, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 973.461708] env[61806]: INFO nova.compute.manager [-] [instance: 676f18b3-3da1-4a53-b96d-fe64ee8f9101] Took 1.23 seconds to deallocate network for instance. [ 973.478742] env[61806]: DEBUG oslo_vmware.api [None req-1e11ef2b-e661-483a-b1ad-9cb57e29c94f tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Task: {'id': session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]52564958-1cad-8a6a-91ac-5ca7cd579cf9, 'name': SearchDatastore_Task, 'duration_secs': 0.011651} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 973.479634] env[61806]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f5d315aa-682c-461e-af06-d58c7da2f30f {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 973.486437] env[61806]: DEBUG oslo_vmware.api [None req-1e11ef2b-e661-483a-b1ad-9cb57e29c94f tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Waiting for the task: (returnval){ [ 973.486437] env[61806]: value = "session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]52affcbc-198b-5509-c52d-cccdcb59eca4" [ 973.486437] env[61806]: _type = "Task" [ 973.486437] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 973.496179] env[61806]: DEBUG oslo_vmware.api [None req-1e11ef2b-e661-483a-b1ad-9cb57e29c94f tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Task: {'id': session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]52affcbc-198b-5509-c52d-cccdcb59eca4, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 973.548467] env[61806]: DEBUG oslo_vmware.api [None req-2c70fcbb-efa7-459d-81c0-a2a8a184893c tempest-AttachVolumeNegativeTest-1683033704 tempest-AttachVolumeNegativeTest-1683033704-project-member] Task: {'id': task-1295004, 'name': ReconfigVM_Task, 'duration_secs': 0.713835} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 973.548782] env[61806]: DEBUG nova.virt.vmwareapi.volumeops [None req-2c70fcbb-efa7-459d-81c0-a2a8a184893c tempest-AttachVolumeNegativeTest-1683033704 tempest-AttachVolumeNegativeTest-1683033704-project-member] [instance: e4638faa-7be1-4909-a595-c437837fa314] Reconfigured VM instance instance-0000005d to attach disk [datastore2] e4638faa-7be1-4909-a595-c437837fa314/e4638faa-7be1-4909-a595-c437837fa314.vmdk or device None with type sparse {{(pid=61806) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 973.549441] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-3ce25de0-1e4e-44a7-83e5-c733fd47c3fc {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 973.557600] env[61806]: DEBUG oslo_vmware.api [None req-2c70fcbb-efa7-459d-81c0-a2a8a184893c tempest-AttachVolumeNegativeTest-1683033704 tempest-AttachVolumeNegativeTest-1683033704-project-member] Waiting for the task: (returnval){ [ 973.557600] env[61806]: value = "task-1295008" [ 973.557600] env[61806]: _type = "Task" [ 973.557600] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 973.566876] env[61806]: DEBUG oslo_vmware.api [None req-2c70fcbb-efa7-459d-81c0-a2a8a184893c tempest-AttachVolumeNegativeTest-1683033704 tempest-AttachVolumeNegativeTest-1683033704-project-member] Task: {'id': task-1295008, 'name': Rename_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 973.649960] env[61806]: DEBUG nova.compute.manager [req-0925f20d-ad37-4584-b3dc-cc359529c06f req-b8ddc78e-54b1-45df-bce9-dcbe8602c549 service nova] [instance: 0bac3048-759a-47d9-a6ef-a0386a49a974] Received event network-vif-deleted-69c51913-810b-4b60-8249-2960056a8035 {{(pid=61806) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 973.650284] env[61806]: INFO nova.compute.manager [req-0925f20d-ad37-4584-b3dc-cc359529c06f req-b8ddc78e-54b1-45df-bce9-dcbe8602c549 service nova] [instance: 0bac3048-759a-47d9-a6ef-a0386a49a974] Neutron deleted interface 69c51913-810b-4b60-8249-2960056a8035; detaching it from the instance and deleting it from the info cache [ 973.650620] env[61806]: DEBUG nova.network.neutron [req-0925f20d-ad37-4584-b3dc-cc359529c06f req-b8ddc78e-54b1-45df-bce9-dcbe8602c549 service nova] [instance: 0bac3048-759a-47d9-a6ef-a0386a49a974] Updating instance_info_cache with network_info: [] {{(pid=61806) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 973.695106] env[61806]: DEBUG nova.scheduler.client.report [None req-d01361ec-1270-4938-a4ab-1319441d7070 tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Inventory has not changed for provider a2858be1-fd22-4e08-979e-87ad25293407 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 138, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61806) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 973.953044] env[61806]: DEBUG oslo_vmware.api [None req-28d96f96-d23f-4c45-9d59-0355de5e0c4f tempest-ServersNegativeTestJSON-891856253 tempest-ServersNegativeTestJSON-891856253-project-member] Task: {'id': task-1295007, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.638172} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 973.953044] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [None req-28d96f96-d23f-4c45-9d59-0355de5e0c4f tempest-ServersNegativeTestJSON-891856253 tempest-ServersNegativeTestJSON-891856253-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa.vmdk to [datastore1] e146f7e5-bbf8-4adf-a2e9-b7e753438976/e146f7e5-bbf8-4adf-a2e9-b7e753438976.vmdk {{(pid=61806) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 973.953044] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-28d96f96-d23f-4c45-9d59-0355de5e0c4f tempest-ServersNegativeTestJSON-891856253 tempest-ServersNegativeTestJSON-891856253-project-member] [instance: e146f7e5-bbf8-4adf-a2e9-b7e753438976] Extending root virtual disk to 1048576 {{(pid=61806) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 973.953536] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-f71846e9-f076-41db-ba5c-e4330c03d392 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 973.960767] env[61806]: DEBUG oslo_vmware.api [None req-28d96f96-d23f-4c45-9d59-0355de5e0c4f tempest-ServersNegativeTestJSON-891856253 tempest-ServersNegativeTestJSON-891856253-project-member] Waiting for the task: (returnval){ [ 973.960767] env[61806]: value = "task-1295009" [ 973.960767] env[61806]: _type = "Task" [ 973.960767] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 973.969554] env[61806]: DEBUG oslo_vmware.api [None req-28d96f96-d23f-4c45-9d59-0355de5e0c4f tempest-ServersNegativeTestJSON-891856253 tempest-ServersNegativeTestJSON-891856253-project-member] Task: {'id': task-1295009, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 973.973716] env[61806]: DEBUG oslo_concurrency.lockutils [None req-bf67b140-cd2e-4476-9b1b-cb960c9fa13c tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 973.997190] env[61806]: DEBUG oslo_vmware.api [None req-1e11ef2b-e661-483a-b1ad-9cb57e29c94f tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Task: {'id': session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]52affcbc-198b-5509-c52d-cccdcb59eca4, 'name': SearchDatastore_Task, 'duration_secs': 0.033121} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 973.997499] env[61806]: DEBUG oslo_concurrency.lockutils [None req-1e11ef2b-e661-483a-b1ad-9cb57e29c94f tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Releasing lock "[datastore1] devstack-image-cache_base/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa.vmdk" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 973.997825] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [None req-1e11ef2b-e661-483a-b1ad-9cb57e29c94f tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa.vmdk to [datastore1] 461f98f2-80ba-4088-9e7f-1ba2a9b0c1d9/461f98f2-80ba-4088-9e7f-1ba2a9b0c1d9.vmdk {{(pid=61806) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 973.998183] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-9866b069-9d11-427a-8ae3-0a888105096d {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 974.005692] env[61806]: DEBUG oslo_vmware.api [None req-1e11ef2b-e661-483a-b1ad-9cb57e29c94f tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Waiting for the task: (returnval){ [ 974.005692] env[61806]: value = "task-1295010" [ 974.005692] env[61806]: _type = "Task" [ 974.005692] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 974.014348] env[61806]: DEBUG oslo_vmware.api [None req-1e11ef2b-e661-483a-b1ad-9cb57e29c94f tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Task: {'id': task-1295010, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 974.069163] env[61806]: DEBUG oslo_vmware.api [None req-2c70fcbb-efa7-459d-81c0-a2a8a184893c tempest-AttachVolumeNegativeTest-1683033704 tempest-AttachVolumeNegativeTest-1683033704-project-member] Task: {'id': task-1295008, 'name': Rename_Task, 'duration_secs': 0.217367} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 974.069629] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [None req-2c70fcbb-efa7-459d-81c0-a2a8a184893c tempest-AttachVolumeNegativeTest-1683033704 tempest-AttachVolumeNegativeTest-1683033704-project-member] [instance: e4638faa-7be1-4909-a595-c437837fa314] Powering on the VM {{(pid=61806) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 974.070034] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-cf41487a-c648-43ff-be44-83cf65eed636 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 974.078911] env[61806]: DEBUG oslo_vmware.api [None req-2c70fcbb-efa7-459d-81c0-a2a8a184893c tempest-AttachVolumeNegativeTest-1683033704 tempest-AttachVolumeNegativeTest-1683033704-project-member] Waiting for the task: (returnval){ [ 974.078911] env[61806]: value = "task-1295011" [ 974.078911] env[61806]: _type = "Task" [ 974.078911] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 974.090184] env[61806]: DEBUG oslo_vmware.api [None req-2c70fcbb-efa7-459d-81c0-a2a8a184893c tempest-AttachVolumeNegativeTest-1683033704 tempest-AttachVolumeNegativeTest-1683033704-project-member] Task: {'id': task-1295011, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 974.124743] env[61806]: DEBUG nova.network.neutron [-] [instance: 0bac3048-759a-47d9-a6ef-a0386a49a974] Updating instance_info_cache with network_info: [] {{(pid=61806) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 974.154528] env[61806]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-0dc55a33-f45b-4a81-af94-07925b73dc63 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 974.165130] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1dac7119-6227-40f5-8aef-2a6ebbbf5ed4 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 974.198413] env[61806]: DEBUG nova.compute.manager [req-0925f20d-ad37-4584-b3dc-cc359529c06f req-b8ddc78e-54b1-45df-bce9-dcbe8602c549 service nova] [instance: 0bac3048-759a-47d9-a6ef-a0386a49a974] Detach interface failed, port_id=69c51913-810b-4b60-8249-2960056a8035, reason: Instance 0bac3048-759a-47d9-a6ef-a0386a49a974 could not be found. {{(pid=61806) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10963}} [ 974.201531] env[61806]: DEBUG oslo_concurrency.lockutils [None req-d01361ec-1270-4938-a4ab-1319441d7070 tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.298s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 974.202127] env[61806]: DEBUG nova.compute.manager [None req-d01361ec-1270-4938-a4ab-1319441d7070 tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] [instance: 962d3d88-03ee-4ff2-9d79-15091671b330] Start building networks asynchronously for instance. {{(pid=61806) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 974.204816] env[61806]: DEBUG oslo_concurrency.lockutils [None req-a06ee197-a4db-47dc-a9f0-aee2aa23de70 tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 1.588s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 974.206252] env[61806]: INFO nova.compute.claims [None req-a06ee197-a4db-47dc-a9f0-aee2aa23de70 tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] [instance: 9357e61f-1628-43bd-ab46-de13c1529f51] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 974.407411] env[61806]: DEBUG oslo_concurrency.lockutils [None req-49ec17b3-08f6-4e32-88fe-0409196f3352 tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] Acquiring lock "refresh_cache-d9caee4a-a00c-4e50-b01c-99a8f796a09f" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 974.407665] env[61806]: DEBUG oslo_concurrency.lockutils [None req-49ec17b3-08f6-4e32-88fe-0409196f3352 tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] Acquired lock "refresh_cache-d9caee4a-a00c-4e50-b01c-99a8f796a09f" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 974.407933] env[61806]: DEBUG nova.network.neutron [None req-49ec17b3-08f6-4e32-88fe-0409196f3352 tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] [instance: d9caee4a-a00c-4e50-b01c-99a8f796a09f] Building network info cache for instance {{(pid=61806) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 974.474869] env[61806]: DEBUG oslo_vmware.api [None req-28d96f96-d23f-4c45-9d59-0355de5e0c4f tempest-ServersNegativeTestJSON-891856253 tempest-ServersNegativeTestJSON-891856253-project-member] Task: {'id': task-1295009, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.074365} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 974.475270] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-28d96f96-d23f-4c45-9d59-0355de5e0c4f tempest-ServersNegativeTestJSON-891856253 tempest-ServersNegativeTestJSON-891856253-project-member] [instance: e146f7e5-bbf8-4adf-a2e9-b7e753438976] Extended root virtual disk {{(pid=61806) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 974.476033] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ea87e6d2-b058-48ba-ab54-57249f0f8925 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 974.501039] env[61806]: DEBUG nova.virt.vmwareapi.volumeops [None req-28d96f96-d23f-4c45-9d59-0355de5e0c4f tempest-ServersNegativeTestJSON-891856253 tempest-ServersNegativeTestJSON-891856253-project-member] [instance: e146f7e5-bbf8-4adf-a2e9-b7e753438976] Reconfiguring VM instance instance-0000005e to attach disk [datastore1] e146f7e5-bbf8-4adf-a2e9-b7e753438976/e146f7e5-bbf8-4adf-a2e9-b7e753438976.vmdk or device None with type sparse {{(pid=61806) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 974.501544] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-400c26c1-eeb1-4044-8b78-e0e3312dcd18 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 974.525399] env[61806]: DEBUG oslo_vmware.api [None req-1e11ef2b-e661-483a-b1ad-9cb57e29c94f tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Task: {'id': task-1295010, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 974.527056] env[61806]: DEBUG oslo_vmware.api [None req-28d96f96-d23f-4c45-9d59-0355de5e0c4f tempest-ServersNegativeTestJSON-891856253 tempest-ServersNegativeTestJSON-891856253-project-member] Waiting for the task: (returnval){ [ 974.527056] env[61806]: value = "task-1295012" [ 974.527056] env[61806]: _type = "Task" [ 974.527056] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 974.535869] env[61806]: DEBUG oslo_vmware.api [None req-28d96f96-d23f-4c45-9d59-0355de5e0c4f tempest-ServersNegativeTestJSON-891856253 tempest-ServersNegativeTestJSON-891856253-project-member] Task: {'id': task-1295012, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 974.591055] env[61806]: DEBUG oslo_vmware.api [None req-2c70fcbb-efa7-459d-81c0-a2a8a184893c tempest-AttachVolumeNegativeTest-1683033704 tempest-AttachVolumeNegativeTest-1683033704-project-member] Task: {'id': task-1295011, 'name': PowerOnVM_Task} progress is 100%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 974.627697] env[61806]: INFO nova.compute.manager [-] [instance: 0bac3048-759a-47d9-a6ef-a0386a49a974] Took 1.25 seconds to deallocate network for instance. [ 974.710616] env[61806]: DEBUG nova.compute.utils [None req-d01361ec-1270-4938-a4ab-1319441d7070 tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Using /dev/sd instead of None {{(pid=61806) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 974.714170] env[61806]: DEBUG nova.compute.manager [None req-d01361ec-1270-4938-a4ab-1319441d7070 tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] [instance: 962d3d88-03ee-4ff2-9d79-15091671b330] Allocating IP information in the background. {{(pid=61806) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 974.714407] env[61806]: DEBUG nova.network.neutron [None req-d01361ec-1270-4938-a4ab-1319441d7070 tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] [instance: 962d3d88-03ee-4ff2-9d79-15091671b330] allocate_for_instance() {{(pid=61806) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 974.770143] env[61806]: DEBUG nova.policy [None req-d01361ec-1270-4938-a4ab-1319441d7070 tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '2b86d596e89645c2ab31332afdcbc3c2', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '7ab95fb72b5d46c3b6c7bebbccf897cd', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61806) authorize /opt/stack/nova/nova/policy.py:201}} [ 974.851646] env[61806]: DEBUG oslo_concurrency.lockutils [None req-57da61d4-139b-46a7-9163-13f9c5735119 tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] Acquiring lock "d9caee4a-a00c-4e50-b01c-99a8f796a09f" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 974.851915] env[61806]: DEBUG oslo_concurrency.lockutils [None req-57da61d4-139b-46a7-9163-13f9c5735119 tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] Lock "d9caee4a-a00c-4e50-b01c-99a8f796a09f" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 974.852206] env[61806]: DEBUG oslo_concurrency.lockutils [None req-57da61d4-139b-46a7-9163-13f9c5735119 tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] Acquiring lock "d9caee4a-a00c-4e50-b01c-99a8f796a09f-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 974.852455] env[61806]: DEBUG oslo_concurrency.lockutils [None req-57da61d4-139b-46a7-9163-13f9c5735119 tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] Lock "d9caee4a-a00c-4e50-b01c-99a8f796a09f-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 974.852644] env[61806]: DEBUG oslo_concurrency.lockutils [None req-57da61d4-139b-46a7-9163-13f9c5735119 tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] Lock "d9caee4a-a00c-4e50-b01c-99a8f796a09f-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 974.855193] env[61806]: INFO nova.compute.manager [None req-57da61d4-139b-46a7-9163-13f9c5735119 tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] [instance: d9caee4a-a00c-4e50-b01c-99a8f796a09f] Terminating instance [ 974.857915] env[61806]: DEBUG nova.compute.manager [None req-57da61d4-139b-46a7-9163-13f9c5735119 tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] [instance: d9caee4a-a00c-4e50-b01c-99a8f796a09f] Start destroying the instance on the hypervisor. {{(pid=61806) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 974.857969] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-57da61d4-139b-46a7-9163-13f9c5735119 tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] [instance: d9caee4a-a00c-4e50-b01c-99a8f796a09f] Destroying instance {{(pid=61806) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 974.859640] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-40b7e677-3a01-4bcf-b56e-b2f9e800e432 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 974.872869] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [None req-57da61d4-139b-46a7-9163-13f9c5735119 tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] [instance: d9caee4a-a00c-4e50-b01c-99a8f796a09f] Powering off the VM {{(pid=61806) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 974.874613] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-0a5b74b0-205b-42e5-898b-e883a7f405b6 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 974.883545] env[61806]: DEBUG oslo_vmware.api [None req-57da61d4-139b-46a7-9163-13f9c5735119 tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] Waiting for the task: (returnval){ [ 974.883545] env[61806]: value = "task-1295013" [ 974.883545] env[61806]: _type = "Task" [ 974.883545] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 974.894062] env[61806]: DEBUG oslo_vmware.api [None req-57da61d4-139b-46a7-9163-13f9c5735119 tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] Task: {'id': task-1295013, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 975.030913] env[61806]: DEBUG oslo_vmware.api [None req-1e11ef2b-e661-483a-b1ad-9cb57e29c94f tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Task: {'id': task-1295010, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 975.041379] env[61806]: DEBUG oslo_vmware.api [None req-28d96f96-d23f-4c45-9d59-0355de5e0c4f tempest-ServersNegativeTestJSON-891856253 tempest-ServersNegativeTestJSON-891856253-project-member] Task: {'id': task-1295012, 'name': ReconfigVM_Task, 'duration_secs': 0.349592} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 975.041379] env[61806]: DEBUG nova.virt.vmwareapi.volumeops [None req-28d96f96-d23f-4c45-9d59-0355de5e0c4f tempest-ServersNegativeTestJSON-891856253 tempest-ServersNegativeTestJSON-891856253-project-member] [instance: e146f7e5-bbf8-4adf-a2e9-b7e753438976] Reconfigured VM instance instance-0000005e to attach disk [datastore1] e146f7e5-bbf8-4adf-a2e9-b7e753438976/e146f7e5-bbf8-4adf-a2e9-b7e753438976.vmdk or device None with type sparse {{(pid=61806) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 975.041379] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-f1c9cb40-b1e8-4c30-9d45-4cc297ca54a0 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 975.051948] env[61806]: DEBUG oslo_vmware.api [None req-28d96f96-d23f-4c45-9d59-0355de5e0c4f tempest-ServersNegativeTestJSON-891856253 tempest-ServersNegativeTestJSON-891856253-project-member] Waiting for the task: (returnval){ [ 975.051948] env[61806]: value = "task-1295014" [ 975.051948] env[61806]: _type = "Task" [ 975.051948] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 975.063223] env[61806]: DEBUG oslo_vmware.api [None req-28d96f96-d23f-4c45-9d59-0355de5e0c4f tempest-ServersNegativeTestJSON-891856253 tempest-ServersNegativeTestJSON-891856253-project-member] Task: {'id': task-1295014, 'name': Rename_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 975.066166] env[61806]: DEBUG nova.network.neutron [None req-d01361ec-1270-4938-a4ab-1319441d7070 tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] [instance: 962d3d88-03ee-4ff2-9d79-15091671b330] Successfully created port: 1a817be6-517c-4be9-8c14-c6f3995f87e0 {{(pid=61806) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 975.094604] env[61806]: DEBUG oslo_vmware.api [None req-2c70fcbb-efa7-459d-81c0-a2a8a184893c tempest-AttachVolumeNegativeTest-1683033704 tempest-AttachVolumeNegativeTest-1683033704-project-member] Task: {'id': task-1295011, 'name': PowerOnVM_Task, 'duration_secs': 0.523776} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 975.094604] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [None req-2c70fcbb-efa7-459d-81c0-a2a8a184893c tempest-AttachVolumeNegativeTest-1683033704 tempest-AttachVolumeNegativeTest-1683033704-project-member] [instance: e4638faa-7be1-4909-a595-c437837fa314] Powered on the VM {{(pid=61806) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 975.094604] env[61806]: INFO nova.compute.manager [None req-2c70fcbb-efa7-459d-81c0-a2a8a184893c tempest-AttachVolumeNegativeTest-1683033704 tempest-AttachVolumeNegativeTest-1683033704-project-member] [instance: e4638faa-7be1-4909-a595-c437837fa314] Took 8.66 seconds to spawn the instance on the hypervisor. [ 975.094604] env[61806]: DEBUG nova.compute.manager [None req-2c70fcbb-efa7-459d-81c0-a2a8a184893c tempest-AttachVolumeNegativeTest-1683033704 tempest-AttachVolumeNegativeTest-1683033704-project-member] [instance: e4638faa-7be1-4909-a595-c437837fa314] Checking state {{(pid=61806) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 975.094604] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0646144b-9d5e-4775-8fb3-9f34350f5bee {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 975.134826] env[61806]: DEBUG oslo_concurrency.lockutils [None req-38f9dcbc-e34b-458d-8ba6-e6b3c0e6b471 tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 975.197124] env[61806]: INFO nova.network.neutron [None req-49ec17b3-08f6-4e32-88fe-0409196f3352 tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] [instance: d9caee4a-a00c-4e50-b01c-99a8f796a09f] Port 7e031e88-6928-472d-a65e-b818d89e6fc5 from network info_cache is no longer associated with instance in Neutron. Removing from network info_cache. [ 975.197511] env[61806]: DEBUG nova.network.neutron [None req-49ec17b3-08f6-4e32-88fe-0409196f3352 tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] [instance: d9caee4a-a00c-4e50-b01c-99a8f796a09f] Updating instance_info_cache with network_info: [{"id": "c725ee28-d7a8-4b31-92bc-f63174564f5e", "address": "fa:16:3e:3a:18:8e", "network": {"id": "a7ffda72-e57c-41fa-a7ed-d7b12084bc9d", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-62794865-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.218", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "6e391e1267014fafbb8f5a3211299819", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8b3cfeb1-f262-4fd9-b506-8e9c0733e2d8", "external-id": "nsx-vlan-transportzone-119", "segmentation_id": 119, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc725ee28-d7", "ovs_interfaceid": "c725ee28-d7a8-4b31-92bc-f63174564f5e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61806) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 975.214927] env[61806]: DEBUG nova.compute.manager [None req-d01361ec-1270-4938-a4ab-1319441d7070 tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] [instance: 962d3d88-03ee-4ff2-9d79-15091671b330] Start building block device mappings for instance. {{(pid=61806) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 975.396827] env[61806]: DEBUG oslo_vmware.api [None req-57da61d4-139b-46a7-9163-13f9c5735119 tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] Task: {'id': task-1295013, 'name': PowerOffVM_Task, 'duration_secs': 0.207214} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 975.397441] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [None req-57da61d4-139b-46a7-9163-13f9c5735119 tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] [instance: d9caee4a-a00c-4e50-b01c-99a8f796a09f] Powered off the VM {{(pid=61806) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 975.397441] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-57da61d4-139b-46a7-9163-13f9c5735119 tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] [instance: d9caee4a-a00c-4e50-b01c-99a8f796a09f] Unregistering the VM {{(pid=61806) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 975.397790] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-cfe7c6d3-d5dc-4a5b-866a-35656c741ced {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 975.411718] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cc9f417d-71f6-4b1d-94bd-10a29d83bf86 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 975.419973] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-db68f913-b28c-4fc9-8137-56d24a604842 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 975.453033] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-792db3e7-a8fe-48d5-a29a-b6914a89e7ab {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 975.461797] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f31afbab-4297-4987-bde1-6fc9a34eebc2 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 975.476364] env[61806]: DEBUG nova.compute.provider_tree [None req-a06ee197-a4db-47dc-a9f0-aee2aa23de70 tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Inventory has not changed in ProviderTree for provider: a2858be1-fd22-4e08-979e-87ad25293407 {{(pid=61806) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 975.528794] env[61806]: DEBUG oslo_vmware.api [None req-1e11ef2b-e661-483a-b1ad-9cb57e29c94f tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Task: {'id': task-1295010, 'name': CopyVirtualDisk_Task, 'duration_secs': 1.049976} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 975.529089] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [None req-1e11ef2b-e661-483a-b1ad-9cb57e29c94f tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa.vmdk to [datastore1] 461f98f2-80ba-4088-9e7f-1ba2a9b0c1d9/461f98f2-80ba-4088-9e7f-1ba2a9b0c1d9.vmdk {{(pid=61806) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 975.529620] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-1e11ef2b-e661-483a-b1ad-9cb57e29c94f tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] [instance: 461f98f2-80ba-4088-9e7f-1ba2a9b0c1d9] Extending root virtual disk to 1048576 {{(pid=61806) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 975.529620] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-c8e93f7b-189d-4d49-8cb7-36e443823578 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 975.536847] env[61806]: DEBUG oslo_vmware.api [None req-1e11ef2b-e661-483a-b1ad-9cb57e29c94f tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Waiting for the task: (returnval){ [ 975.536847] env[61806]: value = "task-1295016" [ 975.536847] env[61806]: _type = "Task" [ 975.536847] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 975.546144] env[61806]: DEBUG oslo_vmware.api [None req-1e11ef2b-e661-483a-b1ad-9cb57e29c94f tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Task: {'id': task-1295016, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 975.562810] env[61806]: DEBUG oslo_vmware.api [None req-28d96f96-d23f-4c45-9d59-0355de5e0c4f tempest-ServersNegativeTestJSON-891856253 tempest-ServersNegativeTestJSON-891856253-project-member] Task: {'id': task-1295014, 'name': Rename_Task, 'duration_secs': 0.150611} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 975.563134] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [None req-28d96f96-d23f-4c45-9d59-0355de5e0c4f tempest-ServersNegativeTestJSON-891856253 tempest-ServersNegativeTestJSON-891856253-project-member] [instance: e146f7e5-bbf8-4adf-a2e9-b7e753438976] Powering on the VM {{(pid=61806) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 975.563416] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-d14dbb72-38a2-41d7-8c15-38e0de67b3b0 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 975.571260] env[61806]: DEBUG oslo_vmware.api [None req-28d96f96-d23f-4c45-9d59-0355de5e0c4f tempest-ServersNegativeTestJSON-891856253 tempest-ServersNegativeTestJSON-891856253-project-member] Waiting for the task: (returnval){ [ 975.571260] env[61806]: value = "task-1295017" [ 975.571260] env[61806]: _type = "Task" [ 975.571260] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 975.581742] env[61806]: DEBUG oslo_vmware.api [None req-28d96f96-d23f-4c45-9d59-0355de5e0c4f tempest-ServersNegativeTestJSON-891856253 tempest-ServersNegativeTestJSON-891856253-project-member] Task: {'id': task-1295017, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 975.601421] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-57da61d4-139b-46a7-9163-13f9c5735119 tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] [instance: d9caee4a-a00c-4e50-b01c-99a8f796a09f] Unregistered the VM {{(pid=61806) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 975.601750] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-57da61d4-139b-46a7-9163-13f9c5735119 tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] [instance: d9caee4a-a00c-4e50-b01c-99a8f796a09f] Deleting contents of the VM from datastore datastore1 {{(pid=61806) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 975.601964] env[61806]: DEBUG nova.virt.vmwareapi.ds_util [None req-57da61d4-139b-46a7-9163-13f9c5735119 tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] Deleting the datastore file [datastore1] d9caee4a-a00c-4e50-b01c-99a8f796a09f {{(pid=61806) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 975.602574] env[61806]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-ae3c70ab-d9ca-44db-b4cb-1c5854248240 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 975.614321] env[61806]: DEBUG oslo_vmware.api [None req-57da61d4-139b-46a7-9163-13f9c5735119 tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] Waiting for the task: (returnval){ [ 975.614321] env[61806]: value = "task-1295018" [ 975.614321] env[61806]: _type = "Task" [ 975.614321] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 975.614504] env[61806]: INFO nova.compute.manager [None req-2c70fcbb-efa7-459d-81c0-a2a8a184893c tempest-AttachVolumeNegativeTest-1683033704 tempest-AttachVolumeNegativeTest-1683033704-project-member] [instance: e4638faa-7be1-4909-a595-c437837fa314] Took 16.32 seconds to build instance. [ 975.624630] env[61806]: DEBUG oslo_vmware.api [None req-57da61d4-139b-46a7-9163-13f9c5735119 tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] Task: {'id': task-1295018, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 975.700632] env[61806]: DEBUG oslo_concurrency.lockutils [None req-49ec17b3-08f6-4e32-88fe-0409196f3352 tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] Releasing lock "refresh_cache-d9caee4a-a00c-4e50-b01c-99a8f796a09f" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 975.980053] env[61806]: DEBUG nova.scheduler.client.report [None req-a06ee197-a4db-47dc-a9f0-aee2aa23de70 tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Inventory has not changed for provider a2858be1-fd22-4e08-979e-87ad25293407 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 138, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61806) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 976.046821] env[61806]: DEBUG oslo_vmware.api [None req-1e11ef2b-e661-483a-b1ad-9cb57e29c94f tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Task: {'id': task-1295016, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.076989} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 976.047204] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-1e11ef2b-e661-483a-b1ad-9cb57e29c94f tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] [instance: 461f98f2-80ba-4088-9e7f-1ba2a9b0c1d9] Extended root virtual disk {{(pid=61806) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 976.047778] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-597df420-2407-40bf-b521-62764426eb5f {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 976.070120] env[61806]: DEBUG nova.virt.vmwareapi.volumeops [None req-1e11ef2b-e661-483a-b1ad-9cb57e29c94f tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] [instance: 461f98f2-80ba-4088-9e7f-1ba2a9b0c1d9] Reconfiguring VM instance instance-0000005b to attach disk [datastore1] 461f98f2-80ba-4088-9e7f-1ba2a9b0c1d9/461f98f2-80ba-4088-9e7f-1ba2a9b0c1d9.vmdk or device None with type sparse {{(pid=61806) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 976.070453] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-007dbbb3-abab-459c-a230-7f7f94f6c04b {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 976.094299] env[61806]: DEBUG oslo_vmware.api [None req-28d96f96-d23f-4c45-9d59-0355de5e0c4f tempest-ServersNegativeTestJSON-891856253 tempest-ServersNegativeTestJSON-891856253-project-member] Task: {'id': task-1295017, 'name': PowerOnVM_Task} progress is 89%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 976.095717] env[61806]: DEBUG oslo_vmware.api [None req-1e11ef2b-e661-483a-b1ad-9cb57e29c94f tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Waiting for the task: (returnval){ [ 976.095717] env[61806]: value = "task-1295019" [ 976.095717] env[61806]: _type = "Task" [ 976.095717] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 976.108454] env[61806]: DEBUG oslo_vmware.api [None req-1e11ef2b-e661-483a-b1ad-9cb57e29c94f tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Task: {'id': task-1295019, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 976.120056] env[61806]: DEBUG oslo_concurrency.lockutils [None req-2c70fcbb-efa7-459d-81c0-a2a8a184893c tempest-AttachVolumeNegativeTest-1683033704 tempest-AttachVolumeNegativeTest-1683033704-project-member] Lock "e4638faa-7be1-4909-a595-c437837fa314" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 17.845s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 976.126748] env[61806]: DEBUG oslo_vmware.api [None req-57da61d4-139b-46a7-9163-13f9c5735119 tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] Task: {'id': task-1295018, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.19675} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 976.127026] env[61806]: DEBUG nova.virt.vmwareapi.ds_util [None req-57da61d4-139b-46a7-9163-13f9c5735119 tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] Deleted the datastore file {{(pid=61806) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 976.127225] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-57da61d4-139b-46a7-9163-13f9c5735119 tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] [instance: d9caee4a-a00c-4e50-b01c-99a8f796a09f] Deleted contents of the VM from datastore datastore1 {{(pid=61806) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 976.127455] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-57da61d4-139b-46a7-9163-13f9c5735119 tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] [instance: d9caee4a-a00c-4e50-b01c-99a8f796a09f] Instance destroyed {{(pid=61806) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 976.127663] env[61806]: INFO nova.compute.manager [None req-57da61d4-139b-46a7-9163-13f9c5735119 tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] [instance: d9caee4a-a00c-4e50-b01c-99a8f796a09f] Took 1.27 seconds to destroy the instance on the hypervisor. [ 976.127912] env[61806]: DEBUG oslo.service.loopingcall [None req-57da61d4-139b-46a7-9163-13f9c5735119 tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61806) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 976.128593] env[61806]: DEBUG nova.compute.manager [-] [instance: d9caee4a-a00c-4e50-b01c-99a8f796a09f] Deallocating network for instance {{(pid=61806) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 976.128593] env[61806]: DEBUG nova.network.neutron [-] [instance: d9caee4a-a00c-4e50-b01c-99a8f796a09f] deallocate_for_instance() {{(pid=61806) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 976.204362] env[61806]: DEBUG oslo_concurrency.lockutils [None req-49ec17b3-08f6-4e32-88fe-0409196f3352 tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] Lock "interface-d9caee4a-a00c-4e50-b01c-99a8f796a09f-7e031e88-6928-472d-a65e-b818d89e6fc5" "released" by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" :: held 9.804s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 976.226109] env[61806]: DEBUG nova.compute.manager [None req-d01361ec-1270-4938-a4ab-1319441d7070 tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] [instance: 962d3d88-03ee-4ff2-9d79-15091671b330] Start spawning the instance on the hypervisor. {{(pid=61806) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 976.254323] env[61806]: DEBUG nova.virt.hardware [None req-d01361ec-1270-4938-a4ab-1319441d7070 tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-15T18:05:10Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-15T18:04:52Z,direct_url=,disk_format='vmdk',id=ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='84ba83fce2c349a6988173c5d6fc3b07',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-15T18:04:53Z,virtual_size=,visibility=), allow threads: False {{(pid=61806) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 976.254638] env[61806]: DEBUG nova.virt.hardware [None req-d01361ec-1270-4938-a4ab-1319441d7070 tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Flavor limits 0:0:0 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 976.254824] env[61806]: DEBUG nova.virt.hardware [None req-d01361ec-1270-4938-a4ab-1319441d7070 tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Image limits 0:0:0 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 976.255037] env[61806]: DEBUG nova.virt.hardware [None req-d01361ec-1270-4938-a4ab-1319441d7070 tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Flavor pref 0:0:0 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 976.255223] env[61806]: DEBUG nova.virt.hardware [None req-d01361ec-1270-4938-a4ab-1319441d7070 tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Image pref 0:0:0 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 976.255391] env[61806]: DEBUG nova.virt.hardware [None req-d01361ec-1270-4938-a4ab-1319441d7070 tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 976.255615] env[61806]: DEBUG nova.virt.hardware [None req-d01361ec-1270-4938-a4ab-1319441d7070 tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61806) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 976.255782] env[61806]: DEBUG nova.virt.hardware [None req-d01361ec-1270-4938-a4ab-1319441d7070 tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61806) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 976.255958] env[61806]: DEBUG nova.virt.hardware [None req-d01361ec-1270-4938-a4ab-1319441d7070 tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Got 1 possible topologies {{(pid=61806) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 976.256149] env[61806]: DEBUG nova.virt.hardware [None req-d01361ec-1270-4938-a4ab-1319441d7070 tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61806) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 976.256330] env[61806]: DEBUG nova.virt.hardware [None req-d01361ec-1270-4938-a4ab-1319441d7070 tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61806) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 976.257783] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4fd55b32-8a9e-488b-ad10-28669d77d555 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 976.267205] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-866467d3-1fc4-4df1-a786-bcc937c3501d {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 976.485709] env[61806]: DEBUG oslo_concurrency.lockutils [None req-a06ee197-a4db-47dc-a9f0-aee2aa23de70 tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.281s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 976.486279] env[61806]: DEBUG nova.compute.manager [None req-a06ee197-a4db-47dc-a9f0-aee2aa23de70 tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] [instance: 9357e61f-1628-43bd-ab46-de13c1529f51] Start building networks asynchronously for instance. {{(pid=61806) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 976.492025] env[61806]: DEBUG oslo_concurrency.lockutils [None req-bf67b140-cd2e-4476-9b1b-cb960c9fa13c tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 2.516s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 976.492025] env[61806]: DEBUG nova.objects.instance [None req-bf67b140-cd2e-4476-9b1b-cb960c9fa13c tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Lazy-loading 'resources' on Instance uuid 676f18b3-3da1-4a53-b96d-fe64ee8f9101 {{(pid=61806) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 976.594836] env[61806]: DEBUG oslo_vmware.api [None req-28d96f96-d23f-4c45-9d59-0355de5e0c4f tempest-ServersNegativeTestJSON-891856253 tempest-ServersNegativeTestJSON-891856253-project-member] Task: {'id': task-1295017, 'name': PowerOnVM_Task, 'duration_secs': 0.810317} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 976.595110] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [None req-28d96f96-d23f-4c45-9d59-0355de5e0c4f tempest-ServersNegativeTestJSON-891856253 tempest-ServersNegativeTestJSON-891856253-project-member] [instance: e146f7e5-bbf8-4adf-a2e9-b7e753438976] Powered on the VM {{(pid=61806) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 976.595353] env[61806]: INFO nova.compute.manager [None req-28d96f96-d23f-4c45-9d59-0355de5e0c4f tempest-ServersNegativeTestJSON-891856253 tempest-ServersNegativeTestJSON-891856253-project-member] [instance: e146f7e5-bbf8-4adf-a2e9-b7e753438976] Took 7.80 seconds to spawn the instance on the hypervisor. [ 976.595554] env[61806]: DEBUG nova.compute.manager [None req-28d96f96-d23f-4c45-9d59-0355de5e0c4f tempest-ServersNegativeTestJSON-891856253 tempest-ServersNegativeTestJSON-891856253-project-member] [instance: e146f7e5-bbf8-4adf-a2e9-b7e753438976] Checking state {{(pid=61806) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 976.596713] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ea4b4a24-3e87-4256-8fbc-64e921a15c01 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 976.607680] env[61806]: DEBUG oslo_vmware.api [None req-1e11ef2b-e661-483a-b1ad-9cb57e29c94f tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Task: {'id': task-1295019, 'name': ReconfigVM_Task, 'duration_secs': 0.417071} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 976.609297] env[61806]: DEBUG nova.virt.vmwareapi.volumeops [None req-1e11ef2b-e661-483a-b1ad-9cb57e29c94f tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] [instance: 461f98f2-80ba-4088-9e7f-1ba2a9b0c1d9] Reconfigured VM instance instance-0000005b to attach disk [datastore1] 461f98f2-80ba-4088-9e7f-1ba2a9b0c1d9/461f98f2-80ba-4088-9e7f-1ba2a9b0c1d9.vmdk or device None with type sparse {{(pid=61806) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 976.612360] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-0ff003c0-a994-45cf-b090-fe274d6544fc {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 976.621418] env[61806]: DEBUG oslo_vmware.api [None req-1e11ef2b-e661-483a-b1ad-9cb57e29c94f tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Waiting for the task: (returnval){ [ 976.621418] env[61806]: value = "task-1295020" [ 976.621418] env[61806]: _type = "Task" [ 976.621418] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 976.629673] env[61806]: DEBUG oslo_vmware.api [None req-1e11ef2b-e661-483a-b1ad-9cb57e29c94f tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Task: {'id': task-1295020, 'name': Rename_Task} progress is 5%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 976.714163] env[61806]: DEBUG nova.compute.manager [req-7d1795db-e742-4cf8-a3b0-7cd46bb25a59 req-aa031661-7b8c-4add-84fb-b4ae3b7c5798 service nova] [instance: 962d3d88-03ee-4ff2-9d79-15091671b330] Received event network-vif-plugged-1a817be6-517c-4be9-8c14-c6f3995f87e0 {{(pid=61806) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 976.714465] env[61806]: DEBUG oslo_concurrency.lockutils [req-7d1795db-e742-4cf8-a3b0-7cd46bb25a59 req-aa031661-7b8c-4add-84fb-b4ae3b7c5798 service nova] Acquiring lock "962d3d88-03ee-4ff2-9d79-15091671b330-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 976.714712] env[61806]: DEBUG oslo_concurrency.lockutils [req-7d1795db-e742-4cf8-a3b0-7cd46bb25a59 req-aa031661-7b8c-4add-84fb-b4ae3b7c5798 service nova] Lock "962d3d88-03ee-4ff2-9d79-15091671b330-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 976.714919] env[61806]: DEBUG oslo_concurrency.lockutils [req-7d1795db-e742-4cf8-a3b0-7cd46bb25a59 req-aa031661-7b8c-4add-84fb-b4ae3b7c5798 service nova] Lock "962d3d88-03ee-4ff2-9d79-15091671b330-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 976.715341] env[61806]: DEBUG nova.compute.manager [req-7d1795db-e742-4cf8-a3b0-7cd46bb25a59 req-aa031661-7b8c-4add-84fb-b4ae3b7c5798 service nova] [instance: 962d3d88-03ee-4ff2-9d79-15091671b330] No waiting events found dispatching network-vif-plugged-1a817be6-517c-4be9-8c14-c6f3995f87e0 {{(pid=61806) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 976.715635] env[61806]: WARNING nova.compute.manager [req-7d1795db-e742-4cf8-a3b0-7cd46bb25a59 req-aa031661-7b8c-4add-84fb-b4ae3b7c5798 service nova] [instance: 962d3d88-03ee-4ff2-9d79-15091671b330] Received unexpected event network-vif-plugged-1a817be6-517c-4be9-8c14-c6f3995f87e0 for instance with vm_state building and task_state spawning. [ 976.946365] env[61806]: DEBUG nova.compute.manager [req-4d86522c-763d-45b4-b6c5-273f37364aab req-bec15160-9f6b-47dd-9cb6-5217ac24f8d4 service nova] [instance: d9caee4a-a00c-4e50-b01c-99a8f796a09f] Received event network-vif-deleted-c725ee28-d7a8-4b31-92bc-f63174564f5e {{(pid=61806) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 976.946365] env[61806]: INFO nova.compute.manager [req-4d86522c-763d-45b4-b6c5-273f37364aab req-bec15160-9f6b-47dd-9cb6-5217ac24f8d4 service nova] [instance: d9caee4a-a00c-4e50-b01c-99a8f796a09f] Neutron deleted interface c725ee28-d7a8-4b31-92bc-f63174564f5e; detaching it from the instance and deleting it from the info cache [ 976.946365] env[61806]: DEBUG nova.network.neutron [req-4d86522c-763d-45b4-b6c5-273f37364aab req-bec15160-9f6b-47dd-9cb6-5217ac24f8d4 service nova] [instance: d9caee4a-a00c-4e50-b01c-99a8f796a09f] Updating instance_info_cache with network_info: [] {{(pid=61806) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 976.947938] env[61806]: DEBUG nova.network.neutron [None req-d01361ec-1270-4938-a4ab-1319441d7070 tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] [instance: 962d3d88-03ee-4ff2-9d79-15091671b330] Successfully updated port: 1a817be6-517c-4be9-8c14-c6f3995f87e0 {{(pid=61806) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 976.950781] env[61806]: DEBUG oslo_concurrency.lockutils [None req-b6079056-6464-46e6-97ab-343b76917c18 tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] Acquiring lock "fd9e520a-2e7b-43d4-b84d-8933c32cc6a6" by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 976.950997] env[61806]: DEBUG oslo_concurrency.lockutils [None req-b6079056-6464-46e6-97ab-343b76917c18 tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] Lock "fd9e520a-2e7b-43d4-b84d-8933c32cc6a6" acquired by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" :: waited 0.000s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 976.951230] env[61806]: INFO nova.compute.manager [None req-b6079056-6464-46e6-97ab-343b76917c18 tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] [instance: fd9e520a-2e7b-43d4-b84d-8933c32cc6a6] Shelving [ 976.992985] env[61806]: DEBUG nova.compute.utils [None req-a06ee197-a4db-47dc-a9f0-aee2aa23de70 tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Using /dev/sd instead of None {{(pid=61806) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 976.997069] env[61806]: DEBUG nova.compute.manager [None req-a06ee197-a4db-47dc-a9f0-aee2aa23de70 tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] [instance: 9357e61f-1628-43bd-ab46-de13c1529f51] Allocating IP information in the background. {{(pid=61806) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 976.997254] env[61806]: DEBUG nova.network.neutron [None req-a06ee197-a4db-47dc-a9f0-aee2aa23de70 tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] [instance: 9357e61f-1628-43bd-ab46-de13c1529f51] allocate_for_instance() {{(pid=61806) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 977.055193] env[61806]: DEBUG nova.policy [None req-a06ee197-a4db-47dc-a9f0-aee2aa23de70 tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'b2631ce1f6dd40a4bfc5da7f90cde260', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '4b9d94c3d684474b99e305fb6cd9b95f', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61806) authorize /opt/stack/nova/nova/policy.py:201}} [ 977.131192] env[61806]: INFO nova.compute.manager [None req-28d96f96-d23f-4c45-9d59-0355de5e0c4f tempest-ServersNegativeTestJSON-891856253 tempest-ServersNegativeTestJSON-891856253-project-member] [instance: e146f7e5-bbf8-4adf-a2e9-b7e753438976] Took 16.29 seconds to build instance. [ 977.137312] env[61806]: DEBUG oslo_vmware.api [None req-1e11ef2b-e661-483a-b1ad-9cb57e29c94f tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Task: {'id': task-1295020, 'name': Rename_Task, 'duration_secs': 0.171789} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 977.137312] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [None req-1e11ef2b-e661-483a-b1ad-9cb57e29c94f tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] [instance: 461f98f2-80ba-4088-9e7f-1ba2a9b0c1d9] Powering on the VM {{(pid=61806) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 977.137312] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-73d675fc-1724-494b-8d1e-fdc6630d3e65 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 977.147075] env[61806]: DEBUG oslo_vmware.api [None req-1e11ef2b-e661-483a-b1ad-9cb57e29c94f tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Waiting for the task: (returnval){ [ 977.147075] env[61806]: value = "task-1295021" [ 977.147075] env[61806]: _type = "Task" [ 977.147075] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 977.157770] env[61806]: DEBUG oslo_vmware.api [None req-1e11ef2b-e661-483a-b1ad-9cb57e29c94f tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Task: {'id': task-1295021, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 977.214828] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-07bacaf6-6112-4a96-8f39-98f33020f440 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 977.223368] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f0716b3d-f4a8-421f-bd1e-9a0c4c7a8a4a {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 977.265300] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-adb8c10c-8dc0-4ad6-9093-5cec8903bf21 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 977.274486] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3b02d466-945c-4fe4-96ee-1c374322996f {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 977.289818] env[61806]: DEBUG nova.compute.provider_tree [None req-bf67b140-cd2e-4476-9b1b-cb960c9fa13c tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Inventory has not changed in ProviderTree for provider: a2858be1-fd22-4e08-979e-87ad25293407 {{(pid=61806) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 977.418073] env[61806]: DEBUG nova.network.neutron [-] [instance: d9caee4a-a00c-4e50-b01c-99a8f796a09f] Updating instance_info_cache with network_info: [] {{(pid=61806) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 977.452623] env[61806]: DEBUG oslo_concurrency.lockutils [None req-d01361ec-1270-4938-a4ab-1319441d7070 tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Acquiring lock "refresh_cache-962d3d88-03ee-4ff2-9d79-15091671b330" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 977.456019] env[61806]: DEBUG oslo_concurrency.lockutils [None req-d01361ec-1270-4938-a4ab-1319441d7070 tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Acquired lock "refresh_cache-962d3d88-03ee-4ff2-9d79-15091671b330" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 977.456019] env[61806]: DEBUG nova.network.neutron [None req-d01361ec-1270-4938-a4ab-1319441d7070 tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] [instance: 962d3d88-03ee-4ff2-9d79-15091671b330] Building network info cache for instance {{(pid=61806) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 977.456019] env[61806]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-686de921-6597-48d9-b0f5-b1ed93d9dcb3 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 977.466052] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [None req-b6079056-6464-46e6-97ab-343b76917c18 tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] [instance: fd9e520a-2e7b-43d4-b84d-8933c32cc6a6] Powering off the VM {{(pid=61806) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 977.466394] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-03082a72-5cf8-44bb-ad1a-f00f532d4e1a {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 977.473497] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-474d8579-8e79-40c6-85e9-3498fbea6f47 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 977.487333] env[61806]: DEBUG oslo_vmware.api [None req-b6079056-6464-46e6-97ab-343b76917c18 tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] Waiting for the task: (returnval){ [ 977.487333] env[61806]: value = "task-1295022" [ 977.487333] env[61806]: _type = "Task" [ 977.487333] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 977.498365] env[61806]: DEBUG nova.compute.manager [None req-a06ee197-a4db-47dc-a9f0-aee2aa23de70 tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] [instance: 9357e61f-1628-43bd-ab46-de13c1529f51] Start building block device mappings for instance. {{(pid=61806) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 977.506019] env[61806]: DEBUG oslo_vmware.api [None req-b6079056-6464-46e6-97ab-343b76917c18 tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] Task: {'id': task-1295022, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 977.514940] env[61806]: DEBUG nova.compute.manager [req-4d86522c-763d-45b4-b6c5-273f37364aab req-bec15160-9f6b-47dd-9cb6-5217ac24f8d4 service nova] [instance: d9caee4a-a00c-4e50-b01c-99a8f796a09f] Detach interface failed, port_id=c725ee28-d7a8-4b31-92bc-f63174564f5e, reason: Instance d9caee4a-a00c-4e50-b01c-99a8f796a09f could not be found. {{(pid=61806) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10963}} [ 977.638940] env[61806]: DEBUG oslo_concurrency.lockutils [None req-28d96f96-d23f-4c45-9d59-0355de5e0c4f tempest-ServersNegativeTestJSON-891856253 tempest-ServersNegativeTestJSON-891856253-project-member] Lock "e146f7e5-bbf8-4adf-a2e9-b7e753438976" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 17.806s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 977.659924] env[61806]: DEBUG oslo_vmware.api [None req-1e11ef2b-e661-483a-b1ad-9cb57e29c94f tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Task: {'id': task-1295021, 'name': PowerOnVM_Task, 'duration_secs': 0.505412} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 977.660280] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [None req-1e11ef2b-e661-483a-b1ad-9cb57e29c94f tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] [instance: 461f98f2-80ba-4088-9e7f-1ba2a9b0c1d9] Powered on the VM {{(pid=61806) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 977.660481] env[61806]: DEBUG nova.compute.manager [None req-1e11ef2b-e661-483a-b1ad-9cb57e29c94f tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] [instance: 461f98f2-80ba-4088-9e7f-1ba2a9b0c1d9] Checking state {{(pid=61806) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 977.661592] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-204fa469-8de9-464f-afd0-0ac481f97fb5 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 977.775879] env[61806]: DEBUG nova.network.neutron [None req-a06ee197-a4db-47dc-a9f0-aee2aa23de70 tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] [instance: 9357e61f-1628-43bd-ab46-de13c1529f51] Successfully created port: a4ca7c56-b6a1-4bd2-bd47-02b0ed4aa097 {{(pid=61806) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 977.792887] env[61806]: DEBUG nova.scheduler.client.report [None req-bf67b140-cd2e-4476-9b1b-cb960c9fa13c tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Inventory has not changed for provider a2858be1-fd22-4e08-979e-87ad25293407 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 138, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61806) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 977.921244] env[61806]: INFO nova.compute.manager [-] [instance: d9caee4a-a00c-4e50-b01c-99a8f796a09f] Took 1.79 seconds to deallocate network for instance. [ 977.992926] env[61806]: DEBUG nova.network.neutron [None req-d01361ec-1270-4938-a4ab-1319441d7070 tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] [instance: 962d3d88-03ee-4ff2-9d79-15091671b330] Instance cache missing network info. {{(pid=61806) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 978.001344] env[61806]: DEBUG oslo_vmware.api [None req-b6079056-6464-46e6-97ab-343b76917c18 tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] Task: {'id': task-1295022, 'name': PowerOffVM_Task, 'duration_secs': 0.337418} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 978.001527] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [None req-b6079056-6464-46e6-97ab-343b76917c18 tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] [instance: fd9e520a-2e7b-43d4-b84d-8933c32cc6a6] Powered off the VM {{(pid=61806) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 978.002978] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-210f29f9-866a-4766-a09c-ac8aef4a1db0 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 978.032706] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-51d3391d-620d-46c1-82d3-d7b1bf3f8284 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 978.049022] env[61806]: DEBUG oslo_concurrency.lockutils [None req-474c365d-f281-4fe7-b609-9d0c04e06393 tempest-ServersNegativeTestJSON-891856253 tempest-ServersNegativeTestJSON-891856253-project-member] Acquiring lock "e146f7e5-bbf8-4adf-a2e9-b7e753438976" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 978.049022] env[61806]: DEBUG oslo_concurrency.lockutils [None req-474c365d-f281-4fe7-b609-9d0c04e06393 tempest-ServersNegativeTestJSON-891856253 tempest-ServersNegativeTestJSON-891856253-project-member] Lock "e146f7e5-bbf8-4adf-a2e9-b7e753438976" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 978.049022] env[61806]: DEBUG oslo_concurrency.lockutils [None req-474c365d-f281-4fe7-b609-9d0c04e06393 tempest-ServersNegativeTestJSON-891856253 tempest-ServersNegativeTestJSON-891856253-project-member] Acquiring lock "e146f7e5-bbf8-4adf-a2e9-b7e753438976-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 978.049022] env[61806]: DEBUG oslo_concurrency.lockutils [None req-474c365d-f281-4fe7-b609-9d0c04e06393 tempest-ServersNegativeTestJSON-891856253 tempest-ServersNegativeTestJSON-891856253-project-member] Lock "e146f7e5-bbf8-4adf-a2e9-b7e753438976-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 978.049022] env[61806]: DEBUG oslo_concurrency.lockutils [None req-474c365d-f281-4fe7-b609-9d0c04e06393 tempest-ServersNegativeTestJSON-891856253 tempest-ServersNegativeTestJSON-891856253-project-member] Lock "e146f7e5-bbf8-4adf-a2e9-b7e753438976-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 978.050597] env[61806]: INFO nova.compute.manager [None req-474c365d-f281-4fe7-b609-9d0c04e06393 tempest-ServersNegativeTestJSON-891856253 tempest-ServersNegativeTestJSON-891856253-project-member] [instance: e146f7e5-bbf8-4adf-a2e9-b7e753438976] Terminating instance [ 978.055030] env[61806]: DEBUG nova.compute.manager [None req-474c365d-f281-4fe7-b609-9d0c04e06393 tempest-ServersNegativeTestJSON-891856253 tempest-ServersNegativeTestJSON-891856253-project-member] [instance: e146f7e5-bbf8-4adf-a2e9-b7e753438976] Start destroying the instance on the hypervisor. {{(pid=61806) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 978.055235] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-474c365d-f281-4fe7-b609-9d0c04e06393 tempest-ServersNegativeTestJSON-891856253 tempest-ServersNegativeTestJSON-891856253-project-member] [instance: e146f7e5-bbf8-4adf-a2e9-b7e753438976] Destroying instance {{(pid=61806) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 978.056384] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cb4b41df-82cb-4661-8654-1cdf806f7def {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 978.065132] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [None req-474c365d-f281-4fe7-b609-9d0c04e06393 tempest-ServersNegativeTestJSON-891856253 tempest-ServersNegativeTestJSON-891856253-project-member] [instance: e146f7e5-bbf8-4adf-a2e9-b7e753438976] Powering off the VM {{(pid=61806) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 978.065489] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-3f0c58f0-f88f-49cc-a1e5-c8b45000dd84 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 978.072934] env[61806]: DEBUG oslo_vmware.api [None req-474c365d-f281-4fe7-b609-9d0c04e06393 tempest-ServersNegativeTestJSON-891856253 tempest-ServersNegativeTestJSON-891856253-project-member] Waiting for the task: (returnval){ [ 978.072934] env[61806]: value = "task-1295023" [ 978.072934] env[61806]: _type = "Task" [ 978.072934] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 978.083183] env[61806]: DEBUG oslo_vmware.api [None req-474c365d-f281-4fe7-b609-9d0c04e06393 tempest-ServersNegativeTestJSON-891856253 tempest-ServersNegativeTestJSON-891856253-project-member] Task: {'id': task-1295023, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 978.156255] env[61806]: DEBUG nova.network.neutron [None req-d01361ec-1270-4938-a4ab-1319441d7070 tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] [instance: 962d3d88-03ee-4ff2-9d79-15091671b330] Updating instance_info_cache with network_info: [{"id": "1a817be6-517c-4be9-8c14-c6f3995f87e0", "address": "fa:16:3e:f7:3d:db", "network": {"id": "b7e78559-d885-4b59-9c7f-9d966b424d04", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-1456721354-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7ab95fb72b5d46c3b6c7bebbccf897cd", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d891a662-7da0-4be5-8d0d-01b4ff70552a", "external-id": "nsx-vlan-transportzone-815", "segmentation_id": 815, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1a817be6-51", "ovs_interfaceid": "1a817be6-517c-4be9-8c14-c6f3995f87e0", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61806) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 978.180891] env[61806]: DEBUG oslo_concurrency.lockutils [None req-1e11ef2b-e661-483a-b1ad-9cb57e29c94f tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 978.299139] env[61806]: DEBUG oslo_concurrency.lockutils [None req-bf67b140-cd2e-4476-9b1b-cb960c9fa13c tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.810s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 978.302314] env[61806]: DEBUG oslo_concurrency.lockutils [None req-38f9dcbc-e34b-458d-8ba6-e6b3c0e6b471 tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 3.167s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 978.302584] env[61806]: DEBUG nova.objects.instance [None req-38f9dcbc-e34b-458d-8ba6-e6b3c0e6b471 tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] Lazy-loading 'resources' on Instance uuid 0bac3048-759a-47d9-a6ef-a0386a49a974 {{(pid=61806) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 978.325668] env[61806]: INFO nova.scheduler.client.report [None req-bf67b140-cd2e-4476-9b1b-cb960c9fa13c tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Deleted allocations for instance 676f18b3-3da1-4a53-b96d-fe64ee8f9101 [ 978.429435] env[61806]: DEBUG oslo_concurrency.lockutils [None req-57da61d4-139b-46a7-9163-13f9c5735119 tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 978.537330] env[61806]: DEBUG nova.compute.manager [None req-a06ee197-a4db-47dc-a9f0-aee2aa23de70 tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] [instance: 9357e61f-1628-43bd-ab46-de13c1529f51] Start spawning the instance on the hypervisor. {{(pid=61806) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 978.544509] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-b6079056-6464-46e6-97ab-343b76917c18 tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] [instance: fd9e520a-2e7b-43d4-b84d-8933c32cc6a6] Creating Snapshot of the VM instance {{(pid=61806) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 978.544841] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-ba909d66-6245-4205-8357-23b1ba1dfab2 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 978.554902] env[61806]: DEBUG oslo_vmware.api [None req-b6079056-6464-46e6-97ab-343b76917c18 tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] Waiting for the task: (returnval){ [ 978.554902] env[61806]: value = "task-1295024" [ 978.554902] env[61806]: _type = "Task" [ 978.554902] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 978.566511] env[61806]: DEBUG oslo_vmware.api [None req-b6079056-6464-46e6-97ab-343b76917c18 tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] Task: {'id': task-1295024, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 978.568879] env[61806]: DEBUG nova.virt.hardware [None req-a06ee197-a4db-47dc-a9f0-aee2aa23de70 tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-15T18:05:10Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-15T18:04:52Z,direct_url=,disk_format='vmdk',id=ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='84ba83fce2c349a6988173c5d6fc3b07',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-15T18:04:53Z,virtual_size=,visibility=), allow threads: False {{(pid=61806) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 978.569130] env[61806]: DEBUG nova.virt.hardware [None req-a06ee197-a4db-47dc-a9f0-aee2aa23de70 tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Flavor limits 0:0:0 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 978.569309] env[61806]: DEBUG nova.virt.hardware [None req-a06ee197-a4db-47dc-a9f0-aee2aa23de70 tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Image limits 0:0:0 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 978.569505] env[61806]: DEBUG nova.virt.hardware [None req-a06ee197-a4db-47dc-a9f0-aee2aa23de70 tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Flavor pref 0:0:0 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 978.569751] env[61806]: DEBUG nova.virt.hardware [None req-a06ee197-a4db-47dc-a9f0-aee2aa23de70 tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Image pref 0:0:0 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 978.569966] env[61806]: DEBUG nova.virt.hardware [None req-a06ee197-a4db-47dc-a9f0-aee2aa23de70 tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 978.570226] env[61806]: DEBUG nova.virt.hardware [None req-a06ee197-a4db-47dc-a9f0-aee2aa23de70 tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61806) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 978.570402] env[61806]: DEBUG nova.virt.hardware [None req-a06ee197-a4db-47dc-a9f0-aee2aa23de70 tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61806) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 978.570582] env[61806]: DEBUG nova.virt.hardware [None req-a06ee197-a4db-47dc-a9f0-aee2aa23de70 tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Got 1 possible topologies {{(pid=61806) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 978.570750] env[61806]: DEBUG nova.virt.hardware [None req-a06ee197-a4db-47dc-a9f0-aee2aa23de70 tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61806) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 978.570925] env[61806]: DEBUG nova.virt.hardware [None req-a06ee197-a4db-47dc-a9f0-aee2aa23de70 tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61806) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 978.571799] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c6d86b8a-6b9a-4175-a108-1fadc6eb01bf {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 978.585510] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ed48dc5f-7457-4110-8b6d-ed6880ecfaab {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 978.589436] env[61806]: DEBUG oslo_vmware.api [None req-474c365d-f281-4fe7-b609-9d0c04e06393 tempest-ServersNegativeTestJSON-891856253 tempest-ServersNegativeTestJSON-891856253-project-member] Task: {'id': task-1295023, 'name': PowerOffVM_Task, 'duration_secs': 0.412103} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 978.589712] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [None req-474c365d-f281-4fe7-b609-9d0c04e06393 tempest-ServersNegativeTestJSON-891856253 tempest-ServersNegativeTestJSON-891856253-project-member] [instance: e146f7e5-bbf8-4adf-a2e9-b7e753438976] Powered off the VM {{(pid=61806) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 978.589886] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-474c365d-f281-4fe7-b609-9d0c04e06393 tempest-ServersNegativeTestJSON-891856253 tempest-ServersNegativeTestJSON-891856253-project-member] [instance: e146f7e5-bbf8-4adf-a2e9-b7e753438976] Unregistering the VM {{(pid=61806) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 978.590507] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-b42c77d3-4e7e-4e75-a347-c9488dd547df {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 978.658737] env[61806]: DEBUG oslo_concurrency.lockutils [None req-d01361ec-1270-4938-a4ab-1319441d7070 tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Releasing lock "refresh_cache-962d3d88-03ee-4ff2-9d79-15091671b330" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 978.659128] env[61806]: DEBUG nova.compute.manager [None req-d01361ec-1270-4938-a4ab-1319441d7070 tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] [instance: 962d3d88-03ee-4ff2-9d79-15091671b330] Instance network_info: |[{"id": "1a817be6-517c-4be9-8c14-c6f3995f87e0", "address": "fa:16:3e:f7:3d:db", "network": {"id": "b7e78559-d885-4b59-9c7f-9d966b424d04", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-1456721354-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7ab95fb72b5d46c3b6c7bebbccf897cd", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d891a662-7da0-4be5-8d0d-01b4ff70552a", "external-id": "nsx-vlan-transportzone-815", "segmentation_id": 815, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1a817be6-51", "ovs_interfaceid": "1a817be6-517c-4be9-8c14-c6f3995f87e0", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61806) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 978.659579] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-d01361ec-1270-4938-a4ab-1319441d7070 tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] [instance: 962d3d88-03ee-4ff2-9d79-15091671b330] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:f7:3d:db', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'd891a662-7da0-4be5-8d0d-01b4ff70552a', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '1a817be6-517c-4be9-8c14-c6f3995f87e0', 'vif_model': 'vmxnet3'}] {{(pid=61806) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 978.667278] env[61806]: DEBUG oslo.service.loopingcall [None req-d01361ec-1270-4938-a4ab-1319441d7070 tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61806) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 978.668640] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 962d3d88-03ee-4ff2-9d79-15091671b330] Creating VM on the ESX host {{(pid=61806) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 978.668925] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-474c365d-f281-4fe7-b609-9d0c04e06393 tempest-ServersNegativeTestJSON-891856253 tempest-ServersNegativeTestJSON-891856253-project-member] [instance: e146f7e5-bbf8-4adf-a2e9-b7e753438976] Unregistered the VM {{(pid=61806) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 978.669135] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-474c365d-f281-4fe7-b609-9d0c04e06393 tempest-ServersNegativeTestJSON-891856253 tempest-ServersNegativeTestJSON-891856253-project-member] [instance: e146f7e5-bbf8-4adf-a2e9-b7e753438976] Deleting contents of the VM from datastore datastore1 {{(pid=61806) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 978.669350] env[61806]: DEBUG nova.virt.vmwareapi.ds_util [None req-474c365d-f281-4fe7-b609-9d0c04e06393 tempest-ServersNegativeTestJSON-891856253 tempest-ServersNegativeTestJSON-891856253-project-member] Deleting the datastore file [datastore1] e146f7e5-bbf8-4adf-a2e9-b7e753438976 {{(pid=61806) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 978.669608] env[61806]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-081e0e83-567c-4e55-9c80-bdb309c39fa2 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 978.683712] env[61806]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-5dcee04d-4d62-4d9b-88a4-0f4e1ac28d58 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 978.690976] env[61806]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 978.690976] env[61806]: value = "task-1295027" [ 978.690976] env[61806]: _type = "Task" [ 978.690976] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 978.692351] env[61806]: DEBUG oslo_vmware.api [None req-474c365d-f281-4fe7-b609-9d0c04e06393 tempest-ServersNegativeTestJSON-891856253 tempest-ServersNegativeTestJSON-891856253-project-member] Waiting for the task: (returnval){ [ 978.692351] env[61806]: value = "task-1295026" [ 978.692351] env[61806]: _type = "Task" [ 978.692351] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 978.703952] env[61806]: DEBUG oslo_vmware.api [-] Task: {'id': task-1295027, 'name': CreateVM_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 978.706967] env[61806]: DEBUG oslo_vmware.api [None req-474c365d-f281-4fe7-b609-9d0c04e06393 tempest-ServersNegativeTestJSON-891856253 tempest-ServersNegativeTestJSON-891856253-project-member] Task: {'id': task-1295026, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 978.752173] env[61806]: DEBUG nova.compute.manager [req-a9da2378-8b8a-4564-90e1-34e0b059e0ce req-29824897-1d26-4e6d-86d6-5b8dfd680b8a service nova] [instance: 962d3d88-03ee-4ff2-9d79-15091671b330] Received event network-changed-1a817be6-517c-4be9-8c14-c6f3995f87e0 {{(pid=61806) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 978.752565] env[61806]: DEBUG nova.compute.manager [req-a9da2378-8b8a-4564-90e1-34e0b059e0ce req-29824897-1d26-4e6d-86d6-5b8dfd680b8a service nova] [instance: 962d3d88-03ee-4ff2-9d79-15091671b330] Refreshing instance network info cache due to event network-changed-1a817be6-517c-4be9-8c14-c6f3995f87e0. {{(pid=61806) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 978.752798] env[61806]: DEBUG oslo_concurrency.lockutils [req-a9da2378-8b8a-4564-90e1-34e0b059e0ce req-29824897-1d26-4e6d-86d6-5b8dfd680b8a service nova] Acquiring lock "refresh_cache-962d3d88-03ee-4ff2-9d79-15091671b330" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 978.752952] env[61806]: DEBUG oslo_concurrency.lockutils [req-a9da2378-8b8a-4564-90e1-34e0b059e0ce req-29824897-1d26-4e6d-86d6-5b8dfd680b8a service nova] Acquired lock "refresh_cache-962d3d88-03ee-4ff2-9d79-15091671b330" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 978.753147] env[61806]: DEBUG nova.network.neutron [req-a9da2378-8b8a-4564-90e1-34e0b059e0ce req-29824897-1d26-4e6d-86d6-5b8dfd680b8a service nova] [instance: 962d3d88-03ee-4ff2-9d79-15091671b330] Refreshing network info cache for port 1a817be6-517c-4be9-8c14-c6f3995f87e0 {{(pid=61806) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 978.835765] env[61806]: DEBUG oslo_concurrency.lockutils [None req-bf67b140-cd2e-4476-9b1b-cb960c9fa13c tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Lock "676f18b3-3da1-4a53-b96d-fe64ee8f9101" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 7.780s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 978.997331] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9a81b5a0-9cfa-4d94-8b82-abbbb13a5a68 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 979.005618] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0457273a-4bd5-4ada-92e3-6e53e43bb3a8 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 979.039507] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bc3b2308-7bfe-4dbc-9596-a466734e7a01 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 979.047880] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-069529ad-e182-4339-b88b-d0a0475bac2a {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 979.063766] env[61806]: DEBUG nova.compute.provider_tree [None req-38f9dcbc-e34b-458d-8ba6-e6b3c0e6b471 tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] Inventory has not changed in ProviderTree for provider: a2858be1-fd22-4e08-979e-87ad25293407 {{(pid=61806) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 979.073488] env[61806]: DEBUG oslo_vmware.api [None req-b6079056-6464-46e6-97ab-343b76917c18 tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] Task: {'id': task-1295024, 'name': CreateSnapshot_Task} progress is 100%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 979.205811] env[61806]: DEBUG oslo_vmware.api [None req-474c365d-f281-4fe7-b609-9d0c04e06393 tempest-ServersNegativeTestJSON-891856253 tempest-ServersNegativeTestJSON-891856253-project-member] Task: {'id': task-1295026, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.200607} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 979.208869] env[61806]: DEBUG nova.virt.vmwareapi.ds_util [None req-474c365d-f281-4fe7-b609-9d0c04e06393 tempest-ServersNegativeTestJSON-891856253 tempest-ServersNegativeTestJSON-891856253-project-member] Deleted the datastore file {{(pid=61806) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 979.209134] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-474c365d-f281-4fe7-b609-9d0c04e06393 tempest-ServersNegativeTestJSON-891856253 tempest-ServersNegativeTestJSON-891856253-project-member] [instance: e146f7e5-bbf8-4adf-a2e9-b7e753438976] Deleted contents of the VM from datastore datastore1 {{(pid=61806) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 979.209352] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-474c365d-f281-4fe7-b609-9d0c04e06393 tempest-ServersNegativeTestJSON-891856253 tempest-ServersNegativeTestJSON-891856253-project-member] [instance: e146f7e5-bbf8-4adf-a2e9-b7e753438976] Instance destroyed {{(pid=61806) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 979.209588] env[61806]: INFO nova.compute.manager [None req-474c365d-f281-4fe7-b609-9d0c04e06393 tempest-ServersNegativeTestJSON-891856253 tempest-ServersNegativeTestJSON-891856253-project-member] [instance: e146f7e5-bbf8-4adf-a2e9-b7e753438976] Took 1.15 seconds to destroy the instance on the hypervisor. [ 979.209910] env[61806]: DEBUG oslo.service.loopingcall [None req-474c365d-f281-4fe7-b609-9d0c04e06393 tempest-ServersNegativeTestJSON-891856253 tempest-ServersNegativeTestJSON-891856253-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61806) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 979.210119] env[61806]: DEBUG oslo_vmware.api [-] Task: {'id': task-1295027, 'name': CreateVM_Task, 'duration_secs': 0.372483} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 979.210368] env[61806]: DEBUG nova.compute.manager [-] [instance: e146f7e5-bbf8-4adf-a2e9-b7e753438976] Deallocating network for instance {{(pid=61806) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 979.210504] env[61806]: DEBUG nova.network.neutron [-] [instance: e146f7e5-bbf8-4adf-a2e9-b7e753438976] deallocate_for_instance() {{(pid=61806) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 979.212371] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 962d3d88-03ee-4ff2-9d79-15091671b330] Created VM on the ESX host {{(pid=61806) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 979.213016] env[61806]: DEBUG oslo_concurrency.lockutils [None req-d01361ec-1270-4938-a4ab-1319441d7070 tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 979.213193] env[61806]: DEBUG oslo_concurrency.lockutils [None req-d01361ec-1270-4938-a4ab-1319441d7070 tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Acquired lock "[datastore1] devstack-image-cache_base/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 979.213510] env[61806]: DEBUG oslo_concurrency.lockutils [None req-d01361ec-1270-4938-a4ab-1319441d7070 tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 979.214788] env[61806]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-cd4fcb74-5f9b-4f45-9359-7b47f333074e {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 979.220226] env[61806]: DEBUG oslo_vmware.api [None req-d01361ec-1270-4938-a4ab-1319441d7070 tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Waiting for the task: (returnval){ [ 979.220226] env[61806]: value = "session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]5220d9cc-ca8d-6f05-ae1f-564cad7c471d" [ 979.220226] env[61806]: _type = "Task" [ 979.220226] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 979.228383] env[61806]: DEBUG oslo_vmware.api [None req-d01361ec-1270-4938-a4ab-1319441d7070 tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Task: {'id': session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]5220d9cc-ca8d-6f05-ae1f-564cad7c471d, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 979.313258] env[61806]: DEBUG nova.network.neutron [None req-a06ee197-a4db-47dc-a9f0-aee2aa23de70 tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] [instance: 9357e61f-1628-43bd-ab46-de13c1529f51] Successfully updated port: a4ca7c56-b6a1-4bd2-bd47-02b0ed4aa097 {{(pid=61806) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 979.569114] env[61806]: DEBUG nova.scheduler.client.report [None req-38f9dcbc-e34b-458d-8ba6-e6b3c0e6b471 tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] Inventory has not changed for provider a2858be1-fd22-4e08-979e-87ad25293407 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 138, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61806) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 979.578045] env[61806]: DEBUG oslo_vmware.api [None req-b6079056-6464-46e6-97ab-343b76917c18 tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] Task: {'id': task-1295024, 'name': CreateSnapshot_Task, 'duration_secs': 0.521311} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 979.578281] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-b6079056-6464-46e6-97ab-343b76917c18 tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] [instance: fd9e520a-2e7b-43d4-b84d-8933c32cc6a6] Created Snapshot of the VM instance {{(pid=61806) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 979.579060] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9b50d568-457a-412b-9ada-88ed6bf423dd {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 979.647299] env[61806]: DEBUG nova.network.neutron [req-a9da2378-8b8a-4564-90e1-34e0b059e0ce req-29824897-1d26-4e6d-86d6-5b8dfd680b8a service nova] [instance: 962d3d88-03ee-4ff2-9d79-15091671b330] Updated VIF entry in instance network info cache for port 1a817be6-517c-4be9-8c14-c6f3995f87e0. {{(pid=61806) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 979.647684] env[61806]: DEBUG nova.network.neutron [req-a9da2378-8b8a-4564-90e1-34e0b059e0ce req-29824897-1d26-4e6d-86d6-5b8dfd680b8a service nova] [instance: 962d3d88-03ee-4ff2-9d79-15091671b330] Updating instance_info_cache with network_info: [{"id": "1a817be6-517c-4be9-8c14-c6f3995f87e0", "address": "fa:16:3e:f7:3d:db", "network": {"id": "b7e78559-d885-4b59-9c7f-9d966b424d04", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-1456721354-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7ab95fb72b5d46c3b6c7bebbccf897cd", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d891a662-7da0-4be5-8d0d-01b4ff70552a", "external-id": "nsx-vlan-transportzone-815", "segmentation_id": 815, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1a817be6-51", "ovs_interfaceid": "1a817be6-517c-4be9-8c14-c6f3995f87e0", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61806) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 979.650609] env[61806]: DEBUG nova.compute.manager [req-9c143e36-931a-4f0a-985d-4a4f9fd66438 req-c48e4d10-9396-47bf-8c03-3089c2f45b3e service nova] [instance: e146f7e5-bbf8-4adf-a2e9-b7e753438976] Received event network-vif-deleted-b68f5483-a11f-4b35-aaa0-2dcd7f50e448 {{(pid=61806) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 979.650810] env[61806]: INFO nova.compute.manager [req-9c143e36-931a-4f0a-985d-4a4f9fd66438 req-c48e4d10-9396-47bf-8c03-3089c2f45b3e service nova] [instance: e146f7e5-bbf8-4adf-a2e9-b7e753438976] Neutron deleted interface b68f5483-a11f-4b35-aaa0-2dcd7f50e448; detaching it from the instance and deleting it from the info cache [ 979.651052] env[61806]: DEBUG nova.network.neutron [req-9c143e36-931a-4f0a-985d-4a4f9fd66438 req-c48e4d10-9396-47bf-8c03-3089c2f45b3e service nova] [instance: e146f7e5-bbf8-4adf-a2e9-b7e753438976] Updating instance_info_cache with network_info: [] {{(pid=61806) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 979.731191] env[61806]: DEBUG oslo_vmware.api [None req-d01361ec-1270-4938-a4ab-1319441d7070 tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Task: {'id': session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]5220d9cc-ca8d-6f05-ae1f-564cad7c471d, 'name': SearchDatastore_Task, 'duration_secs': 0.015425} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 979.731611] env[61806]: DEBUG oslo_concurrency.lockutils [None req-d01361ec-1270-4938-a4ab-1319441d7070 tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Releasing lock "[datastore1] devstack-image-cache_base/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 979.731890] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-d01361ec-1270-4938-a4ab-1319441d7070 tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] [instance: 962d3d88-03ee-4ff2-9d79-15091671b330] Processing image ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa {{(pid=61806) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 979.732236] env[61806]: DEBUG oslo_concurrency.lockutils [None req-d01361ec-1270-4938-a4ab-1319441d7070 tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa.vmdk" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 979.732414] env[61806]: DEBUG oslo_concurrency.lockutils [None req-d01361ec-1270-4938-a4ab-1319441d7070 tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Acquired lock "[datastore1] devstack-image-cache_base/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa.vmdk" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 979.732612] env[61806]: DEBUG nova.virt.vmwareapi.ds_util [None req-d01361ec-1270-4938-a4ab-1319441d7070 tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61806) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 979.732870] env[61806]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-a615a433-4555-4c4b-b57c-b0377e1374cc {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 979.747177] env[61806]: DEBUG nova.virt.vmwareapi.ds_util [None req-d01361ec-1270-4938-a4ab-1319441d7070 tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61806) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 979.747365] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-d01361ec-1270-4938-a4ab-1319441d7070 tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61806) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 979.748089] env[61806]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ae279dc6-4b9e-48d0-9bdb-2832bf8a954e {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 979.753538] env[61806]: DEBUG oslo_vmware.api [None req-d01361ec-1270-4938-a4ab-1319441d7070 tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Waiting for the task: (returnval){ [ 979.753538] env[61806]: value = "session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]52a34fc6-78b9-e770-5d63-1151efd6d346" [ 979.753538] env[61806]: _type = "Task" [ 979.753538] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 979.756606] env[61806]: DEBUG oslo_concurrency.lockutils [None req-eb9ccb65-4d10-45a5-b6b1-254652cf9691 tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Acquiring lock "461f98f2-80ba-4088-9e7f-1ba2a9b0c1d9" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 979.756880] env[61806]: DEBUG oslo_concurrency.lockutils [None req-eb9ccb65-4d10-45a5-b6b1-254652cf9691 tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Lock "461f98f2-80ba-4088-9e7f-1ba2a9b0c1d9" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 979.757135] env[61806]: DEBUG oslo_concurrency.lockutils [None req-eb9ccb65-4d10-45a5-b6b1-254652cf9691 tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Acquiring lock "461f98f2-80ba-4088-9e7f-1ba2a9b0c1d9-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 979.757381] env[61806]: DEBUG oslo_concurrency.lockutils [None req-eb9ccb65-4d10-45a5-b6b1-254652cf9691 tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Lock "461f98f2-80ba-4088-9e7f-1ba2a9b0c1d9-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 979.757608] env[61806]: DEBUG oslo_concurrency.lockutils [None req-eb9ccb65-4d10-45a5-b6b1-254652cf9691 tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Lock "461f98f2-80ba-4088-9e7f-1ba2a9b0c1d9-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 979.759913] env[61806]: INFO nova.compute.manager [None req-eb9ccb65-4d10-45a5-b6b1-254652cf9691 tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] [instance: 461f98f2-80ba-4088-9e7f-1ba2a9b0c1d9] Terminating instance [ 979.761776] env[61806]: DEBUG nova.compute.manager [None req-eb9ccb65-4d10-45a5-b6b1-254652cf9691 tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] [instance: 461f98f2-80ba-4088-9e7f-1ba2a9b0c1d9] Start destroying the instance on the hypervisor. {{(pid=61806) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 979.761990] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-eb9ccb65-4d10-45a5-b6b1-254652cf9691 tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] [instance: 461f98f2-80ba-4088-9e7f-1ba2a9b0c1d9] Destroying instance {{(pid=61806) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 979.762938] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-774bb5b5-dceb-43f2-8883-bd47890e1e66 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 979.768280] env[61806]: DEBUG oslo_vmware.api [None req-d01361ec-1270-4938-a4ab-1319441d7070 tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Task: {'id': session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]52a34fc6-78b9-e770-5d63-1151efd6d346, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 979.773848] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [None req-eb9ccb65-4d10-45a5-b6b1-254652cf9691 tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] [instance: 461f98f2-80ba-4088-9e7f-1ba2a9b0c1d9] Powering off the VM {{(pid=61806) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 979.774105] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-1f277f8c-2e77-409b-a88d-770e08ae37da {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 979.781778] env[61806]: DEBUG oslo_vmware.api [None req-eb9ccb65-4d10-45a5-b6b1-254652cf9691 tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Waiting for the task: (returnval){ [ 979.781778] env[61806]: value = "task-1295028" [ 979.781778] env[61806]: _type = "Task" [ 979.781778] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 979.790289] env[61806]: DEBUG oslo_vmware.api [None req-eb9ccb65-4d10-45a5-b6b1-254652cf9691 tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Task: {'id': task-1295028, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 979.815431] env[61806]: DEBUG oslo_concurrency.lockutils [None req-a06ee197-a4db-47dc-a9f0-aee2aa23de70 tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Acquiring lock "refresh_cache-9357e61f-1628-43bd-ab46-de13c1529f51" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 979.815621] env[61806]: DEBUG oslo_concurrency.lockutils [None req-a06ee197-a4db-47dc-a9f0-aee2aa23de70 tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Acquired lock "refresh_cache-9357e61f-1628-43bd-ab46-de13c1529f51" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 979.815804] env[61806]: DEBUG nova.network.neutron [None req-a06ee197-a4db-47dc-a9f0-aee2aa23de70 tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] [instance: 9357e61f-1628-43bd-ab46-de13c1529f51] Building network info cache for instance {{(pid=61806) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 980.076368] env[61806]: DEBUG oslo_concurrency.lockutils [None req-38f9dcbc-e34b-458d-8ba6-e6b3c0e6b471 tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.775s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 980.078987] env[61806]: DEBUG oslo_concurrency.lockutils [None req-1e11ef2b-e661-483a-b1ad-9cb57e29c94f tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: waited 1.898s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 980.079207] env[61806]: DEBUG nova.objects.instance [None req-1e11ef2b-e661-483a-b1ad-9cb57e29c94f tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] [instance: 461f98f2-80ba-4088-9e7f-1ba2a9b0c1d9] Trying to apply a migration context that does not seem to be set for this instance {{(pid=61806) apply_migration_context /opt/stack/nova/nova/objects/instance.py:1067}} [ 980.099750] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-b6079056-6464-46e6-97ab-343b76917c18 tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] [instance: fd9e520a-2e7b-43d4-b84d-8933c32cc6a6] Creating linked-clone VM from snapshot {{(pid=61806) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 980.100075] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-b83a0ff8-8b2e-46da-95bd-675a82f77cbc {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 980.104360] env[61806]: INFO nova.scheduler.client.report [None req-38f9dcbc-e34b-458d-8ba6-e6b3c0e6b471 tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] Deleted allocations for instance 0bac3048-759a-47d9-a6ef-a0386a49a974 [ 980.114456] env[61806]: DEBUG oslo_vmware.api [None req-b6079056-6464-46e6-97ab-343b76917c18 tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] Waiting for the task: (returnval){ [ 980.114456] env[61806]: value = "task-1295029" [ 980.114456] env[61806]: _type = "Task" [ 980.114456] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 980.121185] env[61806]: DEBUG oslo_vmware.api [None req-b6079056-6464-46e6-97ab-343b76917c18 tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] Task: {'id': task-1295029, 'name': CloneVM_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 980.122545] env[61806]: DEBUG nova.network.neutron [-] [instance: e146f7e5-bbf8-4adf-a2e9-b7e753438976] Updating instance_info_cache with network_info: [] {{(pid=61806) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 980.153278] env[61806]: DEBUG oslo_concurrency.lockutils [req-a9da2378-8b8a-4564-90e1-34e0b059e0ce req-29824897-1d26-4e6d-86d6-5b8dfd680b8a service nova] Releasing lock "refresh_cache-962d3d88-03ee-4ff2-9d79-15091671b330" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 980.153712] env[61806]: DEBUG nova.compute.manager [req-a9da2378-8b8a-4564-90e1-34e0b059e0ce req-29824897-1d26-4e6d-86d6-5b8dfd680b8a service nova] [instance: e4638faa-7be1-4909-a595-c437837fa314] Received event network-changed-bfb37474-8e7a-4820-9045-e1b388cfb87d {{(pid=61806) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 980.154009] env[61806]: DEBUG nova.compute.manager [req-a9da2378-8b8a-4564-90e1-34e0b059e0ce req-29824897-1d26-4e6d-86d6-5b8dfd680b8a service nova] [instance: e4638faa-7be1-4909-a595-c437837fa314] Refreshing instance network info cache due to event network-changed-bfb37474-8e7a-4820-9045-e1b388cfb87d. {{(pid=61806) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 980.154390] env[61806]: DEBUG oslo_concurrency.lockutils [req-a9da2378-8b8a-4564-90e1-34e0b059e0ce req-29824897-1d26-4e6d-86d6-5b8dfd680b8a service nova] Acquiring lock "refresh_cache-e4638faa-7be1-4909-a595-c437837fa314" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 980.154614] env[61806]: DEBUG oslo_concurrency.lockutils [req-a9da2378-8b8a-4564-90e1-34e0b059e0ce req-29824897-1d26-4e6d-86d6-5b8dfd680b8a service nova] Acquired lock "refresh_cache-e4638faa-7be1-4909-a595-c437837fa314" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 980.154802] env[61806]: DEBUG nova.network.neutron [req-a9da2378-8b8a-4564-90e1-34e0b059e0ce req-29824897-1d26-4e6d-86d6-5b8dfd680b8a service nova] [instance: e4638faa-7be1-4909-a595-c437837fa314] Refreshing network info cache for port bfb37474-8e7a-4820-9045-e1b388cfb87d {{(pid=61806) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 980.157143] env[61806]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-2e345341-3c36-438d-80d2-043f4aa716e2 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 980.170156] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f2f7a375-6cd7-4413-8792-daed7680760e {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 980.203707] env[61806]: DEBUG nova.compute.manager [req-9c143e36-931a-4f0a-985d-4a4f9fd66438 req-c48e4d10-9396-47bf-8c03-3089c2f45b3e service nova] [instance: e146f7e5-bbf8-4adf-a2e9-b7e753438976] Detach interface failed, port_id=b68f5483-a11f-4b35-aaa0-2dcd7f50e448, reason: Instance e146f7e5-bbf8-4adf-a2e9-b7e753438976 could not be found. {{(pid=61806) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10963}} [ 980.265078] env[61806]: DEBUG oslo_vmware.api [None req-d01361ec-1270-4938-a4ab-1319441d7070 tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Task: {'id': session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]52a34fc6-78b9-e770-5d63-1151efd6d346, 'name': SearchDatastore_Task, 'duration_secs': 0.021517} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 980.265951] env[61806]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-07cd0297-2d1f-473f-b69e-a77da3ea1aea {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 980.272394] env[61806]: DEBUG oslo_vmware.api [None req-d01361ec-1270-4938-a4ab-1319441d7070 tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Waiting for the task: (returnval){ [ 980.272394] env[61806]: value = "session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]52a81562-ac21-5944-d7ac-a1ba83fa248f" [ 980.272394] env[61806]: _type = "Task" [ 980.272394] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 980.280845] env[61806]: DEBUG oslo_vmware.api [None req-d01361ec-1270-4938-a4ab-1319441d7070 tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Task: {'id': session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]52a81562-ac21-5944-d7ac-a1ba83fa248f, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 980.290472] env[61806]: DEBUG oslo_vmware.api [None req-eb9ccb65-4d10-45a5-b6b1-254652cf9691 tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Task: {'id': task-1295028, 'name': PowerOffVM_Task, 'duration_secs': 0.216226} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 980.290874] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [None req-eb9ccb65-4d10-45a5-b6b1-254652cf9691 tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] [instance: 461f98f2-80ba-4088-9e7f-1ba2a9b0c1d9] Powered off the VM {{(pid=61806) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 980.291167] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-eb9ccb65-4d10-45a5-b6b1-254652cf9691 tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] [instance: 461f98f2-80ba-4088-9e7f-1ba2a9b0c1d9] Unregistering the VM {{(pid=61806) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 980.291478] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-fd0d74ef-93c4-4538-845c-77a8e5313f87 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 980.347105] env[61806]: DEBUG nova.network.neutron [None req-a06ee197-a4db-47dc-a9f0-aee2aa23de70 tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] [instance: 9357e61f-1628-43bd-ab46-de13c1529f51] Instance cache missing network info. {{(pid=61806) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 980.381933] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-eb9ccb65-4d10-45a5-b6b1-254652cf9691 tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] [instance: 461f98f2-80ba-4088-9e7f-1ba2a9b0c1d9] Unregistered the VM {{(pid=61806) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 980.382478] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-eb9ccb65-4d10-45a5-b6b1-254652cf9691 tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] [instance: 461f98f2-80ba-4088-9e7f-1ba2a9b0c1d9] Deleting contents of the VM from datastore datastore1 {{(pid=61806) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 980.382770] env[61806]: DEBUG nova.virt.vmwareapi.ds_util [None req-eb9ccb65-4d10-45a5-b6b1-254652cf9691 tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Deleting the datastore file [datastore1] 461f98f2-80ba-4088-9e7f-1ba2a9b0c1d9 {{(pid=61806) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 980.383198] env[61806]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-532b4f5b-a292-4286-9216-9cf404f11092 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 980.395089] env[61806]: DEBUG oslo_vmware.api [None req-eb9ccb65-4d10-45a5-b6b1-254652cf9691 tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Waiting for the task: (returnval){ [ 980.395089] env[61806]: value = "task-1295031" [ 980.395089] env[61806]: _type = "Task" [ 980.395089] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 980.407123] env[61806]: DEBUG oslo_vmware.api [None req-eb9ccb65-4d10-45a5-b6b1-254652cf9691 tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Task: {'id': task-1295031, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 980.552876] env[61806]: DEBUG nova.network.neutron [None req-a06ee197-a4db-47dc-a9f0-aee2aa23de70 tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] [instance: 9357e61f-1628-43bd-ab46-de13c1529f51] Updating instance_info_cache with network_info: [{"id": "a4ca7c56-b6a1-4bd2-bd47-02b0ed4aa097", "address": "fa:16:3e:66:09:0d", "network": {"id": "46515a4e-610f-42fb-a424-205a03df0640", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-1133623547-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4b9d94c3d684474b99e305fb6cd9b95f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "82ca17df-257e-40e6-9ec9-310ed6f05ccb", "external-id": "nsx-vlan-transportzone-778", "segmentation_id": 778, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa4ca7c56-b6", "ovs_interfaceid": "a4ca7c56-b6a1-4bd2-bd47-02b0ed4aa097", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61806) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 980.612073] env[61806]: DEBUG oslo_concurrency.lockutils [None req-38f9dcbc-e34b-458d-8ba6-e6b3c0e6b471 tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] Lock "0bac3048-759a-47d9-a6ef-a0386a49a974" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 8.363s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 980.623218] env[61806]: DEBUG oslo_vmware.api [None req-b6079056-6464-46e6-97ab-343b76917c18 tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] Task: {'id': task-1295029, 'name': CloneVM_Task} progress is 94%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 980.625234] env[61806]: INFO nova.compute.manager [-] [instance: e146f7e5-bbf8-4adf-a2e9-b7e753438976] Took 1.41 seconds to deallocate network for instance. [ 980.781229] env[61806]: DEBUG nova.compute.manager [req-b412e6c4-fe0f-4d20-b005-9ffd6f04d34a req-5a669e7b-4ef6-497e-bbcc-cb3f7fb924e9 service nova] [instance: 9357e61f-1628-43bd-ab46-de13c1529f51] Received event network-vif-plugged-a4ca7c56-b6a1-4bd2-bd47-02b0ed4aa097 {{(pid=61806) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 980.781484] env[61806]: DEBUG oslo_concurrency.lockutils [req-b412e6c4-fe0f-4d20-b005-9ffd6f04d34a req-5a669e7b-4ef6-497e-bbcc-cb3f7fb924e9 service nova] Acquiring lock "9357e61f-1628-43bd-ab46-de13c1529f51-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 980.781720] env[61806]: DEBUG oslo_concurrency.lockutils [req-b412e6c4-fe0f-4d20-b005-9ffd6f04d34a req-5a669e7b-4ef6-497e-bbcc-cb3f7fb924e9 service nova] Lock "9357e61f-1628-43bd-ab46-de13c1529f51-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 980.781915] env[61806]: DEBUG oslo_concurrency.lockutils [req-b412e6c4-fe0f-4d20-b005-9ffd6f04d34a req-5a669e7b-4ef6-497e-bbcc-cb3f7fb924e9 service nova] Lock "9357e61f-1628-43bd-ab46-de13c1529f51-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 980.782211] env[61806]: DEBUG nova.compute.manager [req-b412e6c4-fe0f-4d20-b005-9ffd6f04d34a req-5a669e7b-4ef6-497e-bbcc-cb3f7fb924e9 service nova] [instance: 9357e61f-1628-43bd-ab46-de13c1529f51] No waiting events found dispatching network-vif-plugged-a4ca7c56-b6a1-4bd2-bd47-02b0ed4aa097 {{(pid=61806) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 980.782476] env[61806]: WARNING nova.compute.manager [req-b412e6c4-fe0f-4d20-b005-9ffd6f04d34a req-5a669e7b-4ef6-497e-bbcc-cb3f7fb924e9 service nova] [instance: 9357e61f-1628-43bd-ab46-de13c1529f51] Received unexpected event network-vif-plugged-a4ca7c56-b6a1-4bd2-bd47-02b0ed4aa097 for instance with vm_state building and task_state spawning. [ 980.782675] env[61806]: DEBUG nova.compute.manager [req-b412e6c4-fe0f-4d20-b005-9ffd6f04d34a req-5a669e7b-4ef6-497e-bbcc-cb3f7fb924e9 service nova] [instance: 9357e61f-1628-43bd-ab46-de13c1529f51] Received event network-changed-a4ca7c56-b6a1-4bd2-bd47-02b0ed4aa097 {{(pid=61806) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 980.782862] env[61806]: DEBUG nova.compute.manager [req-b412e6c4-fe0f-4d20-b005-9ffd6f04d34a req-5a669e7b-4ef6-497e-bbcc-cb3f7fb924e9 service nova] [instance: 9357e61f-1628-43bd-ab46-de13c1529f51] Refreshing instance network info cache due to event network-changed-a4ca7c56-b6a1-4bd2-bd47-02b0ed4aa097. {{(pid=61806) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 980.783064] env[61806]: DEBUG oslo_concurrency.lockutils [req-b412e6c4-fe0f-4d20-b005-9ffd6f04d34a req-5a669e7b-4ef6-497e-bbcc-cb3f7fb924e9 service nova] Acquiring lock "refresh_cache-9357e61f-1628-43bd-ab46-de13c1529f51" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 980.790573] env[61806]: DEBUG oslo_vmware.api [None req-d01361ec-1270-4938-a4ab-1319441d7070 tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Task: {'id': session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]52a81562-ac21-5944-d7ac-a1ba83fa248f, 'name': SearchDatastore_Task, 'duration_secs': 0.01813} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 980.792069] env[61806]: DEBUG oslo_concurrency.lockutils [None req-d01361ec-1270-4938-a4ab-1319441d7070 tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Releasing lock "[datastore1] devstack-image-cache_base/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa.vmdk" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 980.792069] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [None req-d01361ec-1270-4938-a4ab-1319441d7070 tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa.vmdk to [datastore1] 962d3d88-03ee-4ff2-9d79-15091671b330/962d3d88-03ee-4ff2-9d79-15091671b330.vmdk {{(pid=61806) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 980.792069] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-9caa770e-e89c-4659-89e7-4d3476adea8d {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 980.800426] env[61806]: DEBUG oslo_vmware.api [None req-d01361ec-1270-4938-a4ab-1319441d7070 tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Waiting for the task: (returnval){ [ 980.800426] env[61806]: value = "task-1295032" [ 980.800426] env[61806]: _type = "Task" [ 980.800426] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 980.810263] env[61806]: DEBUG oslo_vmware.api [None req-d01361ec-1270-4938-a4ab-1319441d7070 tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Task: {'id': task-1295032, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 980.885392] env[61806]: DEBUG nova.network.neutron [req-a9da2378-8b8a-4564-90e1-34e0b059e0ce req-29824897-1d26-4e6d-86d6-5b8dfd680b8a service nova] [instance: e4638faa-7be1-4909-a595-c437837fa314] Updated VIF entry in instance network info cache for port bfb37474-8e7a-4820-9045-e1b388cfb87d. {{(pid=61806) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 980.885964] env[61806]: DEBUG nova.network.neutron [req-a9da2378-8b8a-4564-90e1-34e0b059e0ce req-29824897-1d26-4e6d-86d6-5b8dfd680b8a service nova] [instance: e4638faa-7be1-4909-a595-c437837fa314] Updating instance_info_cache with network_info: [{"id": "bfb37474-8e7a-4820-9045-e1b388cfb87d", "address": "fa:16:3e:b5:75:56", "network": {"id": "d636ca55-0737-48e4-8d9d-7fe0f797f49e", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-2090334349-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.158", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "22e47882c2c44ce492dff6b1c5d782e2", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6f493cd8-1cb4-42a1-8d56-bfa6ac7cf563", "external-id": "nsx-vlan-transportzone-931", "segmentation_id": 931, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapbfb37474-8e", "ovs_interfaceid": "bfb37474-8e7a-4820-9045-e1b388cfb87d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61806) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 980.909624] env[61806]: DEBUG oslo_vmware.api [None req-eb9ccb65-4d10-45a5-b6b1-254652cf9691 tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Task: {'id': task-1295031, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 981.033301] env[61806]: DEBUG oslo_concurrency.lockutils [None req-7fff0eac-b278-41ae-86ad-6a3d9bc09f29 tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] Acquiring lock "38e96e22-4200-4175-9085-31977357da2d" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 981.033567] env[61806]: DEBUG oslo_concurrency.lockutils [None req-7fff0eac-b278-41ae-86ad-6a3d9bc09f29 tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] Lock "38e96e22-4200-4175-9085-31977357da2d" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 981.033815] env[61806]: DEBUG oslo_concurrency.lockutils [None req-7fff0eac-b278-41ae-86ad-6a3d9bc09f29 tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] Acquiring lock "38e96e22-4200-4175-9085-31977357da2d-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 981.034042] env[61806]: DEBUG oslo_concurrency.lockutils [None req-7fff0eac-b278-41ae-86ad-6a3d9bc09f29 tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] Lock "38e96e22-4200-4175-9085-31977357da2d-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 981.034235] env[61806]: DEBUG oslo_concurrency.lockutils [None req-7fff0eac-b278-41ae-86ad-6a3d9bc09f29 tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] Lock "38e96e22-4200-4175-9085-31977357da2d-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 981.036697] env[61806]: INFO nova.compute.manager [None req-7fff0eac-b278-41ae-86ad-6a3d9bc09f29 tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] [instance: 38e96e22-4200-4175-9085-31977357da2d] Terminating instance [ 981.038794] env[61806]: DEBUG nova.compute.manager [None req-7fff0eac-b278-41ae-86ad-6a3d9bc09f29 tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] [instance: 38e96e22-4200-4175-9085-31977357da2d] Start destroying the instance on the hypervisor. {{(pid=61806) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 981.039030] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-7fff0eac-b278-41ae-86ad-6a3d9bc09f29 tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] [instance: 38e96e22-4200-4175-9085-31977357da2d] Destroying instance {{(pid=61806) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 981.039917] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a8bc0638-902d-4892-bdaf-ae4cc8c57807 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 981.051345] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [None req-7fff0eac-b278-41ae-86ad-6a3d9bc09f29 tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] [instance: 38e96e22-4200-4175-9085-31977357da2d] Powering off the VM {{(pid=61806) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 981.051675] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-a52a8e70-b982-4ae0-9805-72337e4f3655 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 981.055167] env[61806]: DEBUG oslo_concurrency.lockutils [None req-a06ee197-a4db-47dc-a9f0-aee2aa23de70 tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Releasing lock "refresh_cache-9357e61f-1628-43bd-ab46-de13c1529f51" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 981.055540] env[61806]: DEBUG nova.compute.manager [None req-a06ee197-a4db-47dc-a9f0-aee2aa23de70 tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] [instance: 9357e61f-1628-43bd-ab46-de13c1529f51] Instance network_info: |[{"id": "a4ca7c56-b6a1-4bd2-bd47-02b0ed4aa097", "address": "fa:16:3e:66:09:0d", "network": {"id": "46515a4e-610f-42fb-a424-205a03df0640", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-1133623547-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4b9d94c3d684474b99e305fb6cd9b95f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "82ca17df-257e-40e6-9ec9-310ed6f05ccb", "external-id": "nsx-vlan-transportzone-778", "segmentation_id": 778, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa4ca7c56-b6", "ovs_interfaceid": "a4ca7c56-b6a1-4bd2-bd47-02b0ed4aa097", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61806) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 981.055905] env[61806]: DEBUG oslo_concurrency.lockutils [req-b412e6c4-fe0f-4d20-b005-9ffd6f04d34a req-5a669e7b-4ef6-497e-bbcc-cb3f7fb924e9 service nova] Acquired lock "refresh_cache-9357e61f-1628-43bd-ab46-de13c1529f51" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 981.056203] env[61806]: DEBUG nova.network.neutron [req-b412e6c4-fe0f-4d20-b005-9ffd6f04d34a req-5a669e7b-4ef6-497e-bbcc-cb3f7fb924e9 service nova] [instance: 9357e61f-1628-43bd-ab46-de13c1529f51] Refreshing network info cache for port a4ca7c56-b6a1-4bd2-bd47-02b0ed4aa097 {{(pid=61806) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 981.057607] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-a06ee197-a4db-47dc-a9f0-aee2aa23de70 tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] [instance: 9357e61f-1628-43bd-ab46-de13c1529f51] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:66:09:0d', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '82ca17df-257e-40e6-9ec9-310ed6f05ccb', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'a4ca7c56-b6a1-4bd2-bd47-02b0ed4aa097', 'vif_model': 'vmxnet3'}] {{(pid=61806) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 981.065989] env[61806]: DEBUG oslo.service.loopingcall [None req-a06ee197-a4db-47dc-a9f0-aee2aa23de70 tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61806) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 981.070968] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 9357e61f-1628-43bd-ab46-de13c1529f51] Creating VM on the ESX host {{(pid=61806) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 981.071393] env[61806]: DEBUG oslo_vmware.api [None req-7fff0eac-b278-41ae-86ad-6a3d9bc09f29 tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] Waiting for the task: (returnval){ [ 981.071393] env[61806]: value = "task-1295033" [ 981.071393] env[61806]: _type = "Task" [ 981.071393] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 981.072040] env[61806]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-2e6a5a7d-ea66-4b91-a3f5-f188b1cc8695 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 981.092847] env[61806]: DEBUG oslo_concurrency.lockutils [None req-1e11ef2b-e661-483a-b1ad-9cb57e29c94f tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: held 1.014s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 981.094774] env[61806]: DEBUG oslo_concurrency.lockutils [None req-57da61d4-139b-46a7-9163-13f9c5735119 tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 2.665s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 981.095366] env[61806]: DEBUG nova.objects.instance [None req-57da61d4-139b-46a7-9163-13f9c5735119 tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] Lazy-loading 'resources' on Instance uuid d9caee4a-a00c-4e50-b01c-99a8f796a09f {{(pid=61806) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 981.104210] env[61806]: DEBUG oslo_vmware.api [None req-7fff0eac-b278-41ae-86ad-6a3d9bc09f29 tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] Task: {'id': task-1295033, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 981.106196] env[61806]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 981.106196] env[61806]: value = "task-1295034" [ 981.106196] env[61806]: _type = "Task" [ 981.106196] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 981.123455] env[61806]: DEBUG oslo_vmware.api [-] Task: {'id': task-1295034, 'name': CreateVM_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 981.130676] env[61806]: DEBUG oslo_vmware.api [None req-b6079056-6464-46e6-97ab-343b76917c18 tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] Task: {'id': task-1295029, 'name': CloneVM_Task} progress is 95%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 981.131710] env[61806]: DEBUG oslo_concurrency.lockutils [None req-474c365d-f281-4fe7-b609-9d0c04e06393 tempest-ServersNegativeTestJSON-891856253 tempest-ServersNegativeTestJSON-891856253-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 981.316107] env[61806]: DEBUG oslo_vmware.api [None req-d01361ec-1270-4938-a4ab-1319441d7070 tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Task: {'id': task-1295032, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 981.370041] env[61806]: DEBUG nova.network.neutron [req-b412e6c4-fe0f-4d20-b005-9ffd6f04d34a req-5a669e7b-4ef6-497e-bbcc-cb3f7fb924e9 service nova] [instance: 9357e61f-1628-43bd-ab46-de13c1529f51] Updated VIF entry in instance network info cache for port a4ca7c56-b6a1-4bd2-bd47-02b0ed4aa097. {{(pid=61806) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 981.370541] env[61806]: DEBUG nova.network.neutron [req-b412e6c4-fe0f-4d20-b005-9ffd6f04d34a req-5a669e7b-4ef6-497e-bbcc-cb3f7fb924e9 service nova] [instance: 9357e61f-1628-43bd-ab46-de13c1529f51] Updating instance_info_cache with network_info: [{"id": "a4ca7c56-b6a1-4bd2-bd47-02b0ed4aa097", "address": "fa:16:3e:66:09:0d", "network": {"id": "46515a4e-610f-42fb-a424-205a03df0640", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-1133623547-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4b9d94c3d684474b99e305fb6cd9b95f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "82ca17df-257e-40e6-9ec9-310ed6f05ccb", "external-id": "nsx-vlan-transportzone-778", "segmentation_id": 778, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa4ca7c56-b6", "ovs_interfaceid": "a4ca7c56-b6a1-4bd2-bd47-02b0ed4aa097", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61806) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 981.388562] env[61806]: DEBUG oslo_concurrency.lockutils [req-a9da2378-8b8a-4564-90e1-34e0b059e0ce req-29824897-1d26-4e6d-86d6-5b8dfd680b8a service nova] Releasing lock "refresh_cache-e4638faa-7be1-4909-a595-c437837fa314" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 981.407137] env[61806]: DEBUG oslo_vmware.api [None req-eb9ccb65-4d10-45a5-b6b1-254652cf9691 tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Task: {'id': task-1295031, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.533929} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 981.407513] env[61806]: DEBUG nova.virt.vmwareapi.ds_util [None req-eb9ccb65-4d10-45a5-b6b1-254652cf9691 tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Deleted the datastore file {{(pid=61806) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 981.407790] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-eb9ccb65-4d10-45a5-b6b1-254652cf9691 tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] [instance: 461f98f2-80ba-4088-9e7f-1ba2a9b0c1d9] Deleted contents of the VM from datastore datastore1 {{(pid=61806) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 981.408056] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-eb9ccb65-4d10-45a5-b6b1-254652cf9691 tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] [instance: 461f98f2-80ba-4088-9e7f-1ba2a9b0c1d9] Instance destroyed {{(pid=61806) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 981.408314] env[61806]: INFO nova.compute.manager [None req-eb9ccb65-4d10-45a5-b6b1-254652cf9691 tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] [instance: 461f98f2-80ba-4088-9e7f-1ba2a9b0c1d9] Took 1.65 seconds to destroy the instance on the hypervisor. [ 981.408650] env[61806]: DEBUG oslo.service.loopingcall [None req-eb9ccb65-4d10-45a5-b6b1-254652cf9691 tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61806) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 981.408905] env[61806]: DEBUG nova.compute.manager [-] [instance: 461f98f2-80ba-4088-9e7f-1ba2a9b0c1d9] Deallocating network for instance {{(pid=61806) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 981.409099] env[61806]: DEBUG nova.network.neutron [-] [instance: 461f98f2-80ba-4088-9e7f-1ba2a9b0c1d9] deallocate_for_instance() {{(pid=61806) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 981.601534] env[61806]: DEBUG oslo_vmware.api [None req-7fff0eac-b278-41ae-86ad-6a3d9bc09f29 tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] Task: {'id': task-1295033, 'name': PowerOffVM_Task, 'duration_secs': 0.420413} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 981.602108] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [None req-7fff0eac-b278-41ae-86ad-6a3d9bc09f29 tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] [instance: 38e96e22-4200-4175-9085-31977357da2d] Powered off the VM {{(pid=61806) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 981.602330] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-7fff0eac-b278-41ae-86ad-6a3d9bc09f29 tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] [instance: 38e96e22-4200-4175-9085-31977357da2d] Unregistering the VM {{(pid=61806) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 981.602751] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-6bb3d975-963a-43cd-91ff-ea307a75b340 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 981.617232] env[61806]: DEBUG oslo_vmware.api [-] Task: {'id': task-1295034, 'name': CreateVM_Task} progress is 25%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 981.626885] env[61806]: DEBUG oslo_vmware.api [None req-b6079056-6464-46e6-97ab-343b76917c18 tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] Task: {'id': task-1295029, 'name': CloneVM_Task, 'duration_secs': 1.342401} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 981.630806] env[61806]: INFO nova.virt.vmwareapi.vmops [None req-b6079056-6464-46e6-97ab-343b76917c18 tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] [instance: fd9e520a-2e7b-43d4-b84d-8933c32cc6a6] Created linked-clone VM from snapshot [ 981.631553] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-45569798-53ac-43a1-bee0-e1960cc1d371 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 981.641333] env[61806]: DEBUG nova.virt.vmwareapi.images [None req-b6079056-6464-46e6-97ab-343b76917c18 tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] [instance: fd9e520a-2e7b-43d4-b84d-8933c32cc6a6] Uploading image b08c43d9-cfd5-4a1a-b99d-0aebb718b592 {{(pid=61806) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:442}} [ 981.672760] env[61806]: DEBUG oslo_vmware.rw_handles [None req-b6079056-6464-46e6-97ab-343b76917c18 tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] Creating HttpNfcLease lease for exporting VM: (result){ [ 981.672760] env[61806]: value = "vm-277734" [ 981.672760] env[61806]: _type = "VirtualMachine" [ 981.672760] env[61806]: }. {{(pid=61806) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:478}} [ 981.673158] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ExportVm with opID=oslo.vmware-67bd9322-3896-4d03-8822-e7012b2d4b1b {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 981.679850] env[61806]: DEBUG nova.compute.manager [req-fda1a621-b410-4c9c-873f-8b20d09d37c2 req-79aaa05f-5556-4c8c-af1f-7149daa6607f service nova] [instance: 461f98f2-80ba-4088-9e7f-1ba2a9b0c1d9] Received event network-vif-deleted-91000a2c-19c5-404d-90b3-367698085aec {{(pid=61806) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 981.680164] env[61806]: INFO nova.compute.manager [req-fda1a621-b410-4c9c-873f-8b20d09d37c2 req-79aaa05f-5556-4c8c-af1f-7149daa6607f service nova] [instance: 461f98f2-80ba-4088-9e7f-1ba2a9b0c1d9] Neutron deleted interface 91000a2c-19c5-404d-90b3-367698085aec; detaching it from the instance and deleting it from the info cache [ 981.680290] env[61806]: DEBUG nova.network.neutron [req-fda1a621-b410-4c9c-873f-8b20d09d37c2 req-79aaa05f-5556-4c8c-af1f-7149daa6607f service nova] [instance: 461f98f2-80ba-4088-9e7f-1ba2a9b0c1d9] Updating instance_info_cache with network_info: [] {{(pid=61806) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 981.683230] env[61806]: DEBUG oslo_vmware.rw_handles [None req-b6079056-6464-46e6-97ab-343b76917c18 tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] Lease: (returnval){ [ 981.683230] env[61806]: value = "session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]52fcfa07-66a4-ac99-991d-0e1ef2d08ce0" [ 981.683230] env[61806]: _type = "HttpNfcLease" [ 981.683230] env[61806]: } obtained for exporting VM: (result){ [ 981.683230] env[61806]: value = "vm-277734" [ 981.683230] env[61806]: _type = "VirtualMachine" [ 981.683230] env[61806]: }. {{(pid=61806) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:481}} [ 981.683605] env[61806]: DEBUG oslo_vmware.api [None req-b6079056-6464-46e6-97ab-343b76917c18 tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] Waiting for the lease: (returnval){ [ 981.683605] env[61806]: value = "session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]52fcfa07-66a4-ac99-991d-0e1ef2d08ce0" [ 981.683605] env[61806]: _type = "HttpNfcLease" [ 981.683605] env[61806]: } to be ready. {{(pid=61806) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 981.693941] env[61806]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 981.693941] env[61806]: value = "session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]52fcfa07-66a4-ac99-991d-0e1ef2d08ce0" [ 981.693941] env[61806]: _type = "HttpNfcLease" [ 981.693941] env[61806]: } is ready. {{(pid=61806) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 981.694932] env[61806]: DEBUG oslo_vmware.rw_handles [None req-b6079056-6464-46e6-97ab-343b76917c18 tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 981.694932] env[61806]: value = "session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]52fcfa07-66a4-ac99-991d-0e1ef2d08ce0" [ 981.694932] env[61806]: _type = "HttpNfcLease" [ 981.694932] env[61806]: }. {{(pid=61806) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:486}} [ 981.695796] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c085502f-01f7-4828-99fa-399750187ebb {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 981.709170] env[61806]: DEBUG oslo_vmware.rw_handles [None req-b6079056-6464-46e6-97ab-343b76917c18 tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] Found VMDK URL: https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52335490-af02-ad6e-9173-20b0caccf108/disk-0.vmdk from lease info. {{(pid=61806) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 981.709402] env[61806]: DEBUG oslo_vmware.rw_handles [None req-b6079056-6464-46e6-97ab-343b76917c18 tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] Opening URL: https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52335490-af02-ad6e-9173-20b0caccf108/disk-0.vmdk for reading. {{(pid=61806) _create_read_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:99}} [ 981.710758] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-7fff0eac-b278-41ae-86ad-6a3d9bc09f29 tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] [instance: 38e96e22-4200-4175-9085-31977357da2d] Unregistered the VM {{(pid=61806) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 981.710984] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-7fff0eac-b278-41ae-86ad-6a3d9bc09f29 tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] [instance: 38e96e22-4200-4175-9085-31977357da2d] Deleting contents of the VM from datastore datastore1 {{(pid=61806) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 981.711200] env[61806]: DEBUG nova.virt.vmwareapi.ds_util [None req-7fff0eac-b278-41ae-86ad-6a3d9bc09f29 tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] Deleting the datastore file [datastore1] 38e96e22-4200-4175-9085-31977357da2d {{(pid=61806) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 981.712239] env[61806]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-24e2c5ef-58fe-4c8f-94c9-4057023fdefc {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 981.777657] env[61806]: DEBUG oslo_vmware.api [None req-7fff0eac-b278-41ae-86ad-6a3d9bc09f29 tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] Waiting for the task: (returnval){ [ 981.777657] env[61806]: value = "task-1295037" [ 981.777657] env[61806]: _type = "Task" [ 981.777657] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 981.791738] env[61806]: DEBUG oslo_vmware.api [None req-7fff0eac-b278-41ae-86ad-6a3d9bc09f29 tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] Task: {'id': task-1295037, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 981.803941] env[61806]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-5f05cd10-ba92-4f39-bce6-a41d52dd6e7a {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 981.817882] env[61806]: DEBUG oslo_vmware.api [None req-d01361ec-1270-4938-a4ab-1319441d7070 tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Task: {'id': task-1295032, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.545973} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 981.819824] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [None req-d01361ec-1270-4938-a4ab-1319441d7070 tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa.vmdk to [datastore1] 962d3d88-03ee-4ff2-9d79-15091671b330/962d3d88-03ee-4ff2-9d79-15091671b330.vmdk {{(pid=61806) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 981.820089] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-d01361ec-1270-4938-a4ab-1319441d7070 tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] [instance: 962d3d88-03ee-4ff2-9d79-15091671b330] Extending root virtual disk to 1048576 {{(pid=61806) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 981.820843] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-9cf34f33-8c78-49b2-8013-6b642c77edcf {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 981.832429] env[61806]: DEBUG oslo_vmware.api [None req-d01361ec-1270-4938-a4ab-1319441d7070 tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Waiting for the task: (returnval){ [ 981.832429] env[61806]: value = "task-1295038" [ 981.832429] env[61806]: _type = "Task" [ 981.832429] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 981.848402] env[61806]: DEBUG oslo_vmware.api [None req-d01361ec-1270-4938-a4ab-1319441d7070 tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Task: {'id': task-1295038, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 981.863048] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fc42d7f3-2adc-4f86-b2f8-fa23e2b66fb0 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 981.870955] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f5dfb969-d7ac-4e3b-9fc5-18595b5b2779 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 981.874799] env[61806]: DEBUG oslo_concurrency.lockutils [req-b412e6c4-fe0f-4d20-b005-9ffd6f04d34a req-5a669e7b-4ef6-497e-bbcc-cb3f7fb924e9 service nova] Releasing lock "refresh_cache-9357e61f-1628-43bd-ab46-de13c1529f51" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 981.904679] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8a29de10-c087-4bae-930d-4c1e1c96c922 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 981.915618] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ac189a02-fd87-4e7d-8546-787d741c98cc {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 981.929636] env[61806]: DEBUG nova.compute.provider_tree [None req-57da61d4-139b-46a7-9163-13f9c5735119 tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] Inventory has not changed in ProviderTree for provider: a2858be1-fd22-4e08-979e-87ad25293407 {{(pid=61806) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 982.117976] env[61806]: DEBUG oslo_vmware.api [-] Task: {'id': task-1295034, 'name': CreateVM_Task, 'duration_secs': 0.693151} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 982.118408] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 9357e61f-1628-43bd-ab46-de13c1529f51] Created VM on the ESX host {{(pid=61806) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 982.119057] env[61806]: DEBUG oslo_concurrency.lockutils [None req-a06ee197-a4db-47dc-a9f0-aee2aa23de70 tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 982.119057] env[61806]: DEBUG oslo_concurrency.lockutils [None req-a06ee197-a4db-47dc-a9f0-aee2aa23de70 tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Acquired lock "[datastore1] devstack-image-cache_base/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 982.119407] env[61806]: DEBUG oslo_concurrency.lockutils [None req-a06ee197-a4db-47dc-a9f0-aee2aa23de70 tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 982.119630] env[61806]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-845a24f6-b835-4211-9852-2ee3f963374c {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 982.126506] env[61806]: DEBUG oslo_vmware.api [None req-a06ee197-a4db-47dc-a9f0-aee2aa23de70 tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Waiting for the task: (returnval){ [ 982.126506] env[61806]: value = "session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]52eca845-224c-4c6d-05ca-7410737d1f7d" [ 982.126506] env[61806]: _type = "Task" [ 982.126506] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 982.137962] env[61806]: DEBUG oslo_vmware.api [None req-a06ee197-a4db-47dc-a9f0-aee2aa23de70 tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Task: {'id': session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]52eca845-224c-4c6d-05ca-7410737d1f7d, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 982.145064] env[61806]: DEBUG nova.network.neutron [-] [instance: 461f98f2-80ba-4088-9e7f-1ba2a9b0c1d9] Updating instance_info_cache with network_info: [] {{(pid=61806) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 982.185111] env[61806]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-164f3ef6-61be-4171-bb7a-150ae8cfc497 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 982.196704] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-362e6269-351c-4994-a2e9-a1794291289f {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 982.230704] env[61806]: DEBUG nova.compute.manager [req-fda1a621-b410-4c9c-873f-8b20d09d37c2 req-79aaa05f-5556-4c8c-af1f-7149daa6607f service nova] [instance: 461f98f2-80ba-4088-9e7f-1ba2a9b0c1d9] Detach interface failed, port_id=91000a2c-19c5-404d-90b3-367698085aec, reason: Instance 461f98f2-80ba-4088-9e7f-1ba2a9b0c1d9 could not be found. {{(pid=61806) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10963}} [ 982.288394] env[61806]: DEBUG oslo_vmware.api [None req-7fff0eac-b278-41ae-86ad-6a3d9bc09f29 tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] Task: {'id': task-1295037, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.455285} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 982.288612] env[61806]: DEBUG nova.virt.vmwareapi.ds_util [None req-7fff0eac-b278-41ae-86ad-6a3d9bc09f29 tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] Deleted the datastore file {{(pid=61806) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 982.288707] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-7fff0eac-b278-41ae-86ad-6a3d9bc09f29 tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] [instance: 38e96e22-4200-4175-9085-31977357da2d] Deleted contents of the VM from datastore datastore1 {{(pid=61806) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 982.288894] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-7fff0eac-b278-41ae-86ad-6a3d9bc09f29 tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] [instance: 38e96e22-4200-4175-9085-31977357da2d] Instance destroyed {{(pid=61806) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 982.289091] env[61806]: INFO nova.compute.manager [None req-7fff0eac-b278-41ae-86ad-6a3d9bc09f29 tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] [instance: 38e96e22-4200-4175-9085-31977357da2d] Took 1.25 seconds to destroy the instance on the hypervisor. [ 982.289871] env[61806]: DEBUG oslo.service.loopingcall [None req-7fff0eac-b278-41ae-86ad-6a3d9bc09f29 tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61806) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 982.289871] env[61806]: DEBUG nova.compute.manager [-] [instance: 38e96e22-4200-4175-9085-31977357da2d] Deallocating network for instance {{(pid=61806) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 982.289871] env[61806]: DEBUG nova.network.neutron [-] [instance: 38e96e22-4200-4175-9085-31977357da2d] deallocate_for_instance() {{(pid=61806) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 982.344394] env[61806]: DEBUG oslo_vmware.api [None req-d01361ec-1270-4938-a4ab-1319441d7070 tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Task: {'id': task-1295038, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.086501} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 982.344823] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-d01361ec-1270-4938-a4ab-1319441d7070 tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] [instance: 962d3d88-03ee-4ff2-9d79-15091671b330] Extended root virtual disk {{(pid=61806) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 982.345803] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e9e63b3c-1bcf-41e3-a784-2a588108c3a4 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 982.379983] env[61806]: DEBUG nova.virt.vmwareapi.volumeops [None req-d01361ec-1270-4938-a4ab-1319441d7070 tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] [instance: 962d3d88-03ee-4ff2-9d79-15091671b330] Reconfiguring VM instance instance-0000005f to attach disk [datastore1] 962d3d88-03ee-4ff2-9d79-15091671b330/962d3d88-03ee-4ff2-9d79-15091671b330.vmdk or device None with type sparse {{(pid=61806) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 982.382973] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-004801da-805b-43e3-a9b4-335af6826086 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 982.405940] env[61806]: DEBUG oslo_vmware.api [None req-d01361ec-1270-4938-a4ab-1319441d7070 tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Waiting for the task: (returnval){ [ 982.405940] env[61806]: value = "task-1295039" [ 982.405940] env[61806]: _type = "Task" [ 982.405940] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 982.415966] env[61806]: DEBUG oslo_vmware.api [None req-d01361ec-1270-4938-a4ab-1319441d7070 tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Task: {'id': task-1295039, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 982.434806] env[61806]: DEBUG nova.scheduler.client.report [None req-57da61d4-139b-46a7-9163-13f9c5735119 tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] Inventory has not changed for provider a2858be1-fd22-4e08-979e-87ad25293407 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 138, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61806) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 982.639303] env[61806]: DEBUG oslo_vmware.api [None req-a06ee197-a4db-47dc-a9f0-aee2aa23de70 tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Task: {'id': session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]52eca845-224c-4c6d-05ca-7410737d1f7d, 'name': SearchDatastore_Task, 'duration_secs': 0.021819} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 982.639766] env[61806]: DEBUG oslo_concurrency.lockutils [None req-a06ee197-a4db-47dc-a9f0-aee2aa23de70 tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Releasing lock "[datastore1] devstack-image-cache_base/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 982.640083] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-a06ee197-a4db-47dc-a9f0-aee2aa23de70 tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] [instance: 9357e61f-1628-43bd-ab46-de13c1529f51] Processing image ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa {{(pid=61806) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 982.640800] env[61806]: DEBUG oslo_concurrency.lockutils [None req-a06ee197-a4db-47dc-a9f0-aee2aa23de70 tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa.vmdk" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 982.640800] env[61806]: DEBUG oslo_concurrency.lockutils [None req-a06ee197-a4db-47dc-a9f0-aee2aa23de70 tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Acquired lock "[datastore1] devstack-image-cache_base/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa.vmdk" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 982.640800] env[61806]: DEBUG nova.virt.vmwareapi.ds_util [None req-a06ee197-a4db-47dc-a9f0-aee2aa23de70 tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61806) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 982.641085] env[61806]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-692540d3-12e6-4e63-bcce-dce63f5a3d79 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 982.647356] env[61806]: INFO nova.compute.manager [-] [instance: 461f98f2-80ba-4088-9e7f-1ba2a9b0c1d9] Took 1.24 seconds to deallocate network for instance. [ 982.653882] env[61806]: DEBUG nova.virt.vmwareapi.ds_util [None req-a06ee197-a4db-47dc-a9f0-aee2aa23de70 tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61806) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 982.654192] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-a06ee197-a4db-47dc-a9f0-aee2aa23de70 tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61806) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 982.655054] env[61806]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-fe305ac2-ed41-4342-bf7f-3c57e4e85d77 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 982.661178] env[61806]: DEBUG oslo_vmware.api [None req-a06ee197-a4db-47dc-a9f0-aee2aa23de70 tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Waiting for the task: (returnval){ [ 982.661178] env[61806]: value = "session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]52d1f2bd-e69f-8ada-6cb7-8f6c7900e348" [ 982.661178] env[61806]: _type = "Task" [ 982.661178] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 982.671675] env[61806]: DEBUG oslo_vmware.api [None req-a06ee197-a4db-47dc-a9f0-aee2aa23de70 tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Task: {'id': session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]52d1f2bd-e69f-8ada-6cb7-8f6c7900e348, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 982.919246] env[61806]: DEBUG oslo_vmware.api [None req-d01361ec-1270-4938-a4ab-1319441d7070 tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Task: {'id': task-1295039, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 982.942651] env[61806]: DEBUG oslo_concurrency.lockutils [None req-57da61d4-139b-46a7-9163-13f9c5735119 tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.848s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 982.946082] env[61806]: DEBUG oslo_concurrency.lockutils [None req-474c365d-f281-4fe7-b609-9d0c04e06393 tempest-ServersNegativeTestJSON-891856253 tempest-ServersNegativeTestJSON-891856253-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 1.814s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 982.946543] env[61806]: DEBUG nova.objects.instance [None req-474c365d-f281-4fe7-b609-9d0c04e06393 tempest-ServersNegativeTestJSON-891856253 tempest-ServersNegativeTestJSON-891856253-project-member] Lazy-loading 'resources' on Instance uuid e146f7e5-bbf8-4adf-a2e9-b7e753438976 {{(pid=61806) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 982.968431] env[61806]: INFO nova.scheduler.client.report [None req-57da61d4-139b-46a7-9163-13f9c5735119 tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] Deleted allocations for instance d9caee4a-a00c-4e50-b01c-99a8f796a09f [ 983.119306] env[61806]: DEBUG nova.network.neutron [-] [instance: 38e96e22-4200-4175-9085-31977357da2d] Updating instance_info_cache with network_info: [] {{(pid=61806) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 983.155544] env[61806]: DEBUG oslo_concurrency.lockutils [None req-eb9ccb65-4d10-45a5-b6b1-254652cf9691 tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 983.172408] env[61806]: DEBUG oslo_vmware.api [None req-a06ee197-a4db-47dc-a9f0-aee2aa23de70 tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Task: {'id': session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]52d1f2bd-e69f-8ada-6cb7-8f6c7900e348, 'name': SearchDatastore_Task, 'duration_secs': 0.014642} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 983.173529] env[61806]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-9bc936f0-1f2a-4d62-a8bb-ce7572021401 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 983.179105] env[61806]: DEBUG oslo_vmware.api [None req-a06ee197-a4db-47dc-a9f0-aee2aa23de70 tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Waiting for the task: (returnval){ [ 983.179105] env[61806]: value = "session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]521f0883-bdfe-f784-6be3-c44926e1b587" [ 983.179105] env[61806]: _type = "Task" [ 983.179105] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 983.187069] env[61806]: DEBUG oslo_vmware.api [None req-a06ee197-a4db-47dc-a9f0-aee2aa23de70 tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Task: {'id': session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]521f0883-bdfe-f784-6be3-c44926e1b587, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 983.418263] env[61806]: DEBUG oslo_vmware.api [None req-d01361ec-1270-4938-a4ab-1319441d7070 tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Task: {'id': task-1295039, 'name': ReconfigVM_Task, 'duration_secs': 0.624441} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 983.419076] env[61806]: DEBUG nova.virt.vmwareapi.volumeops [None req-d01361ec-1270-4938-a4ab-1319441d7070 tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] [instance: 962d3d88-03ee-4ff2-9d79-15091671b330] Reconfigured VM instance instance-0000005f to attach disk [datastore1] 962d3d88-03ee-4ff2-9d79-15091671b330/962d3d88-03ee-4ff2-9d79-15091671b330.vmdk or device None with type sparse {{(pid=61806) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 983.419591] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-ff96028b-7beb-4132-b307-461c59967bbf {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 983.428510] env[61806]: DEBUG oslo_vmware.api [None req-d01361ec-1270-4938-a4ab-1319441d7070 tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Waiting for the task: (returnval){ [ 983.428510] env[61806]: value = "task-1295040" [ 983.428510] env[61806]: _type = "Task" [ 983.428510] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 983.438668] env[61806]: DEBUG oslo_vmware.api [None req-d01361ec-1270-4938-a4ab-1319441d7070 tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Task: {'id': task-1295040, 'name': Rename_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 983.478827] env[61806]: DEBUG oslo_concurrency.lockutils [None req-57da61d4-139b-46a7-9163-13f9c5735119 tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] Lock "d9caee4a-a00c-4e50-b01c-99a8f796a09f" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 8.627s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 983.605455] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d96f2cad-ed79-4a87-bef0-d95030dbc3e5 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 983.614287] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-27ac640c-fcbe-4ce0-bd23-f1316f8387ee {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 983.647279] env[61806]: INFO nova.compute.manager [-] [instance: 38e96e22-4200-4175-9085-31977357da2d] Took 1.36 seconds to deallocate network for instance. [ 983.650324] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-86392663-5b2d-4d7e-86ec-2a544f80926b {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 983.664590] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5b2dfeaf-0bf5-4d5b-9974-03eb38d92fb9 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 983.681435] env[61806]: DEBUG nova.compute.provider_tree [None req-474c365d-f281-4fe7-b609-9d0c04e06393 tempest-ServersNegativeTestJSON-891856253 tempest-ServersNegativeTestJSON-891856253-project-member] Inventory has not changed in ProviderTree for provider: a2858be1-fd22-4e08-979e-87ad25293407 {{(pid=61806) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 983.694265] env[61806]: DEBUG oslo_vmware.api [None req-a06ee197-a4db-47dc-a9f0-aee2aa23de70 tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Task: {'id': session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]521f0883-bdfe-f784-6be3-c44926e1b587, 'name': SearchDatastore_Task, 'duration_secs': 0.010792} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 983.694671] env[61806]: DEBUG oslo_concurrency.lockutils [None req-a06ee197-a4db-47dc-a9f0-aee2aa23de70 tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Releasing lock "[datastore1] devstack-image-cache_base/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa.vmdk" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 983.694984] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [None req-a06ee197-a4db-47dc-a9f0-aee2aa23de70 tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa.vmdk to [datastore1] 9357e61f-1628-43bd-ab46-de13c1529f51/9357e61f-1628-43bd-ab46-de13c1529f51.vmdk {{(pid=61806) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 983.695553] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-2bde7210-6b98-4a03-9049-89104d2b255c {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 983.704031] env[61806]: DEBUG oslo_vmware.api [None req-a06ee197-a4db-47dc-a9f0-aee2aa23de70 tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Waiting for the task: (returnval){ [ 983.704031] env[61806]: value = "task-1295041" [ 983.704031] env[61806]: _type = "Task" [ 983.704031] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 983.709645] env[61806]: DEBUG nova.compute.manager [req-330827f7-c6c7-4bb8-8443-50191942625f req-c24ad1e7-e5df-4937-956c-9f6631e82056 service nova] [instance: 38e96e22-4200-4175-9085-31977357da2d] Received event network-vif-deleted-85d68e9b-a57d-4078-baf3-cd7c620f9366 {{(pid=61806) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 983.716415] env[61806]: DEBUG oslo_vmware.api [None req-a06ee197-a4db-47dc-a9f0-aee2aa23de70 tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Task: {'id': task-1295041, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 983.835021] env[61806]: DEBUG oslo_concurrency.lockutils [None req-5dadd265-b355-4de5-b44a-eef29dd0feac tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] Acquiring lock "6f6ba57a-e2d6-4749-a53a-e263861cb1c0" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 983.835417] env[61806]: DEBUG oslo_concurrency.lockutils [None req-5dadd265-b355-4de5-b44a-eef29dd0feac tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] Lock "6f6ba57a-e2d6-4749-a53a-e263861cb1c0" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 983.835674] env[61806]: DEBUG oslo_concurrency.lockutils [None req-5dadd265-b355-4de5-b44a-eef29dd0feac tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] Acquiring lock "6f6ba57a-e2d6-4749-a53a-e263861cb1c0-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 983.836046] env[61806]: DEBUG oslo_concurrency.lockutils [None req-5dadd265-b355-4de5-b44a-eef29dd0feac tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] Lock "6f6ba57a-e2d6-4749-a53a-e263861cb1c0-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 983.836120] env[61806]: DEBUG oslo_concurrency.lockutils [None req-5dadd265-b355-4de5-b44a-eef29dd0feac tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] Lock "6f6ba57a-e2d6-4749-a53a-e263861cb1c0-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 983.838974] env[61806]: INFO nova.compute.manager [None req-5dadd265-b355-4de5-b44a-eef29dd0feac tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] [instance: 6f6ba57a-e2d6-4749-a53a-e263861cb1c0] Terminating instance [ 983.840994] env[61806]: DEBUG nova.compute.manager [None req-5dadd265-b355-4de5-b44a-eef29dd0feac tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] [instance: 6f6ba57a-e2d6-4749-a53a-e263861cb1c0] Start destroying the instance on the hypervisor. {{(pid=61806) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 983.841250] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-5dadd265-b355-4de5-b44a-eef29dd0feac tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] [instance: 6f6ba57a-e2d6-4749-a53a-e263861cb1c0] Destroying instance {{(pid=61806) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 983.842188] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fd4e0539-d887-4b4b-94c1-8a0388c106cf {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 983.851561] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [None req-5dadd265-b355-4de5-b44a-eef29dd0feac tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] [instance: 6f6ba57a-e2d6-4749-a53a-e263861cb1c0] Powering off the VM {{(pid=61806) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 983.851898] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-7eb5073a-0ed2-4a0d-83d1-0c88bbf67683 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 983.859251] env[61806]: DEBUG oslo_vmware.api [None req-5dadd265-b355-4de5-b44a-eef29dd0feac tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] Waiting for the task: (returnval){ [ 983.859251] env[61806]: value = "task-1295042" [ 983.859251] env[61806]: _type = "Task" [ 983.859251] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 983.868925] env[61806]: DEBUG oslo_vmware.api [None req-5dadd265-b355-4de5-b44a-eef29dd0feac tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] Task: {'id': task-1295042, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 983.941419] env[61806]: DEBUG oslo_vmware.api [None req-d01361ec-1270-4938-a4ab-1319441d7070 tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Task: {'id': task-1295040, 'name': Rename_Task, 'duration_secs': 0.242955} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 983.941832] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [None req-d01361ec-1270-4938-a4ab-1319441d7070 tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] [instance: 962d3d88-03ee-4ff2-9d79-15091671b330] Powering on the VM {{(pid=61806) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 983.942283] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-74bf93ba-43c5-46fc-b2c2-18161e968a17 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 983.951926] env[61806]: DEBUG oslo_vmware.api [None req-d01361ec-1270-4938-a4ab-1319441d7070 tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Waiting for the task: (returnval){ [ 983.951926] env[61806]: value = "task-1295043" [ 983.951926] env[61806]: _type = "Task" [ 983.951926] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 983.964412] env[61806]: DEBUG oslo_vmware.api [None req-d01361ec-1270-4938-a4ab-1319441d7070 tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Task: {'id': task-1295043, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 984.158436] env[61806]: DEBUG oslo_concurrency.lockutils [None req-7fff0eac-b278-41ae-86ad-6a3d9bc09f29 tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 984.187875] env[61806]: DEBUG nova.scheduler.client.report [None req-474c365d-f281-4fe7-b609-9d0c04e06393 tempest-ServersNegativeTestJSON-891856253 tempest-ServersNegativeTestJSON-891856253-project-member] Inventory has not changed for provider a2858be1-fd22-4e08-979e-87ad25293407 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 138, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61806) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 984.219070] env[61806]: DEBUG oslo_vmware.api [None req-a06ee197-a4db-47dc-a9f0-aee2aa23de70 tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Task: {'id': task-1295041, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 984.371542] env[61806]: DEBUG oslo_vmware.api [None req-5dadd265-b355-4de5-b44a-eef29dd0feac tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] Task: {'id': task-1295042, 'name': PowerOffVM_Task, 'duration_secs': 0.329104} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 984.371986] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [None req-5dadd265-b355-4de5-b44a-eef29dd0feac tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] [instance: 6f6ba57a-e2d6-4749-a53a-e263861cb1c0] Powered off the VM {{(pid=61806) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 984.372316] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-5dadd265-b355-4de5-b44a-eef29dd0feac tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] [instance: 6f6ba57a-e2d6-4749-a53a-e263861cb1c0] Unregistering the VM {{(pid=61806) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 984.372701] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-7c9f7bd6-a1cb-4471-997b-b07f6d423c55 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 984.458386] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-5dadd265-b355-4de5-b44a-eef29dd0feac tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] [instance: 6f6ba57a-e2d6-4749-a53a-e263861cb1c0] Unregistered the VM {{(pid=61806) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 984.458657] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-5dadd265-b355-4de5-b44a-eef29dd0feac tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] [instance: 6f6ba57a-e2d6-4749-a53a-e263861cb1c0] Deleting contents of the VM from datastore datastore2 {{(pid=61806) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 984.458950] env[61806]: DEBUG nova.virt.vmwareapi.ds_util [None req-5dadd265-b355-4de5-b44a-eef29dd0feac tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] Deleting the datastore file [datastore2] 6f6ba57a-e2d6-4749-a53a-e263861cb1c0 {{(pid=61806) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 984.459657] env[61806]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-e1e8d2d3-4d18-44ab-b199-c44c558c849e {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 984.466468] env[61806]: DEBUG oslo_vmware.api [None req-d01361ec-1270-4938-a4ab-1319441d7070 tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Task: {'id': task-1295043, 'name': PowerOnVM_Task} progress is 89%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 984.472054] env[61806]: DEBUG oslo_vmware.api [None req-5dadd265-b355-4de5-b44a-eef29dd0feac tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] Waiting for the task: (returnval){ [ 984.472054] env[61806]: value = "task-1295045" [ 984.472054] env[61806]: _type = "Task" [ 984.472054] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 984.480123] env[61806]: DEBUG oslo_vmware.api [None req-5dadd265-b355-4de5-b44a-eef29dd0feac tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] Task: {'id': task-1295045, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 984.693692] env[61806]: DEBUG oslo_concurrency.lockutils [None req-474c365d-f281-4fe7-b609-9d0c04e06393 tempest-ServersNegativeTestJSON-891856253 tempest-ServersNegativeTestJSON-891856253-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.748s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 984.696123] env[61806]: DEBUG oslo_concurrency.lockutils [None req-eb9ccb65-4d10-45a5-b6b1-254652cf9691 tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 1.541s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 984.696370] env[61806]: DEBUG nova.objects.instance [None req-eb9ccb65-4d10-45a5-b6b1-254652cf9691 tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Lazy-loading 'resources' on Instance uuid 461f98f2-80ba-4088-9e7f-1ba2a9b0c1d9 {{(pid=61806) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 984.718309] env[61806]: DEBUG oslo_vmware.api [None req-a06ee197-a4db-47dc-a9f0-aee2aa23de70 tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Task: {'id': task-1295041, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.562493} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 984.719360] env[61806]: INFO nova.scheduler.client.report [None req-474c365d-f281-4fe7-b609-9d0c04e06393 tempest-ServersNegativeTestJSON-891856253 tempest-ServersNegativeTestJSON-891856253-project-member] Deleted allocations for instance e146f7e5-bbf8-4adf-a2e9-b7e753438976 [ 984.720670] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [None req-a06ee197-a4db-47dc-a9f0-aee2aa23de70 tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa.vmdk to [datastore1] 9357e61f-1628-43bd-ab46-de13c1529f51/9357e61f-1628-43bd-ab46-de13c1529f51.vmdk {{(pid=61806) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 984.720790] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-a06ee197-a4db-47dc-a9f0-aee2aa23de70 tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] [instance: 9357e61f-1628-43bd-ab46-de13c1529f51] Extending root virtual disk to 1048576 {{(pid=61806) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 984.724044] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-d316294b-7b6a-40a0-bc87-f2d93a36ac2e {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 984.733049] env[61806]: DEBUG oslo_vmware.api [None req-a06ee197-a4db-47dc-a9f0-aee2aa23de70 tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Waiting for the task: (returnval){ [ 984.733049] env[61806]: value = "task-1295046" [ 984.733049] env[61806]: _type = "Task" [ 984.733049] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 984.741651] env[61806]: DEBUG oslo_vmware.api [None req-a06ee197-a4db-47dc-a9f0-aee2aa23de70 tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Task: {'id': task-1295046, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 984.963364] env[61806]: DEBUG oslo_vmware.api [None req-d01361ec-1270-4938-a4ab-1319441d7070 tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Task: {'id': task-1295043, 'name': PowerOnVM_Task, 'duration_secs': 0.592061} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 984.963665] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [None req-d01361ec-1270-4938-a4ab-1319441d7070 tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] [instance: 962d3d88-03ee-4ff2-9d79-15091671b330] Powered on the VM {{(pid=61806) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 984.963875] env[61806]: INFO nova.compute.manager [None req-d01361ec-1270-4938-a4ab-1319441d7070 tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] [instance: 962d3d88-03ee-4ff2-9d79-15091671b330] Took 8.74 seconds to spawn the instance on the hypervisor. [ 984.964072] env[61806]: DEBUG nova.compute.manager [None req-d01361ec-1270-4938-a4ab-1319441d7070 tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] [instance: 962d3d88-03ee-4ff2-9d79-15091671b330] Checking state {{(pid=61806) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 984.964913] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3acddbe7-09db-4a9e-9222-909776a78c63 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 984.983093] env[61806]: DEBUG oslo_vmware.api [None req-5dadd265-b355-4de5-b44a-eef29dd0feac tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] Task: {'id': task-1295045, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.340246} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 984.983093] env[61806]: DEBUG nova.virt.vmwareapi.ds_util [None req-5dadd265-b355-4de5-b44a-eef29dd0feac tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] Deleted the datastore file {{(pid=61806) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 984.983093] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-5dadd265-b355-4de5-b44a-eef29dd0feac tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] [instance: 6f6ba57a-e2d6-4749-a53a-e263861cb1c0] Deleted contents of the VM from datastore datastore2 {{(pid=61806) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 984.983093] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-5dadd265-b355-4de5-b44a-eef29dd0feac tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] [instance: 6f6ba57a-e2d6-4749-a53a-e263861cb1c0] Instance destroyed {{(pid=61806) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 984.983385] env[61806]: INFO nova.compute.manager [None req-5dadd265-b355-4de5-b44a-eef29dd0feac tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] [instance: 6f6ba57a-e2d6-4749-a53a-e263861cb1c0] Took 1.14 seconds to destroy the instance on the hypervisor. [ 984.983457] env[61806]: DEBUG oslo.service.loopingcall [None req-5dadd265-b355-4de5-b44a-eef29dd0feac tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61806) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 984.983652] env[61806]: DEBUG nova.compute.manager [-] [instance: 6f6ba57a-e2d6-4749-a53a-e263861cb1c0] Deallocating network for instance {{(pid=61806) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 984.983750] env[61806]: DEBUG nova.network.neutron [-] [instance: 6f6ba57a-e2d6-4749-a53a-e263861cb1c0] deallocate_for_instance() {{(pid=61806) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 985.231434] env[61806]: DEBUG oslo_concurrency.lockutils [None req-474c365d-f281-4fe7-b609-9d0c04e06393 tempest-ServersNegativeTestJSON-891856253 tempest-ServersNegativeTestJSON-891856253-project-member] Lock "e146f7e5-bbf8-4adf-a2e9-b7e753438976" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 7.184s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 985.243821] env[61806]: DEBUG oslo_vmware.api [None req-a06ee197-a4db-47dc-a9f0-aee2aa23de70 tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Task: {'id': task-1295046, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.079477} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 985.246437] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-a06ee197-a4db-47dc-a9f0-aee2aa23de70 tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] [instance: 9357e61f-1628-43bd-ab46-de13c1529f51] Extended root virtual disk {{(pid=61806) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 985.247512] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-668ba2b9-6d52-4fa6-9ac0-90025eade3b3 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 985.272017] env[61806]: DEBUG nova.virt.vmwareapi.volumeops [None req-a06ee197-a4db-47dc-a9f0-aee2aa23de70 tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] [instance: 9357e61f-1628-43bd-ab46-de13c1529f51] Reconfiguring VM instance instance-00000060 to attach disk [datastore1] 9357e61f-1628-43bd-ab46-de13c1529f51/9357e61f-1628-43bd-ab46-de13c1529f51.vmdk or device None with type sparse {{(pid=61806) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 985.274943] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-17a56971-9b46-4ecb-b00c-9e312721988d {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 985.295414] env[61806]: DEBUG oslo_vmware.api [None req-a06ee197-a4db-47dc-a9f0-aee2aa23de70 tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Waiting for the task: (returnval){ [ 985.295414] env[61806]: value = "task-1295047" [ 985.295414] env[61806]: _type = "Task" [ 985.295414] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 985.307976] env[61806]: DEBUG oslo_vmware.api [None req-a06ee197-a4db-47dc-a9f0-aee2aa23de70 tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Task: {'id': task-1295047, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 985.374486] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ac51395f-018a-425a-8d67-38f5c58c4b99 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 985.382643] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d9409aef-1d9f-402e-9e80-f74a4ae9d121 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 985.415553] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-49f132ed-05f7-4984-967b-04169b60e69a {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 985.423948] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-10510c44-a254-4e82-abc8-6b4da725d46a {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 985.438839] env[61806]: DEBUG nova.compute.provider_tree [None req-eb9ccb65-4d10-45a5-b6b1-254652cf9691 tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Inventory has not changed in ProviderTree for provider: a2858be1-fd22-4e08-979e-87ad25293407 {{(pid=61806) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 985.488108] env[61806]: INFO nova.compute.manager [None req-d01361ec-1270-4938-a4ab-1319441d7070 tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] [instance: 962d3d88-03ee-4ff2-9d79-15091671b330] Took 18.32 seconds to build instance. [ 985.687944] env[61806]: DEBUG nova.network.neutron [-] [instance: 6f6ba57a-e2d6-4749-a53a-e263861cb1c0] Updating instance_info_cache with network_info: [] {{(pid=61806) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 985.732209] env[61806]: DEBUG nova.compute.manager [req-1d523475-844a-4fc1-93fb-376bdaa6f946 req-d4e39461-9469-4918-ab58-a51274698c88 service nova] [instance: 6f6ba57a-e2d6-4749-a53a-e263861cb1c0] Received event network-vif-deleted-013c9f0d-65b1-4279-a823-a7f1c14dfb91 {{(pid=61806) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 985.806805] env[61806]: DEBUG oslo_vmware.api [None req-a06ee197-a4db-47dc-a9f0-aee2aa23de70 tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Task: {'id': task-1295047, 'name': ReconfigVM_Task, 'duration_secs': 0.356125} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 985.807200] env[61806]: DEBUG nova.virt.vmwareapi.volumeops [None req-a06ee197-a4db-47dc-a9f0-aee2aa23de70 tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] [instance: 9357e61f-1628-43bd-ab46-de13c1529f51] Reconfigured VM instance instance-00000060 to attach disk [datastore1] 9357e61f-1628-43bd-ab46-de13c1529f51/9357e61f-1628-43bd-ab46-de13c1529f51.vmdk or device None with type sparse {{(pid=61806) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 985.807772] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-27750f37-68d5-4158-9223-8569c6fb30e7 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 985.816061] env[61806]: DEBUG oslo_vmware.api [None req-a06ee197-a4db-47dc-a9f0-aee2aa23de70 tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Waiting for the task: (returnval){ [ 985.816061] env[61806]: value = "task-1295048" [ 985.816061] env[61806]: _type = "Task" [ 985.816061] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 985.825279] env[61806]: DEBUG oslo_vmware.api [None req-a06ee197-a4db-47dc-a9f0-aee2aa23de70 tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Task: {'id': task-1295048, 'name': Rename_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 985.942782] env[61806]: DEBUG nova.scheduler.client.report [None req-eb9ccb65-4d10-45a5-b6b1-254652cf9691 tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Inventory has not changed for provider a2858be1-fd22-4e08-979e-87ad25293407 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 138, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61806) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 985.990855] env[61806]: DEBUG oslo_concurrency.lockutils [None req-d01361ec-1270-4938-a4ab-1319441d7070 tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Lock "962d3d88-03ee-4ff2-9d79-15091671b330" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 19.826s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 986.191425] env[61806]: INFO nova.compute.manager [-] [instance: 6f6ba57a-e2d6-4749-a53a-e263861cb1c0] Took 1.21 seconds to deallocate network for instance. [ 986.328836] env[61806]: DEBUG oslo_vmware.api [None req-a06ee197-a4db-47dc-a9f0-aee2aa23de70 tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Task: {'id': task-1295048, 'name': Rename_Task, 'duration_secs': 0.164683} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 986.329216] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [None req-a06ee197-a4db-47dc-a9f0-aee2aa23de70 tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] [instance: 9357e61f-1628-43bd-ab46-de13c1529f51] Powering on the VM {{(pid=61806) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 986.329373] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-475eb202-966c-4e84-9686-2f9e62d1bdb6 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 986.337790] env[61806]: DEBUG oslo_vmware.api [None req-a06ee197-a4db-47dc-a9f0-aee2aa23de70 tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Waiting for the task: (returnval){ [ 986.337790] env[61806]: value = "task-1295049" [ 986.337790] env[61806]: _type = "Task" [ 986.337790] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 986.346307] env[61806]: DEBUG oslo_vmware.api [None req-a06ee197-a4db-47dc-a9f0-aee2aa23de70 tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Task: {'id': task-1295049, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 986.437918] env[61806]: DEBUG oslo_concurrency.lockutils [None req-8b43c00c-1795-4a6c-a629-06587eb2119f tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Acquiring lock "962d3d88-03ee-4ff2-9d79-15091671b330" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 986.438334] env[61806]: DEBUG oslo_concurrency.lockutils [None req-8b43c00c-1795-4a6c-a629-06587eb2119f tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Lock "962d3d88-03ee-4ff2-9d79-15091671b330" acquired by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: waited 0.001s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 986.438577] env[61806]: DEBUG nova.compute.manager [None req-8b43c00c-1795-4a6c-a629-06587eb2119f tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] [instance: 962d3d88-03ee-4ff2-9d79-15091671b330] Checking state {{(pid=61806) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 986.439734] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a2219fb4-1052-4180-8b00-12b487a1a893 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 986.448734] env[61806]: DEBUG oslo_concurrency.lockutils [None req-eb9ccb65-4d10-45a5-b6b1-254652cf9691 tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.752s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 986.450896] env[61806]: DEBUG nova.compute.manager [None req-8b43c00c-1795-4a6c-a629-06587eb2119f tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] [instance: 962d3d88-03ee-4ff2-9d79-15091671b330] Stopping instance; current vm_state: active, current task_state: powering-off, current DB power_state: 1, current VM power_state: 1 {{(pid=61806) do_stop_instance /opt/stack/nova/nova/compute/manager.py:3368}} [ 986.451468] env[61806]: DEBUG nova.objects.instance [None req-8b43c00c-1795-4a6c-a629-06587eb2119f tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Lazy-loading 'flavor' on Instance uuid 962d3d88-03ee-4ff2-9d79-15091671b330 {{(pid=61806) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 986.453048] env[61806]: DEBUG oslo_concurrency.lockutils [None req-7fff0eac-b278-41ae-86ad-6a3d9bc09f29 tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 2.295s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 986.453311] env[61806]: DEBUG nova.objects.instance [None req-7fff0eac-b278-41ae-86ad-6a3d9bc09f29 tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] Lazy-loading 'resources' on Instance uuid 38e96e22-4200-4175-9085-31977357da2d {{(pid=61806) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 986.480048] env[61806]: INFO nova.scheduler.client.report [None req-eb9ccb65-4d10-45a5-b6b1-254652cf9691 tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Deleted allocations for instance 461f98f2-80ba-4088-9e7f-1ba2a9b0c1d9 [ 986.698359] env[61806]: DEBUG oslo_concurrency.lockutils [None req-5dadd265-b355-4de5-b44a-eef29dd0feac tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 986.848761] env[61806]: DEBUG oslo_vmware.api [None req-a06ee197-a4db-47dc-a9f0-aee2aa23de70 tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Task: {'id': task-1295049, 'name': PowerOnVM_Task, 'duration_secs': 0.46038} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 986.849055] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [None req-a06ee197-a4db-47dc-a9f0-aee2aa23de70 tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] [instance: 9357e61f-1628-43bd-ab46-de13c1529f51] Powered on the VM {{(pid=61806) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 986.849267] env[61806]: INFO nova.compute.manager [None req-a06ee197-a4db-47dc-a9f0-aee2aa23de70 tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] [instance: 9357e61f-1628-43bd-ab46-de13c1529f51] Took 8.31 seconds to spawn the instance on the hypervisor. [ 986.849452] env[61806]: DEBUG nova.compute.manager [None req-a06ee197-a4db-47dc-a9f0-aee2aa23de70 tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] [instance: 9357e61f-1628-43bd-ab46-de13c1529f51] Checking state {{(pid=61806) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 986.850361] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ec3a0fc9-9a10-4e60-a643-4edb9c788812 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 986.959704] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [None req-8b43c00c-1795-4a6c-a629-06587eb2119f tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] [instance: 962d3d88-03ee-4ff2-9d79-15091671b330] Powering off the VM {{(pid=61806) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 986.960438] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-b8a34203-9407-4978-b4db-4a1f7d6e642b {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 986.971220] env[61806]: DEBUG oslo_vmware.api [None req-8b43c00c-1795-4a6c-a629-06587eb2119f tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Waiting for the task: (returnval){ [ 986.971220] env[61806]: value = "task-1295050" [ 986.971220] env[61806]: _type = "Task" [ 986.971220] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 986.984886] env[61806]: DEBUG oslo_vmware.api [None req-8b43c00c-1795-4a6c-a629-06587eb2119f tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Task: {'id': task-1295050, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 986.994203] env[61806]: DEBUG oslo_concurrency.lockutils [None req-eb9ccb65-4d10-45a5-b6b1-254652cf9691 tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Lock "461f98f2-80ba-4088-9e7f-1ba2a9b0c1d9" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 7.237s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 987.098196] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ca5ce92c-82fb-4b60-bd81-2128e60a5927 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 987.106439] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8f74db00-9fa7-418f-9ec4-dbc1be816f23 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 987.140464] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d0d58614-9605-4d85-bab8-4c24588da74d {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 987.149853] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eb7627c5-cc4f-457d-ae23-1189dc39fa2d {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 987.165660] env[61806]: DEBUG nova.compute.provider_tree [None req-7fff0eac-b278-41ae-86ad-6a3d9bc09f29 tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] Inventory has not changed in ProviderTree for provider: a2858be1-fd22-4e08-979e-87ad25293407 {{(pid=61806) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 987.371092] env[61806]: INFO nova.compute.manager [None req-a06ee197-a4db-47dc-a9f0-aee2aa23de70 tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] [instance: 9357e61f-1628-43bd-ab46-de13c1529f51] Took 14.77 seconds to build instance. [ 987.482272] env[61806]: DEBUG oslo_vmware.api [None req-8b43c00c-1795-4a6c-a629-06587eb2119f tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Task: {'id': task-1295050, 'name': PowerOffVM_Task, 'duration_secs': 0.224308} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 987.482608] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [None req-8b43c00c-1795-4a6c-a629-06587eb2119f tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] [instance: 962d3d88-03ee-4ff2-9d79-15091671b330] Powered off the VM {{(pid=61806) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 987.482803] env[61806]: DEBUG nova.compute.manager [None req-8b43c00c-1795-4a6c-a629-06587eb2119f tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] [instance: 962d3d88-03ee-4ff2-9d79-15091671b330] Checking state {{(pid=61806) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 987.483630] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-21b1831a-e5f2-4c2e-8379-bc1b9c16fc22 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 987.669366] env[61806]: DEBUG nova.scheduler.client.report [None req-7fff0eac-b278-41ae-86ad-6a3d9bc09f29 tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] Inventory has not changed for provider a2858be1-fd22-4e08-979e-87ad25293407 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 138, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61806) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 987.873804] env[61806]: DEBUG oslo_concurrency.lockutils [None req-a06ee197-a4db-47dc-a9f0-aee2aa23de70 tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Lock "9357e61f-1628-43bd-ab46-de13c1529f51" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 16.286s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 987.996689] env[61806]: DEBUG oslo_concurrency.lockutils [None req-8b43c00c-1795-4a6c-a629-06587eb2119f tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Lock "962d3d88-03ee-4ff2-9d79-15091671b330" "released" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: held 1.558s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 988.175862] env[61806]: DEBUG oslo_concurrency.lockutils [None req-7fff0eac-b278-41ae-86ad-6a3d9bc09f29 tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.723s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 988.179027] env[61806]: DEBUG oslo_concurrency.lockutils [None req-5dadd265-b355-4de5-b44a-eef29dd0feac tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 1.480s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 988.179027] env[61806]: DEBUG nova.objects.instance [None req-5dadd265-b355-4de5-b44a-eef29dd0feac tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] Lazy-loading 'resources' on Instance uuid 6f6ba57a-e2d6-4749-a53a-e263861cb1c0 {{(pid=61806) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 988.195303] env[61806]: INFO nova.scheduler.client.report [None req-7fff0eac-b278-41ae-86ad-6a3d9bc09f29 tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] Deleted allocations for instance 38e96e22-4200-4175-9085-31977357da2d [ 988.707143] env[61806]: DEBUG oslo_concurrency.lockutils [None req-7fff0eac-b278-41ae-86ad-6a3d9bc09f29 tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] Lock "38e96e22-4200-4175-9085-31977357da2d" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 7.673s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 988.828371] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5fe394ce-453e-48aa-8537-0d881dff24b6 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 988.833742] env[61806]: DEBUG nova.compute.manager [None req-d41c382b-3631-43f7-9094-5acd8818b680 tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] [instance: 67d1bdd8-cf8f-4fb5-bd3c-42008e4d60e3] Stashing vm_state: active {{(pid=61806) _prep_resize /opt/stack/nova/nova/compute/manager.py:5624}} [ 988.839796] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-491810b3-f954-4f95-a9ce-1660acffd860 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 988.871623] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-67e7425d-5e09-43d3-9f68-d28674e1e782 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 988.880671] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-248c7b53-3ca0-4789-b0f0-f6f3eb93f18f {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 988.896432] env[61806]: DEBUG nova.compute.provider_tree [None req-5dadd265-b355-4de5-b44a-eef29dd0feac tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] Inventory has not changed in ProviderTree for provider: a2858be1-fd22-4e08-979e-87ad25293407 {{(pid=61806) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 988.918300] env[61806]: DEBUG oslo_concurrency.lockutils [None req-907fa7f6-24e7-45d8-9e7f-b2cf924bf805 tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Acquiring lock "962d3d88-03ee-4ff2-9d79-15091671b330" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 988.918625] env[61806]: DEBUG oslo_concurrency.lockutils [None req-907fa7f6-24e7-45d8-9e7f-b2cf924bf805 tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Lock "962d3d88-03ee-4ff2-9d79-15091671b330" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 988.918867] env[61806]: DEBUG oslo_concurrency.lockutils [None req-907fa7f6-24e7-45d8-9e7f-b2cf924bf805 tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Acquiring lock "962d3d88-03ee-4ff2-9d79-15091671b330-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 988.919105] env[61806]: DEBUG oslo_concurrency.lockutils [None req-907fa7f6-24e7-45d8-9e7f-b2cf924bf805 tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Lock "962d3d88-03ee-4ff2-9d79-15091671b330-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 988.919297] env[61806]: DEBUG oslo_concurrency.lockutils [None req-907fa7f6-24e7-45d8-9e7f-b2cf924bf805 tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Lock "962d3d88-03ee-4ff2-9d79-15091671b330-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 988.921481] env[61806]: INFO nova.compute.manager [None req-907fa7f6-24e7-45d8-9e7f-b2cf924bf805 tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] [instance: 962d3d88-03ee-4ff2-9d79-15091671b330] Terminating instance [ 988.923346] env[61806]: DEBUG nova.compute.manager [None req-907fa7f6-24e7-45d8-9e7f-b2cf924bf805 tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] [instance: 962d3d88-03ee-4ff2-9d79-15091671b330] Start destroying the instance on the hypervisor. {{(pid=61806) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 988.923557] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-907fa7f6-24e7-45d8-9e7f-b2cf924bf805 tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] [instance: 962d3d88-03ee-4ff2-9d79-15091671b330] Destroying instance {{(pid=61806) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 988.924439] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-149ce1dd-297d-45c3-a9f0-1f2097967fbd {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 988.933034] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-907fa7f6-24e7-45d8-9e7f-b2cf924bf805 tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] [instance: 962d3d88-03ee-4ff2-9d79-15091671b330] Unregistering the VM {{(pid=61806) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 988.933301] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-f399ba72-a4af-41a7-b091-79cd1deb3aa6 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 989.005661] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-907fa7f6-24e7-45d8-9e7f-b2cf924bf805 tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] [instance: 962d3d88-03ee-4ff2-9d79-15091671b330] Unregistered the VM {{(pid=61806) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 989.005661] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-907fa7f6-24e7-45d8-9e7f-b2cf924bf805 tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] [instance: 962d3d88-03ee-4ff2-9d79-15091671b330] Deleting contents of the VM from datastore datastore1 {{(pid=61806) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 989.005661] env[61806]: DEBUG nova.virt.vmwareapi.ds_util [None req-907fa7f6-24e7-45d8-9e7f-b2cf924bf805 tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Deleting the datastore file [datastore1] 962d3d88-03ee-4ff2-9d79-15091671b330 {{(pid=61806) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 989.005661] env[61806]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-370d26ab-e543-410a-ac4d-eb182675dc94 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 989.016350] env[61806]: DEBUG oslo_vmware.api [None req-907fa7f6-24e7-45d8-9e7f-b2cf924bf805 tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Waiting for the task: (returnval){ [ 989.016350] env[61806]: value = "task-1295052" [ 989.016350] env[61806]: _type = "Task" [ 989.016350] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 989.025384] env[61806]: DEBUG oslo_vmware.api [None req-907fa7f6-24e7-45d8-9e7f-b2cf924bf805 tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Task: {'id': task-1295052, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 989.138846] env[61806]: DEBUG nova.compute.manager [None req-1c35eaa5-2854-4e3e-bc2c-4ef5a33b9a12 tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] [instance: 9357e61f-1628-43bd-ab46-de13c1529f51] Stashing vm_state: active {{(pid=61806) _prep_resize /opt/stack/nova/nova/compute/manager.py:5624}} [ 989.356699] env[61806]: DEBUG oslo_concurrency.lockutils [None req-d41c382b-3631-43f7-9094-5acd8818b680 tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 989.400839] env[61806]: DEBUG nova.scheduler.client.report [None req-5dadd265-b355-4de5-b44a-eef29dd0feac tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] Inventory has not changed for provider a2858be1-fd22-4e08-979e-87ad25293407 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 138, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61806) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 989.526299] env[61806]: DEBUG oslo_vmware.api [None req-907fa7f6-24e7-45d8-9e7f-b2cf924bf805 tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Task: {'id': task-1295052, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.285856} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 989.526684] env[61806]: DEBUG nova.virt.vmwareapi.ds_util [None req-907fa7f6-24e7-45d8-9e7f-b2cf924bf805 tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Deleted the datastore file {{(pid=61806) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 989.527012] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-907fa7f6-24e7-45d8-9e7f-b2cf924bf805 tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] [instance: 962d3d88-03ee-4ff2-9d79-15091671b330] Deleted contents of the VM from datastore datastore1 {{(pid=61806) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 989.527295] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-907fa7f6-24e7-45d8-9e7f-b2cf924bf805 tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] [instance: 962d3d88-03ee-4ff2-9d79-15091671b330] Instance destroyed {{(pid=61806) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 989.527585] env[61806]: INFO nova.compute.manager [None req-907fa7f6-24e7-45d8-9e7f-b2cf924bf805 tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] [instance: 962d3d88-03ee-4ff2-9d79-15091671b330] Took 0.60 seconds to destroy the instance on the hypervisor. [ 989.527934] env[61806]: DEBUG oslo.service.loopingcall [None req-907fa7f6-24e7-45d8-9e7f-b2cf924bf805 tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61806) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 989.528266] env[61806]: DEBUG nova.compute.manager [-] [instance: 962d3d88-03ee-4ff2-9d79-15091671b330] Deallocating network for instance {{(pid=61806) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 989.528423] env[61806]: DEBUG nova.network.neutron [-] [instance: 962d3d88-03ee-4ff2-9d79-15091671b330] deallocate_for_instance() {{(pid=61806) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 989.671534] env[61806]: DEBUG oslo_concurrency.lockutils [None req-1c35eaa5-2854-4e3e-bc2c-4ef5a33b9a12 tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 989.908475] env[61806]: DEBUG oslo_concurrency.lockutils [None req-5dadd265-b355-4de5-b44a-eef29dd0feac tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.730s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 989.911016] env[61806]: DEBUG oslo_concurrency.lockutils [None req-d41c382b-3631-43f7-9094-5acd8818b680 tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: waited 0.554s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 989.927047] env[61806]: INFO nova.scheduler.client.report [None req-5dadd265-b355-4de5-b44a-eef29dd0feac tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] Deleted allocations for instance 6f6ba57a-e2d6-4749-a53a-e263861cb1c0 [ 989.974954] env[61806]: DEBUG nova.compute.manager [req-e9b8cce1-2ca9-485d-a669-eb38f4cb0be5 req-b79baf5d-269a-4c60-aa7b-936d63467f9b service nova] [instance: 962d3d88-03ee-4ff2-9d79-15091671b330] Received event network-vif-deleted-1a817be6-517c-4be9-8c14-c6f3995f87e0 {{(pid=61806) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 989.974954] env[61806]: INFO nova.compute.manager [req-e9b8cce1-2ca9-485d-a669-eb38f4cb0be5 req-b79baf5d-269a-4c60-aa7b-936d63467f9b service nova] [instance: 962d3d88-03ee-4ff2-9d79-15091671b330] Neutron deleted interface 1a817be6-517c-4be9-8c14-c6f3995f87e0; detaching it from the instance and deleting it from the info cache [ 989.975114] env[61806]: DEBUG nova.network.neutron [req-e9b8cce1-2ca9-485d-a669-eb38f4cb0be5 req-b79baf5d-269a-4c60-aa7b-936d63467f9b service nova] [instance: 962d3d88-03ee-4ff2-9d79-15091671b330] Updating instance_info_cache with network_info: [] {{(pid=61806) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 990.189694] env[61806]: DEBUG oslo_vmware.rw_handles [None req-b6079056-6464-46e6-97ab-343b76917c18 tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] Getting lease state for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52335490-af02-ad6e-9173-20b0caccf108/disk-0.vmdk. {{(pid=61806) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 990.190639] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-76143966-ce8d-4f10-8d5e-3fecb5bdab56 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 990.197702] env[61806]: DEBUG oslo_vmware.rw_handles [None req-b6079056-6464-46e6-97ab-343b76917c18 tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] Lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52335490-af02-ad6e-9173-20b0caccf108/disk-0.vmdk is in state: ready. {{(pid=61806) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 990.197890] env[61806]: ERROR oslo_vmware.rw_handles [None req-b6079056-6464-46e6-97ab-343b76917c18 tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] Aborting lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52335490-af02-ad6e-9173-20b0caccf108/disk-0.vmdk due to incomplete transfer. [ 990.198150] env[61806]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseAbort with opID=oslo.vmware-5f81bd07-a48e-4ba3-89cd-88d2d35140bc {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 990.206215] env[61806]: DEBUG oslo_vmware.rw_handles [None req-b6079056-6464-46e6-97ab-343b76917c18 tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] Closed VMDK read handle for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52335490-af02-ad6e-9173-20b0caccf108/disk-0.vmdk. {{(pid=61806) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:735}} [ 990.206435] env[61806]: DEBUG nova.virt.vmwareapi.images [None req-b6079056-6464-46e6-97ab-343b76917c18 tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] [instance: fd9e520a-2e7b-43d4-b84d-8933c32cc6a6] Uploaded image b08c43d9-cfd5-4a1a-b99d-0aebb718b592 to the Glance image server {{(pid=61806) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:473}} [ 990.209313] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [None req-b6079056-6464-46e6-97ab-343b76917c18 tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] [instance: fd9e520a-2e7b-43d4-b84d-8933c32cc6a6] Destroying the VM {{(pid=61806) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1358}} [ 990.209584] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-162caaa0-bb1d-4863-8785-c88993079bff {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 990.216492] env[61806]: DEBUG oslo_vmware.api [None req-b6079056-6464-46e6-97ab-343b76917c18 tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] Waiting for the task: (returnval){ [ 990.216492] env[61806]: value = "task-1295053" [ 990.216492] env[61806]: _type = "Task" [ 990.216492] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 990.228623] env[61806]: DEBUG oslo_concurrency.lockutils [None req-b4773569-1b9a-41b1-9eb3-234618dc8825 tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] Acquiring lock "fe1d4ace-7de3-423b-b423-9d7397aa2413" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 990.228858] env[61806]: DEBUG oslo_concurrency.lockutils [None req-b4773569-1b9a-41b1-9eb3-234618dc8825 tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] Lock "fe1d4ace-7de3-423b-b423-9d7397aa2413" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 990.230026] env[61806]: DEBUG oslo_vmware.api [None req-b6079056-6464-46e6-97ab-343b76917c18 tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] Task: {'id': task-1295053, 'name': Destroy_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 990.377170] env[61806]: DEBUG nova.network.neutron [-] [instance: 962d3d88-03ee-4ff2-9d79-15091671b330] Updating instance_info_cache with network_info: [] {{(pid=61806) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 990.416249] env[61806]: INFO nova.compute.claims [None req-d41c382b-3631-43f7-9094-5acd8818b680 tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] [instance: 67d1bdd8-cf8f-4fb5-bd3c-42008e4d60e3] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 990.434810] env[61806]: DEBUG oslo_concurrency.lockutils [None req-5dadd265-b355-4de5-b44a-eef29dd0feac tempest-AttachInterfacesTestJSON-1989978232 tempest-AttachInterfacesTestJSON-1989978232-project-member] Lock "6f6ba57a-e2d6-4749-a53a-e263861cb1c0" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 6.599s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 990.477812] env[61806]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-9ef44cb1-fb20-4f69-8571-957316a9d5b2 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 990.489265] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9bcd08d1-5d50-4054-b75a-a428b062c782 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 990.522614] env[61806]: DEBUG nova.compute.manager [req-e9b8cce1-2ca9-485d-a669-eb38f4cb0be5 req-b79baf5d-269a-4c60-aa7b-936d63467f9b service nova] [instance: 962d3d88-03ee-4ff2-9d79-15091671b330] Detach interface failed, port_id=1a817be6-517c-4be9-8c14-c6f3995f87e0, reason: Instance 962d3d88-03ee-4ff2-9d79-15091671b330 could not be found. {{(pid=61806) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10963}} [ 990.726944] env[61806]: DEBUG oslo_vmware.api [None req-b6079056-6464-46e6-97ab-343b76917c18 tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] Task: {'id': task-1295053, 'name': Destroy_Task, 'duration_secs': 0.366762} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 990.727234] env[61806]: INFO nova.virt.vmwareapi.vm_util [None req-b6079056-6464-46e6-97ab-343b76917c18 tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] [instance: fd9e520a-2e7b-43d4-b84d-8933c32cc6a6] Destroyed the VM [ 990.727494] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-b6079056-6464-46e6-97ab-343b76917c18 tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] [instance: fd9e520a-2e7b-43d4-b84d-8933c32cc6a6] Deleting Snapshot of the VM instance {{(pid=61806) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 990.727801] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-77a6f6a9-63a8-465a-8fc5-c656eb62be20 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 990.730909] env[61806]: DEBUG nova.compute.manager [None req-b4773569-1b9a-41b1-9eb3-234618dc8825 tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] [instance: fe1d4ace-7de3-423b-b423-9d7397aa2413] Starting instance... {{(pid=61806) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 990.734968] env[61806]: DEBUG oslo_vmware.api [None req-b6079056-6464-46e6-97ab-343b76917c18 tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] Waiting for the task: (returnval){ [ 990.734968] env[61806]: value = "task-1295054" [ 990.734968] env[61806]: _type = "Task" [ 990.734968] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 990.743263] env[61806]: DEBUG oslo_vmware.api [None req-b6079056-6464-46e6-97ab-343b76917c18 tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] Task: {'id': task-1295054, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 990.879437] env[61806]: INFO nova.compute.manager [-] [instance: 962d3d88-03ee-4ff2-9d79-15091671b330] Took 1.35 seconds to deallocate network for instance. [ 990.923717] env[61806]: INFO nova.compute.resource_tracker [None req-d41c382b-3631-43f7-9094-5acd8818b680 tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] [instance: 67d1bdd8-cf8f-4fb5-bd3c-42008e4d60e3] Updating resource usage from migration ca0a088e-384c-40b9-a197-357d189bc64d [ 991.072035] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a58f9a45-4b4a-4ec2-8996-d0cd70361688 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 991.080153] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fba4a6ed-21e5-43c3-8eb5-2269f7269024 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 991.110102] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-99838480-5115-4a83-a2c3-012953daab7a {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 991.118295] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-244d41b8-cd31-48e9-9974-d7e580a1b05a {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 991.863252] env[61806]: DEBUG oslo_concurrency.lockutils [None req-907fa7f6-24e7-45d8-9e7f-b2cf924bf805 tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 991.863748] env[61806]: DEBUG nova.compute.provider_tree [None req-d41c382b-3631-43f7-9094-5acd8818b680 tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Inventory has not changed in ProviderTree for provider: a2858be1-fd22-4e08-979e-87ad25293407 {{(pid=61806) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 991.877493] env[61806]: DEBUG oslo_vmware.api [None req-b6079056-6464-46e6-97ab-343b76917c18 tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] Task: {'id': task-1295054, 'name': RemoveSnapshot_Task, 'duration_secs': 0.409473} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 991.877670] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-b6079056-6464-46e6-97ab-343b76917c18 tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] [instance: fd9e520a-2e7b-43d4-b84d-8933c32cc6a6] Deleted Snapshot of the VM instance {{(pid=61806) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 991.880060] env[61806]: DEBUG nova.compute.manager [None req-b6079056-6464-46e6-97ab-343b76917c18 tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] [instance: fd9e520a-2e7b-43d4-b84d-8933c32cc6a6] Checking state {{(pid=61806) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 991.880060] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-932d66d5-510e-4826-8955-bff025a3811a {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 991.883261] env[61806]: DEBUG oslo_concurrency.lockutils [None req-b4773569-1b9a-41b1-9eb3-234618dc8825 tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 992.371865] env[61806]: DEBUG nova.scheduler.client.report [None req-d41c382b-3631-43f7-9094-5acd8818b680 tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Inventory has not changed for provider a2858be1-fd22-4e08-979e-87ad25293407 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 138, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61806) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 992.391297] env[61806]: INFO nova.compute.manager [None req-b6079056-6464-46e6-97ab-343b76917c18 tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] [instance: fd9e520a-2e7b-43d4-b84d-8933c32cc6a6] Shelve offloading [ 992.394794] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [None req-b6079056-6464-46e6-97ab-343b76917c18 tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] [instance: fd9e520a-2e7b-43d4-b84d-8933c32cc6a6] Powering off the VM {{(pid=61806) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 992.395056] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-6c6cad3c-b9aa-4e20-a038-e19d90a8986e {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 992.405641] env[61806]: DEBUG oslo_vmware.api [None req-b6079056-6464-46e6-97ab-343b76917c18 tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] Waiting for the task: (returnval){ [ 992.405641] env[61806]: value = "task-1295055" [ 992.405641] env[61806]: _type = "Task" [ 992.405641] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 992.416381] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [None req-b6079056-6464-46e6-97ab-343b76917c18 tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] [instance: fd9e520a-2e7b-43d4-b84d-8933c32cc6a6] VM already powered off {{(pid=61806) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1509}} [ 992.416588] env[61806]: DEBUG nova.compute.manager [None req-b6079056-6464-46e6-97ab-343b76917c18 tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] [instance: fd9e520a-2e7b-43d4-b84d-8933c32cc6a6] Checking state {{(pid=61806) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 992.417378] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-324531f3-c510-4714-9afe-04d58a147a39 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 992.424174] env[61806]: DEBUG oslo_concurrency.lockutils [None req-b6079056-6464-46e6-97ab-343b76917c18 tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] Acquiring lock "refresh_cache-fd9e520a-2e7b-43d4-b84d-8933c32cc6a6" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 992.424467] env[61806]: DEBUG oslo_concurrency.lockutils [None req-b6079056-6464-46e6-97ab-343b76917c18 tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] Acquired lock "refresh_cache-fd9e520a-2e7b-43d4-b84d-8933c32cc6a6" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 992.424740] env[61806]: DEBUG nova.network.neutron [None req-b6079056-6464-46e6-97ab-343b76917c18 tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] [instance: fd9e520a-2e7b-43d4-b84d-8933c32cc6a6] Building network info cache for instance {{(pid=61806) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 992.878061] env[61806]: DEBUG oslo_concurrency.lockutils [None req-d41c382b-3631-43f7-9094-5acd8818b680 tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: held 2.967s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 992.878169] env[61806]: INFO nova.compute.manager [None req-d41c382b-3631-43f7-9094-5acd8818b680 tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] [instance: 67d1bdd8-cf8f-4fb5-bd3c-42008e4d60e3] Migrating [ 992.891735] env[61806]: DEBUG oslo_concurrency.lockutils [None req-1c35eaa5-2854-4e3e-bc2c-4ef5a33b9a12 tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: waited 3.221s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 993.252228] env[61806]: DEBUG nova.network.neutron [None req-b6079056-6464-46e6-97ab-343b76917c18 tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] [instance: fd9e520a-2e7b-43d4-b84d-8933c32cc6a6] Updating instance_info_cache with network_info: [{"id": "670f75e3-58f9-493b-9148-416e2e9cea5d", "address": "fa:16:3e:76:be:e1", "network": {"id": "a9d993bc-0817-4ee6-ab6e-1ef9432a110f", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-939246736-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.226", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "bbbb09a49203421d807f733616c854df", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ad4c015b-4a8b-46ca-9556-74bad8db9fb3", "external-id": "nsx-vlan-transportzone-246", "segmentation_id": 246, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap670f75e3-58", "ovs_interfaceid": "670f75e3-58f9-493b-9148-416e2e9cea5d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61806) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 993.397813] env[61806]: INFO nova.compute.claims [None req-1c35eaa5-2854-4e3e-bc2c-4ef5a33b9a12 tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] [instance: 9357e61f-1628-43bd-ab46-de13c1529f51] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 993.402361] env[61806]: DEBUG oslo_concurrency.lockutils [None req-d41c382b-3631-43f7-9094-5acd8818b680 tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Acquiring lock "refresh_cache-67d1bdd8-cf8f-4fb5-bd3c-42008e4d60e3" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 993.402636] env[61806]: DEBUG oslo_concurrency.lockutils [None req-d41c382b-3631-43f7-9094-5acd8818b680 tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Acquired lock "refresh_cache-67d1bdd8-cf8f-4fb5-bd3c-42008e4d60e3" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 993.402842] env[61806]: DEBUG nova.network.neutron [None req-d41c382b-3631-43f7-9094-5acd8818b680 tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] [instance: 67d1bdd8-cf8f-4fb5-bd3c-42008e4d60e3] Building network info cache for instance {{(pid=61806) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 993.754817] env[61806]: DEBUG oslo_concurrency.lockutils [None req-b6079056-6464-46e6-97ab-343b76917c18 tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] Releasing lock "refresh_cache-fd9e520a-2e7b-43d4-b84d-8933c32cc6a6" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 993.906521] env[61806]: INFO nova.compute.resource_tracker [None req-1c35eaa5-2854-4e3e-bc2c-4ef5a33b9a12 tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] [instance: 9357e61f-1628-43bd-ab46-de13c1529f51] Updating resource usage from migration ce240126-0844-456c-b7bf-b9d90e170466 [ 994.105257] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ff94ed0b-7989-44a2-820e-ab4f687f77c1 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 994.114252] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-107648d4-87a5-4ee9-a1f5-65787f999c4b {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 994.146907] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d1b7626e-09ef-444b-8285-e62b23620186 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 994.150400] env[61806]: DEBUG nova.compute.manager [req-a1984e43-7e06-4be2-9ba2-45048efac651 req-9dda38cf-8b61-4bec-a212-07b8dd64e9da service nova] [instance: fd9e520a-2e7b-43d4-b84d-8933c32cc6a6] Received event network-vif-unplugged-670f75e3-58f9-493b-9148-416e2e9cea5d {{(pid=61806) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 994.150619] env[61806]: DEBUG oslo_concurrency.lockutils [req-a1984e43-7e06-4be2-9ba2-45048efac651 req-9dda38cf-8b61-4bec-a212-07b8dd64e9da service nova] Acquiring lock "fd9e520a-2e7b-43d4-b84d-8933c32cc6a6-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 994.150824] env[61806]: DEBUG oslo_concurrency.lockutils [req-a1984e43-7e06-4be2-9ba2-45048efac651 req-9dda38cf-8b61-4bec-a212-07b8dd64e9da service nova] Lock "fd9e520a-2e7b-43d4-b84d-8933c32cc6a6-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 994.150992] env[61806]: DEBUG oslo_concurrency.lockutils [req-a1984e43-7e06-4be2-9ba2-45048efac651 req-9dda38cf-8b61-4bec-a212-07b8dd64e9da service nova] Lock "fd9e520a-2e7b-43d4-b84d-8933c32cc6a6-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 994.151176] env[61806]: DEBUG nova.compute.manager [req-a1984e43-7e06-4be2-9ba2-45048efac651 req-9dda38cf-8b61-4bec-a212-07b8dd64e9da service nova] [instance: fd9e520a-2e7b-43d4-b84d-8933c32cc6a6] No waiting events found dispatching network-vif-unplugged-670f75e3-58f9-493b-9148-416e2e9cea5d {{(pid=61806) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 994.151349] env[61806]: WARNING nova.compute.manager [req-a1984e43-7e06-4be2-9ba2-45048efac651 req-9dda38cf-8b61-4bec-a212-07b8dd64e9da service nova] [instance: fd9e520a-2e7b-43d4-b84d-8933c32cc6a6] Received unexpected event network-vif-unplugged-670f75e3-58f9-493b-9148-416e2e9cea5d for instance with vm_state shelved and task_state shelving_offloading. [ 994.157739] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7d03c64b-55e6-4e89-89d0-bc7a344a671c {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 994.173817] env[61806]: DEBUG nova.compute.provider_tree [None req-1c35eaa5-2854-4e3e-bc2c-4ef5a33b9a12 tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Inventory has not changed in ProviderTree for provider: a2858be1-fd22-4e08-979e-87ad25293407 {{(pid=61806) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 994.360299] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-b6079056-6464-46e6-97ab-343b76917c18 tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] [instance: fd9e520a-2e7b-43d4-b84d-8933c32cc6a6] Destroying instance {{(pid=61806) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 994.361211] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e6161051-3179-4a3c-8517-027e7238cc6e {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 994.370399] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-b6079056-6464-46e6-97ab-343b76917c18 tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] [instance: fd9e520a-2e7b-43d4-b84d-8933c32cc6a6] Unregistering the VM {{(pid=61806) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 994.370678] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-aba63656-65e0-4a8a-90d2-d1c2a7a25b92 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 994.380346] env[61806]: DEBUG nova.network.neutron [None req-d41c382b-3631-43f7-9094-5acd8818b680 tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] [instance: 67d1bdd8-cf8f-4fb5-bd3c-42008e4d60e3] Updating instance_info_cache with network_info: [{"id": "859fb723-4a1c-4ea8-bd7c-0ed308f3a9e3", "address": "fa:16:3e:3c:40:d6", "network": {"id": "746a03b6-c1a6-4d27-98b9-0bee6376634f", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-1971382587-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.245", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c1f2d161c50f4b70987b90e6364c3779", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6a31b224-77d7-48a4-af87-312758611b8e", "external-id": "nsx-vlan-transportzone-761", "segmentation_id": 761, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap859fb723-4a", "ovs_interfaceid": "859fb723-4a1c-4ea8-bd7c-0ed308f3a9e3", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61806) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 994.446710] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-b6079056-6464-46e6-97ab-343b76917c18 tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] [instance: fd9e520a-2e7b-43d4-b84d-8933c32cc6a6] Unregistered the VM {{(pid=61806) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 994.447240] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-b6079056-6464-46e6-97ab-343b76917c18 tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] [instance: fd9e520a-2e7b-43d4-b84d-8933c32cc6a6] Deleting contents of the VM from datastore datastore2 {{(pid=61806) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 994.447766] env[61806]: DEBUG nova.virt.vmwareapi.ds_util [None req-b6079056-6464-46e6-97ab-343b76917c18 tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] Deleting the datastore file [datastore2] fd9e520a-2e7b-43d4-b84d-8933c32cc6a6 {{(pid=61806) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 994.448078] env[61806]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-c6679706-a2cf-459e-82ad-1afb64113c92 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 994.456120] env[61806]: DEBUG oslo_vmware.api [None req-b6079056-6464-46e6-97ab-343b76917c18 tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] Waiting for the task: (returnval){ [ 994.456120] env[61806]: value = "task-1295057" [ 994.456120] env[61806]: _type = "Task" [ 994.456120] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 994.467941] env[61806]: DEBUG oslo_vmware.api [None req-b6079056-6464-46e6-97ab-343b76917c18 tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] Task: {'id': task-1295057, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 994.677080] env[61806]: DEBUG nova.scheduler.client.report [None req-1c35eaa5-2854-4e3e-bc2c-4ef5a33b9a12 tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Inventory has not changed for provider a2858be1-fd22-4e08-979e-87ad25293407 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 138, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61806) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 994.883631] env[61806]: DEBUG oslo_concurrency.lockutils [None req-d41c382b-3631-43f7-9094-5acd8818b680 tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Releasing lock "refresh_cache-67d1bdd8-cf8f-4fb5-bd3c-42008e4d60e3" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 994.967686] env[61806]: DEBUG oslo_vmware.api [None req-b6079056-6464-46e6-97ab-343b76917c18 tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] Task: {'id': task-1295057, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.15823} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 994.967686] env[61806]: DEBUG nova.virt.vmwareapi.ds_util [None req-b6079056-6464-46e6-97ab-343b76917c18 tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] Deleted the datastore file {{(pid=61806) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 994.967686] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-b6079056-6464-46e6-97ab-343b76917c18 tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] [instance: fd9e520a-2e7b-43d4-b84d-8933c32cc6a6] Deleted contents of the VM from datastore datastore2 {{(pid=61806) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 994.967900] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-b6079056-6464-46e6-97ab-343b76917c18 tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] [instance: fd9e520a-2e7b-43d4-b84d-8933c32cc6a6] Instance destroyed {{(pid=61806) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 995.000205] env[61806]: INFO nova.scheduler.client.report [None req-b6079056-6464-46e6-97ab-343b76917c18 tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] Deleted allocations for instance fd9e520a-2e7b-43d4-b84d-8933c32cc6a6 [ 995.184246] env[61806]: DEBUG oslo_concurrency.lockutils [None req-1c35eaa5-2854-4e3e-bc2c-4ef5a33b9a12 tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: held 2.292s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 995.184561] env[61806]: INFO nova.compute.manager [None req-1c35eaa5-2854-4e3e-bc2c-4ef5a33b9a12 tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] [instance: 9357e61f-1628-43bd-ab46-de13c1529f51] Migrating [ 995.191883] env[61806]: DEBUG oslo_concurrency.lockutils [None req-907fa7f6-24e7-45d8-9e7f-b2cf924bf805 tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 3.329s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 995.192122] env[61806]: DEBUG nova.objects.instance [None req-907fa7f6-24e7-45d8-9e7f-b2cf924bf805 tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Lazy-loading 'resources' on Instance uuid 962d3d88-03ee-4ff2-9d79-15091671b330 {{(pid=61806) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 995.505897] env[61806]: DEBUG oslo_concurrency.lockutils [None req-b6079056-6464-46e6-97ab-343b76917c18 tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 995.714999] env[61806]: DEBUG oslo_concurrency.lockutils [None req-1c35eaa5-2854-4e3e-bc2c-4ef5a33b9a12 tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Acquiring lock "refresh_cache-9357e61f-1628-43bd-ab46-de13c1529f51" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 995.714999] env[61806]: DEBUG oslo_concurrency.lockutils [None req-1c35eaa5-2854-4e3e-bc2c-4ef5a33b9a12 tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Acquired lock "refresh_cache-9357e61f-1628-43bd-ab46-de13c1529f51" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 995.714999] env[61806]: DEBUG nova.network.neutron [None req-1c35eaa5-2854-4e3e-bc2c-4ef5a33b9a12 tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] [instance: 9357e61f-1628-43bd-ab46-de13c1529f51] Building network info cache for instance {{(pid=61806) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 995.860923] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-75a43102-e750-4d2d-909b-b322eb9215d9 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 995.869950] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7bba8a75-d18f-4314-90d2-6ef4ae8973a8 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 995.914118] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-20a370a2-4518-4997-baaf-cd59909beda9 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 995.920530] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fe5f026d-383a-4d79-bc9d-72471b37f5c5 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 995.936140] env[61806]: DEBUG nova.compute.provider_tree [None req-907fa7f6-24e7-45d8-9e7f-b2cf924bf805 tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Inventory has not changed in ProviderTree for provider: a2858be1-fd22-4e08-979e-87ad25293407 {{(pid=61806) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 996.273822] env[61806]: DEBUG nova.compute.manager [req-c2d91fc2-4de8-4cfb-88b5-fcf63103ec86 req-66728731-80ed-43d5-99c8-553860a18dd4 service nova] [instance: fd9e520a-2e7b-43d4-b84d-8933c32cc6a6] Received event network-changed-670f75e3-58f9-493b-9148-416e2e9cea5d {{(pid=61806) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 996.273951] env[61806]: DEBUG nova.compute.manager [req-c2d91fc2-4de8-4cfb-88b5-fcf63103ec86 req-66728731-80ed-43d5-99c8-553860a18dd4 service nova] [instance: fd9e520a-2e7b-43d4-b84d-8933c32cc6a6] Refreshing instance network info cache due to event network-changed-670f75e3-58f9-493b-9148-416e2e9cea5d. {{(pid=61806) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 996.274190] env[61806]: DEBUG oslo_concurrency.lockutils [req-c2d91fc2-4de8-4cfb-88b5-fcf63103ec86 req-66728731-80ed-43d5-99c8-553860a18dd4 service nova] Acquiring lock "refresh_cache-fd9e520a-2e7b-43d4-b84d-8933c32cc6a6" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 996.274342] env[61806]: DEBUG oslo_concurrency.lockutils [req-c2d91fc2-4de8-4cfb-88b5-fcf63103ec86 req-66728731-80ed-43d5-99c8-553860a18dd4 service nova] Acquired lock "refresh_cache-fd9e520a-2e7b-43d4-b84d-8933c32cc6a6" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 996.274536] env[61806]: DEBUG nova.network.neutron [req-c2d91fc2-4de8-4cfb-88b5-fcf63103ec86 req-66728731-80ed-43d5-99c8-553860a18dd4 service nova] [instance: fd9e520a-2e7b-43d4-b84d-8933c32cc6a6] Refreshing network info cache for port 670f75e3-58f9-493b-9148-416e2e9cea5d {{(pid=61806) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 996.415936] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-63c41608-b252-4371-b640-15b9305cbab4 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 996.447247] env[61806]: DEBUG nova.scheduler.client.report [None req-907fa7f6-24e7-45d8-9e7f-b2cf924bf805 tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Inventory has not changed for provider a2858be1-fd22-4e08-979e-87ad25293407 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 138, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61806) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 996.451089] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-d41c382b-3631-43f7-9094-5acd8818b680 tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] [instance: 67d1bdd8-cf8f-4fb5-bd3c-42008e4d60e3] Updating instance '67d1bdd8-cf8f-4fb5-bd3c-42008e4d60e3' progress to 0 {{(pid=61806) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 996.549634] env[61806]: DEBUG nova.network.neutron [None req-1c35eaa5-2854-4e3e-bc2c-4ef5a33b9a12 tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] [instance: 9357e61f-1628-43bd-ab46-de13c1529f51] Updating instance_info_cache with network_info: [{"id": "a4ca7c56-b6a1-4bd2-bd47-02b0ed4aa097", "address": "fa:16:3e:66:09:0d", "network": {"id": "46515a4e-610f-42fb-a424-205a03df0640", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-1133623547-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4b9d94c3d684474b99e305fb6cd9b95f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "82ca17df-257e-40e6-9ec9-310ed6f05ccb", "external-id": "nsx-vlan-transportzone-778", "segmentation_id": 778, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa4ca7c56-b6", "ovs_interfaceid": "a4ca7c56-b6a1-4bd2-bd47-02b0ed4aa097", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61806) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 996.957511] env[61806]: DEBUG oslo_concurrency.lockutils [None req-907fa7f6-24e7-45d8-9e7f-b2cf924bf805 tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.764s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 996.960873] env[61806]: DEBUG oslo_concurrency.lockutils [None req-b4773569-1b9a-41b1-9eb3-234618dc8825 tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 5.077s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 996.962794] env[61806]: INFO nova.compute.claims [None req-b4773569-1b9a-41b1-9eb3-234618dc8825 tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] [instance: fe1d4ace-7de3-423b-b423-9d7397aa2413] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 996.968635] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [None req-d41c382b-3631-43f7-9094-5acd8818b680 tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] [instance: 67d1bdd8-cf8f-4fb5-bd3c-42008e4d60e3] Powering off the VM {{(pid=61806) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 996.969678] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-7cbb27af-940a-4ace-80af-683291023cff {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 996.978156] env[61806]: DEBUG oslo_vmware.api [None req-d41c382b-3631-43f7-9094-5acd8818b680 tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Waiting for the task: (returnval){ [ 996.978156] env[61806]: value = "task-1295058" [ 996.978156] env[61806]: _type = "Task" [ 996.978156] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 996.988379] env[61806]: DEBUG oslo_vmware.api [None req-d41c382b-3631-43f7-9094-5acd8818b680 tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Task: {'id': task-1295058, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 996.991214] env[61806]: INFO nova.scheduler.client.report [None req-907fa7f6-24e7-45d8-9e7f-b2cf924bf805 tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Deleted allocations for instance 962d3d88-03ee-4ff2-9d79-15091671b330 [ 997.046499] env[61806]: DEBUG nova.network.neutron [req-c2d91fc2-4de8-4cfb-88b5-fcf63103ec86 req-66728731-80ed-43d5-99c8-553860a18dd4 service nova] [instance: fd9e520a-2e7b-43d4-b84d-8933c32cc6a6] Updated VIF entry in instance network info cache for port 670f75e3-58f9-493b-9148-416e2e9cea5d. {{(pid=61806) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 997.046905] env[61806]: DEBUG nova.network.neutron [req-c2d91fc2-4de8-4cfb-88b5-fcf63103ec86 req-66728731-80ed-43d5-99c8-553860a18dd4 service nova] [instance: fd9e520a-2e7b-43d4-b84d-8933c32cc6a6] Updating instance_info_cache with network_info: [{"id": "670f75e3-58f9-493b-9148-416e2e9cea5d", "address": "fa:16:3e:76:be:e1", "network": {"id": "a9d993bc-0817-4ee6-ab6e-1ef9432a110f", "bridge": null, "label": "tempest-AttachVolumeShelveTestJSON-939246736-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.226", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "bbbb09a49203421d807f733616c854df", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "unbound", "details": {}, "devname": "tap670f75e3-58", "ovs_interfaceid": null, "qbh_params": null, "qbg_params": null, "active": false, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61806) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 997.051231] env[61806]: DEBUG oslo_concurrency.lockutils [None req-1c35eaa5-2854-4e3e-bc2c-4ef5a33b9a12 tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Releasing lock "refresh_cache-9357e61f-1628-43bd-ab46-de13c1529f51" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 997.493393] env[61806]: DEBUG oslo_vmware.api [None req-d41c382b-3631-43f7-9094-5acd8818b680 tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Task: {'id': task-1295058, 'name': PowerOffVM_Task, 'duration_secs': 0.299684} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 997.493720] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [None req-d41c382b-3631-43f7-9094-5acd8818b680 tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] [instance: 67d1bdd8-cf8f-4fb5-bd3c-42008e4d60e3] Powered off the VM {{(pid=61806) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 997.493925] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-d41c382b-3631-43f7-9094-5acd8818b680 tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] [instance: 67d1bdd8-cf8f-4fb5-bd3c-42008e4d60e3] Updating instance '67d1bdd8-cf8f-4fb5-bd3c-42008e4d60e3' progress to 17 {{(pid=61806) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 997.505133] env[61806]: DEBUG oslo_concurrency.lockutils [None req-907fa7f6-24e7-45d8-9e7f-b2cf924bf805 tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Lock "962d3d88-03ee-4ff2-9d79-15091671b330" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 8.586s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 997.551168] env[61806]: DEBUG oslo_concurrency.lockutils [req-c2d91fc2-4de8-4cfb-88b5-fcf63103ec86 req-66728731-80ed-43d5-99c8-553860a18dd4 service nova] Releasing lock "refresh_cache-fd9e520a-2e7b-43d4-b84d-8933c32cc6a6" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 997.616797] env[61806]: DEBUG oslo_concurrency.lockutils [None req-de9a979c-45ed-463a-871e-b7dda8c366a2 tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] Acquiring lock "fd9e520a-2e7b-43d4-b84d-8933c32cc6a6" by "nova.compute.manager.ComputeManager.unshelve_instance..do_unshelve_instance" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 998.009025] env[61806]: DEBUG nova.virt.hardware [None req-d41c382b-3631-43f7-9094-5acd8818b680 tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-15T18:05:12Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='84',id=12,is_public=True,memory_mb=256,name='m1.micro',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=61806) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 998.009025] env[61806]: DEBUG nova.virt.hardware [None req-d41c382b-3631-43f7-9094-5acd8818b680 tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Flavor limits 0:0:0 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 998.009025] env[61806]: DEBUG nova.virt.hardware [None req-d41c382b-3631-43f7-9094-5acd8818b680 tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Image limits 0:0:0 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 998.009025] env[61806]: DEBUG nova.virt.hardware [None req-d41c382b-3631-43f7-9094-5acd8818b680 tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Flavor pref 0:0:0 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 998.009025] env[61806]: DEBUG nova.virt.hardware [None req-d41c382b-3631-43f7-9094-5acd8818b680 tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Image pref 0:0:0 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 998.009651] env[61806]: DEBUG nova.virt.hardware [None req-d41c382b-3631-43f7-9094-5acd8818b680 tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 998.010498] env[61806]: DEBUG nova.virt.hardware [None req-d41c382b-3631-43f7-9094-5acd8818b680 tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61806) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 998.010498] env[61806]: DEBUG nova.virt.hardware [None req-d41c382b-3631-43f7-9094-5acd8818b680 tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61806) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 998.010857] env[61806]: DEBUG nova.virt.hardware [None req-d41c382b-3631-43f7-9094-5acd8818b680 tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Got 1 possible topologies {{(pid=61806) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 998.011849] env[61806]: DEBUG nova.virt.hardware [None req-d41c382b-3631-43f7-9094-5acd8818b680 tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61806) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 998.011849] env[61806]: DEBUG nova.virt.hardware [None req-d41c382b-3631-43f7-9094-5acd8818b680 tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61806) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 998.020311] env[61806]: DEBUG oslo_concurrency.lockutils [None req-2bebc059-5179-46cc-9f09-746fe48209df tempest-ServersNegativeTestJSON-891856253 tempest-ServersNegativeTestJSON-891856253-project-member] Acquiring lock "bdbaef31-7d37-40cb-ac3e-ad152b37e2d2" by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 998.020728] env[61806]: DEBUG oslo_concurrency.lockutils [None req-2bebc059-5179-46cc-9f09-746fe48209df tempest-ServersNegativeTestJSON-891856253 tempest-ServersNegativeTestJSON-891856253-project-member] Lock "bdbaef31-7d37-40cb-ac3e-ad152b37e2d2" acquired by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" :: waited 0.001s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 998.021166] env[61806]: INFO nova.compute.manager [None req-2bebc059-5179-46cc-9f09-746fe48209df tempest-ServersNegativeTestJSON-891856253 tempest-ServersNegativeTestJSON-891856253-project-member] [instance: bdbaef31-7d37-40cb-ac3e-ad152b37e2d2] Shelving [ 998.023348] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-e5c6df89-30c4-46b2-960b-4fdeb641a0a9 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 998.043541] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [None req-2bebc059-5179-46cc-9f09-746fe48209df tempest-ServersNegativeTestJSON-891856253 tempest-ServersNegativeTestJSON-891856253-project-member] [instance: bdbaef31-7d37-40cb-ac3e-ad152b37e2d2] Powering off the VM {{(pid=61806) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 998.044175] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-b9a7ae87-21cf-4633-80bc-9567b14f171e {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 998.053019] env[61806]: DEBUG oslo_vmware.api [None req-2bebc059-5179-46cc-9f09-746fe48209df tempest-ServersNegativeTestJSON-891856253 tempest-ServersNegativeTestJSON-891856253-project-member] Waiting for the task: (returnval){ [ 998.053019] env[61806]: value = "task-1295060" [ 998.053019] env[61806]: _type = "Task" [ 998.053019] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 998.053019] env[61806]: DEBUG oslo_vmware.api [None req-d41c382b-3631-43f7-9094-5acd8818b680 tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Waiting for the task: (returnval){ [ 998.053019] env[61806]: value = "task-1295059" [ 998.053019] env[61806]: _type = "Task" [ 998.053019] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 998.066567] env[61806]: DEBUG oslo_vmware.api [None req-2bebc059-5179-46cc-9f09-746fe48209df tempest-ServersNegativeTestJSON-891856253 tempest-ServersNegativeTestJSON-891856253-project-member] Task: {'id': task-1295060, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 998.072081] env[61806]: DEBUG oslo_vmware.api [None req-d41c382b-3631-43f7-9094-5acd8818b680 tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Task: {'id': task-1295059, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 998.215540] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2509922e-2335-441b-95bd-f4526ab94d91 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 998.223355] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-335498f3-feb5-4d6d-9c8a-faf8a1a95fd0 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 998.255719] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a09282ca-a0f4-426f-9bfa-3f37c40be192 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 998.267017] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8719ba6d-837c-4510-b73a-99d5479c3a37 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 998.281201] env[61806]: DEBUG nova.compute.provider_tree [None req-b4773569-1b9a-41b1-9eb3-234618dc8825 tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] Inventory has not changed in ProviderTree for provider: a2858be1-fd22-4e08-979e-87ad25293407 {{(pid=61806) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 998.497221] env[61806]: DEBUG oslo_concurrency.lockutils [None req-8755c097-96dd-4051-93b1-330ffc947718 tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Acquiring lock "71999b04-bf6e-4e4c-87c9-d5ec1932a80f" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 998.497445] env[61806]: DEBUG oslo_concurrency.lockutils [None req-8755c097-96dd-4051-93b1-330ffc947718 tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Lock "71999b04-bf6e-4e4c-87c9-d5ec1932a80f" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 998.565043] env[61806]: DEBUG oslo_vmware.api [None req-2bebc059-5179-46cc-9f09-746fe48209df tempest-ServersNegativeTestJSON-891856253 tempest-ServersNegativeTestJSON-891856253-project-member] Task: {'id': task-1295060, 'name': PowerOffVM_Task, 'duration_secs': 0.253959} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 998.567954] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [None req-2bebc059-5179-46cc-9f09-746fe48209df tempest-ServersNegativeTestJSON-891856253 tempest-ServersNegativeTestJSON-891856253-project-member] [instance: bdbaef31-7d37-40cb-ac3e-ad152b37e2d2] Powered off the VM {{(pid=61806) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 998.570061] env[61806]: DEBUG oslo_vmware.api [None req-d41c382b-3631-43f7-9094-5acd8818b680 tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Task: {'id': task-1295059, 'name': ReconfigVM_Task, 'duration_secs': 0.227016} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 998.571145] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-844cb113-f985-45a9-9615-921604f61143 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 998.573512] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-d41c382b-3631-43f7-9094-5acd8818b680 tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] [instance: 67d1bdd8-cf8f-4fb5-bd3c-42008e4d60e3] Updating instance '67d1bdd8-cf8f-4fb5-bd3c-42008e4d60e3' progress to 33 {{(pid=61806) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 998.577313] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1442ac08-3666-4760-a486-d7a6444e23dd {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 998.594719] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-1c35eaa5-2854-4e3e-bc2c-4ef5a33b9a12 tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] [instance: 9357e61f-1628-43bd-ab46-de13c1529f51] Updating instance '9357e61f-1628-43bd-ab46-de13c1529f51' progress to 0 {{(pid=61806) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 998.612715] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f2d0f092-9758-4d9a-b3c3-eca4697e2406 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 998.783871] env[61806]: DEBUG nova.scheduler.client.report [None req-b4773569-1b9a-41b1-9eb3-234618dc8825 tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] Inventory has not changed for provider a2858be1-fd22-4e08-979e-87ad25293407 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 138, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61806) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 998.796429] env[61806]: DEBUG oslo_concurrency.lockutils [None req-4b7244a6-0e60-4b9b-ac2e-5c7b0ff2b8c7 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Acquiring lock "d0669232-6f33-4b30-97c2-2e4239af1ad0" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 998.796676] env[61806]: DEBUG oslo_concurrency.lockutils [None req-4b7244a6-0e60-4b9b-ac2e-5c7b0ff2b8c7 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Lock "d0669232-6f33-4b30-97c2-2e4239af1ad0" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 998.999781] env[61806]: DEBUG nova.compute.manager [None req-8755c097-96dd-4051-93b1-330ffc947718 tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] [instance: 71999b04-bf6e-4e4c-87c9-d5ec1932a80f] Starting instance... {{(pid=61806) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 999.082560] env[61806]: DEBUG nova.virt.hardware [None req-d41c382b-3631-43f7-9094-5acd8818b680 tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-15T18:05:10Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=61806) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 999.082877] env[61806]: DEBUG nova.virt.hardware [None req-d41c382b-3631-43f7-9094-5acd8818b680 tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Flavor limits 0:0:0 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 999.083072] env[61806]: DEBUG nova.virt.hardware [None req-d41c382b-3631-43f7-9094-5acd8818b680 tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Image limits 0:0:0 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 999.083299] env[61806]: DEBUG nova.virt.hardware [None req-d41c382b-3631-43f7-9094-5acd8818b680 tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Flavor pref 0:0:0 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 999.083478] env[61806]: DEBUG nova.virt.hardware [None req-d41c382b-3631-43f7-9094-5acd8818b680 tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Image pref 0:0:0 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 999.083641] env[61806]: DEBUG nova.virt.hardware [None req-d41c382b-3631-43f7-9094-5acd8818b680 tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 999.083877] env[61806]: DEBUG nova.virt.hardware [None req-d41c382b-3631-43f7-9094-5acd8818b680 tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61806) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 999.084065] env[61806]: DEBUG nova.virt.hardware [None req-d41c382b-3631-43f7-9094-5acd8818b680 tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61806) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 999.084246] env[61806]: DEBUG nova.virt.hardware [None req-d41c382b-3631-43f7-9094-5acd8818b680 tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Got 1 possible topologies {{(pid=61806) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 999.084416] env[61806]: DEBUG nova.virt.hardware [None req-d41c382b-3631-43f7-9094-5acd8818b680 tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61806) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 999.084674] env[61806]: DEBUG nova.virt.hardware [None req-d41c382b-3631-43f7-9094-5acd8818b680 tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61806) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 999.090103] env[61806]: DEBUG nova.virt.vmwareapi.volumeops [None req-d41c382b-3631-43f7-9094-5acd8818b680 tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] [instance: 67d1bdd8-cf8f-4fb5-bd3c-42008e4d60e3] Reconfiguring VM instance instance-00000041 to detach disk 2000 {{(pid=61806) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 999.090399] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-f3b13b63-62d3-473c-993f-7d3f769a6d16 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 999.111142] env[61806]: DEBUG oslo_vmware.api [None req-d41c382b-3631-43f7-9094-5acd8818b680 tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Waiting for the task: (returnval){ [ 999.111142] env[61806]: value = "task-1295061" [ 999.111142] env[61806]: _type = "Task" [ 999.111142] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 999.115912] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [None req-1c35eaa5-2854-4e3e-bc2c-4ef5a33b9a12 tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] [instance: 9357e61f-1628-43bd-ab46-de13c1529f51] Powering off the VM {{(pid=61806) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 999.116195] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-16897e8c-e147-4c00-8e45-83ad40b5726c {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 999.124300] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-2bebc059-5179-46cc-9f09-746fe48209df tempest-ServersNegativeTestJSON-891856253 tempest-ServersNegativeTestJSON-891856253-project-member] [instance: bdbaef31-7d37-40cb-ac3e-ad152b37e2d2] Creating Snapshot of the VM instance {{(pid=61806) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 999.124638] env[61806]: DEBUG oslo_vmware.api [None req-d41c382b-3631-43f7-9094-5acd8818b680 tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Task: {'id': task-1295061, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 999.125926] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-1f7970a4-0040-4387-9032-731a24bf206f {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 999.127849] env[61806]: DEBUG oslo_vmware.api [None req-1c35eaa5-2854-4e3e-bc2c-4ef5a33b9a12 tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Waiting for the task: (returnval){ [ 999.127849] env[61806]: value = "task-1295062" [ 999.127849] env[61806]: _type = "Task" [ 999.127849] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 999.133161] env[61806]: DEBUG oslo_vmware.api [None req-2bebc059-5179-46cc-9f09-746fe48209df tempest-ServersNegativeTestJSON-891856253 tempest-ServersNegativeTestJSON-891856253-project-member] Waiting for the task: (returnval){ [ 999.133161] env[61806]: value = "task-1295063" [ 999.133161] env[61806]: _type = "Task" [ 999.133161] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 999.139074] env[61806]: DEBUG oslo_vmware.api [None req-1c35eaa5-2854-4e3e-bc2c-4ef5a33b9a12 tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Task: {'id': task-1295062, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 999.146494] env[61806]: DEBUG oslo_vmware.api [None req-2bebc059-5179-46cc-9f09-746fe48209df tempest-ServersNegativeTestJSON-891856253 tempest-ServersNegativeTestJSON-891856253-project-member] Task: {'id': task-1295063, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 999.289294] env[61806]: DEBUG oslo_concurrency.lockutils [None req-b4773569-1b9a-41b1-9eb3-234618dc8825 tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.328s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 999.290510] env[61806]: DEBUG nova.compute.manager [None req-b4773569-1b9a-41b1-9eb3-234618dc8825 tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] [instance: fe1d4ace-7de3-423b-b423-9d7397aa2413] Start building networks asynchronously for instance. {{(pid=61806) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 999.292735] env[61806]: DEBUG oslo_concurrency.lockutils [None req-b6079056-6464-46e6-97ab-343b76917c18 tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 3.787s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 999.292933] env[61806]: DEBUG nova.objects.instance [None req-b6079056-6464-46e6-97ab-343b76917c18 tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] Lazy-loading 'resources' on Instance uuid fd9e520a-2e7b-43d4-b84d-8933c32cc6a6 {{(pid=61806) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 999.299041] env[61806]: DEBUG nova.compute.manager [None req-4b7244a6-0e60-4b9b-ac2e-5c7b0ff2b8c7 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] [instance: d0669232-6f33-4b30-97c2-2e4239af1ad0] Starting instance... {{(pid=61806) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 999.529333] env[61806]: DEBUG oslo_concurrency.lockutils [None req-8755c097-96dd-4051-93b1-330ffc947718 tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 999.621232] env[61806]: DEBUG oslo_vmware.api [None req-d41c382b-3631-43f7-9094-5acd8818b680 tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Task: {'id': task-1295061, 'name': ReconfigVM_Task, 'duration_secs': 0.235951} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 999.621560] env[61806]: DEBUG nova.virt.vmwareapi.volumeops [None req-d41c382b-3631-43f7-9094-5acd8818b680 tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] [instance: 67d1bdd8-cf8f-4fb5-bd3c-42008e4d60e3] Reconfigured VM instance instance-00000041 to detach disk 2000 {{(pid=61806) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 999.622299] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-04efa36e-bf88-4f4f-9df9-4b44763c4f1d {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 999.645784] env[61806]: DEBUG nova.virt.vmwareapi.volumeops [None req-d41c382b-3631-43f7-9094-5acd8818b680 tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] [instance: 67d1bdd8-cf8f-4fb5-bd3c-42008e4d60e3] Reconfiguring VM instance instance-00000041 to attach disk [datastore1] 67d1bdd8-cf8f-4fb5-bd3c-42008e4d60e3/67d1bdd8-cf8f-4fb5-bd3c-42008e4d60e3.vmdk or device None with type thin {{(pid=61806) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 999.651429] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-1862a393-b3ba-45e7-8011-d948f1d8d1cb {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 999.670082] env[61806]: DEBUG oslo_vmware.api [None req-2bebc059-5179-46cc-9f09-746fe48209df tempest-ServersNegativeTestJSON-891856253 tempest-ServersNegativeTestJSON-891856253-project-member] Task: {'id': task-1295063, 'name': CreateSnapshot_Task} progress is 100%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 999.673856] env[61806]: DEBUG oslo_vmware.api [None req-d41c382b-3631-43f7-9094-5acd8818b680 tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Waiting for the task: (returnval){ [ 999.673856] env[61806]: value = "task-1295064" [ 999.673856] env[61806]: _type = "Task" [ 999.673856] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 999.674107] env[61806]: DEBUG oslo_vmware.api [None req-1c35eaa5-2854-4e3e-bc2c-4ef5a33b9a12 tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Task: {'id': task-1295062, 'name': PowerOffVM_Task, 'duration_secs': 0.220855} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 999.674406] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [None req-1c35eaa5-2854-4e3e-bc2c-4ef5a33b9a12 tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] [instance: 9357e61f-1628-43bd-ab46-de13c1529f51] Powered off the VM {{(pid=61806) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 999.674618] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-1c35eaa5-2854-4e3e-bc2c-4ef5a33b9a12 tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] [instance: 9357e61f-1628-43bd-ab46-de13c1529f51] Updating instance '9357e61f-1628-43bd-ab46-de13c1529f51' progress to 17 {{(pid=61806) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 999.686811] env[61806]: DEBUG oslo_vmware.api [None req-d41c382b-3631-43f7-9094-5acd8818b680 tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Task: {'id': task-1295064, 'name': ReconfigVM_Task} progress is 10%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 999.796063] env[61806]: DEBUG nova.compute.utils [None req-b4773569-1b9a-41b1-9eb3-234618dc8825 tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] Using /dev/sd instead of None {{(pid=61806) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 999.797664] env[61806]: DEBUG nova.objects.instance [None req-b6079056-6464-46e6-97ab-343b76917c18 tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] Lazy-loading 'numa_topology' on Instance uuid fd9e520a-2e7b-43d4-b84d-8933c32cc6a6 {{(pid=61806) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 999.798729] env[61806]: DEBUG nova.compute.manager [None req-b4773569-1b9a-41b1-9eb3-234618dc8825 tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] [instance: fe1d4ace-7de3-423b-b423-9d7397aa2413] Allocating IP information in the background. {{(pid=61806) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 999.798906] env[61806]: DEBUG nova.network.neutron [None req-b4773569-1b9a-41b1-9eb3-234618dc8825 tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] [instance: fe1d4ace-7de3-423b-b423-9d7397aa2413] allocate_for_instance() {{(pid=61806) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 999.819985] env[61806]: DEBUG oslo_concurrency.lockutils [None req-4b7244a6-0e60-4b9b-ac2e-5c7b0ff2b8c7 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 999.852763] env[61806]: DEBUG nova.policy [None req-b4773569-1b9a-41b1-9eb3-234618dc8825 tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '4c5d4c6e1ad844a585c08c4644a60bbf', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '73ebcc4d79d248efb653a084a5e44302', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61806) authorize /opt/stack/nova/nova/policy.py:201}} [ 1000.117877] env[61806]: DEBUG nova.network.neutron [None req-b4773569-1b9a-41b1-9eb3-234618dc8825 tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] [instance: fe1d4ace-7de3-423b-b423-9d7397aa2413] Successfully created port: b2369483-32e0-4f80-be14-b45402c39094 {{(pid=61806) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1000.154107] env[61806]: DEBUG oslo_vmware.api [None req-2bebc059-5179-46cc-9f09-746fe48209df tempest-ServersNegativeTestJSON-891856253 tempest-ServersNegativeTestJSON-891856253-project-member] Task: {'id': task-1295063, 'name': CreateSnapshot_Task, 'duration_secs': 0.583745} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1000.154474] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-2bebc059-5179-46cc-9f09-746fe48209df tempest-ServersNegativeTestJSON-891856253 tempest-ServersNegativeTestJSON-891856253-project-member] [instance: bdbaef31-7d37-40cb-ac3e-ad152b37e2d2] Created Snapshot of the VM instance {{(pid=61806) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 1000.155380] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5c953f39-2963-4573-bf92-1d03ee9a4f55 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1000.187210] env[61806]: DEBUG nova.virt.hardware [None req-1c35eaa5-2854-4e3e-bc2c-4ef5a33b9a12 tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-15T18:05:12Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='84',id=12,is_public=True,memory_mb=256,name='m1.micro',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=61806) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1000.187468] env[61806]: DEBUG nova.virt.hardware [None req-1c35eaa5-2854-4e3e-bc2c-4ef5a33b9a12 tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Flavor limits 0:0:0 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1000.187637] env[61806]: DEBUG nova.virt.hardware [None req-1c35eaa5-2854-4e3e-bc2c-4ef5a33b9a12 tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Image limits 0:0:0 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1000.187828] env[61806]: DEBUG nova.virt.hardware [None req-1c35eaa5-2854-4e3e-bc2c-4ef5a33b9a12 tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Flavor pref 0:0:0 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1000.187982] env[61806]: DEBUG nova.virt.hardware [None req-1c35eaa5-2854-4e3e-bc2c-4ef5a33b9a12 tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Image pref 0:0:0 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1000.188149] env[61806]: DEBUG nova.virt.hardware [None req-1c35eaa5-2854-4e3e-bc2c-4ef5a33b9a12 tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1000.188362] env[61806]: DEBUG nova.virt.hardware [None req-1c35eaa5-2854-4e3e-bc2c-4ef5a33b9a12 tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61806) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1000.188540] env[61806]: DEBUG nova.virt.hardware [None req-1c35eaa5-2854-4e3e-bc2c-4ef5a33b9a12 tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61806) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1000.188713] env[61806]: DEBUG nova.virt.hardware [None req-1c35eaa5-2854-4e3e-bc2c-4ef5a33b9a12 tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Got 1 possible topologies {{(pid=61806) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1000.188912] env[61806]: DEBUG nova.virt.hardware [None req-1c35eaa5-2854-4e3e-bc2c-4ef5a33b9a12 tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61806) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1000.189104] env[61806]: DEBUG nova.virt.hardware [None req-1c35eaa5-2854-4e3e-bc2c-4ef5a33b9a12 tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61806) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1000.194149] env[61806]: DEBUG oslo_vmware.api [None req-d41c382b-3631-43f7-9094-5acd8818b680 tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Task: {'id': task-1295064, 'name': ReconfigVM_Task, 'duration_secs': 0.288987} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1000.194401] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-4ad441b6-02ce-4ff7-89c8-6f5ec18f008f {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1000.205456] env[61806]: DEBUG nova.virt.vmwareapi.volumeops [None req-d41c382b-3631-43f7-9094-5acd8818b680 tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] [instance: 67d1bdd8-cf8f-4fb5-bd3c-42008e4d60e3] Reconfigured VM instance instance-00000041 to attach disk [datastore1] 67d1bdd8-cf8f-4fb5-bd3c-42008e4d60e3/67d1bdd8-cf8f-4fb5-bd3c-42008e4d60e3.vmdk or device None with type thin {{(pid=61806) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1000.205713] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-d41c382b-3631-43f7-9094-5acd8818b680 tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] [instance: 67d1bdd8-cf8f-4fb5-bd3c-42008e4d60e3] Updating instance '67d1bdd8-cf8f-4fb5-bd3c-42008e4d60e3' progress to 50 {{(pid=61806) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 1000.218172] env[61806]: DEBUG oslo_vmware.api [None req-1c35eaa5-2854-4e3e-bc2c-4ef5a33b9a12 tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Waiting for the task: (returnval){ [ 1000.218172] env[61806]: value = "task-1295065" [ 1000.218172] env[61806]: _type = "Task" [ 1000.218172] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1000.227204] env[61806]: DEBUG oslo_vmware.api [None req-1c35eaa5-2854-4e3e-bc2c-4ef5a33b9a12 tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Task: {'id': task-1295065, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1000.299694] env[61806]: DEBUG nova.compute.manager [None req-b4773569-1b9a-41b1-9eb3-234618dc8825 tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] [instance: fe1d4ace-7de3-423b-b423-9d7397aa2413] Start building block device mappings for instance. {{(pid=61806) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1000.302721] env[61806]: DEBUG nova.objects.base [None req-b6079056-6464-46e6-97ab-343b76917c18 tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] Object Instance lazy-loaded attributes: resources,numa_topology {{(pid=61806) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 1000.456368] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6ad1b2f2-b294-414f-aea8-423a27a2b5b9 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1000.464898] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-67bd3775-f308-4b7e-9dd0-4c4bf78c5baf {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1000.499702] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8257c4b4-e9b9-4b64-8273-6d9c6a73c7b3 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1000.509055] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-272d7d43-b6bb-462d-9bbe-7f33b504b07d {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1000.523048] env[61806]: DEBUG nova.compute.provider_tree [None req-b6079056-6464-46e6-97ab-343b76917c18 tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] Inventory has not changed in ProviderTree for provider: a2858be1-fd22-4e08-979e-87ad25293407 {{(pid=61806) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1000.674611] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-2bebc059-5179-46cc-9f09-746fe48209df tempest-ServersNegativeTestJSON-891856253 tempest-ServersNegativeTestJSON-891856253-project-member] [instance: bdbaef31-7d37-40cb-ac3e-ad152b37e2d2] Creating linked-clone VM from snapshot {{(pid=61806) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 1000.675041] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-74612a43-be11-46c7-b7aa-9b2e648ae747 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1000.684173] env[61806]: DEBUG oslo_vmware.api [None req-2bebc059-5179-46cc-9f09-746fe48209df tempest-ServersNegativeTestJSON-891856253 tempest-ServersNegativeTestJSON-891856253-project-member] Waiting for the task: (returnval){ [ 1000.684173] env[61806]: value = "task-1295066" [ 1000.684173] env[61806]: _type = "Task" [ 1000.684173] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1000.692464] env[61806]: DEBUG oslo_vmware.api [None req-2bebc059-5179-46cc-9f09-746fe48209df tempest-ServersNegativeTestJSON-891856253 tempest-ServersNegativeTestJSON-891856253-project-member] Task: {'id': task-1295066, 'name': CloneVM_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1000.711989] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-71acaaea-e084-4ed5-ae96-19c76cc4445e {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1000.733900] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8330d09c-b320-4392-8085-ab41ceeca84b {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1000.742531] env[61806]: DEBUG oslo_vmware.api [None req-1c35eaa5-2854-4e3e-bc2c-4ef5a33b9a12 tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Task: {'id': task-1295065, 'name': ReconfigVM_Task, 'duration_secs': 0.216843} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1000.754644] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-1c35eaa5-2854-4e3e-bc2c-4ef5a33b9a12 tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] [instance: 9357e61f-1628-43bd-ab46-de13c1529f51] Updating instance '9357e61f-1628-43bd-ab46-de13c1529f51' progress to 33 {{(pid=61806) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 1000.758568] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-d41c382b-3631-43f7-9094-5acd8818b680 tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] [instance: 67d1bdd8-cf8f-4fb5-bd3c-42008e4d60e3] Updating instance '67d1bdd8-cf8f-4fb5-bd3c-42008e4d60e3' progress to 67 {{(pid=61806) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 1001.026293] env[61806]: DEBUG nova.scheduler.client.report [None req-b6079056-6464-46e6-97ab-343b76917c18 tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] Inventory has not changed for provider a2858be1-fd22-4e08-979e-87ad25293407 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 138, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61806) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1001.194770] env[61806]: DEBUG oslo_vmware.api [None req-2bebc059-5179-46cc-9f09-746fe48209df tempest-ServersNegativeTestJSON-891856253 tempest-ServersNegativeTestJSON-891856253-project-member] Task: {'id': task-1295066, 'name': CloneVM_Task} progress is 94%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1001.268443] env[61806]: DEBUG nova.virt.hardware [None req-1c35eaa5-2854-4e3e-bc2c-4ef5a33b9a12 tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-15T18:05:10Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=61806) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1001.268673] env[61806]: DEBUG nova.virt.hardware [None req-1c35eaa5-2854-4e3e-bc2c-4ef5a33b9a12 tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Flavor limits 0:0:0 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1001.268673] env[61806]: DEBUG nova.virt.hardware [None req-1c35eaa5-2854-4e3e-bc2c-4ef5a33b9a12 tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Image limits 0:0:0 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1001.268806] env[61806]: DEBUG nova.virt.hardware [None req-1c35eaa5-2854-4e3e-bc2c-4ef5a33b9a12 tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Flavor pref 0:0:0 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1001.268962] env[61806]: DEBUG nova.virt.hardware [None req-1c35eaa5-2854-4e3e-bc2c-4ef5a33b9a12 tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Image pref 0:0:0 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1001.269133] env[61806]: DEBUG nova.virt.hardware [None req-1c35eaa5-2854-4e3e-bc2c-4ef5a33b9a12 tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1001.269347] env[61806]: DEBUG nova.virt.hardware [None req-1c35eaa5-2854-4e3e-bc2c-4ef5a33b9a12 tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61806) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1001.269514] env[61806]: DEBUG nova.virt.hardware [None req-1c35eaa5-2854-4e3e-bc2c-4ef5a33b9a12 tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61806) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1001.269683] env[61806]: DEBUG nova.virt.hardware [None req-1c35eaa5-2854-4e3e-bc2c-4ef5a33b9a12 tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Got 1 possible topologies {{(pid=61806) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1001.269910] env[61806]: DEBUG nova.virt.hardware [None req-1c35eaa5-2854-4e3e-bc2c-4ef5a33b9a12 tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61806) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1001.270163] env[61806]: DEBUG nova.virt.hardware [None req-1c35eaa5-2854-4e3e-bc2c-4ef5a33b9a12 tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61806) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1001.275420] env[61806]: DEBUG nova.virt.vmwareapi.volumeops [None req-1c35eaa5-2854-4e3e-bc2c-4ef5a33b9a12 tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] [instance: 9357e61f-1628-43bd-ab46-de13c1529f51] Reconfiguring VM instance instance-00000060 to detach disk 2000 {{(pid=61806) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1001.275726] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-a85b187a-3fe6-46e2-b3f4-25e5fe2b5f2b {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1001.295019] env[61806]: DEBUG oslo_vmware.api [None req-1c35eaa5-2854-4e3e-bc2c-4ef5a33b9a12 tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Waiting for the task: (returnval){ [ 1001.295019] env[61806]: value = "task-1295067" [ 1001.295019] env[61806]: _type = "Task" [ 1001.295019] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1001.303606] env[61806]: DEBUG oslo_vmware.api [None req-1c35eaa5-2854-4e3e-bc2c-4ef5a33b9a12 tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Task: {'id': task-1295067, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1001.311807] env[61806]: DEBUG nova.compute.manager [None req-b4773569-1b9a-41b1-9eb3-234618dc8825 tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] [instance: fe1d4ace-7de3-423b-b423-9d7397aa2413] Start spawning the instance on the hypervisor. {{(pid=61806) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1001.330190] env[61806]: DEBUG nova.network.neutron [None req-d41c382b-3631-43f7-9094-5acd8818b680 tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] [instance: 67d1bdd8-cf8f-4fb5-bd3c-42008e4d60e3] Port 859fb723-4a1c-4ea8-bd7c-0ed308f3a9e3 binding to destination host cpu-1 is already ACTIVE {{(pid=61806) migrate_instance_start /opt/stack/nova/nova/network/neutron.py:3171}} [ 1001.339725] env[61806]: DEBUG nova.virt.hardware [None req-b4773569-1b9a-41b1-9eb3-234618dc8825 tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-15T18:05:10Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-15T18:04:52Z,direct_url=,disk_format='vmdk',id=ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='84ba83fce2c349a6988173c5d6fc3b07',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-15T18:04:53Z,virtual_size=,visibility=), allow threads: False {{(pid=61806) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1001.340024] env[61806]: DEBUG nova.virt.hardware [None req-b4773569-1b9a-41b1-9eb3-234618dc8825 tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] Flavor limits 0:0:0 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1001.340219] env[61806]: DEBUG nova.virt.hardware [None req-b4773569-1b9a-41b1-9eb3-234618dc8825 tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] Image limits 0:0:0 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1001.340404] env[61806]: DEBUG nova.virt.hardware [None req-b4773569-1b9a-41b1-9eb3-234618dc8825 tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] Flavor pref 0:0:0 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1001.340557] env[61806]: DEBUG nova.virt.hardware [None req-b4773569-1b9a-41b1-9eb3-234618dc8825 tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] Image pref 0:0:0 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1001.340711] env[61806]: DEBUG nova.virt.hardware [None req-b4773569-1b9a-41b1-9eb3-234618dc8825 tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1001.340944] env[61806]: DEBUG nova.virt.hardware [None req-b4773569-1b9a-41b1-9eb3-234618dc8825 tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61806) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1001.341140] env[61806]: DEBUG nova.virt.hardware [None req-b4773569-1b9a-41b1-9eb3-234618dc8825 tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61806) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1001.341393] env[61806]: DEBUG nova.virt.hardware [None req-b4773569-1b9a-41b1-9eb3-234618dc8825 tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] Got 1 possible topologies {{(pid=61806) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1001.341585] env[61806]: DEBUG nova.virt.hardware [None req-b4773569-1b9a-41b1-9eb3-234618dc8825 tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61806) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1001.341772] env[61806]: DEBUG nova.virt.hardware [None req-b4773569-1b9a-41b1-9eb3-234618dc8825 tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61806) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1001.342666] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eaea285a-c7f3-462b-a239-39d9a0a483fb {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1001.351925] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d7aa1e3b-5e0c-4267-8396-23cdbc96d36f {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1001.531736] env[61806]: DEBUG oslo_concurrency.lockutils [None req-b6079056-6464-46e6-97ab-343b76917c18 tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.239s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1001.539121] env[61806]: DEBUG oslo_concurrency.lockutils [None req-8755c097-96dd-4051-93b1-330ffc947718 tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 2.007s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1001.540695] env[61806]: INFO nova.compute.claims [None req-8755c097-96dd-4051-93b1-330ffc947718 tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] [instance: 71999b04-bf6e-4e4c-87c9-d5ec1932a80f] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1001.698577] env[61806]: DEBUG oslo_vmware.api [None req-2bebc059-5179-46cc-9f09-746fe48209df tempest-ServersNegativeTestJSON-891856253 tempest-ServersNegativeTestJSON-891856253-project-member] Task: {'id': task-1295066, 'name': CloneVM_Task} progress is 95%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1001.780714] env[61806]: DEBUG nova.compute.manager [req-274cb097-f02d-4db6-a660-8a630ff63598 req-59ed89a7-d187-4c60-a269-04b2758b77ab service nova] [instance: fe1d4ace-7de3-423b-b423-9d7397aa2413] Received event network-vif-plugged-b2369483-32e0-4f80-be14-b45402c39094 {{(pid=61806) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1001.781017] env[61806]: DEBUG oslo_concurrency.lockutils [req-274cb097-f02d-4db6-a660-8a630ff63598 req-59ed89a7-d187-4c60-a269-04b2758b77ab service nova] Acquiring lock "fe1d4ace-7de3-423b-b423-9d7397aa2413-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1001.781271] env[61806]: DEBUG oslo_concurrency.lockutils [req-274cb097-f02d-4db6-a660-8a630ff63598 req-59ed89a7-d187-4c60-a269-04b2758b77ab service nova] Lock "fe1d4ace-7de3-423b-b423-9d7397aa2413-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1001.781471] env[61806]: DEBUG oslo_concurrency.lockutils [req-274cb097-f02d-4db6-a660-8a630ff63598 req-59ed89a7-d187-4c60-a269-04b2758b77ab service nova] Lock "fe1d4ace-7de3-423b-b423-9d7397aa2413-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1001.781651] env[61806]: DEBUG nova.compute.manager [req-274cb097-f02d-4db6-a660-8a630ff63598 req-59ed89a7-d187-4c60-a269-04b2758b77ab service nova] [instance: fe1d4ace-7de3-423b-b423-9d7397aa2413] No waiting events found dispatching network-vif-plugged-b2369483-32e0-4f80-be14-b45402c39094 {{(pid=61806) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1001.781773] env[61806]: WARNING nova.compute.manager [req-274cb097-f02d-4db6-a660-8a630ff63598 req-59ed89a7-d187-4c60-a269-04b2758b77ab service nova] [instance: fe1d4ace-7de3-423b-b423-9d7397aa2413] Received unexpected event network-vif-plugged-b2369483-32e0-4f80-be14-b45402c39094 for instance with vm_state building and task_state spawning. [ 1001.805473] env[61806]: DEBUG oslo_vmware.api [None req-1c35eaa5-2854-4e3e-bc2c-4ef5a33b9a12 tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Task: {'id': task-1295067, 'name': ReconfigVM_Task, 'duration_secs': 0.288696} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1001.805770] env[61806]: DEBUG nova.virt.vmwareapi.volumeops [None req-1c35eaa5-2854-4e3e-bc2c-4ef5a33b9a12 tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] [instance: 9357e61f-1628-43bd-ab46-de13c1529f51] Reconfigured VM instance instance-00000060 to detach disk 2000 {{(pid=61806) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1001.807056] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-533a72ec-81aa-4c7c-8fdb-f91ade647441 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1001.829928] env[61806]: DEBUG nova.virt.vmwareapi.volumeops [None req-1c35eaa5-2854-4e3e-bc2c-4ef5a33b9a12 tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] [instance: 9357e61f-1628-43bd-ab46-de13c1529f51] Reconfiguring VM instance instance-00000060 to attach disk [datastore1] 9357e61f-1628-43bd-ab46-de13c1529f51/9357e61f-1628-43bd-ab46-de13c1529f51.vmdk or device None with type thin {{(pid=61806) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1001.830246] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-029ccb93-11f1-40d0-bd48-ec22603f7669 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1001.853340] env[61806]: DEBUG oslo_vmware.api [None req-1c35eaa5-2854-4e3e-bc2c-4ef5a33b9a12 tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Waiting for the task: (returnval){ [ 1001.853340] env[61806]: value = "task-1295068" [ 1001.853340] env[61806]: _type = "Task" [ 1001.853340] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1001.864300] env[61806]: DEBUG oslo_vmware.api [None req-1c35eaa5-2854-4e3e-bc2c-4ef5a33b9a12 tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Task: {'id': task-1295068, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1001.901086] env[61806]: DEBUG nova.network.neutron [None req-b4773569-1b9a-41b1-9eb3-234618dc8825 tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] [instance: fe1d4ace-7de3-423b-b423-9d7397aa2413] Successfully updated port: b2369483-32e0-4f80-be14-b45402c39094 {{(pid=61806) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1002.050265] env[61806]: DEBUG oslo_concurrency.lockutils [None req-b6079056-6464-46e6-97ab-343b76917c18 tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] Lock "fd9e520a-2e7b-43d4-b84d-8933c32cc6a6" "released" by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" :: held 25.099s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1002.051413] env[61806]: DEBUG oslo_concurrency.lockutils [None req-de9a979c-45ed-463a-871e-b7dda8c366a2 tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] Lock "fd9e520a-2e7b-43d4-b84d-8933c32cc6a6" acquired by "nova.compute.manager.ComputeManager.unshelve_instance..do_unshelve_instance" :: waited 4.435s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1002.051540] env[61806]: INFO nova.compute.manager [None req-de9a979c-45ed-463a-871e-b7dda8c366a2 tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] [instance: fd9e520a-2e7b-43d4-b84d-8933c32cc6a6] Unshelving [ 1002.195440] env[61806]: DEBUG oslo_vmware.api [None req-2bebc059-5179-46cc-9f09-746fe48209df tempest-ServersNegativeTestJSON-891856253 tempest-ServersNegativeTestJSON-891856253-project-member] Task: {'id': task-1295066, 'name': CloneVM_Task, 'duration_secs': 1.402676} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1002.195748] env[61806]: INFO nova.virt.vmwareapi.vmops [None req-2bebc059-5179-46cc-9f09-746fe48209df tempest-ServersNegativeTestJSON-891856253 tempest-ServersNegativeTestJSON-891856253-project-member] [instance: bdbaef31-7d37-40cb-ac3e-ad152b37e2d2] Created linked-clone VM from snapshot [ 1002.196524] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8583a149-b37c-4ebe-85ab-6c228c00e860 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1002.204521] env[61806]: DEBUG nova.virt.vmwareapi.images [None req-2bebc059-5179-46cc-9f09-746fe48209df tempest-ServersNegativeTestJSON-891856253 tempest-ServersNegativeTestJSON-891856253-project-member] [instance: bdbaef31-7d37-40cb-ac3e-ad152b37e2d2] Uploading image 20c8ea6e-0db5-4788-9e34-92c1d63bb7b8 {{(pid=61806) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:442}} [ 1002.224397] env[61806]: DEBUG oslo_vmware.rw_handles [None req-2bebc059-5179-46cc-9f09-746fe48209df tempest-ServersNegativeTestJSON-891856253 tempest-ServersNegativeTestJSON-891856253-project-member] Creating HttpNfcLease lease for exporting VM: (result){ [ 1002.224397] env[61806]: value = "vm-277737" [ 1002.224397] env[61806]: _type = "VirtualMachine" [ 1002.224397] env[61806]: }. {{(pid=61806) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:478}} [ 1002.224684] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ExportVm with opID=oslo.vmware-7f465147-5a87-4dea-8a06-fa6ca9992416 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1002.232709] env[61806]: DEBUG oslo_vmware.rw_handles [None req-2bebc059-5179-46cc-9f09-746fe48209df tempest-ServersNegativeTestJSON-891856253 tempest-ServersNegativeTestJSON-891856253-project-member] Lease: (returnval){ [ 1002.232709] env[61806]: value = "session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]5242c2b9-14c4-57ac-e4ab-e5ca38633943" [ 1002.232709] env[61806]: _type = "HttpNfcLease" [ 1002.232709] env[61806]: } obtained for exporting VM: (result){ [ 1002.232709] env[61806]: value = "vm-277737" [ 1002.232709] env[61806]: _type = "VirtualMachine" [ 1002.232709] env[61806]: }. {{(pid=61806) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:481}} [ 1002.233032] env[61806]: DEBUG oslo_vmware.api [None req-2bebc059-5179-46cc-9f09-746fe48209df tempest-ServersNegativeTestJSON-891856253 tempest-ServersNegativeTestJSON-891856253-project-member] Waiting for the lease: (returnval){ [ 1002.233032] env[61806]: value = "session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]5242c2b9-14c4-57ac-e4ab-e5ca38633943" [ 1002.233032] env[61806]: _type = "HttpNfcLease" [ 1002.233032] env[61806]: } to be ready. {{(pid=61806) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 1002.239593] env[61806]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1002.239593] env[61806]: value = "session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]5242c2b9-14c4-57ac-e4ab-e5ca38633943" [ 1002.239593] env[61806]: _type = "HttpNfcLease" [ 1002.239593] env[61806]: } is initializing. {{(pid=61806) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 1002.363164] env[61806]: DEBUG oslo_concurrency.lockutils [None req-d41c382b-3631-43f7-9094-5acd8818b680 tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Acquiring lock "67d1bdd8-cf8f-4fb5-bd3c-42008e4d60e3-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1002.363406] env[61806]: DEBUG oslo_concurrency.lockutils [None req-d41c382b-3631-43f7-9094-5acd8818b680 tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Lock "67d1bdd8-cf8f-4fb5-bd3c-42008e4d60e3-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1002.363586] env[61806]: DEBUG oslo_concurrency.lockutils [None req-d41c382b-3631-43f7-9094-5acd8818b680 tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Lock "67d1bdd8-cf8f-4fb5-bd3c-42008e4d60e3-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1002.375729] env[61806]: DEBUG oslo_vmware.api [None req-1c35eaa5-2854-4e3e-bc2c-4ef5a33b9a12 tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Task: {'id': task-1295068, 'name': ReconfigVM_Task, 'duration_secs': 0.320891} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1002.376316] env[61806]: DEBUG nova.virt.vmwareapi.volumeops [None req-1c35eaa5-2854-4e3e-bc2c-4ef5a33b9a12 tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] [instance: 9357e61f-1628-43bd-ab46-de13c1529f51] Reconfigured VM instance instance-00000060 to attach disk [datastore1] 9357e61f-1628-43bd-ab46-de13c1529f51/9357e61f-1628-43bd-ab46-de13c1529f51.vmdk or device None with type thin {{(pid=61806) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1002.376617] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-1c35eaa5-2854-4e3e-bc2c-4ef5a33b9a12 tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] [instance: 9357e61f-1628-43bd-ab46-de13c1529f51] Updating instance '9357e61f-1628-43bd-ab46-de13c1529f51' progress to 50 {{(pid=61806) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 1002.403498] env[61806]: DEBUG oslo_concurrency.lockutils [None req-b4773569-1b9a-41b1-9eb3-234618dc8825 tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] Acquiring lock "refresh_cache-fe1d4ace-7de3-423b-b423-9d7397aa2413" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1002.403691] env[61806]: DEBUG oslo_concurrency.lockutils [None req-b4773569-1b9a-41b1-9eb3-234618dc8825 tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] Acquired lock "refresh_cache-fe1d4ace-7de3-423b-b423-9d7397aa2413" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1002.403783] env[61806]: DEBUG nova.network.neutron [None req-b4773569-1b9a-41b1-9eb3-234618dc8825 tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] [instance: fe1d4ace-7de3-423b-b423-9d7397aa2413] Building network info cache for instance {{(pid=61806) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1002.680890] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9e706fdd-6dd5-4272-8fc1-14794d41e18a {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1002.688552] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d229c356-b5c8-42f3-8654-46c37182a41c {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1002.719167] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3640b48b-5c31-47c0-968a-4e54f3f3c3d8 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1002.726200] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-966ce8a6-385c-42a8-9f51-88393c0f4811 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1002.740739] env[61806]: DEBUG nova.compute.provider_tree [None req-8755c097-96dd-4051-93b1-330ffc947718 tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Inventory has not changed in ProviderTree for provider: a2858be1-fd22-4e08-979e-87ad25293407 {{(pid=61806) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1002.748292] env[61806]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1002.748292] env[61806]: value = "session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]5242c2b9-14c4-57ac-e4ab-e5ca38633943" [ 1002.748292] env[61806]: _type = "HttpNfcLease" [ 1002.748292] env[61806]: } is ready. {{(pid=61806) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 1002.748523] env[61806]: DEBUG oslo_vmware.rw_handles [None req-2bebc059-5179-46cc-9f09-746fe48209df tempest-ServersNegativeTestJSON-891856253 tempest-ServersNegativeTestJSON-891856253-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 1002.748523] env[61806]: value = "session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]5242c2b9-14c4-57ac-e4ab-e5ca38633943" [ 1002.748523] env[61806]: _type = "HttpNfcLease" [ 1002.748523] env[61806]: }. {{(pid=61806) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:486}} [ 1002.749432] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0c9f9bf2-7a66-418b-9be9-13910a98c641 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1002.756978] env[61806]: DEBUG oslo_vmware.rw_handles [None req-2bebc059-5179-46cc-9f09-746fe48209df tempest-ServersNegativeTestJSON-891856253 tempest-ServersNegativeTestJSON-891856253-project-member] Found VMDK URL: https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/5272df73-a1bf-0cd6-7e38-e3301b18dad4/disk-0.vmdk from lease info. {{(pid=61806) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 1002.757179] env[61806]: DEBUG oslo_vmware.rw_handles [None req-2bebc059-5179-46cc-9f09-746fe48209df tempest-ServersNegativeTestJSON-891856253 tempest-ServersNegativeTestJSON-891856253-project-member] Opening URL: https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/5272df73-a1bf-0cd6-7e38-e3301b18dad4/disk-0.vmdk for reading. {{(pid=61806) _create_read_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:99}} [ 1002.852039] env[61806]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-179faac3-7771-4495-8653-43b3f8e59c6e {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1002.888060] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d4bd5fb4-e5de-4416-af0b-ac59da5d24b0 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1002.909468] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-66a987af-d10f-42e2-8cf3-2d2092f91f21 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1002.927815] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-1c35eaa5-2854-4e3e-bc2c-4ef5a33b9a12 tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] [instance: 9357e61f-1628-43bd-ab46-de13c1529f51] Updating instance '9357e61f-1628-43bd-ab46-de13c1529f51' progress to 67 {{(pid=61806) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 1002.940889] env[61806]: DEBUG nova.network.neutron [None req-b4773569-1b9a-41b1-9eb3-234618dc8825 tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] [instance: fe1d4ace-7de3-423b-b423-9d7397aa2413] Instance cache missing network info. {{(pid=61806) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 1003.062612] env[61806]: DEBUG nova.compute.utils [None req-de9a979c-45ed-463a-871e-b7dda8c366a2 tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] Using /dev/sd instead of None {{(pid=61806) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1003.082584] env[61806]: DEBUG nova.network.neutron [None req-b4773569-1b9a-41b1-9eb3-234618dc8825 tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] [instance: fe1d4ace-7de3-423b-b423-9d7397aa2413] Updating instance_info_cache with network_info: [{"id": "b2369483-32e0-4f80-be14-b45402c39094", "address": "fa:16:3e:ba:4d:9f", "network": {"id": "f280a01b-7000-4e54-ab9a-20eab156a410", "bridge": "br-int", "label": "tempest-ServersTestJSON-2080718146-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "73ebcc4d79d248efb653a084a5e44302", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "43798f54-0c07-4417-a23f-58bb6b7e204b", "external-id": "nsx-vlan-transportzone-571", "segmentation_id": 571, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb2369483-32", "ovs_interfaceid": "b2369483-32e0-4f80-be14-b45402c39094", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61806) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1003.246979] env[61806]: DEBUG nova.scheduler.client.report [None req-8755c097-96dd-4051-93b1-330ffc947718 tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Inventory has not changed for provider a2858be1-fd22-4e08-979e-87ad25293407 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 138, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61806) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1003.402472] env[61806]: DEBUG oslo_concurrency.lockutils [None req-d41c382b-3631-43f7-9094-5acd8818b680 tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Acquiring lock "refresh_cache-67d1bdd8-cf8f-4fb5-bd3c-42008e4d60e3" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1003.402698] env[61806]: DEBUG oslo_concurrency.lockutils [None req-d41c382b-3631-43f7-9094-5acd8818b680 tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Acquired lock "refresh_cache-67d1bdd8-cf8f-4fb5-bd3c-42008e4d60e3" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1003.402894] env[61806]: DEBUG nova.network.neutron [None req-d41c382b-3631-43f7-9094-5acd8818b680 tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] [instance: 67d1bdd8-cf8f-4fb5-bd3c-42008e4d60e3] Building network info cache for instance {{(pid=61806) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1003.471918] env[61806]: DEBUG nova.network.neutron [None req-1c35eaa5-2854-4e3e-bc2c-4ef5a33b9a12 tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] [instance: 9357e61f-1628-43bd-ab46-de13c1529f51] Port a4ca7c56-b6a1-4bd2-bd47-02b0ed4aa097 binding to destination host cpu-1 is already ACTIVE {{(pid=61806) migrate_instance_start /opt/stack/nova/nova/network/neutron.py:3171}} [ 1003.565704] env[61806]: INFO nova.virt.block_device [None req-de9a979c-45ed-463a-871e-b7dda8c366a2 tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] [instance: fd9e520a-2e7b-43d4-b84d-8933c32cc6a6] Booting with volume dd35bef0-37b9-4821-9a9d-7f5c7789746c at /dev/sdb [ 1003.585387] env[61806]: DEBUG oslo_concurrency.lockutils [None req-b4773569-1b9a-41b1-9eb3-234618dc8825 tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] Releasing lock "refresh_cache-fe1d4ace-7de3-423b-b423-9d7397aa2413" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1003.585509] env[61806]: DEBUG nova.compute.manager [None req-b4773569-1b9a-41b1-9eb3-234618dc8825 tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] [instance: fe1d4ace-7de3-423b-b423-9d7397aa2413] Instance network_info: |[{"id": "b2369483-32e0-4f80-be14-b45402c39094", "address": "fa:16:3e:ba:4d:9f", "network": {"id": "f280a01b-7000-4e54-ab9a-20eab156a410", "bridge": "br-int", "label": "tempest-ServersTestJSON-2080718146-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "73ebcc4d79d248efb653a084a5e44302", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "43798f54-0c07-4417-a23f-58bb6b7e204b", "external-id": "nsx-vlan-transportzone-571", "segmentation_id": 571, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb2369483-32", "ovs_interfaceid": "b2369483-32e0-4f80-be14-b45402c39094", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61806) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1003.586143] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-b4773569-1b9a-41b1-9eb3-234618dc8825 tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] [instance: fe1d4ace-7de3-423b-b423-9d7397aa2413] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:ba:4d:9f', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '43798f54-0c07-4417-a23f-58bb6b7e204b', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'b2369483-32e0-4f80-be14-b45402c39094', 'vif_model': 'vmxnet3'}] {{(pid=61806) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1003.593982] env[61806]: DEBUG oslo.service.loopingcall [None req-b4773569-1b9a-41b1-9eb3-234618dc8825 tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61806) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1003.594517] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: fe1d4ace-7de3-423b-b423-9d7397aa2413] Creating VM on the ESX host {{(pid=61806) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 1003.594854] env[61806]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-662364ea-8da5-4805-8854-82e0a0c0476d {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1003.611775] env[61806]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-52530f80-6f58-41ed-a229-e35c432418fb {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1003.621890] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-50e30528-1379-4b29-96f2-077e9c920d33 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1003.634232] env[61806]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1003.634232] env[61806]: value = "task-1295070" [ 1003.634232] env[61806]: _type = "Task" [ 1003.634232] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1003.644201] env[61806]: DEBUG oslo_vmware.api [-] Task: {'id': task-1295070, 'name': CreateVM_Task} progress is 6%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1003.653651] env[61806]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-50c53652-f416-4422-842d-4ed9fda0df52 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1003.666361] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0f917eba-5480-4702-9ed8-2ae5654beda0 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1003.698455] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-75eca55e-b29f-43b1-83e1-056984090a0d {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1003.706659] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b1c30af5-531f-4240-a935-3ecf9479fc4a {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1003.722218] env[61806]: DEBUG nova.virt.block_device [None req-de9a979c-45ed-463a-871e-b7dda8c366a2 tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] [instance: fd9e520a-2e7b-43d4-b84d-8933c32cc6a6] Updating existing volume attachment record: 60011335-d621-48ac-bdd9-7da611ca0e30 {{(pid=61806) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 1003.753013] env[61806]: DEBUG oslo_concurrency.lockutils [None req-8755c097-96dd-4051-93b1-330ffc947718 tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.217s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1003.753752] env[61806]: DEBUG nova.compute.manager [None req-8755c097-96dd-4051-93b1-330ffc947718 tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] [instance: 71999b04-bf6e-4e4c-87c9-d5ec1932a80f] Start building networks asynchronously for instance. {{(pid=61806) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1003.757169] env[61806]: DEBUG oslo_concurrency.lockutils [None req-4b7244a6-0e60-4b9b-ac2e-5c7b0ff2b8c7 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 3.937s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1003.758917] env[61806]: INFO nova.compute.claims [None req-4b7244a6-0e60-4b9b-ac2e-5c7b0ff2b8c7 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] [instance: d0669232-6f33-4b30-97c2-2e4239af1ad0] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1003.810683] env[61806]: DEBUG nova.compute.manager [req-f922106e-cae8-4d51-ad99-2257479ea742 req-28d90ba6-6274-44f3-b7b7-b5664f6d4a13 service nova] [instance: fe1d4ace-7de3-423b-b423-9d7397aa2413] Received event network-changed-b2369483-32e0-4f80-be14-b45402c39094 {{(pid=61806) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1003.810987] env[61806]: DEBUG nova.compute.manager [req-f922106e-cae8-4d51-ad99-2257479ea742 req-28d90ba6-6274-44f3-b7b7-b5664f6d4a13 service nova] [instance: fe1d4ace-7de3-423b-b423-9d7397aa2413] Refreshing instance network info cache due to event network-changed-b2369483-32e0-4f80-be14-b45402c39094. {{(pid=61806) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 1003.811316] env[61806]: DEBUG oslo_concurrency.lockutils [req-f922106e-cae8-4d51-ad99-2257479ea742 req-28d90ba6-6274-44f3-b7b7-b5664f6d4a13 service nova] Acquiring lock "refresh_cache-fe1d4ace-7de3-423b-b423-9d7397aa2413" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1003.811545] env[61806]: DEBUG oslo_concurrency.lockutils [req-f922106e-cae8-4d51-ad99-2257479ea742 req-28d90ba6-6274-44f3-b7b7-b5664f6d4a13 service nova] Acquired lock "refresh_cache-fe1d4ace-7de3-423b-b423-9d7397aa2413" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1003.811736] env[61806]: DEBUG nova.network.neutron [req-f922106e-cae8-4d51-ad99-2257479ea742 req-28d90ba6-6274-44f3-b7b7-b5664f6d4a13 service nova] [instance: fe1d4ace-7de3-423b-b423-9d7397aa2413] Refreshing network info cache for port b2369483-32e0-4f80-be14-b45402c39094 {{(pid=61806) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1004.145623] env[61806]: DEBUG oslo_vmware.api [-] Task: {'id': task-1295070, 'name': CreateVM_Task, 'duration_secs': 0.445365} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1004.145820] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: fe1d4ace-7de3-423b-b423-9d7397aa2413] Created VM on the ESX host {{(pid=61806) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 1004.146576] env[61806]: DEBUG oslo_concurrency.lockutils [None req-b4773569-1b9a-41b1-9eb3-234618dc8825 tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1004.146868] env[61806]: DEBUG oslo_concurrency.lockutils [None req-b4773569-1b9a-41b1-9eb3-234618dc8825 tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] Acquired lock "[datastore2] devstack-image-cache_base/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1004.147413] env[61806]: DEBUG oslo_concurrency.lockutils [None req-b4773569-1b9a-41b1-9eb3-234618dc8825 tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1004.147733] env[61806]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-8a0d4a47-afd4-44cf-a5f7-1a76a8fd0b7e {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1004.152707] env[61806]: DEBUG oslo_vmware.api [None req-b4773569-1b9a-41b1-9eb3-234618dc8825 tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] Waiting for the task: (returnval){ [ 1004.152707] env[61806]: value = "session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]52cc895f-84b9-75b5-8186-9409eaf77e2d" [ 1004.152707] env[61806]: _type = "Task" [ 1004.152707] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1004.161858] env[61806]: DEBUG oslo_vmware.api [None req-b4773569-1b9a-41b1-9eb3-234618dc8825 tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] Task: {'id': session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]52cc895f-84b9-75b5-8186-9409eaf77e2d, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1004.199431] env[61806]: DEBUG nova.network.neutron [None req-d41c382b-3631-43f7-9094-5acd8818b680 tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] [instance: 67d1bdd8-cf8f-4fb5-bd3c-42008e4d60e3] Updating instance_info_cache with network_info: [{"id": "859fb723-4a1c-4ea8-bd7c-0ed308f3a9e3", "address": "fa:16:3e:3c:40:d6", "network": {"id": "746a03b6-c1a6-4d27-98b9-0bee6376634f", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-1971382587-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.245", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c1f2d161c50f4b70987b90e6364c3779", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6a31b224-77d7-48a4-af87-312758611b8e", "external-id": "nsx-vlan-transportzone-761", "segmentation_id": 761, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap859fb723-4a", "ovs_interfaceid": "859fb723-4a1c-4ea8-bd7c-0ed308f3a9e3", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61806) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1004.265030] env[61806]: DEBUG nova.compute.utils [None req-8755c097-96dd-4051-93b1-330ffc947718 tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Using /dev/sd instead of None {{(pid=61806) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1004.268970] env[61806]: DEBUG nova.compute.manager [None req-8755c097-96dd-4051-93b1-330ffc947718 tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] [instance: 71999b04-bf6e-4e4c-87c9-d5ec1932a80f] Allocating IP information in the background. {{(pid=61806) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1004.269250] env[61806]: DEBUG nova.network.neutron [None req-8755c097-96dd-4051-93b1-330ffc947718 tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] [instance: 71999b04-bf6e-4e4c-87c9-d5ec1932a80f] allocate_for_instance() {{(pid=61806) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 1004.356371] env[61806]: DEBUG nova.policy [None req-8755c097-96dd-4051-93b1-330ffc947718 tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '2b86d596e89645c2ab31332afdcbc3c2', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '7ab95fb72b5d46c3b6c7bebbccf897cd', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61806) authorize /opt/stack/nova/nova/policy.py:201}} [ 1004.499462] env[61806]: DEBUG oslo_concurrency.lockutils [None req-1c35eaa5-2854-4e3e-bc2c-4ef5a33b9a12 tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Acquiring lock "9357e61f-1628-43bd-ab46-de13c1529f51-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1004.499797] env[61806]: DEBUG oslo_concurrency.lockutils [None req-1c35eaa5-2854-4e3e-bc2c-4ef5a33b9a12 tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Lock "9357e61f-1628-43bd-ab46-de13c1529f51-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1004.500080] env[61806]: DEBUG oslo_concurrency.lockutils [None req-1c35eaa5-2854-4e3e-bc2c-4ef5a33b9a12 tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Lock "9357e61f-1628-43bd-ab46-de13c1529f51-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1004.664467] env[61806]: DEBUG oslo_vmware.api [None req-b4773569-1b9a-41b1-9eb3-234618dc8825 tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] Task: {'id': session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]52cc895f-84b9-75b5-8186-9409eaf77e2d, 'name': SearchDatastore_Task, 'duration_secs': 0.012816} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1004.664918] env[61806]: DEBUG oslo_concurrency.lockutils [None req-b4773569-1b9a-41b1-9eb3-234618dc8825 tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] Releasing lock "[datastore2] devstack-image-cache_base/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1004.665257] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-b4773569-1b9a-41b1-9eb3-234618dc8825 tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] [instance: fe1d4ace-7de3-423b-b423-9d7397aa2413] Processing image ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa {{(pid=61806) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1004.665519] env[61806]: DEBUG oslo_concurrency.lockutils [None req-b4773569-1b9a-41b1-9eb3-234618dc8825 tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa.vmdk" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1004.665751] env[61806]: DEBUG oslo_concurrency.lockutils [None req-b4773569-1b9a-41b1-9eb3-234618dc8825 tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] Acquired lock "[datastore2] devstack-image-cache_base/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa.vmdk" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1004.665877] env[61806]: DEBUG nova.virt.vmwareapi.ds_util [None req-b4773569-1b9a-41b1-9eb3-234618dc8825 tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61806) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1004.666599] env[61806]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-f8b5ef09-858b-4719-8243-96057c088d41 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1004.676489] env[61806]: DEBUG nova.virt.vmwareapi.ds_util [None req-b4773569-1b9a-41b1-9eb3-234618dc8825 tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61806) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1004.676759] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-b4773569-1b9a-41b1-9eb3-234618dc8825 tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61806) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 1004.677895] env[61806]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-15839fae-4ce1-4c4a-bae6-e5fe24fe687d {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1004.680969] env[61806]: DEBUG nova.network.neutron [req-f922106e-cae8-4d51-ad99-2257479ea742 req-28d90ba6-6274-44f3-b7b7-b5664f6d4a13 service nova] [instance: fe1d4ace-7de3-423b-b423-9d7397aa2413] Updated VIF entry in instance network info cache for port b2369483-32e0-4f80-be14-b45402c39094. {{(pid=61806) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1004.681447] env[61806]: DEBUG nova.network.neutron [req-f922106e-cae8-4d51-ad99-2257479ea742 req-28d90ba6-6274-44f3-b7b7-b5664f6d4a13 service nova] [instance: fe1d4ace-7de3-423b-b423-9d7397aa2413] Updating instance_info_cache with network_info: [{"id": "b2369483-32e0-4f80-be14-b45402c39094", "address": "fa:16:3e:ba:4d:9f", "network": {"id": "f280a01b-7000-4e54-ab9a-20eab156a410", "bridge": "br-int", "label": "tempest-ServersTestJSON-2080718146-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "73ebcc4d79d248efb653a084a5e44302", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "43798f54-0c07-4417-a23f-58bb6b7e204b", "external-id": "nsx-vlan-transportzone-571", "segmentation_id": 571, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb2369483-32", "ovs_interfaceid": "b2369483-32e0-4f80-be14-b45402c39094", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61806) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1004.686928] env[61806]: DEBUG oslo_vmware.api [None req-b4773569-1b9a-41b1-9eb3-234618dc8825 tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] Waiting for the task: (returnval){ [ 1004.686928] env[61806]: value = "session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]5281d1fe-12c4-2652-1182-8076f67d0cb8" [ 1004.686928] env[61806]: _type = "Task" [ 1004.686928] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1004.694710] env[61806]: DEBUG oslo_vmware.api [None req-b4773569-1b9a-41b1-9eb3-234618dc8825 tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] Task: {'id': session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]5281d1fe-12c4-2652-1182-8076f67d0cb8, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1004.701825] env[61806]: DEBUG oslo_concurrency.lockutils [None req-d41c382b-3631-43f7-9094-5acd8818b680 tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Releasing lock "refresh_cache-67d1bdd8-cf8f-4fb5-bd3c-42008e4d60e3" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1004.770201] env[61806]: DEBUG nova.compute.manager [None req-8755c097-96dd-4051-93b1-330ffc947718 tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] [instance: 71999b04-bf6e-4e4c-87c9-d5ec1932a80f] Start building block device mappings for instance. {{(pid=61806) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1004.784472] env[61806]: DEBUG nova.network.neutron [None req-8755c097-96dd-4051-93b1-330ffc947718 tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] [instance: 71999b04-bf6e-4e4c-87c9-d5ec1932a80f] Successfully created port: 927351ab-9ed5-41c7-beea-77033dc7601b {{(pid=61806) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1004.932862] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b832896b-a2d9-4c64-bc57-accd1b7be401 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1004.941631] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-95d8baae-a248-44f4-adfd-dd678eb14c14 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1004.973803] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-88b74ed2-5d78-4534-b893-d910c798d072 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1004.982828] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6eef7482-e6f9-4305-8d34-5bc8791e686c {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1004.997156] env[61806]: DEBUG nova.compute.provider_tree [None req-4b7244a6-0e60-4b9b-ac2e-5c7b0ff2b8c7 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Inventory has not changed in ProviderTree for provider: a2858be1-fd22-4e08-979e-87ad25293407 {{(pid=61806) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1005.185249] env[61806]: DEBUG oslo_concurrency.lockutils [req-f922106e-cae8-4d51-ad99-2257479ea742 req-28d90ba6-6274-44f3-b7b7-b5664f6d4a13 service nova] Releasing lock "refresh_cache-fe1d4ace-7de3-423b-b423-9d7397aa2413" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1005.197545] env[61806]: DEBUG oslo_vmware.api [None req-b4773569-1b9a-41b1-9eb3-234618dc8825 tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] Task: {'id': session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]5281d1fe-12c4-2652-1182-8076f67d0cb8, 'name': SearchDatastore_Task, 'duration_secs': 0.016715} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1005.198578] env[61806]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-dfc446f8-c710-4058-ad94-193b86c9ee23 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1005.204855] env[61806]: DEBUG oslo_vmware.api [None req-b4773569-1b9a-41b1-9eb3-234618dc8825 tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] Waiting for the task: (returnval){ [ 1005.204855] env[61806]: value = "session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]52bdc3f9-de0a-2b6e-6385-4a2ab195694a" [ 1005.204855] env[61806]: _type = "Task" [ 1005.204855] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1005.216966] env[61806]: DEBUG oslo_vmware.api [None req-b4773569-1b9a-41b1-9eb3-234618dc8825 tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] Task: {'id': session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]52bdc3f9-de0a-2b6e-6385-4a2ab195694a, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1005.228183] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d0af4a92-dfec-42ba-af1d-b167d8744d00 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1005.248835] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-283603c1-a1d4-471e-954d-16ce59f3b76a {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1005.256374] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-d41c382b-3631-43f7-9094-5acd8818b680 tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] [instance: 67d1bdd8-cf8f-4fb5-bd3c-42008e4d60e3] Updating instance '67d1bdd8-cf8f-4fb5-bd3c-42008e4d60e3' progress to 83 {{(pid=61806) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 1005.500461] env[61806]: DEBUG nova.scheduler.client.report [None req-4b7244a6-0e60-4b9b-ac2e-5c7b0ff2b8c7 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Inventory has not changed for provider a2858be1-fd22-4e08-979e-87ad25293407 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 138, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61806) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1005.547020] env[61806]: DEBUG oslo_concurrency.lockutils [None req-1c35eaa5-2854-4e3e-bc2c-4ef5a33b9a12 tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Acquiring lock "refresh_cache-9357e61f-1628-43bd-ab46-de13c1529f51" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1005.547201] env[61806]: DEBUG oslo_concurrency.lockutils [None req-1c35eaa5-2854-4e3e-bc2c-4ef5a33b9a12 tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Acquired lock "refresh_cache-9357e61f-1628-43bd-ab46-de13c1529f51" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1005.547417] env[61806]: DEBUG nova.network.neutron [None req-1c35eaa5-2854-4e3e-bc2c-4ef5a33b9a12 tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] [instance: 9357e61f-1628-43bd-ab46-de13c1529f51] Building network info cache for instance {{(pid=61806) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1005.716072] env[61806]: DEBUG oslo_vmware.api [None req-b4773569-1b9a-41b1-9eb3-234618dc8825 tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] Task: {'id': session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]52bdc3f9-de0a-2b6e-6385-4a2ab195694a, 'name': SearchDatastore_Task, 'duration_secs': 0.017479} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1005.716430] env[61806]: DEBUG oslo_concurrency.lockutils [None req-b4773569-1b9a-41b1-9eb3-234618dc8825 tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] Releasing lock "[datastore2] devstack-image-cache_base/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa.vmdk" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1005.716729] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [None req-b4773569-1b9a-41b1-9eb3-234618dc8825 tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa.vmdk to [datastore2] fe1d4ace-7de3-423b-b423-9d7397aa2413/fe1d4ace-7de3-423b-b423-9d7397aa2413.vmdk {{(pid=61806) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 1005.717080] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-7e85ec0c-8ad5-4779-bf13-cf2745624dec {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1005.726998] env[61806]: DEBUG oslo_vmware.api [None req-b4773569-1b9a-41b1-9eb3-234618dc8825 tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] Waiting for the task: (returnval){ [ 1005.726998] env[61806]: value = "task-1295074" [ 1005.726998] env[61806]: _type = "Task" [ 1005.726998] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1005.735450] env[61806]: DEBUG oslo_vmware.api [None req-b4773569-1b9a-41b1-9eb3-234618dc8825 tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] Task: {'id': task-1295074, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1005.763136] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [None req-d41c382b-3631-43f7-9094-5acd8818b680 tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] [instance: 67d1bdd8-cf8f-4fb5-bd3c-42008e4d60e3] Powering on the VM {{(pid=61806) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1005.763136] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-a8a00517-9778-40ca-84a0-1e4c866d0e41 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1005.771503] env[61806]: DEBUG oslo_vmware.api [None req-d41c382b-3631-43f7-9094-5acd8818b680 tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Waiting for the task: (returnval){ [ 1005.771503] env[61806]: value = "task-1295075" [ 1005.771503] env[61806]: _type = "Task" [ 1005.771503] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1005.780441] env[61806]: DEBUG oslo_vmware.api [None req-d41c382b-3631-43f7-9094-5acd8818b680 tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Task: {'id': task-1295075, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1005.783990] env[61806]: DEBUG nova.compute.manager [None req-8755c097-96dd-4051-93b1-330ffc947718 tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] [instance: 71999b04-bf6e-4e4c-87c9-d5ec1932a80f] Start spawning the instance on the hypervisor. {{(pid=61806) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1005.812267] env[61806]: DEBUG nova.virt.hardware [None req-8755c097-96dd-4051-93b1-330ffc947718 tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-15T18:05:10Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-15T18:04:52Z,direct_url=,disk_format='vmdk',id=ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='84ba83fce2c349a6988173c5d6fc3b07',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-15T18:04:53Z,virtual_size=,visibility=), allow threads: False {{(pid=61806) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1005.812597] env[61806]: DEBUG nova.virt.hardware [None req-8755c097-96dd-4051-93b1-330ffc947718 tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Flavor limits 0:0:0 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1005.812862] env[61806]: DEBUG nova.virt.hardware [None req-8755c097-96dd-4051-93b1-330ffc947718 tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Image limits 0:0:0 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1005.813151] env[61806]: DEBUG nova.virt.hardware [None req-8755c097-96dd-4051-93b1-330ffc947718 tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Flavor pref 0:0:0 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1005.813365] env[61806]: DEBUG nova.virt.hardware [None req-8755c097-96dd-4051-93b1-330ffc947718 tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Image pref 0:0:0 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1005.813529] env[61806]: DEBUG nova.virt.hardware [None req-8755c097-96dd-4051-93b1-330ffc947718 tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1005.814474] env[61806]: DEBUG nova.virt.hardware [None req-8755c097-96dd-4051-93b1-330ffc947718 tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61806) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1005.814474] env[61806]: DEBUG nova.virt.hardware [None req-8755c097-96dd-4051-93b1-330ffc947718 tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61806) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1005.814474] env[61806]: DEBUG nova.virt.hardware [None req-8755c097-96dd-4051-93b1-330ffc947718 tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Got 1 possible topologies {{(pid=61806) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1005.814474] env[61806]: DEBUG nova.virt.hardware [None req-8755c097-96dd-4051-93b1-330ffc947718 tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61806) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1005.814736] env[61806]: DEBUG nova.virt.hardware [None req-8755c097-96dd-4051-93b1-330ffc947718 tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61806) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1005.815646] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4d90b869-ca94-4ffa-bf39-19449ee6a823 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1005.825677] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8e8b62af-1bad-4723-95d4-68c41ed27335 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1006.007999] env[61806]: DEBUG oslo_concurrency.lockutils [None req-4b7244a6-0e60-4b9b-ac2e-5c7b0ff2b8c7 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.251s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1006.008626] env[61806]: DEBUG nova.compute.manager [None req-4b7244a6-0e60-4b9b-ac2e-5c7b0ff2b8c7 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] [instance: d0669232-6f33-4b30-97c2-2e4239af1ad0] Start building networks asynchronously for instance. {{(pid=61806) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1006.241561] env[61806]: DEBUG oslo_vmware.api [None req-b4773569-1b9a-41b1-9eb3-234618dc8825 tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] Task: {'id': task-1295074, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1006.283956] env[61806]: DEBUG oslo_vmware.api [None req-d41c382b-3631-43f7-9094-5acd8818b680 tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Task: {'id': task-1295075, 'name': PowerOnVM_Task} progress is 89%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1006.442367] env[61806]: DEBUG nova.compute.manager [req-6904f178-0629-41c7-a2ec-6b31d2b51d6f req-5e995dc7-9ed1-4d8f-b04e-8bb7b6381d10 service nova] [instance: 71999b04-bf6e-4e4c-87c9-d5ec1932a80f] Received event network-vif-plugged-927351ab-9ed5-41c7-beea-77033dc7601b {{(pid=61806) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1006.442601] env[61806]: DEBUG oslo_concurrency.lockutils [req-6904f178-0629-41c7-a2ec-6b31d2b51d6f req-5e995dc7-9ed1-4d8f-b04e-8bb7b6381d10 service nova] Acquiring lock "71999b04-bf6e-4e4c-87c9-d5ec1932a80f-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1006.442847] env[61806]: DEBUG oslo_concurrency.lockutils [req-6904f178-0629-41c7-a2ec-6b31d2b51d6f req-5e995dc7-9ed1-4d8f-b04e-8bb7b6381d10 service nova] Lock "71999b04-bf6e-4e4c-87c9-d5ec1932a80f-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1006.443233] env[61806]: DEBUG oslo_concurrency.lockutils [req-6904f178-0629-41c7-a2ec-6b31d2b51d6f req-5e995dc7-9ed1-4d8f-b04e-8bb7b6381d10 service nova] Lock "71999b04-bf6e-4e4c-87c9-d5ec1932a80f-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1006.443474] env[61806]: DEBUG nova.compute.manager [req-6904f178-0629-41c7-a2ec-6b31d2b51d6f req-5e995dc7-9ed1-4d8f-b04e-8bb7b6381d10 service nova] [instance: 71999b04-bf6e-4e4c-87c9-d5ec1932a80f] No waiting events found dispatching network-vif-plugged-927351ab-9ed5-41c7-beea-77033dc7601b {{(pid=61806) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1006.443660] env[61806]: WARNING nova.compute.manager [req-6904f178-0629-41c7-a2ec-6b31d2b51d6f req-5e995dc7-9ed1-4d8f-b04e-8bb7b6381d10 service nova] [instance: 71999b04-bf6e-4e4c-87c9-d5ec1932a80f] Received unexpected event network-vif-plugged-927351ab-9ed5-41c7-beea-77033dc7601b for instance with vm_state building and task_state spawning. [ 1006.476594] env[61806]: DEBUG nova.network.neutron [None req-1c35eaa5-2854-4e3e-bc2c-4ef5a33b9a12 tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] [instance: 9357e61f-1628-43bd-ab46-de13c1529f51] Updating instance_info_cache with network_info: [{"id": "a4ca7c56-b6a1-4bd2-bd47-02b0ed4aa097", "address": "fa:16:3e:66:09:0d", "network": {"id": "46515a4e-610f-42fb-a424-205a03df0640", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-1133623547-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4b9d94c3d684474b99e305fb6cd9b95f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "82ca17df-257e-40e6-9ec9-310ed6f05ccb", "external-id": "nsx-vlan-transportzone-778", "segmentation_id": 778, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa4ca7c56-b6", "ovs_interfaceid": "a4ca7c56-b6a1-4bd2-bd47-02b0ed4aa097", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61806) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1006.514403] env[61806]: DEBUG nova.compute.utils [None req-4b7244a6-0e60-4b9b-ac2e-5c7b0ff2b8c7 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Using /dev/sd instead of None {{(pid=61806) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1006.515984] env[61806]: DEBUG nova.compute.manager [None req-4b7244a6-0e60-4b9b-ac2e-5c7b0ff2b8c7 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] [instance: d0669232-6f33-4b30-97c2-2e4239af1ad0] Allocating IP information in the background. {{(pid=61806) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1006.516169] env[61806]: DEBUG nova.network.neutron [None req-4b7244a6-0e60-4b9b-ac2e-5c7b0ff2b8c7 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] [instance: d0669232-6f33-4b30-97c2-2e4239af1ad0] allocate_for_instance() {{(pid=61806) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 1006.596803] env[61806]: DEBUG nova.policy [None req-4b7244a6-0e60-4b9b-ac2e-5c7b0ff2b8c7 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '772f4c6d731c4d5daea11b37a0f6718b', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '1c55cc0455584fbb8cda88d33b8f8b7e', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61806) authorize /opt/stack/nova/nova/policy.py:201}} [ 1006.646508] env[61806]: DEBUG nova.network.neutron [None req-8755c097-96dd-4051-93b1-330ffc947718 tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] [instance: 71999b04-bf6e-4e4c-87c9-d5ec1932a80f] Successfully updated port: 927351ab-9ed5-41c7-beea-77033dc7601b {{(pid=61806) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1006.737697] env[61806]: DEBUG oslo_vmware.api [None req-b4773569-1b9a-41b1-9eb3-234618dc8825 tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] Task: {'id': task-1295074, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.585408} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1006.737986] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [None req-b4773569-1b9a-41b1-9eb3-234618dc8825 tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa.vmdk to [datastore2] fe1d4ace-7de3-423b-b423-9d7397aa2413/fe1d4ace-7de3-423b-b423-9d7397aa2413.vmdk {{(pid=61806) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 1006.738285] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-b4773569-1b9a-41b1-9eb3-234618dc8825 tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] [instance: fe1d4ace-7de3-423b-b423-9d7397aa2413] Extending root virtual disk to 1048576 {{(pid=61806) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1006.738550] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-e1f01725-90a0-4cf6-9817-7b225ac313ba {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1006.745480] env[61806]: DEBUG oslo_vmware.api [None req-b4773569-1b9a-41b1-9eb3-234618dc8825 tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] Waiting for the task: (returnval){ [ 1006.745480] env[61806]: value = "task-1295077" [ 1006.745480] env[61806]: _type = "Task" [ 1006.745480] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1006.753228] env[61806]: DEBUG oslo_vmware.api [None req-b4773569-1b9a-41b1-9eb3-234618dc8825 tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] Task: {'id': task-1295077, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1006.781166] env[61806]: DEBUG oslo_vmware.api [None req-d41c382b-3631-43f7-9094-5acd8818b680 tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Task: {'id': task-1295075, 'name': PowerOnVM_Task, 'duration_secs': 0.622027} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1006.781520] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [None req-d41c382b-3631-43f7-9094-5acd8818b680 tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] [instance: 67d1bdd8-cf8f-4fb5-bd3c-42008e4d60e3] Powered on the VM {{(pid=61806) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1006.781641] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-d41c382b-3631-43f7-9094-5acd8818b680 tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] [instance: 67d1bdd8-cf8f-4fb5-bd3c-42008e4d60e3] Updating instance '67d1bdd8-cf8f-4fb5-bd3c-42008e4d60e3' progress to 100 {{(pid=61806) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 1006.895451] env[61806]: DEBUG nova.network.neutron [None req-4b7244a6-0e60-4b9b-ac2e-5c7b0ff2b8c7 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] [instance: d0669232-6f33-4b30-97c2-2e4239af1ad0] Successfully created port: f958e9a3-f8d6-4665-a59f-876aab66edfb {{(pid=61806) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1006.979735] env[61806]: DEBUG oslo_concurrency.lockutils [None req-1c35eaa5-2854-4e3e-bc2c-4ef5a33b9a12 tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Releasing lock "refresh_cache-9357e61f-1628-43bd-ab46-de13c1529f51" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1007.019793] env[61806]: DEBUG nova.compute.manager [None req-4b7244a6-0e60-4b9b-ac2e-5c7b0ff2b8c7 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] [instance: d0669232-6f33-4b30-97c2-2e4239af1ad0] Start building block device mappings for instance. {{(pid=61806) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1007.149701] env[61806]: DEBUG oslo_concurrency.lockutils [None req-8755c097-96dd-4051-93b1-330ffc947718 tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Acquiring lock "refresh_cache-71999b04-bf6e-4e4c-87c9-d5ec1932a80f" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1007.149864] env[61806]: DEBUG oslo_concurrency.lockutils [None req-8755c097-96dd-4051-93b1-330ffc947718 tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Acquired lock "refresh_cache-71999b04-bf6e-4e4c-87c9-d5ec1932a80f" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1007.150136] env[61806]: DEBUG nova.network.neutron [None req-8755c097-96dd-4051-93b1-330ffc947718 tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] [instance: 71999b04-bf6e-4e4c-87c9-d5ec1932a80f] Building network info cache for instance {{(pid=61806) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1007.255618] env[61806]: DEBUG oslo_vmware.api [None req-b4773569-1b9a-41b1-9eb3-234618dc8825 tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] Task: {'id': task-1295077, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.277855} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1007.255929] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-b4773569-1b9a-41b1-9eb3-234618dc8825 tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] [instance: fe1d4ace-7de3-423b-b423-9d7397aa2413] Extended root virtual disk {{(pid=61806) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1007.256781] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a2ff771a-9326-4d6c-aea5-0136904d3b0a {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1007.279613] env[61806]: DEBUG nova.virt.vmwareapi.volumeops [None req-b4773569-1b9a-41b1-9eb3-234618dc8825 tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] [instance: fe1d4ace-7de3-423b-b423-9d7397aa2413] Reconfiguring VM instance instance-00000061 to attach disk [datastore2] fe1d4ace-7de3-423b-b423-9d7397aa2413/fe1d4ace-7de3-423b-b423-9d7397aa2413.vmdk or device None with type sparse {{(pid=61806) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1007.279909] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-96fed49b-a54f-4474-9920-6bf4ed1c9cc8 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1007.306603] env[61806]: DEBUG oslo_vmware.api [None req-b4773569-1b9a-41b1-9eb3-234618dc8825 tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] Waiting for the task: (returnval){ [ 1007.306603] env[61806]: value = "task-1295078" [ 1007.306603] env[61806]: _type = "Task" [ 1007.306603] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1007.318140] env[61806]: DEBUG oslo_vmware.api [None req-b4773569-1b9a-41b1-9eb3-234618dc8825 tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] Task: {'id': task-1295078, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1007.508989] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-664c1ff6-57a3-4218-8d35-ae24f8f946e0 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1007.532612] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9741978b-881e-4c5e-91cc-49baa10c27eb {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1007.541419] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-1c35eaa5-2854-4e3e-bc2c-4ef5a33b9a12 tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] [instance: 9357e61f-1628-43bd-ab46-de13c1529f51] Updating instance '9357e61f-1628-43bd-ab46-de13c1529f51' progress to 83 {{(pid=61806) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 1007.683745] env[61806]: DEBUG nova.network.neutron [None req-8755c097-96dd-4051-93b1-330ffc947718 tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] [instance: 71999b04-bf6e-4e4c-87c9-d5ec1932a80f] Instance cache missing network info. {{(pid=61806) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 1007.825413] env[61806]: DEBUG oslo_vmware.api [None req-b4773569-1b9a-41b1-9eb3-234618dc8825 tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] Task: {'id': task-1295078, 'name': ReconfigVM_Task, 'duration_secs': 0.489018} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1007.825772] env[61806]: DEBUG nova.virt.vmwareapi.volumeops [None req-b4773569-1b9a-41b1-9eb3-234618dc8825 tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] [instance: fe1d4ace-7de3-423b-b423-9d7397aa2413] Reconfigured VM instance instance-00000061 to attach disk [datastore2] fe1d4ace-7de3-423b-b423-9d7397aa2413/fe1d4ace-7de3-423b-b423-9d7397aa2413.vmdk or device None with type sparse {{(pid=61806) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1007.826440] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-6f97d465-a6f9-45eb-9218-91ce5104c8dc {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1007.834596] env[61806]: DEBUG oslo_vmware.api [None req-b4773569-1b9a-41b1-9eb3-234618dc8825 tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] Waiting for the task: (returnval){ [ 1007.834596] env[61806]: value = "task-1295079" [ 1007.834596] env[61806]: _type = "Task" [ 1007.834596] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1007.843597] env[61806]: DEBUG oslo_vmware.api [None req-b4773569-1b9a-41b1-9eb3-234618dc8825 tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] Task: {'id': task-1295079, 'name': Rename_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1007.853930] env[61806]: DEBUG nova.network.neutron [None req-8755c097-96dd-4051-93b1-330ffc947718 tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] [instance: 71999b04-bf6e-4e4c-87c9-d5ec1932a80f] Updating instance_info_cache with network_info: [{"id": "927351ab-9ed5-41c7-beea-77033dc7601b", "address": "fa:16:3e:66:b1:8e", "network": {"id": "b7e78559-d885-4b59-9c7f-9d966b424d04", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-1456721354-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7ab95fb72b5d46c3b6c7bebbccf897cd", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d891a662-7da0-4be5-8d0d-01b4ff70552a", "external-id": "nsx-vlan-transportzone-815", "segmentation_id": 815, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap927351ab-9e", "ovs_interfaceid": "927351ab-9ed5-41c7-beea-77033dc7601b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61806) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1008.037492] env[61806]: DEBUG nova.compute.manager [None req-4b7244a6-0e60-4b9b-ac2e-5c7b0ff2b8c7 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] [instance: d0669232-6f33-4b30-97c2-2e4239af1ad0] Start spawning the instance on the hypervisor. {{(pid=61806) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1008.048802] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [None req-1c35eaa5-2854-4e3e-bc2c-4ef5a33b9a12 tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] [instance: 9357e61f-1628-43bd-ab46-de13c1529f51] Powering on the VM {{(pid=61806) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1008.049830] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-58f9d796-067c-42f5-9ded-e77414b7acd7 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1008.057876] env[61806]: DEBUG oslo_vmware.api [None req-1c35eaa5-2854-4e3e-bc2c-4ef5a33b9a12 tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Waiting for the task: (returnval){ [ 1008.057876] env[61806]: value = "task-1295080" [ 1008.057876] env[61806]: _type = "Task" [ 1008.057876] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1008.065909] env[61806]: DEBUG nova.virt.hardware [None req-4b7244a6-0e60-4b9b-ac2e-5c7b0ff2b8c7 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-15T18:05:10Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-15T18:04:52Z,direct_url=,disk_format='vmdk',id=ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='84ba83fce2c349a6988173c5d6fc3b07',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-15T18:04:53Z,virtual_size=,visibility=), allow threads: False {{(pid=61806) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1008.066197] env[61806]: DEBUG nova.virt.hardware [None req-4b7244a6-0e60-4b9b-ac2e-5c7b0ff2b8c7 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Flavor limits 0:0:0 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1008.066362] env[61806]: DEBUG nova.virt.hardware [None req-4b7244a6-0e60-4b9b-ac2e-5c7b0ff2b8c7 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Image limits 0:0:0 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1008.066548] env[61806]: DEBUG nova.virt.hardware [None req-4b7244a6-0e60-4b9b-ac2e-5c7b0ff2b8c7 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Flavor pref 0:0:0 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1008.066703] env[61806]: DEBUG nova.virt.hardware [None req-4b7244a6-0e60-4b9b-ac2e-5c7b0ff2b8c7 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Image pref 0:0:0 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1008.066871] env[61806]: DEBUG nova.virt.hardware [None req-4b7244a6-0e60-4b9b-ac2e-5c7b0ff2b8c7 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1008.067149] env[61806]: DEBUG nova.virt.hardware [None req-4b7244a6-0e60-4b9b-ac2e-5c7b0ff2b8c7 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61806) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1008.067337] env[61806]: DEBUG nova.virt.hardware [None req-4b7244a6-0e60-4b9b-ac2e-5c7b0ff2b8c7 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61806) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1008.067518] env[61806]: DEBUG nova.virt.hardware [None req-4b7244a6-0e60-4b9b-ac2e-5c7b0ff2b8c7 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Got 1 possible topologies {{(pid=61806) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1008.067689] env[61806]: DEBUG nova.virt.hardware [None req-4b7244a6-0e60-4b9b-ac2e-5c7b0ff2b8c7 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61806) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1008.067873] env[61806]: DEBUG nova.virt.hardware [None req-4b7244a6-0e60-4b9b-ac2e-5c7b0ff2b8c7 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61806) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1008.068691] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-732cb286-b1bf-4ec8-9a35-92f815031f2b {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1008.074317] env[61806]: DEBUG oslo_vmware.api [None req-1c35eaa5-2854-4e3e-bc2c-4ef5a33b9a12 tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Task: {'id': task-1295080, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1008.080081] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-90727289-4ac7-4507-96a8-cd95050c3497 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1008.349424] env[61806]: DEBUG oslo_vmware.api [None req-b4773569-1b9a-41b1-9eb3-234618dc8825 tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] Task: {'id': task-1295079, 'name': Rename_Task, 'duration_secs': 0.182414} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1008.350247] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [None req-b4773569-1b9a-41b1-9eb3-234618dc8825 tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] [instance: fe1d4ace-7de3-423b-b423-9d7397aa2413] Powering on the VM {{(pid=61806) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1008.350247] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-b1719f6b-b213-4da5-8d30-4f2afa5000c2 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1008.357844] env[61806]: DEBUG oslo_concurrency.lockutils [None req-8755c097-96dd-4051-93b1-330ffc947718 tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Releasing lock "refresh_cache-71999b04-bf6e-4e4c-87c9-d5ec1932a80f" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1008.357844] env[61806]: DEBUG nova.compute.manager [None req-8755c097-96dd-4051-93b1-330ffc947718 tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] [instance: 71999b04-bf6e-4e4c-87c9-d5ec1932a80f] Instance network_info: |[{"id": "927351ab-9ed5-41c7-beea-77033dc7601b", "address": "fa:16:3e:66:b1:8e", "network": {"id": "b7e78559-d885-4b59-9c7f-9d966b424d04", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-1456721354-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7ab95fb72b5d46c3b6c7bebbccf897cd", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d891a662-7da0-4be5-8d0d-01b4ff70552a", "external-id": "nsx-vlan-transportzone-815", "segmentation_id": 815, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap927351ab-9e", "ovs_interfaceid": "927351ab-9ed5-41c7-beea-77033dc7601b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61806) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1008.359914] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-8755c097-96dd-4051-93b1-330ffc947718 tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] [instance: 71999b04-bf6e-4e4c-87c9-d5ec1932a80f] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:66:b1:8e', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'd891a662-7da0-4be5-8d0d-01b4ff70552a', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '927351ab-9ed5-41c7-beea-77033dc7601b', 'vif_model': 'vmxnet3'}] {{(pid=61806) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1008.370099] env[61806]: DEBUG oslo.service.loopingcall [None req-8755c097-96dd-4051-93b1-330ffc947718 tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61806) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1008.370756] env[61806]: DEBUG oslo_vmware.api [None req-b4773569-1b9a-41b1-9eb3-234618dc8825 tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] Waiting for the task: (returnval){ [ 1008.370756] env[61806]: value = "task-1295081" [ 1008.370756] env[61806]: _type = "Task" [ 1008.370756] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1008.371590] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 71999b04-bf6e-4e4c-87c9-d5ec1932a80f] Creating VM on the ESX host {{(pid=61806) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 1008.372073] env[61806]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-bc6c8352-4341-4e5c-94f9-a2606912cefc {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1008.402064] env[61806]: DEBUG oslo_vmware.api [None req-b4773569-1b9a-41b1-9eb3-234618dc8825 tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] Task: {'id': task-1295081, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1008.407073] env[61806]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1008.407073] env[61806]: value = "task-1295082" [ 1008.407073] env[61806]: _type = "Task" [ 1008.407073] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1008.416455] env[61806]: DEBUG oslo_vmware.api [-] Task: {'id': task-1295082, 'name': CreateVM_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1008.449249] env[61806]: DEBUG nova.network.neutron [None req-4b7244a6-0e60-4b9b-ac2e-5c7b0ff2b8c7 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] [instance: d0669232-6f33-4b30-97c2-2e4239af1ad0] Successfully updated port: f958e9a3-f8d6-4665-a59f-876aab66edfb {{(pid=61806) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1008.483809] env[61806]: DEBUG nova.compute.manager [req-bf432983-ec3d-463a-934e-f98a5b6d151b req-976c8dfe-17a8-4ac4-affe-8048f87a1df9 service nova] [instance: 71999b04-bf6e-4e4c-87c9-d5ec1932a80f] Received event network-changed-927351ab-9ed5-41c7-beea-77033dc7601b {{(pid=61806) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1008.486893] env[61806]: DEBUG nova.compute.manager [req-bf432983-ec3d-463a-934e-f98a5b6d151b req-976c8dfe-17a8-4ac4-affe-8048f87a1df9 service nova] [instance: 71999b04-bf6e-4e4c-87c9-d5ec1932a80f] Refreshing instance network info cache due to event network-changed-927351ab-9ed5-41c7-beea-77033dc7601b. {{(pid=61806) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 1008.487318] env[61806]: DEBUG oslo_concurrency.lockutils [req-bf432983-ec3d-463a-934e-f98a5b6d151b req-976c8dfe-17a8-4ac4-affe-8048f87a1df9 service nova] Acquiring lock "refresh_cache-71999b04-bf6e-4e4c-87c9-d5ec1932a80f" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1008.487927] env[61806]: DEBUG oslo_concurrency.lockutils [req-bf432983-ec3d-463a-934e-f98a5b6d151b req-976c8dfe-17a8-4ac4-affe-8048f87a1df9 service nova] Acquired lock "refresh_cache-71999b04-bf6e-4e4c-87c9-d5ec1932a80f" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1008.487927] env[61806]: DEBUG nova.network.neutron [req-bf432983-ec3d-463a-934e-f98a5b6d151b req-976c8dfe-17a8-4ac4-affe-8048f87a1df9 service nova] [instance: 71999b04-bf6e-4e4c-87c9-d5ec1932a80f] Refreshing network info cache for port 927351ab-9ed5-41c7-beea-77033dc7601b {{(pid=61806) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1008.571577] env[61806]: DEBUG oslo_vmware.api [None req-1c35eaa5-2854-4e3e-bc2c-4ef5a33b9a12 tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Task: {'id': task-1295080, 'name': PowerOnVM_Task} progress is 100%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1008.589077] env[61806]: DEBUG oslo_concurrency.lockutils [None req-a22e273c-c0ad-47bf-9acf-b3a1dd13247a tempest-AttachVolumeNegativeTest-1683033704 tempest-AttachVolumeNegativeTest-1683033704-project-member] Acquiring lock "e4638faa-7be1-4909-a595-c437837fa314" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1008.589319] env[61806]: DEBUG oslo_concurrency.lockutils [None req-a22e273c-c0ad-47bf-9acf-b3a1dd13247a tempest-AttachVolumeNegativeTest-1683033704 tempest-AttachVolumeNegativeTest-1683033704-project-member] Lock "e4638faa-7be1-4909-a595-c437837fa314" acquired by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: waited 0.000s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1008.696586] env[61806]: DEBUG oslo_concurrency.lockutils [None req-9bf95660-956b-4332-b2b6-3c7de9ec6c45 tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Acquiring lock "67d1bdd8-cf8f-4fb5-bd3c-42008e4d60e3" by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1008.696941] env[61806]: DEBUG oslo_concurrency.lockutils [None req-9bf95660-956b-4332-b2b6-3c7de9ec6c45 tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Lock "67d1bdd8-cf8f-4fb5-bd3c-42008e4d60e3" acquired by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" :: waited 0.001s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1008.697204] env[61806]: DEBUG nova.compute.manager [None req-9bf95660-956b-4332-b2b6-3c7de9ec6c45 tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] [instance: 67d1bdd8-cf8f-4fb5-bd3c-42008e4d60e3] Going to confirm migration 2 {{(pid=61806) do_confirm_resize /opt/stack/nova/nova/compute/manager.py:4783}} [ 1008.881964] env[61806]: DEBUG oslo_vmware.api [None req-b4773569-1b9a-41b1-9eb3-234618dc8825 tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] Task: {'id': task-1295081, 'name': PowerOnVM_Task} progress is 89%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1008.916094] env[61806]: DEBUG oslo_vmware.api [-] Task: {'id': task-1295082, 'name': CreateVM_Task, 'duration_secs': 0.432109} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1008.916094] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 71999b04-bf6e-4e4c-87c9-d5ec1932a80f] Created VM on the ESX host {{(pid=61806) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 1008.916773] env[61806]: DEBUG oslo_concurrency.lockutils [None req-8755c097-96dd-4051-93b1-330ffc947718 tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1008.917061] env[61806]: DEBUG oslo_concurrency.lockutils [None req-8755c097-96dd-4051-93b1-330ffc947718 tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Acquired lock "[datastore2] devstack-image-cache_base/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1008.917460] env[61806]: DEBUG oslo_concurrency.lockutils [None req-8755c097-96dd-4051-93b1-330ffc947718 tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1008.917757] env[61806]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f6f8e6fc-71e6-4064-a6f6-89cdddaca2df {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1008.923607] env[61806]: DEBUG oslo_vmware.api [None req-8755c097-96dd-4051-93b1-330ffc947718 tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Waiting for the task: (returnval){ [ 1008.923607] env[61806]: value = "session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]5270e1a7-3c15-f23a-d950-216fb86d9d02" [ 1008.923607] env[61806]: _type = "Task" [ 1008.923607] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1008.934416] env[61806]: DEBUG oslo_vmware.api [None req-8755c097-96dd-4051-93b1-330ffc947718 tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Task: {'id': session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]5270e1a7-3c15-f23a-d950-216fb86d9d02, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1008.952434] env[61806]: DEBUG oslo_concurrency.lockutils [None req-4b7244a6-0e60-4b9b-ac2e-5c7b0ff2b8c7 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Acquiring lock "refresh_cache-d0669232-6f33-4b30-97c2-2e4239af1ad0" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1008.952736] env[61806]: DEBUG oslo_concurrency.lockutils [None req-4b7244a6-0e60-4b9b-ac2e-5c7b0ff2b8c7 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Acquired lock "refresh_cache-d0669232-6f33-4b30-97c2-2e4239af1ad0" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1008.952985] env[61806]: DEBUG nova.network.neutron [None req-4b7244a6-0e60-4b9b-ac2e-5c7b0ff2b8c7 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] [instance: d0669232-6f33-4b30-97c2-2e4239af1ad0] Building network info cache for instance {{(pid=61806) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1009.073778] env[61806]: DEBUG oslo_vmware.api [None req-1c35eaa5-2854-4e3e-bc2c-4ef5a33b9a12 tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Task: {'id': task-1295080, 'name': PowerOnVM_Task, 'duration_secs': 0.56946} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1009.074184] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [None req-1c35eaa5-2854-4e3e-bc2c-4ef5a33b9a12 tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] [instance: 9357e61f-1628-43bd-ab46-de13c1529f51] Powered on the VM {{(pid=61806) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1009.074579] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-1c35eaa5-2854-4e3e-bc2c-4ef5a33b9a12 tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] [instance: 9357e61f-1628-43bd-ab46-de13c1529f51] Updating instance '9357e61f-1628-43bd-ab46-de13c1529f51' progress to 100 {{(pid=61806) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 1009.092455] env[61806]: DEBUG nova.compute.utils [None req-a22e273c-c0ad-47bf-9acf-b3a1dd13247a tempest-AttachVolumeNegativeTest-1683033704 tempest-AttachVolumeNegativeTest-1683033704-project-member] Using /dev/sd instead of None {{(pid=61806) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1009.320635] env[61806]: DEBUG oslo_concurrency.lockutils [None req-9bf95660-956b-4332-b2b6-3c7de9ec6c45 tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Acquiring lock "refresh_cache-67d1bdd8-cf8f-4fb5-bd3c-42008e4d60e3" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1009.320841] env[61806]: DEBUG oslo_concurrency.lockutils [None req-9bf95660-956b-4332-b2b6-3c7de9ec6c45 tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Acquired lock "refresh_cache-67d1bdd8-cf8f-4fb5-bd3c-42008e4d60e3" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1009.321144] env[61806]: DEBUG nova.network.neutron [None req-9bf95660-956b-4332-b2b6-3c7de9ec6c45 tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] [instance: 67d1bdd8-cf8f-4fb5-bd3c-42008e4d60e3] Building network info cache for instance {{(pid=61806) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1009.321379] env[61806]: DEBUG nova.objects.instance [None req-9bf95660-956b-4332-b2b6-3c7de9ec6c45 tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Lazy-loading 'info_cache' on Instance uuid 67d1bdd8-cf8f-4fb5-bd3c-42008e4d60e3 {{(pid=61806) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1009.333552] env[61806]: DEBUG oslo_concurrency.lockutils [None req-de9a979c-45ed-463a-871e-b7dda8c366a2 tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1009.333817] env[61806]: DEBUG oslo_concurrency.lockutils [None req-de9a979c-45ed-463a-871e-b7dda8c366a2 tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.000s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1009.334044] env[61806]: DEBUG nova.objects.instance [None req-de9a979c-45ed-463a-871e-b7dda8c366a2 tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] Lazy-loading 'pci_requests' on Instance uuid fd9e520a-2e7b-43d4-b84d-8933c32cc6a6 {{(pid=61806) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1009.362828] env[61806]: DEBUG nova.network.neutron [req-bf432983-ec3d-463a-934e-f98a5b6d151b req-976c8dfe-17a8-4ac4-affe-8048f87a1df9 service nova] [instance: 71999b04-bf6e-4e4c-87c9-d5ec1932a80f] Updated VIF entry in instance network info cache for port 927351ab-9ed5-41c7-beea-77033dc7601b. {{(pid=61806) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1009.363236] env[61806]: DEBUG nova.network.neutron [req-bf432983-ec3d-463a-934e-f98a5b6d151b req-976c8dfe-17a8-4ac4-affe-8048f87a1df9 service nova] [instance: 71999b04-bf6e-4e4c-87c9-d5ec1932a80f] Updating instance_info_cache with network_info: [{"id": "927351ab-9ed5-41c7-beea-77033dc7601b", "address": "fa:16:3e:66:b1:8e", "network": {"id": "b7e78559-d885-4b59-9c7f-9d966b424d04", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-1456721354-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7ab95fb72b5d46c3b6c7bebbccf897cd", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d891a662-7da0-4be5-8d0d-01b4ff70552a", "external-id": "nsx-vlan-transportzone-815", "segmentation_id": 815, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap927351ab-9e", "ovs_interfaceid": "927351ab-9ed5-41c7-beea-77033dc7601b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61806) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1009.382912] env[61806]: DEBUG oslo_vmware.api [None req-b4773569-1b9a-41b1-9eb3-234618dc8825 tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] Task: {'id': task-1295081, 'name': PowerOnVM_Task, 'duration_secs': 0.720269} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1009.383521] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [None req-b4773569-1b9a-41b1-9eb3-234618dc8825 tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] [instance: fe1d4ace-7de3-423b-b423-9d7397aa2413] Powered on the VM {{(pid=61806) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1009.383615] env[61806]: INFO nova.compute.manager [None req-b4773569-1b9a-41b1-9eb3-234618dc8825 tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] [instance: fe1d4ace-7de3-423b-b423-9d7397aa2413] Took 8.07 seconds to spawn the instance on the hypervisor. [ 1009.383737] env[61806]: DEBUG nova.compute.manager [None req-b4773569-1b9a-41b1-9eb3-234618dc8825 tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] [instance: fe1d4ace-7de3-423b-b423-9d7397aa2413] Checking state {{(pid=61806) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1009.384545] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ff17c45c-485c-44a1-88a0-d02118ac88f6 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1009.436303] env[61806]: DEBUG oslo_vmware.api [None req-8755c097-96dd-4051-93b1-330ffc947718 tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Task: {'id': session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]5270e1a7-3c15-f23a-d950-216fb86d9d02, 'name': SearchDatastore_Task, 'duration_secs': 0.01365} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1009.436613] env[61806]: DEBUG oslo_concurrency.lockutils [None req-8755c097-96dd-4051-93b1-330ffc947718 tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Releasing lock "[datastore2] devstack-image-cache_base/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1009.436856] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-8755c097-96dd-4051-93b1-330ffc947718 tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] [instance: 71999b04-bf6e-4e4c-87c9-d5ec1932a80f] Processing image ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa {{(pid=61806) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1009.437117] env[61806]: DEBUG oslo_concurrency.lockutils [None req-8755c097-96dd-4051-93b1-330ffc947718 tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa.vmdk" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1009.437275] env[61806]: DEBUG oslo_concurrency.lockutils [None req-8755c097-96dd-4051-93b1-330ffc947718 tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Acquired lock "[datastore2] devstack-image-cache_base/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa.vmdk" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1009.437461] env[61806]: DEBUG nova.virt.vmwareapi.ds_util [None req-8755c097-96dd-4051-93b1-330ffc947718 tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61806) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1009.437724] env[61806]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-d07636e3-69ec-445c-93af-5c3a3e83abd6 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1009.447875] env[61806]: DEBUG nova.virt.vmwareapi.ds_util [None req-8755c097-96dd-4051-93b1-330ffc947718 tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61806) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1009.448085] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-8755c097-96dd-4051-93b1-330ffc947718 tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61806) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 1009.448890] env[61806]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-a32f535e-7322-4c63-a986-28def8f84336 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1009.457388] env[61806]: DEBUG oslo_vmware.api [None req-8755c097-96dd-4051-93b1-330ffc947718 tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Waiting for the task: (returnval){ [ 1009.457388] env[61806]: value = "session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]52535aa6-ba53-61b1-d824-be482d92acf6" [ 1009.457388] env[61806]: _type = "Task" [ 1009.457388] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1009.467450] env[61806]: DEBUG oslo_vmware.api [None req-8755c097-96dd-4051-93b1-330ffc947718 tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Task: {'id': session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]52535aa6-ba53-61b1-d824-be482d92acf6, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1009.489298] env[61806]: DEBUG nova.network.neutron [None req-4b7244a6-0e60-4b9b-ac2e-5c7b0ff2b8c7 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] [instance: d0669232-6f33-4b30-97c2-2e4239af1ad0] Instance cache missing network info. {{(pid=61806) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 1009.597431] env[61806]: DEBUG oslo_concurrency.lockutils [None req-a22e273c-c0ad-47bf-9acf-b3a1dd13247a tempest-AttachVolumeNegativeTest-1683033704 tempest-AttachVolumeNegativeTest-1683033704-project-member] Lock "e4638faa-7be1-4909-a595-c437837fa314" "released" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: held 1.008s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1009.663645] env[61806]: DEBUG nova.network.neutron [None req-4b7244a6-0e60-4b9b-ac2e-5c7b0ff2b8c7 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] [instance: d0669232-6f33-4b30-97c2-2e4239af1ad0] Updating instance_info_cache with network_info: [{"id": "f958e9a3-f8d6-4665-a59f-876aab66edfb", "address": "fa:16:3e:fc:93:23", "network": {"id": "33bf5b21-82b0-432e-ae3d-9c2ca75161d0", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-1294159336-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "1c55cc0455584fbb8cda88d33b8f8b7e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "5b0fa7a2-ebd9-4788-8904-7bf250ce466c", "external-id": "nsx-vlan-transportzone-669", "segmentation_id": 669, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf958e9a3-f8", "ovs_interfaceid": "f958e9a3-f8d6-4665-a59f-876aab66edfb", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61806) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1009.838143] env[61806]: DEBUG nova.objects.instance [None req-de9a979c-45ed-463a-871e-b7dda8c366a2 tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] Lazy-loading 'numa_topology' on Instance uuid fd9e520a-2e7b-43d4-b84d-8933c32cc6a6 {{(pid=61806) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1009.865360] env[61806]: DEBUG oslo_concurrency.lockutils [req-bf432983-ec3d-463a-934e-f98a5b6d151b req-976c8dfe-17a8-4ac4-affe-8048f87a1df9 service nova] Releasing lock "refresh_cache-71999b04-bf6e-4e4c-87c9-d5ec1932a80f" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1009.866030] env[61806]: DEBUG nova.compute.manager [req-bf432983-ec3d-463a-934e-f98a5b6d151b req-976c8dfe-17a8-4ac4-affe-8048f87a1df9 service nova] [instance: d0669232-6f33-4b30-97c2-2e4239af1ad0] Received event network-vif-plugged-f958e9a3-f8d6-4665-a59f-876aab66edfb {{(pid=61806) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1009.866030] env[61806]: DEBUG oslo_concurrency.lockutils [req-bf432983-ec3d-463a-934e-f98a5b6d151b req-976c8dfe-17a8-4ac4-affe-8048f87a1df9 service nova] Acquiring lock "d0669232-6f33-4b30-97c2-2e4239af1ad0-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1009.866184] env[61806]: DEBUG oslo_concurrency.lockutils [req-bf432983-ec3d-463a-934e-f98a5b6d151b req-976c8dfe-17a8-4ac4-affe-8048f87a1df9 service nova] Lock "d0669232-6f33-4b30-97c2-2e4239af1ad0-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1009.866229] env[61806]: DEBUG oslo_concurrency.lockutils [req-bf432983-ec3d-463a-934e-f98a5b6d151b req-976c8dfe-17a8-4ac4-affe-8048f87a1df9 service nova] Lock "d0669232-6f33-4b30-97c2-2e4239af1ad0-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1009.866399] env[61806]: DEBUG nova.compute.manager [req-bf432983-ec3d-463a-934e-f98a5b6d151b req-976c8dfe-17a8-4ac4-affe-8048f87a1df9 service nova] [instance: d0669232-6f33-4b30-97c2-2e4239af1ad0] No waiting events found dispatching network-vif-plugged-f958e9a3-f8d6-4665-a59f-876aab66edfb {{(pid=61806) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1009.866575] env[61806]: WARNING nova.compute.manager [req-bf432983-ec3d-463a-934e-f98a5b6d151b req-976c8dfe-17a8-4ac4-affe-8048f87a1df9 service nova] [instance: d0669232-6f33-4b30-97c2-2e4239af1ad0] Received unexpected event network-vif-plugged-f958e9a3-f8d6-4665-a59f-876aab66edfb for instance with vm_state building and task_state spawning. [ 1009.866747] env[61806]: DEBUG nova.compute.manager [req-bf432983-ec3d-463a-934e-f98a5b6d151b req-976c8dfe-17a8-4ac4-affe-8048f87a1df9 service nova] [instance: d0669232-6f33-4b30-97c2-2e4239af1ad0] Received event network-changed-f958e9a3-f8d6-4665-a59f-876aab66edfb {{(pid=61806) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1009.866906] env[61806]: DEBUG nova.compute.manager [req-bf432983-ec3d-463a-934e-f98a5b6d151b req-976c8dfe-17a8-4ac4-affe-8048f87a1df9 service nova] [instance: d0669232-6f33-4b30-97c2-2e4239af1ad0] Refreshing instance network info cache due to event network-changed-f958e9a3-f8d6-4665-a59f-876aab66edfb. {{(pid=61806) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 1009.867137] env[61806]: DEBUG oslo_concurrency.lockutils [req-bf432983-ec3d-463a-934e-f98a5b6d151b req-976c8dfe-17a8-4ac4-affe-8048f87a1df9 service nova] Acquiring lock "refresh_cache-d0669232-6f33-4b30-97c2-2e4239af1ad0" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1009.901460] env[61806]: INFO nova.compute.manager [None req-b4773569-1b9a-41b1-9eb3-234618dc8825 tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] [instance: fe1d4ace-7de3-423b-b423-9d7397aa2413] Took 18.04 seconds to build instance. [ 1009.971051] env[61806]: DEBUG oslo_vmware.api [None req-8755c097-96dd-4051-93b1-330ffc947718 tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Task: {'id': session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]52535aa6-ba53-61b1-d824-be482d92acf6, 'name': SearchDatastore_Task, 'duration_secs': 0.01388} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1009.972048] env[61806]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-a6403673-26d7-4ea4-9d02-ccf3f4e806d2 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1009.978626] env[61806]: DEBUG oslo_vmware.api [None req-8755c097-96dd-4051-93b1-330ffc947718 tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Waiting for the task: (returnval){ [ 1009.978626] env[61806]: value = "session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]5264c158-ae03-10fa-d6e9-641ae6944a41" [ 1009.978626] env[61806]: _type = "Task" [ 1009.978626] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1009.988018] env[61806]: DEBUG oslo_vmware.api [None req-8755c097-96dd-4051-93b1-330ffc947718 tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Task: {'id': session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]5264c158-ae03-10fa-d6e9-641ae6944a41, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1010.167286] env[61806]: DEBUG oslo_concurrency.lockutils [None req-4b7244a6-0e60-4b9b-ac2e-5c7b0ff2b8c7 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Releasing lock "refresh_cache-d0669232-6f33-4b30-97c2-2e4239af1ad0" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1010.167614] env[61806]: DEBUG nova.compute.manager [None req-4b7244a6-0e60-4b9b-ac2e-5c7b0ff2b8c7 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] [instance: d0669232-6f33-4b30-97c2-2e4239af1ad0] Instance network_info: |[{"id": "f958e9a3-f8d6-4665-a59f-876aab66edfb", "address": "fa:16:3e:fc:93:23", "network": {"id": "33bf5b21-82b0-432e-ae3d-9c2ca75161d0", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-1294159336-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "1c55cc0455584fbb8cda88d33b8f8b7e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "5b0fa7a2-ebd9-4788-8904-7bf250ce466c", "external-id": "nsx-vlan-transportzone-669", "segmentation_id": 669, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf958e9a3-f8", "ovs_interfaceid": "f958e9a3-f8d6-4665-a59f-876aab66edfb", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61806) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1010.168850] env[61806]: DEBUG oslo_concurrency.lockutils [req-bf432983-ec3d-463a-934e-f98a5b6d151b req-976c8dfe-17a8-4ac4-affe-8048f87a1df9 service nova] Acquired lock "refresh_cache-d0669232-6f33-4b30-97c2-2e4239af1ad0" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1010.169085] env[61806]: DEBUG nova.network.neutron [req-bf432983-ec3d-463a-934e-f98a5b6d151b req-976c8dfe-17a8-4ac4-affe-8048f87a1df9 service nova] [instance: d0669232-6f33-4b30-97c2-2e4239af1ad0] Refreshing network info cache for port f958e9a3-f8d6-4665-a59f-876aab66edfb {{(pid=61806) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1010.170315] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-4b7244a6-0e60-4b9b-ac2e-5c7b0ff2b8c7 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] [instance: d0669232-6f33-4b30-97c2-2e4239af1ad0] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:fc:93:23', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '5b0fa7a2-ebd9-4788-8904-7bf250ce466c', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'f958e9a3-f8d6-4665-a59f-876aab66edfb', 'vif_model': 'vmxnet3'}] {{(pid=61806) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1010.178081] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [None req-4b7244a6-0e60-4b9b-ac2e-5c7b0ff2b8c7 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Creating folder: Project (1c55cc0455584fbb8cda88d33b8f8b7e). Parent ref: group-v277609. {{(pid=61806) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 1010.181315] env[61806]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-7fcec3b4-6ff4-4d6c-b20f-871b8844664c {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1010.195595] env[61806]: INFO nova.virt.vmwareapi.vm_util [None req-4b7244a6-0e60-4b9b-ac2e-5c7b0ff2b8c7 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Created folder: Project (1c55cc0455584fbb8cda88d33b8f8b7e) in parent group-v277609. [ 1010.195835] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [None req-4b7244a6-0e60-4b9b-ac2e-5c7b0ff2b8c7 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Creating folder: Instances. Parent ref: group-v277742. {{(pid=61806) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 1010.196121] env[61806]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-4101b2a0-269a-408b-80f7-cd641b2b8d79 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1010.207138] env[61806]: INFO nova.virt.vmwareapi.vm_util [None req-4b7244a6-0e60-4b9b-ac2e-5c7b0ff2b8c7 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Created folder: Instances in parent group-v277742. [ 1010.207356] env[61806]: DEBUG oslo.service.loopingcall [None req-4b7244a6-0e60-4b9b-ac2e-5c7b0ff2b8c7 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61806) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1010.207556] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: d0669232-6f33-4b30-97c2-2e4239af1ad0] Creating VM on the ESX host {{(pid=61806) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 1010.207769] env[61806]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-6f62b9f4-b626-43f4-954b-61032abef4ab {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1010.229214] env[61806]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1010.229214] env[61806]: value = "task-1295085" [ 1010.229214] env[61806]: _type = "Task" [ 1010.229214] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1010.238147] env[61806]: DEBUG oslo_vmware.api [-] Task: {'id': task-1295085, 'name': CreateVM_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1010.340565] env[61806]: INFO nova.compute.claims [None req-de9a979c-45ed-463a-871e-b7dda8c366a2 tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] [instance: fd9e520a-2e7b-43d4-b84d-8933c32cc6a6] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1010.403415] env[61806]: DEBUG oslo_concurrency.lockutils [None req-b4773569-1b9a-41b1-9eb3-234618dc8825 tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] Lock "fe1d4ace-7de3-423b-b423-9d7397aa2413" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 20.174s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1010.491133] env[61806]: DEBUG oslo_vmware.api [None req-8755c097-96dd-4051-93b1-330ffc947718 tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Task: {'id': session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]5264c158-ae03-10fa-d6e9-641ae6944a41, 'name': SearchDatastore_Task, 'duration_secs': 0.017677} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1010.491492] env[61806]: DEBUG oslo_concurrency.lockutils [None req-8755c097-96dd-4051-93b1-330ffc947718 tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Releasing lock "[datastore2] devstack-image-cache_base/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa.vmdk" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1010.491924] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [None req-8755c097-96dd-4051-93b1-330ffc947718 tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa.vmdk to [datastore2] 71999b04-bf6e-4e4c-87c9-d5ec1932a80f/71999b04-bf6e-4e4c-87c9-d5ec1932a80f.vmdk {{(pid=61806) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 1010.492183] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-07f44ea2-7dcd-4230-bc32-08d788f1c0f1 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1010.509768] env[61806]: DEBUG oslo_vmware.api [None req-8755c097-96dd-4051-93b1-330ffc947718 tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Waiting for the task: (returnval){ [ 1010.509768] env[61806]: value = "task-1295086" [ 1010.509768] env[61806]: _type = "Task" [ 1010.509768] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1010.519756] env[61806]: DEBUG oslo_vmware.api [None req-8755c097-96dd-4051-93b1-330ffc947718 tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Task: {'id': task-1295086, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1010.549902] env[61806]: DEBUG nova.network.neutron [req-bf432983-ec3d-463a-934e-f98a5b6d151b req-976c8dfe-17a8-4ac4-affe-8048f87a1df9 service nova] [instance: d0669232-6f33-4b30-97c2-2e4239af1ad0] Updated VIF entry in instance network info cache for port f958e9a3-f8d6-4665-a59f-876aab66edfb. {{(pid=61806) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1010.549902] env[61806]: DEBUG nova.network.neutron [req-bf432983-ec3d-463a-934e-f98a5b6d151b req-976c8dfe-17a8-4ac4-affe-8048f87a1df9 service nova] [instance: d0669232-6f33-4b30-97c2-2e4239af1ad0] Updating instance_info_cache with network_info: [{"id": "f958e9a3-f8d6-4665-a59f-876aab66edfb", "address": "fa:16:3e:fc:93:23", "network": {"id": "33bf5b21-82b0-432e-ae3d-9c2ca75161d0", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-1294159336-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "1c55cc0455584fbb8cda88d33b8f8b7e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "5b0fa7a2-ebd9-4788-8904-7bf250ce466c", "external-id": "nsx-vlan-transportzone-669", "segmentation_id": 669, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf958e9a3-f8", "ovs_interfaceid": "f958e9a3-f8d6-4665-a59f-876aab66edfb", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61806) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1010.659786] env[61806]: DEBUG oslo_concurrency.lockutils [None req-a22e273c-c0ad-47bf-9acf-b3a1dd13247a tempest-AttachVolumeNegativeTest-1683033704 tempest-AttachVolumeNegativeTest-1683033704-project-member] Acquiring lock "e4638faa-7be1-4909-a595-c437837fa314" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1010.660099] env[61806]: DEBUG oslo_concurrency.lockutils [None req-a22e273c-c0ad-47bf-9acf-b3a1dd13247a tempest-AttachVolumeNegativeTest-1683033704 tempest-AttachVolumeNegativeTest-1683033704-project-member] Lock "e4638faa-7be1-4909-a595-c437837fa314" acquired by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: waited 0.000s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1010.660335] env[61806]: INFO nova.compute.manager [None req-a22e273c-c0ad-47bf-9acf-b3a1dd13247a tempest-AttachVolumeNegativeTest-1683033704 tempest-AttachVolumeNegativeTest-1683033704-project-member] [instance: e4638faa-7be1-4909-a595-c437837fa314] Attaching volume 1492823b-b43f-42fd-bc67-94fe82531df9 to /dev/sdb [ 1010.700481] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6a82f9ae-352b-4901-9887-2fb5dd0c77ec {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1010.707332] env[61806]: DEBUG nova.network.neutron [None req-9bf95660-956b-4332-b2b6-3c7de9ec6c45 tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] [instance: 67d1bdd8-cf8f-4fb5-bd3c-42008e4d60e3] Updating instance_info_cache with network_info: [{"id": "859fb723-4a1c-4ea8-bd7c-0ed308f3a9e3", "address": "fa:16:3e:3c:40:d6", "network": {"id": "746a03b6-c1a6-4d27-98b9-0bee6376634f", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-1971382587-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.245", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c1f2d161c50f4b70987b90e6364c3779", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6a31b224-77d7-48a4-af87-312758611b8e", "external-id": "nsx-vlan-transportzone-761", "segmentation_id": 761, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap859fb723-4a", "ovs_interfaceid": "859fb723-4a1c-4ea8-bd7c-0ed308f3a9e3", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61806) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1010.714232] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-08bf4f91-6449-4cdf-8c7a-58b823a8bfa6 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1010.731960] env[61806]: DEBUG nova.virt.block_device [None req-a22e273c-c0ad-47bf-9acf-b3a1dd13247a tempest-AttachVolumeNegativeTest-1683033704 tempest-AttachVolumeNegativeTest-1683033704-project-member] [instance: e4638faa-7be1-4909-a595-c437837fa314] Updating existing volume attachment record: a4e57c26-7fdd-4215-a1e2-4265f17fcf05 {{(pid=61806) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 1010.744947] env[61806]: DEBUG oslo_vmware.api [-] Task: {'id': task-1295085, 'name': CreateVM_Task, 'duration_secs': 0.370619} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1010.744947] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: d0669232-6f33-4b30-97c2-2e4239af1ad0] Created VM on the ESX host {{(pid=61806) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 1010.745448] env[61806]: DEBUG oslo_concurrency.lockutils [None req-4b7244a6-0e60-4b9b-ac2e-5c7b0ff2b8c7 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1010.745635] env[61806]: DEBUG oslo_concurrency.lockutils [None req-4b7244a6-0e60-4b9b-ac2e-5c7b0ff2b8c7 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Acquired lock "[datastore2] devstack-image-cache_base/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1010.746048] env[61806]: DEBUG oslo_concurrency.lockutils [None req-4b7244a6-0e60-4b9b-ac2e-5c7b0ff2b8c7 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1010.746927] env[61806]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-003534b0-da60-41e1-b5cd-54fae505b9e1 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1010.752357] env[61806]: DEBUG oslo_vmware.api [None req-4b7244a6-0e60-4b9b-ac2e-5c7b0ff2b8c7 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Waiting for the task: (returnval){ [ 1010.752357] env[61806]: value = "session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]52511701-4b54-a571-bae8-e7159549e946" [ 1010.752357] env[61806]: _type = "Task" [ 1010.752357] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1010.762352] env[61806]: DEBUG oslo_vmware.api [None req-4b7244a6-0e60-4b9b-ac2e-5c7b0ff2b8c7 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Task: {'id': session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]52511701-4b54-a571-bae8-e7159549e946, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1010.825727] env[61806]: DEBUG oslo_concurrency.lockutils [None req-8b3f01e9-87f4-442a-a746-46f4f6533d96 tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] Acquiring lock "fe1d4ace-7de3-423b-b423-9d7397aa2413" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1010.826071] env[61806]: DEBUG oslo_concurrency.lockutils [None req-8b3f01e9-87f4-442a-a746-46f4f6533d96 tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] Lock "fe1d4ace-7de3-423b-b423-9d7397aa2413" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1010.826348] env[61806]: DEBUG oslo_concurrency.lockutils [None req-8b3f01e9-87f4-442a-a746-46f4f6533d96 tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] Acquiring lock "fe1d4ace-7de3-423b-b423-9d7397aa2413-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1010.826582] env[61806]: DEBUG oslo_concurrency.lockutils [None req-8b3f01e9-87f4-442a-a746-46f4f6533d96 tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] Lock "fe1d4ace-7de3-423b-b423-9d7397aa2413-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1010.826792] env[61806]: DEBUG oslo_concurrency.lockutils [None req-8b3f01e9-87f4-442a-a746-46f4f6533d96 tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] Lock "fe1d4ace-7de3-423b-b423-9d7397aa2413-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1010.831199] env[61806]: INFO nova.compute.manager [None req-8b3f01e9-87f4-442a-a746-46f4f6533d96 tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] [instance: fe1d4ace-7de3-423b-b423-9d7397aa2413] Terminating instance [ 1010.844031] env[61806]: DEBUG nova.compute.manager [None req-8b3f01e9-87f4-442a-a746-46f4f6533d96 tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] [instance: fe1d4ace-7de3-423b-b423-9d7397aa2413] Start destroying the instance on the hypervisor. {{(pid=61806) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1010.844282] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-8b3f01e9-87f4-442a-a746-46f4f6533d96 tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] [instance: fe1d4ace-7de3-423b-b423-9d7397aa2413] Destroying instance {{(pid=61806) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1010.848204] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-78a99a64-af47-418b-b225-d011a63bd379 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1010.860356] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [None req-8b3f01e9-87f4-442a-a746-46f4f6533d96 tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] [instance: fe1d4ace-7de3-423b-b423-9d7397aa2413] Powering off the VM {{(pid=61806) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1010.860356] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-7a0f809a-93cf-4928-a3b6-80111753f6f4 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1010.869175] env[61806]: DEBUG oslo_vmware.api [None req-8b3f01e9-87f4-442a-a746-46f4f6533d96 tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] Waiting for the task: (returnval){ [ 1010.869175] env[61806]: value = "task-1295087" [ 1010.869175] env[61806]: _type = "Task" [ 1010.869175] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1010.887027] env[61806]: DEBUG oslo_vmware.api [None req-8b3f01e9-87f4-442a-a746-46f4f6533d96 tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] Task: {'id': task-1295087, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1011.022261] env[61806]: DEBUG oslo_vmware.api [None req-8755c097-96dd-4051-93b1-330ffc947718 tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Task: {'id': task-1295086, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1011.052656] env[61806]: DEBUG oslo_concurrency.lockutils [req-bf432983-ec3d-463a-934e-f98a5b6d151b req-976c8dfe-17a8-4ac4-affe-8048f87a1df9 service nova] Releasing lock "refresh_cache-d0669232-6f33-4b30-97c2-2e4239af1ad0" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1011.211288] env[61806]: DEBUG oslo_concurrency.lockutils [None req-9bf95660-956b-4332-b2b6-3c7de9ec6c45 tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Releasing lock "refresh_cache-67d1bdd8-cf8f-4fb5-bd3c-42008e4d60e3" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1011.211288] env[61806]: DEBUG nova.objects.instance [None req-9bf95660-956b-4332-b2b6-3c7de9ec6c45 tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Lazy-loading 'migration_context' on Instance uuid 67d1bdd8-cf8f-4fb5-bd3c-42008e4d60e3 {{(pid=61806) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1011.264230] env[61806]: DEBUG oslo_vmware.api [None req-4b7244a6-0e60-4b9b-ac2e-5c7b0ff2b8c7 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Task: {'id': session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]52511701-4b54-a571-bae8-e7159549e946, 'name': SearchDatastore_Task, 'duration_secs': 0.067874} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1011.264566] env[61806]: DEBUG oslo_concurrency.lockutils [None req-4b7244a6-0e60-4b9b-ac2e-5c7b0ff2b8c7 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Releasing lock "[datastore2] devstack-image-cache_base/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1011.264854] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-4b7244a6-0e60-4b9b-ac2e-5c7b0ff2b8c7 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] [instance: d0669232-6f33-4b30-97c2-2e4239af1ad0] Processing image ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa {{(pid=61806) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1011.265123] env[61806]: DEBUG oslo_concurrency.lockutils [None req-4b7244a6-0e60-4b9b-ac2e-5c7b0ff2b8c7 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa.vmdk" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1011.265284] env[61806]: DEBUG oslo_concurrency.lockutils [None req-4b7244a6-0e60-4b9b-ac2e-5c7b0ff2b8c7 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Acquired lock "[datastore2] devstack-image-cache_base/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa.vmdk" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1011.265473] env[61806]: DEBUG nova.virt.vmwareapi.ds_util [None req-4b7244a6-0e60-4b9b-ac2e-5c7b0ff2b8c7 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61806) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1011.265780] env[61806]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-ca647ce3-9e9a-4619-a3d4-75be5546ede3 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1011.275482] env[61806]: DEBUG nova.virt.vmwareapi.ds_util [None req-4b7244a6-0e60-4b9b-ac2e-5c7b0ff2b8c7 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61806) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1011.275690] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-4b7244a6-0e60-4b9b-ac2e-5c7b0ff2b8c7 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61806) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 1011.276406] env[61806]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-9702c65e-d9be-439c-b58b-5a2a504a1bc0 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1011.281995] env[61806]: DEBUG oslo_vmware.api [None req-4b7244a6-0e60-4b9b-ac2e-5c7b0ff2b8c7 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Waiting for the task: (returnval){ [ 1011.281995] env[61806]: value = "session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]521444a1-53e2-a648-4ff8-2531a5a31a25" [ 1011.281995] env[61806]: _type = "Task" [ 1011.281995] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1011.290174] env[61806]: DEBUG oslo_vmware.api [None req-4b7244a6-0e60-4b9b-ac2e-5c7b0ff2b8c7 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Task: {'id': session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]521444a1-53e2-a648-4ff8-2531a5a31a25, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1011.380708] env[61806]: DEBUG oslo_concurrency.lockutils [None req-666a872d-3efd-4dd5-b0c9-c25d09581b9c tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Acquiring lock "9357e61f-1628-43bd-ab46-de13c1529f51" by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1011.381072] env[61806]: DEBUG oslo_concurrency.lockutils [None req-666a872d-3efd-4dd5-b0c9-c25d09581b9c tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Lock "9357e61f-1628-43bd-ab46-de13c1529f51" acquired by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" :: waited 0.000s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1011.381236] env[61806]: DEBUG nova.compute.manager [None req-666a872d-3efd-4dd5-b0c9-c25d09581b9c tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] [instance: 9357e61f-1628-43bd-ab46-de13c1529f51] Going to confirm migration 3 {{(pid=61806) do_confirm_resize /opt/stack/nova/nova/compute/manager.py:4783}} [ 1011.382692] env[61806]: DEBUG oslo_vmware.api [None req-8b3f01e9-87f4-442a-a746-46f4f6533d96 tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] Task: {'id': task-1295087, 'name': PowerOffVM_Task, 'duration_secs': 0.367213} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1011.382934] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [None req-8b3f01e9-87f4-442a-a746-46f4f6533d96 tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] [instance: fe1d4ace-7de3-423b-b423-9d7397aa2413] Powered off the VM {{(pid=61806) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1011.383154] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-8b3f01e9-87f4-442a-a746-46f4f6533d96 tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] [instance: fe1d4ace-7de3-423b-b423-9d7397aa2413] Unregistering the VM {{(pid=61806) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1011.383436] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-ff0a9986-4a8a-461d-903f-47e573bbaf70 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1011.469921] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-8b3f01e9-87f4-442a-a746-46f4f6533d96 tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] [instance: fe1d4ace-7de3-423b-b423-9d7397aa2413] Unregistered the VM {{(pid=61806) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1011.470372] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-8b3f01e9-87f4-442a-a746-46f4f6533d96 tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] [instance: fe1d4ace-7de3-423b-b423-9d7397aa2413] Deleting contents of the VM from datastore datastore2 {{(pid=61806) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1011.470461] env[61806]: DEBUG nova.virt.vmwareapi.ds_util [None req-8b3f01e9-87f4-442a-a746-46f4f6533d96 tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] Deleting the datastore file [datastore2] fe1d4ace-7de3-423b-b423-9d7397aa2413 {{(pid=61806) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1011.473755] env[61806]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-c914ee97-1839-4a8f-b1ea-195170b7ca92 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1011.481836] env[61806]: DEBUG oslo_vmware.api [None req-8b3f01e9-87f4-442a-a746-46f4f6533d96 tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] Waiting for the task: (returnval){ [ 1011.481836] env[61806]: value = "task-1295090" [ 1011.481836] env[61806]: _type = "Task" [ 1011.481836] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1011.495357] env[61806]: DEBUG oslo_vmware.api [None req-8b3f01e9-87f4-442a-a746-46f4f6533d96 tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] Task: {'id': task-1295090, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1011.519996] env[61806]: DEBUG oslo_vmware.api [None req-8755c097-96dd-4051-93b1-330ffc947718 tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Task: {'id': task-1295086, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.558904} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1011.520305] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [None req-8755c097-96dd-4051-93b1-330ffc947718 tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa.vmdk to [datastore2] 71999b04-bf6e-4e4c-87c9-d5ec1932a80f/71999b04-bf6e-4e4c-87c9-d5ec1932a80f.vmdk {{(pid=61806) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 1011.520518] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-8755c097-96dd-4051-93b1-330ffc947718 tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] [instance: 71999b04-bf6e-4e4c-87c9-d5ec1932a80f] Extending root virtual disk to 1048576 {{(pid=61806) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1011.520775] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-80eeacac-8455-485c-83fe-e6ec09925edf {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1011.528321] env[61806]: DEBUG oslo_vmware.api [None req-8755c097-96dd-4051-93b1-330ffc947718 tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Waiting for the task: (returnval){ [ 1011.528321] env[61806]: value = "task-1295091" [ 1011.528321] env[61806]: _type = "Task" [ 1011.528321] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1011.539498] env[61806]: DEBUG oslo_vmware.api [None req-8755c097-96dd-4051-93b1-330ffc947718 tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Task: {'id': task-1295091, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1011.616115] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-07a59cb7-3574-46b2-a425-59445d24e202 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1011.624413] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ae32bcdf-e976-42bc-b437-7f5b0edbe83b {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1011.654861] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4df7107e-3e8b-4ea1-b9e7-4f63b3d11adf {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1011.663327] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-00655566-40be-4a62-9304-9f4a047c81b5 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1011.677905] env[61806]: DEBUG nova.compute.provider_tree [None req-de9a979c-45ed-463a-871e-b7dda8c366a2 tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] Inventory has not changed in ProviderTree for provider: a2858be1-fd22-4e08-979e-87ad25293407 {{(pid=61806) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1011.714067] env[61806]: DEBUG nova.objects.base [None req-9bf95660-956b-4332-b2b6-3c7de9ec6c45 tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Object Instance<67d1bdd8-cf8f-4fb5-bd3c-42008e4d60e3> lazy-loaded attributes: info_cache,migration_context {{(pid=61806) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 1011.715213] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-25a4c5e9-fa03-475a-bd8d-fc75a6bbba2b {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1011.736244] env[61806]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-4505daa4-a853-4563-b3a0-7ac067c594a6 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1011.742921] env[61806]: DEBUG oslo_vmware.api [None req-9bf95660-956b-4332-b2b6-3c7de9ec6c45 tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Waiting for the task: (returnval){ [ 1011.742921] env[61806]: value = "session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]521b6e09-cc3f-b5ab-f784-2c8fccf31c22" [ 1011.742921] env[61806]: _type = "Task" [ 1011.742921] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1011.751721] env[61806]: DEBUG oslo_vmware.api [None req-9bf95660-956b-4332-b2b6-3c7de9ec6c45 tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Task: {'id': session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]521b6e09-cc3f-b5ab-f784-2c8fccf31c22, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1011.792465] env[61806]: DEBUG oslo_vmware.api [None req-4b7244a6-0e60-4b9b-ac2e-5c7b0ff2b8c7 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Task: {'id': session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]521444a1-53e2-a648-4ff8-2531a5a31a25, 'name': SearchDatastore_Task, 'duration_secs': 0.010815} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1011.793432] env[61806]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d77d9f07-773a-4079-861a-80150eb8f629 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1011.798938] env[61806]: DEBUG oslo_vmware.api [None req-4b7244a6-0e60-4b9b-ac2e-5c7b0ff2b8c7 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Waiting for the task: (returnval){ [ 1011.798938] env[61806]: value = "session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]523889f4-4c00-2c0a-50fc-b9e2577abfd0" [ 1011.798938] env[61806]: _type = "Task" [ 1011.798938] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1011.808064] env[61806]: DEBUG oslo_vmware.api [None req-4b7244a6-0e60-4b9b-ac2e-5c7b0ff2b8c7 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Task: {'id': session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]523889f4-4c00-2c0a-50fc-b9e2577abfd0, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1011.922736] env[61806]: DEBUG oslo_concurrency.lockutils [None req-666a872d-3efd-4dd5-b0c9-c25d09581b9c tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Acquiring lock "refresh_cache-9357e61f-1628-43bd-ab46-de13c1529f51" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1011.922976] env[61806]: DEBUG oslo_concurrency.lockutils [None req-666a872d-3efd-4dd5-b0c9-c25d09581b9c tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Acquired lock "refresh_cache-9357e61f-1628-43bd-ab46-de13c1529f51" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1011.923220] env[61806]: DEBUG nova.network.neutron [None req-666a872d-3efd-4dd5-b0c9-c25d09581b9c tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] [instance: 9357e61f-1628-43bd-ab46-de13c1529f51] Building network info cache for instance {{(pid=61806) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1011.923388] env[61806]: DEBUG nova.objects.instance [None req-666a872d-3efd-4dd5-b0c9-c25d09581b9c tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Lazy-loading 'info_cache' on Instance uuid 9357e61f-1628-43bd-ab46-de13c1529f51 {{(pid=61806) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1011.993462] env[61806]: DEBUG oslo_vmware.api [None req-8b3f01e9-87f4-442a-a746-46f4f6533d96 tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] Task: {'id': task-1295090, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1012.038289] env[61806]: DEBUG oslo_vmware.api [None req-8755c097-96dd-4051-93b1-330ffc947718 tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Task: {'id': task-1295091, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.069517} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1012.038594] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-8755c097-96dd-4051-93b1-330ffc947718 tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] [instance: 71999b04-bf6e-4e4c-87c9-d5ec1932a80f] Extended root virtual disk {{(pid=61806) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1012.039405] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-561807a6-7e9c-45fc-bcbb-069b2bacd990 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1012.061236] env[61806]: DEBUG nova.virt.vmwareapi.volumeops [None req-8755c097-96dd-4051-93b1-330ffc947718 tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] [instance: 71999b04-bf6e-4e4c-87c9-d5ec1932a80f] Reconfiguring VM instance instance-00000062 to attach disk [datastore2] 71999b04-bf6e-4e4c-87c9-d5ec1932a80f/71999b04-bf6e-4e4c-87c9-d5ec1932a80f.vmdk or device None with type sparse {{(pid=61806) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1012.061528] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-2da22409-e789-4edd-87c7-1893e5f6ea42 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1012.081289] env[61806]: DEBUG oslo_vmware.api [None req-8755c097-96dd-4051-93b1-330ffc947718 tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Waiting for the task: (returnval){ [ 1012.081289] env[61806]: value = "task-1295092" [ 1012.081289] env[61806]: _type = "Task" [ 1012.081289] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1012.089590] env[61806]: DEBUG oslo_vmware.api [None req-8755c097-96dd-4051-93b1-330ffc947718 tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Task: {'id': task-1295092, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1012.181105] env[61806]: DEBUG nova.scheduler.client.report [None req-de9a979c-45ed-463a-871e-b7dda8c366a2 tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] Inventory has not changed for provider a2858be1-fd22-4e08-979e-87ad25293407 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 138, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61806) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1012.253784] env[61806]: DEBUG oslo_vmware.api [None req-9bf95660-956b-4332-b2b6-3c7de9ec6c45 tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Task: {'id': session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]521b6e09-cc3f-b5ab-f784-2c8fccf31c22, 'name': SearchDatastore_Task, 'duration_secs': 0.011335} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1012.254312] env[61806]: DEBUG oslo_concurrency.lockutils [None req-9bf95660-956b-4332-b2b6-3c7de9ec6c45 tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1012.311424] env[61806]: DEBUG oslo_vmware.api [None req-4b7244a6-0e60-4b9b-ac2e-5c7b0ff2b8c7 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Task: {'id': session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]523889f4-4c00-2c0a-50fc-b9e2577abfd0, 'name': SearchDatastore_Task, 'duration_secs': 0.02085} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1012.311794] env[61806]: DEBUG oslo_concurrency.lockutils [None req-4b7244a6-0e60-4b9b-ac2e-5c7b0ff2b8c7 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Releasing lock "[datastore2] devstack-image-cache_base/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa.vmdk" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1012.312156] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [None req-4b7244a6-0e60-4b9b-ac2e-5c7b0ff2b8c7 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa.vmdk to [datastore2] d0669232-6f33-4b30-97c2-2e4239af1ad0/d0669232-6f33-4b30-97c2-2e4239af1ad0.vmdk {{(pid=61806) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 1012.312481] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-eda12acc-195c-4c77-9431-d2e62c16b630 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1012.321531] env[61806]: DEBUG oslo_vmware.api [None req-4b7244a6-0e60-4b9b-ac2e-5c7b0ff2b8c7 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Waiting for the task: (returnval){ [ 1012.321531] env[61806]: value = "task-1295093" [ 1012.321531] env[61806]: _type = "Task" [ 1012.321531] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1012.331370] env[61806]: DEBUG oslo_vmware.api [None req-4b7244a6-0e60-4b9b-ac2e-5c7b0ff2b8c7 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Task: {'id': task-1295093, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1012.493644] env[61806]: DEBUG oslo_vmware.api [None req-8b3f01e9-87f4-442a-a746-46f4f6533d96 tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] Task: {'id': task-1295090, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.519727} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1012.494013] env[61806]: DEBUG nova.virt.vmwareapi.ds_util [None req-8b3f01e9-87f4-442a-a746-46f4f6533d96 tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] Deleted the datastore file {{(pid=61806) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1012.494289] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-8b3f01e9-87f4-442a-a746-46f4f6533d96 tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] [instance: fe1d4ace-7de3-423b-b423-9d7397aa2413] Deleted contents of the VM from datastore datastore2 {{(pid=61806) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1012.494539] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-8b3f01e9-87f4-442a-a746-46f4f6533d96 tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] [instance: fe1d4ace-7de3-423b-b423-9d7397aa2413] Instance destroyed {{(pid=61806) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1012.494794] env[61806]: INFO nova.compute.manager [None req-8b3f01e9-87f4-442a-a746-46f4f6533d96 tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] [instance: fe1d4ace-7de3-423b-b423-9d7397aa2413] Took 1.65 seconds to destroy the instance on the hypervisor. [ 1012.495106] env[61806]: DEBUG oslo.service.loopingcall [None req-8b3f01e9-87f4-442a-a746-46f4f6533d96 tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61806) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1012.495344] env[61806]: DEBUG nova.compute.manager [-] [instance: fe1d4ace-7de3-423b-b423-9d7397aa2413] Deallocating network for instance {{(pid=61806) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1012.495440] env[61806]: DEBUG nova.network.neutron [-] [instance: fe1d4ace-7de3-423b-b423-9d7397aa2413] deallocate_for_instance() {{(pid=61806) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1012.594484] env[61806]: DEBUG oslo_vmware.api [None req-8755c097-96dd-4051-93b1-330ffc947718 tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Task: {'id': task-1295092, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1012.688036] env[61806]: DEBUG oslo_concurrency.lockutils [None req-de9a979c-45ed-463a-871e-b7dda8c366a2 tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 3.352s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1012.689345] env[61806]: DEBUG oslo_concurrency.lockutils [None req-9bf95660-956b-4332-b2b6-3c7de9ec6c45 tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" :: waited 0.435s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1012.767237] env[61806]: INFO nova.network.neutron [None req-de9a979c-45ed-463a-871e-b7dda8c366a2 tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] [instance: fd9e520a-2e7b-43d4-b84d-8933c32cc6a6] Updating port 670f75e3-58f9-493b-9148-416e2e9cea5d with attributes {'binding:host_id': 'cpu-1', 'device_owner': 'compute:nova'} [ 1012.839177] env[61806]: DEBUG oslo_vmware.api [None req-4b7244a6-0e60-4b9b-ac2e-5c7b0ff2b8c7 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Task: {'id': task-1295093, 'name': CopyVirtualDisk_Task} progress is 25%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1012.854324] env[61806]: DEBUG nova.compute.manager [req-43c55f3a-b1dd-4069-a484-e89438d81b50 req-bc0e006c-8361-4af3-a0c0-5141b5891f9a service nova] [instance: fe1d4ace-7de3-423b-b423-9d7397aa2413] Received event network-vif-deleted-b2369483-32e0-4f80-be14-b45402c39094 {{(pid=61806) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1012.854324] env[61806]: INFO nova.compute.manager [req-43c55f3a-b1dd-4069-a484-e89438d81b50 req-bc0e006c-8361-4af3-a0c0-5141b5891f9a service nova] [instance: fe1d4ace-7de3-423b-b423-9d7397aa2413] Neutron deleted interface b2369483-32e0-4f80-be14-b45402c39094; detaching it from the instance and deleting it from the info cache [ 1012.854324] env[61806]: DEBUG nova.network.neutron [req-43c55f3a-b1dd-4069-a484-e89438d81b50 req-bc0e006c-8361-4af3-a0c0-5141b5891f9a service nova] [instance: fe1d4ace-7de3-423b-b423-9d7397aa2413] Updating instance_info_cache with network_info: [] {{(pid=61806) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1013.094314] env[61806]: DEBUG oslo_vmware.api [None req-8755c097-96dd-4051-93b1-330ffc947718 tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Task: {'id': task-1295092, 'name': ReconfigVM_Task, 'duration_secs': 0.791613} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1013.094689] env[61806]: DEBUG nova.virt.vmwareapi.volumeops [None req-8755c097-96dd-4051-93b1-330ffc947718 tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] [instance: 71999b04-bf6e-4e4c-87c9-d5ec1932a80f] Reconfigured VM instance instance-00000062 to attach disk [datastore2] 71999b04-bf6e-4e4c-87c9-d5ec1932a80f/71999b04-bf6e-4e4c-87c9-d5ec1932a80f.vmdk or device None with type sparse {{(pid=61806) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1013.095443] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-ffb91466-6359-4d73-aca8-356708d88651 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1013.102839] env[61806]: DEBUG oslo_vmware.api [None req-8755c097-96dd-4051-93b1-330ffc947718 tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Waiting for the task: (returnval){ [ 1013.102839] env[61806]: value = "task-1295095" [ 1013.102839] env[61806]: _type = "Task" [ 1013.102839] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1013.115455] env[61806]: DEBUG oslo_vmware.api [None req-8755c097-96dd-4051-93b1-330ffc947718 tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Task: {'id': task-1295095, 'name': Rename_Task} progress is 5%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1013.283344] env[61806]: DEBUG nova.network.neutron [None req-666a872d-3efd-4dd5-b0c9-c25d09581b9c tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] [instance: 9357e61f-1628-43bd-ab46-de13c1529f51] Updating instance_info_cache with network_info: [{"id": "a4ca7c56-b6a1-4bd2-bd47-02b0ed4aa097", "address": "fa:16:3e:66:09:0d", "network": {"id": "46515a4e-610f-42fb-a424-205a03df0640", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-1133623547-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4b9d94c3d684474b99e305fb6cd9b95f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "82ca17df-257e-40e6-9ec9-310ed6f05ccb", "external-id": "nsx-vlan-transportzone-778", "segmentation_id": 778, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa4ca7c56-b6", "ovs_interfaceid": "a4ca7c56-b6a1-4bd2-bd47-02b0ed4aa097", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61806) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1013.323528] env[61806]: DEBUG nova.network.neutron [-] [instance: fe1d4ace-7de3-423b-b423-9d7397aa2413] Updating instance_info_cache with network_info: [] {{(pid=61806) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1013.339058] env[61806]: DEBUG oslo_vmware.api [None req-4b7244a6-0e60-4b9b-ac2e-5c7b0ff2b8c7 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Task: {'id': task-1295093, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.697436} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1013.339394] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [None req-4b7244a6-0e60-4b9b-ac2e-5c7b0ff2b8c7 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa.vmdk to [datastore2] d0669232-6f33-4b30-97c2-2e4239af1ad0/d0669232-6f33-4b30-97c2-2e4239af1ad0.vmdk {{(pid=61806) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 1013.339619] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-4b7244a6-0e60-4b9b-ac2e-5c7b0ff2b8c7 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] [instance: d0669232-6f33-4b30-97c2-2e4239af1ad0] Extending root virtual disk to 1048576 {{(pid=61806) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1013.340127] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-a517b7dd-d9bb-4f9a-9345-7a3691ae1970 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1013.348644] env[61806]: DEBUG oslo_vmware.api [None req-4b7244a6-0e60-4b9b-ac2e-5c7b0ff2b8c7 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Waiting for the task: (returnval){ [ 1013.348644] env[61806]: value = "task-1295096" [ 1013.348644] env[61806]: _type = "Task" [ 1013.348644] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1013.358601] env[61806]: DEBUG oslo_vmware.api [None req-4b7244a6-0e60-4b9b-ac2e-5c7b0ff2b8c7 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Task: {'id': task-1295096, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1013.361888] env[61806]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-e4a62d53-3917-4ded-8644-a1be70b3048d {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1013.377147] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7be29588-6613-495b-bbd3-1f7d037d5ce5 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1013.397041] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a97fb44b-e6ad-4c6f-a651-6bb64bda9eb4 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1013.401586] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b1ae4397-9292-47e4-a97c-92177f859bde {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1013.441730] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-19fc34c1-4148-4999-b250-534a413c7183 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1013.444581] env[61806]: DEBUG nova.compute.manager [req-43c55f3a-b1dd-4069-a484-e89438d81b50 req-bc0e006c-8361-4af3-a0c0-5141b5891f9a service nova] [instance: fe1d4ace-7de3-423b-b423-9d7397aa2413] Detach interface failed, port_id=b2369483-32e0-4f80-be14-b45402c39094, reason: Instance fe1d4ace-7de3-423b-b423-9d7397aa2413 could not be found. {{(pid=61806) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10963}} [ 1013.451409] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-979d0e80-0e17-4ccf-a7cb-a02f9caf041b {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1013.467791] env[61806]: DEBUG nova.compute.provider_tree [None req-9bf95660-956b-4332-b2b6-3c7de9ec6c45 tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Inventory has not changed in ProviderTree for provider: a2858be1-fd22-4e08-979e-87ad25293407 {{(pid=61806) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1013.616712] env[61806]: DEBUG oslo_vmware.api [None req-8755c097-96dd-4051-93b1-330ffc947718 tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Task: {'id': task-1295095, 'name': Rename_Task} progress is 14%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1013.790500] env[61806]: DEBUG oslo_concurrency.lockutils [None req-666a872d-3efd-4dd5-b0c9-c25d09581b9c tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Releasing lock "refresh_cache-9357e61f-1628-43bd-ab46-de13c1529f51" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1013.790682] env[61806]: DEBUG nova.objects.instance [None req-666a872d-3efd-4dd5-b0c9-c25d09581b9c tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Lazy-loading 'migration_context' on Instance uuid 9357e61f-1628-43bd-ab46-de13c1529f51 {{(pid=61806) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1013.829439] env[61806]: INFO nova.compute.manager [-] [instance: fe1d4ace-7de3-423b-b423-9d7397aa2413] Took 1.33 seconds to deallocate network for instance. [ 1013.859671] env[61806]: DEBUG oslo_vmware.api [None req-4b7244a6-0e60-4b9b-ac2e-5c7b0ff2b8c7 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Task: {'id': task-1295096, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.141323} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1013.859966] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-4b7244a6-0e60-4b9b-ac2e-5c7b0ff2b8c7 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] [instance: d0669232-6f33-4b30-97c2-2e4239af1ad0] Extended root virtual disk {{(pid=61806) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1013.860776] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ebdc298f-3c57-4922-acd5-d893e6f53317 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1013.884362] env[61806]: DEBUG nova.virt.vmwareapi.volumeops [None req-4b7244a6-0e60-4b9b-ac2e-5c7b0ff2b8c7 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] [instance: d0669232-6f33-4b30-97c2-2e4239af1ad0] Reconfiguring VM instance instance-00000063 to attach disk [datastore2] d0669232-6f33-4b30-97c2-2e4239af1ad0/d0669232-6f33-4b30-97c2-2e4239af1ad0.vmdk or device None with type sparse {{(pid=61806) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1013.885416] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-cf8dc4c4-df44-46b0-b890-91a885ad16e6 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1013.906716] env[61806]: DEBUG oslo_vmware.api [None req-4b7244a6-0e60-4b9b-ac2e-5c7b0ff2b8c7 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Waiting for the task: (returnval){ [ 1013.906716] env[61806]: value = "task-1295097" [ 1013.906716] env[61806]: _type = "Task" [ 1013.906716] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1013.916091] env[61806]: DEBUG oslo_vmware.api [None req-4b7244a6-0e60-4b9b-ac2e-5c7b0ff2b8c7 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Task: {'id': task-1295097, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1013.972641] env[61806]: DEBUG nova.scheduler.client.report [None req-9bf95660-956b-4332-b2b6-3c7de9ec6c45 tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Inventory has not changed for provider a2858be1-fd22-4e08-979e-87ad25293407 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 138, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61806) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1014.116604] env[61806]: DEBUG oslo_vmware.api [None req-8755c097-96dd-4051-93b1-330ffc947718 tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Task: {'id': task-1295095, 'name': Rename_Task} progress is 14%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1014.293275] env[61806]: DEBUG nova.objects.base [None req-666a872d-3efd-4dd5-b0c9-c25d09581b9c tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Object Instance<9357e61f-1628-43bd-ab46-de13c1529f51> lazy-loaded attributes: info_cache,migration_context {{(pid=61806) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 1014.294203] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cb8455b8-0df9-4622-b941-f5f28b8b1aba {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1014.314242] env[61806]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-23490925-0ed9-49f5-8cd6-e84df41c29d8 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1014.320497] env[61806]: DEBUG oslo_vmware.api [None req-666a872d-3efd-4dd5-b0c9-c25d09581b9c tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Waiting for the task: (returnval){ [ 1014.320497] env[61806]: value = "session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]5248ae4e-ba6b-e2f8-b709-bd69908d18ea" [ 1014.320497] env[61806]: _type = "Task" [ 1014.320497] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1014.329302] env[61806]: DEBUG oslo_vmware.api [None req-666a872d-3efd-4dd5-b0c9-c25d09581b9c tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Task: {'id': session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]5248ae4e-ba6b-e2f8-b709-bd69908d18ea, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1014.335772] env[61806]: DEBUG oslo_concurrency.lockutils [None req-8b3f01e9-87f4-442a-a746-46f4f6533d96 tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1014.392629] env[61806]: DEBUG oslo_vmware.rw_handles [None req-2bebc059-5179-46cc-9f09-746fe48209df tempest-ServersNegativeTestJSON-891856253 tempest-ServersNegativeTestJSON-891856253-project-member] Getting lease state for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/5272df73-a1bf-0cd6-7e38-e3301b18dad4/disk-0.vmdk. {{(pid=61806) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 1014.393632] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-331dcc51-b105-4b37-851b-95bbb94990b7 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1014.400523] env[61806]: DEBUG oslo_vmware.rw_handles [None req-2bebc059-5179-46cc-9f09-746fe48209df tempest-ServersNegativeTestJSON-891856253 tempest-ServersNegativeTestJSON-891856253-project-member] Lease for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/5272df73-a1bf-0cd6-7e38-e3301b18dad4/disk-0.vmdk is in state: ready. {{(pid=61806) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 1014.400699] env[61806]: ERROR oslo_vmware.rw_handles [None req-2bebc059-5179-46cc-9f09-746fe48209df tempest-ServersNegativeTestJSON-891856253 tempest-ServersNegativeTestJSON-891856253-project-member] Aborting lease for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/5272df73-a1bf-0cd6-7e38-e3301b18dad4/disk-0.vmdk due to incomplete transfer. [ 1014.400933] env[61806]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseAbort with opID=oslo.vmware-ac0a0a86-1d51-4953-944e-749feeab65f1 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1014.408509] env[61806]: DEBUG oslo_vmware.rw_handles [None req-2bebc059-5179-46cc-9f09-746fe48209df tempest-ServersNegativeTestJSON-891856253 tempest-ServersNegativeTestJSON-891856253-project-member] Closed VMDK read handle for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/5272df73-a1bf-0cd6-7e38-e3301b18dad4/disk-0.vmdk. {{(pid=61806) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:735}} [ 1014.408745] env[61806]: DEBUG nova.virt.vmwareapi.images [None req-2bebc059-5179-46cc-9f09-746fe48209df tempest-ServersNegativeTestJSON-891856253 tempest-ServersNegativeTestJSON-891856253-project-member] [instance: bdbaef31-7d37-40cb-ac3e-ad152b37e2d2] Uploaded image 20c8ea6e-0db5-4788-9e34-92c1d63bb7b8 to the Glance image server {{(pid=61806) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:473}} [ 1014.411051] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [None req-2bebc059-5179-46cc-9f09-746fe48209df tempest-ServersNegativeTestJSON-891856253 tempest-ServersNegativeTestJSON-891856253-project-member] [instance: bdbaef31-7d37-40cb-ac3e-ad152b37e2d2] Destroying the VM {{(pid=61806) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1358}} [ 1014.414367] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-9f0ee7f0-03fb-446c-9d06-71d961b2dc94 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1014.421034] env[61806]: DEBUG oslo_vmware.api [None req-4b7244a6-0e60-4b9b-ac2e-5c7b0ff2b8c7 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Task: {'id': task-1295097, 'name': ReconfigVM_Task, 'duration_secs': 0.298528} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1014.422264] env[61806]: DEBUG nova.virt.vmwareapi.volumeops [None req-4b7244a6-0e60-4b9b-ac2e-5c7b0ff2b8c7 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] [instance: d0669232-6f33-4b30-97c2-2e4239af1ad0] Reconfigured VM instance instance-00000063 to attach disk [datastore2] d0669232-6f33-4b30-97c2-2e4239af1ad0/d0669232-6f33-4b30-97c2-2e4239af1ad0.vmdk or device None with type sparse {{(pid=61806) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1014.422952] env[61806]: DEBUG oslo_vmware.api [None req-2bebc059-5179-46cc-9f09-746fe48209df tempest-ServersNegativeTestJSON-891856253 tempest-ServersNegativeTestJSON-891856253-project-member] Waiting for the task: (returnval){ [ 1014.422952] env[61806]: value = "task-1295098" [ 1014.422952] env[61806]: _type = "Task" [ 1014.422952] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1014.423157] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-29b63b49-a8d8-40a6-b2bb-62e3593dbcab {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1014.434604] env[61806]: DEBUG oslo_vmware.api [None req-2bebc059-5179-46cc-9f09-746fe48209df tempest-ServersNegativeTestJSON-891856253 tempest-ServersNegativeTestJSON-891856253-project-member] Task: {'id': task-1295098, 'name': Destroy_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1014.435945] env[61806]: DEBUG oslo_vmware.api [None req-4b7244a6-0e60-4b9b-ac2e-5c7b0ff2b8c7 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Waiting for the task: (returnval){ [ 1014.435945] env[61806]: value = "task-1295099" [ 1014.435945] env[61806]: _type = "Task" [ 1014.435945] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1014.564692] env[61806]: DEBUG oslo_concurrency.lockutils [None req-de9a979c-45ed-463a-871e-b7dda8c366a2 tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] Acquiring lock "refresh_cache-fd9e520a-2e7b-43d4-b84d-8933c32cc6a6" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1014.565142] env[61806]: DEBUG oslo_concurrency.lockutils [None req-de9a979c-45ed-463a-871e-b7dda8c366a2 tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] Acquired lock "refresh_cache-fd9e520a-2e7b-43d4-b84d-8933c32cc6a6" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1014.566910] env[61806]: DEBUG nova.network.neutron [None req-de9a979c-45ed-463a-871e-b7dda8c366a2 tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] [instance: fd9e520a-2e7b-43d4-b84d-8933c32cc6a6] Building network info cache for instance {{(pid=61806) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1014.617413] env[61806]: DEBUG oslo_vmware.api [None req-8755c097-96dd-4051-93b1-330ffc947718 tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Task: {'id': task-1295095, 'name': Rename_Task, 'duration_secs': 1.176651} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1014.617413] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [None req-8755c097-96dd-4051-93b1-330ffc947718 tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] [instance: 71999b04-bf6e-4e4c-87c9-d5ec1932a80f] Powering on the VM {{(pid=61806) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1014.617413] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-814a7024-bde5-43d5-820e-1b6be0fc44ea {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1014.623454] env[61806]: DEBUG oslo_vmware.api [None req-8755c097-96dd-4051-93b1-330ffc947718 tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Waiting for the task: (returnval){ [ 1014.623454] env[61806]: value = "task-1295100" [ 1014.623454] env[61806]: _type = "Task" [ 1014.623454] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1014.631866] env[61806]: DEBUG oslo_vmware.api [None req-8755c097-96dd-4051-93b1-330ffc947718 tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Task: {'id': task-1295100, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1014.831809] env[61806]: DEBUG oslo_vmware.api [None req-666a872d-3efd-4dd5-b0c9-c25d09581b9c tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Task: {'id': session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]5248ae4e-ba6b-e2f8-b709-bd69908d18ea, 'name': SearchDatastore_Task, 'duration_secs': 0.017612} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1014.832126] env[61806]: DEBUG oslo_concurrency.lockutils [None req-666a872d-3efd-4dd5-b0c9-c25d09581b9c tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1014.880824] env[61806]: DEBUG nova.compute.manager [req-80d50b1a-908a-40d8-9c33-8b23f3ea7bc9 req-a8255949-5019-475a-bf9c-592897441883 service nova] [instance: fd9e520a-2e7b-43d4-b84d-8933c32cc6a6] Received event network-vif-plugged-670f75e3-58f9-493b-9148-416e2e9cea5d {{(pid=61806) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1014.881131] env[61806]: DEBUG oslo_concurrency.lockutils [req-80d50b1a-908a-40d8-9c33-8b23f3ea7bc9 req-a8255949-5019-475a-bf9c-592897441883 service nova] Acquiring lock "fd9e520a-2e7b-43d4-b84d-8933c32cc6a6-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1014.881304] env[61806]: DEBUG oslo_concurrency.lockutils [req-80d50b1a-908a-40d8-9c33-8b23f3ea7bc9 req-a8255949-5019-475a-bf9c-592897441883 service nova] Lock "fd9e520a-2e7b-43d4-b84d-8933c32cc6a6-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1014.881482] env[61806]: DEBUG oslo_concurrency.lockutils [req-80d50b1a-908a-40d8-9c33-8b23f3ea7bc9 req-a8255949-5019-475a-bf9c-592897441883 service nova] Lock "fd9e520a-2e7b-43d4-b84d-8933c32cc6a6-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1014.881657] env[61806]: DEBUG nova.compute.manager [req-80d50b1a-908a-40d8-9c33-8b23f3ea7bc9 req-a8255949-5019-475a-bf9c-592897441883 service nova] [instance: fd9e520a-2e7b-43d4-b84d-8933c32cc6a6] No waiting events found dispatching network-vif-plugged-670f75e3-58f9-493b-9148-416e2e9cea5d {{(pid=61806) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1014.881830] env[61806]: WARNING nova.compute.manager [req-80d50b1a-908a-40d8-9c33-8b23f3ea7bc9 req-a8255949-5019-475a-bf9c-592897441883 service nova] [instance: fd9e520a-2e7b-43d4-b84d-8933c32cc6a6] Received unexpected event network-vif-plugged-670f75e3-58f9-493b-9148-416e2e9cea5d for instance with vm_state shelved_offloaded and task_state spawning. [ 1014.881998] env[61806]: DEBUG nova.compute.manager [req-80d50b1a-908a-40d8-9c33-8b23f3ea7bc9 req-a8255949-5019-475a-bf9c-592897441883 service nova] [instance: fd9e520a-2e7b-43d4-b84d-8933c32cc6a6] Received event network-changed-670f75e3-58f9-493b-9148-416e2e9cea5d {{(pid=61806) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1014.882176] env[61806]: DEBUG nova.compute.manager [req-80d50b1a-908a-40d8-9c33-8b23f3ea7bc9 req-a8255949-5019-475a-bf9c-592897441883 service nova] [instance: fd9e520a-2e7b-43d4-b84d-8933c32cc6a6] Refreshing instance network info cache due to event network-changed-670f75e3-58f9-493b-9148-416e2e9cea5d. {{(pid=61806) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 1014.882351] env[61806]: DEBUG oslo_concurrency.lockutils [req-80d50b1a-908a-40d8-9c33-8b23f3ea7bc9 req-a8255949-5019-475a-bf9c-592897441883 service nova] Acquiring lock "refresh_cache-fd9e520a-2e7b-43d4-b84d-8933c32cc6a6" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1014.935078] env[61806]: DEBUG oslo_vmware.api [None req-2bebc059-5179-46cc-9f09-746fe48209df tempest-ServersNegativeTestJSON-891856253 tempest-ServersNegativeTestJSON-891856253-project-member] Task: {'id': task-1295098, 'name': Destroy_Task} progress is 33%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1014.946180] env[61806]: DEBUG oslo_vmware.api [None req-4b7244a6-0e60-4b9b-ac2e-5c7b0ff2b8c7 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Task: {'id': task-1295099, 'name': Rename_Task, 'duration_secs': 0.164445} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1014.946826] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [None req-4b7244a6-0e60-4b9b-ac2e-5c7b0ff2b8c7 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] [instance: d0669232-6f33-4b30-97c2-2e4239af1ad0] Powering on the VM {{(pid=61806) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1014.947102] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-dbb22003-9435-49dd-879f-725dc3ba2872 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1014.954316] env[61806]: DEBUG oslo_vmware.api [None req-4b7244a6-0e60-4b9b-ac2e-5c7b0ff2b8c7 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Waiting for the task: (returnval){ [ 1014.954316] env[61806]: value = "task-1295101" [ 1014.954316] env[61806]: _type = "Task" [ 1014.954316] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1014.963192] env[61806]: DEBUG oslo_vmware.api [None req-4b7244a6-0e60-4b9b-ac2e-5c7b0ff2b8c7 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Task: {'id': task-1295101, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1014.985305] env[61806]: DEBUG oslo_concurrency.lockutils [None req-9bf95660-956b-4332-b2b6-3c7de9ec6c45 tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" :: held 2.295s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1014.988117] env[61806]: DEBUG oslo_concurrency.lockutils [None req-8b3f01e9-87f4-442a-a746-46f4f6533d96 tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.652s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1014.988428] env[61806]: DEBUG nova.objects.instance [None req-8b3f01e9-87f4-442a-a746-46f4f6533d96 tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] Lazy-loading 'resources' on Instance uuid fe1d4ace-7de3-423b-b423-9d7397aa2413 {{(pid=61806) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1015.134565] env[61806]: DEBUG oslo_vmware.api [None req-8755c097-96dd-4051-93b1-330ffc947718 tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Task: {'id': task-1295100, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1015.284107] env[61806]: DEBUG nova.virt.vmwareapi.volumeops [None req-a22e273c-c0ad-47bf-9acf-b3a1dd13247a tempest-AttachVolumeNegativeTest-1683033704 tempest-AttachVolumeNegativeTest-1683033704-project-member] [instance: e4638faa-7be1-4909-a595-c437837fa314] Volume attach. Driver type: vmdk {{(pid=61806) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 1015.284384] env[61806]: DEBUG nova.virt.vmwareapi.volumeops [None req-a22e273c-c0ad-47bf-9acf-b3a1dd13247a tempest-AttachVolumeNegativeTest-1683033704 tempest-AttachVolumeNegativeTest-1683033704-project-member] [instance: e4638faa-7be1-4909-a595-c437837fa314] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-277745', 'volume_id': '1492823b-b43f-42fd-bc67-94fe82531df9', 'name': 'volume-1492823b-b43f-42fd-bc67-94fe82531df9', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'e4638faa-7be1-4909-a595-c437837fa314', 'attached_at': '', 'detached_at': '', 'volume_id': '1492823b-b43f-42fd-bc67-94fe82531df9', 'serial': '1492823b-b43f-42fd-bc67-94fe82531df9'} {{(pid=61806) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 1015.285773] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f3b00989-4b4e-40c5-9a4d-7e5784da8044 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1015.302785] env[61806]: DEBUG nova.network.neutron [None req-de9a979c-45ed-463a-871e-b7dda8c366a2 tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] [instance: fd9e520a-2e7b-43d4-b84d-8933c32cc6a6] Updating instance_info_cache with network_info: [{"id": "670f75e3-58f9-493b-9148-416e2e9cea5d", "address": "fa:16:3e:76:be:e1", "network": {"id": "a9d993bc-0817-4ee6-ab6e-1ef9432a110f", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-939246736-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.226", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "bbbb09a49203421d807f733616c854df", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ad4c015b-4a8b-46ca-9556-74bad8db9fb3", "external-id": "nsx-vlan-transportzone-246", "segmentation_id": 246, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap670f75e3-58", "ovs_interfaceid": "670f75e3-58f9-493b-9148-416e2e9cea5d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61806) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1015.304572] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fd6d6a22-f461-4543-abba-a17fc8a2a93e {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1015.332314] env[61806]: DEBUG nova.virt.vmwareapi.volumeops [None req-a22e273c-c0ad-47bf-9acf-b3a1dd13247a tempest-AttachVolumeNegativeTest-1683033704 tempest-AttachVolumeNegativeTest-1683033704-project-member] [instance: e4638faa-7be1-4909-a595-c437837fa314] Reconfiguring VM instance instance-0000005d to attach disk [datastore2] volume-1492823b-b43f-42fd-bc67-94fe82531df9/volume-1492823b-b43f-42fd-bc67-94fe82531df9.vmdk or device None with type thin {{(pid=61806) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1015.333335] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-603d1cbd-0781-4184-89f2-a95516d924f9 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1015.353745] env[61806]: DEBUG oslo_vmware.api [None req-a22e273c-c0ad-47bf-9acf-b3a1dd13247a tempest-AttachVolumeNegativeTest-1683033704 tempest-AttachVolumeNegativeTest-1683033704-project-member] Waiting for the task: (returnval){ [ 1015.353745] env[61806]: value = "task-1295102" [ 1015.353745] env[61806]: _type = "Task" [ 1015.353745] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1015.363124] env[61806]: DEBUG oslo_vmware.api [None req-a22e273c-c0ad-47bf-9acf-b3a1dd13247a tempest-AttachVolumeNegativeTest-1683033704 tempest-AttachVolumeNegativeTest-1683033704-project-member] Task: {'id': task-1295102, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1015.436808] env[61806]: DEBUG oslo_vmware.api [None req-2bebc059-5179-46cc-9f09-746fe48209df tempest-ServersNegativeTestJSON-891856253 tempest-ServersNegativeTestJSON-891856253-project-member] Task: {'id': task-1295098, 'name': Destroy_Task, 'duration_secs': 0.867064} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1015.437129] env[61806]: INFO nova.virt.vmwareapi.vm_util [None req-2bebc059-5179-46cc-9f09-746fe48209df tempest-ServersNegativeTestJSON-891856253 tempest-ServersNegativeTestJSON-891856253-project-member] [instance: bdbaef31-7d37-40cb-ac3e-ad152b37e2d2] Destroyed the VM [ 1015.437380] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-2bebc059-5179-46cc-9f09-746fe48209df tempest-ServersNegativeTestJSON-891856253 tempest-ServersNegativeTestJSON-891856253-project-member] [instance: bdbaef31-7d37-40cb-ac3e-ad152b37e2d2] Deleting Snapshot of the VM instance {{(pid=61806) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 1015.437652] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-94b03a01-dc51-4468-b5cf-950de8e0cd1b {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1015.445311] env[61806]: DEBUG oslo_vmware.api [None req-2bebc059-5179-46cc-9f09-746fe48209df tempest-ServersNegativeTestJSON-891856253 tempest-ServersNegativeTestJSON-891856253-project-member] Waiting for the task: (returnval){ [ 1015.445311] env[61806]: value = "task-1295103" [ 1015.445311] env[61806]: _type = "Task" [ 1015.445311] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1015.455747] env[61806]: DEBUG oslo_vmware.api [None req-2bebc059-5179-46cc-9f09-746fe48209df tempest-ServersNegativeTestJSON-891856253 tempest-ServersNegativeTestJSON-891856253-project-member] Task: {'id': task-1295103, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1015.465743] env[61806]: DEBUG oslo_vmware.api [None req-4b7244a6-0e60-4b9b-ac2e-5c7b0ff2b8c7 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Task: {'id': task-1295101, 'name': PowerOnVM_Task} progress is 89%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1015.562859] env[61806]: INFO nova.scheduler.client.report [None req-9bf95660-956b-4332-b2b6-3c7de9ec6c45 tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Deleted allocation for migration ca0a088e-384c-40b9-a197-357d189bc64d [ 1015.638623] env[61806]: DEBUG oslo_vmware.api [None req-8755c097-96dd-4051-93b1-330ffc947718 tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Task: {'id': task-1295100, 'name': PowerOnVM_Task, 'duration_secs': 0.958118} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1015.638917] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [None req-8755c097-96dd-4051-93b1-330ffc947718 tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] [instance: 71999b04-bf6e-4e4c-87c9-d5ec1932a80f] Powered on the VM {{(pid=61806) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1015.639148] env[61806]: INFO nova.compute.manager [None req-8755c097-96dd-4051-93b1-330ffc947718 tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] [instance: 71999b04-bf6e-4e4c-87c9-d5ec1932a80f] Took 9.86 seconds to spawn the instance on the hypervisor. [ 1015.639350] env[61806]: DEBUG nova.compute.manager [None req-8755c097-96dd-4051-93b1-330ffc947718 tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] [instance: 71999b04-bf6e-4e4c-87c9-d5ec1932a80f] Checking state {{(pid=61806) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1015.641076] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8bd7e866-c53c-471f-9fa4-2bdb6d29662d {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1015.662704] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bd2407d1-eeb2-4234-9115-b82336ffad50 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1015.671520] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0ffa62da-44c1-4f62-80ee-6cbbddd709f0 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1015.702903] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2bdc280d-d1cf-4c68-b473-087ff222c854 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1015.712054] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-21f71833-604d-4f58-a283-642c5796a23b {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1015.728150] env[61806]: DEBUG nova.compute.provider_tree [None req-8b3f01e9-87f4-442a-a746-46f4f6533d96 tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] Inventory has not changed in ProviderTree for provider: a2858be1-fd22-4e08-979e-87ad25293407 {{(pid=61806) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1015.808152] env[61806]: DEBUG oslo_concurrency.lockutils [None req-de9a979c-45ed-463a-871e-b7dda8c366a2 tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] Releasing lock "refresh_cache-fd9e520a-2e7b-43d4-b84d-8933c32cc6a6" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1015.811047] env[61806]: DEBUG oslo_concurrency.lockutils [req-80d50b1a-908a-40d8-9c33-8b23f3ea7bc9 req-a8255949-5019-475a-bf9c-592897441883 service nova] Acquired lock "refresh_cache-fd9e520a-2e7b-43d4-b84d-8933c32cc6a6" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1015.811047] env[61806]: DEBUG nova.network.neutron [req-80d50b1a-908a-40d8-9c33-8b23f3ea7bc9 req-a8255949-5019-475a-bf9c-592897441883 service nova] [instance: fd9e520a-2e7b-43d4-b84d-8933c32cc6a6] Refreshing network info cache for port 670f75e3-58f9-493b-9148-416e2e9cea5d {{(pid=61806) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1015.835172] env[61806]: DEBUG nova.virt.hardware [None req-de9a979c-45ed-463a-871e-b7dda8c366a2 tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-15T18:05:10Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='0ce783a032dc53852c0b308ade1b217b',container_format='bare',created_at=2024-10-15T18:14:37Z,direct_url=,disk_format='vmdk',id=b08c43d9-cfd5-4a1a-b99d-0aebb718b592,min_disk=1,min_ram=0,name='tempest-AttachVolumeShelveTestJSON-server-1698015101-shelved',owner='bbbb09a49203421d807f733616c854df',properties=ImageMetaProps,protected=,size=31668736,status='active',tags=,updated_at=2024-10-15T18:14:50Z,virtual_size=,visibility=), allow threads: False {{(pid=61806) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1015.835461] env[61806]: DEBUG nova.virt.hardware [None req-de9a979c-45ed-463a-871e-b7dda8c366a2 tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] Flavor limits 0:0:0 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1015.835652] env[61806]: DEBUG nova.virt.hardware [None req-de9a979c-45ed-463a-871e-b7dda8c366a2 tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] Image limits 0:0:0 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1015.835945] env[61806]: DEBUG nova.virt.hardware [None req-de9a979c-45ed-463a-871e-b7dda8c366a2 tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] Flavor pref 0:0:0 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1015.836149] env[61806]: DEBUG nova.virt.hardware [None req-de9a979c-45ed-463a-871e-b7dda8c366a2 tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] Image pref 0:0:0 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1015.836321] env[61806]: DEBUG nova.virt.hardware [None req-de9a979c-45ed-463a-871e-b7dda8c366a2 tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1015.836565] env[61806]: DEBUG nova.virt.hardware [None req-de9a979c-45ed-463a-871e-b7dda8c366a2 tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61806) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1015.836735] env[61806]: DEBUG nova.virt.hardware [None req-de9a979c-45ed-463a-871e-b7dda8c366a2 tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61806) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1015.836916] env[61806]: DEBUG nova.virt.hardware [None req-de9a979c-45ed-463a-871e-b7dda8c366a2 tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] Got 1 possible topologies {{(pid=61806) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1015.837118] env[61806]: DEBUG nova.virt.hardware [None req-de9a979c-45ed-463a-871e-b7dda8c366a2 tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61806) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1015.837320] env[61806]: DEBUG nova.virt.hardware [None req-de9a979c-45ed-463a-871e-b7dda8c366a2 tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61806) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1015.838229] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0fffbed5-2f96-4267-942c-8138a580b83e {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1015.848104] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8882dee7-f98c-4c93-86f4-7bec5c4228d7 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1015.864787] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-de9a979c-45ed-463a-871e-b7dda8c366a2 tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] [instance: fd9e520a-2e7b-43d4-b84d-8933c32cc6a6] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:76:be:e1', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'ad4c015b-4a8b-46ca-9556-74bad8db9fb3', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '670f75e3-58f9-493b-9148-416e2e9cea5d', 'vif_model': 'vmxnet3'}] {{(pid=61806) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1015.872082] env[61806]: DEBUG oslo.service.loopingcall [None req-de9a979c-45ed-463a-871e-b7dda8c366a2 tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61806) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1015.875193] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: fd9e520a-2e7b-43d4-b84d-8933c32cc6a6] Creating VM on the ESX host {{(pid=61806) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 1015.875446] env[61806]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-27d9a436-2f59-46a1-b942-07b95b5e90a9 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1015.895630] env[61806]: DEBUG oslo_vmware.api [None req-a22e273c-c0ad-47bf-9acf-b3a1dd13247a tempest-AttachVolumeNegativeTest-1683033704 tempest-AttachVolumeNegativeTest-1683033704-project-member] Task: {'id': task-1295102, 'name': ReconfigVM_Task, 'duration_secs': 0.411044} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1015.896956] env[61806]: DEBUG nova.virt.vmwareapi.volumeops [None req-a22e273c-c0ad-47bf-9acf-b3a1dd13247a tempest-AttachVolumeNegativeTest-1683033704 tempest-AttachVolumeNegativeTest-1683033704-project-member] [instance: e4638faa-7be1-4909-a595-c437837fa314] Reconfigured VM instance instance-0000005d to attach disk [datastore2] volume-1492823b-b43f-42fd-bc67-94fe82531df9/volume-1492823b-b43f-42fd-bc67-94fe82531df9.vmdk or device None with type thin {{(pid=61806) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1015.901926] env[61806]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1015.901926] env[61806]: value = "task-1295104" [ 1015.901926] env[61806]: _type = "Task" [ 1015.901926] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1015.902159] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-2ac2b212-61e9-426b-8d82-31b6c1f313a6 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1015.924961] env[61806]: DEBUG oslo_vmware.api [-] Task: {'id': task-1295104, 'name': CreateVM_Task} progress is 6%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1015.926572] env[61806]: DEBUG oslo_vmware.api [None req-a22e273c-c0ad-47bf-9acf-b3a1dd13247a tempest-AttachVolumeNegativeTest-1683033704 tempest-AttachVolumeNegativeTest-1683033704-project-member] Waiting for the task: (returnval){ [ 1015.926572] env[61806]: value = "task-1295105" [ 1015.926572] env[61806]: _type = "Task" [ 1015.926572] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1015.938774] env[61806]: DEBUG oslo_vmware.api [None req-a22e273c-c0ad-47bf-9acf-b3a1dd13247a tempest-AttachVolumeNegativeTest-1683033704 tempest-AttachVolumeNegativeTest-1683033704-project-member] Task: {'id': task-1295105, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1015.958261] env[61806]: DEBUG oslo_vmware.api [None req-2bebc059-5179-46cc-9f09-746fe48209df tempest-ServersNegativeTestJSON-891856253 tempest-ServersNegativeTestJSON-891856253-project-member] Task: {'id': task-1295103, 'name': RemoveSnapshot_Task} progress is 100%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1015.969833] env[61806]: DEBUG oslo_vmware.api [None req-4b7244a6-0e60-4b9b-ac2e-5c7b0ff2b8c7 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Task: {'id': task-1295101, 'name': PowerOnVM_Task, 'duration_secs': 0.585127} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1015.970249] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [None req-4b7244a6-0e60-4b9b-ac2e-5c7b0ff2b8c7 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] [instance: d0669232-6f33-4b30-97c2-2e4239af1ad0] Powered on the VM {{(pid=61806) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1015.970563] env[61806]: INFO nova.compute.manager [None req-4b7244a6-0e60-4b9b-ac2e-5c7b0ff2b8c7 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] [instance: d0669232-6f33-4b30-97c2-2e4239af1ad0] Took 7.93 seconds to spawn the instance on the hypervisor. [ 1015.970813] env[61806]: DEBUG nova.compute.manager [None req-4b7244a6-0e60-4b9b-ac2e-5c7b0ff2b8c7 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] [instance: d0669232-6f33-4b30-97c2-2e4239af1ad0] Checking state {{(pid=61806) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1015.971904] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bd4d063e-bf23-4f5b-9c8c-32bad88168dd {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1016.070366] env[61806]: DEBUG oslo_concurrency.lockutils [None req-9bf95660-956b-4332-b2b6-3c7de9ec6c45 tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Lock "67d1bdd8-cf8f-4fb5-bd3c-42008e4d60e3" "released" by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" :: held 7.373s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1016.166053] env[61806]: INFO nova.compute.manager [None req-8755c097-96dd-4051-93b1-330ffc947718 tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] [instance: 71999b04-bf6e-4e4c-87c9-d5ec1932a80f] Took 16.66 seconds to build instance. [ 1016.231521] env[61806]: DEBUG nova.scheduler.client.report [None req-8b3f01e9-87f4-442a-a746-46f4f6533d96 tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] Inventory has not changed for provider a2858be1-fd22-4e08-979e-87ad25293407 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 138, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61806) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1016.432655] env[61806]: DEBUG oslo_vmware.api [-] Task: {'id': task-1295104, 'name': CreateVM_Task} progress is 99%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1016.445493] env[61806]: DEBUG oslo_vmware.api [None req-a22e273c-c0ad-47bf-9acf-b3a1dd13247a tempest-AttachVolumeNegativeTest-1683033704 tempest-AttachVolumeNegativeTest-1683033704-project-member] Task: {'id': task-1295105, 'name': ReconfigVM_Task, 'duration_secs': 0.237259} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1016.448165] env[61806]: DEBUG nova.virt.vmwareapi.volumeops [None req-a22e273c-c0ad-47bf-9acf-b3a1dd13247a tempest-AttachVolumeNegativeTest-1683033704 tempest-AttachVolumeNegativeTest-1683033704-project-member] [instance: e4638faa-7be1-4909-a595-c437837fa314] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-277745', 'volume_id': '1492823b-b43f-42fd-bc67-94fe82531df9', 'name': 'volume-1492823b-b43f-42fd-bc67-94fe82531df9', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'e4638faa-7be1-4909-a595-c437837fa314', 'attached_at': '', 'detached_at': '', 'volume_id': '1492823b-b43f-42fd-bc67-94fe82531df9', 'serial': '1492823b-b43f-42fd-bc67-94fe82531df9'} {{(pid=61806) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 1016.459568] env[61806]: DEBUG oslo_vmware.api [None req-2bebc059-5179-46cc-9f09-746fe48209df tempest-ServersNegativeTestJSON-891856253 tempest-ServersNegativeTestJSON-891856253-project-member] Task: {'id': task-1295103, 'name': RemoveSnapshot_Task, 'duration_secs': 0.701503} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1016.459870] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-2bebc059-5179-46cc-9f09-746fe48209df tempest-ServersNegativeTestJSON-891856253 tempest-ServersNegativeTestJSON-891856253-project-member] [instance: bdbaef31-7d37-40cb-ac3e-ad152b37e2d2] Deleted Snapshot of the VM instance {{(pid=61806) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 1016.460179] env[61806]: DEBUG nova.compute.manager [None req-2bebc059-5179-46cc-9f09-746fe48209df tempest-ServersNegativeTestJSON-891856253 tempest-ServersNegativeTestJSON-891856253-project-member] [instance: bdbaef31-7d37-40cb-ac3e-ad152b37e2d2] Checking state {{(pid=61806) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1016.460965] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a59b05ef-7786-4058-9f3c-d2a820011af5 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1016.493334] env[61806]: INFO nova.compute.manager [None req-4b7244a6-0e60-4b9b-ac2e-5c7b0ff2b8c7 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] [instance: d0669232-6f33-4b30-97c2-2e4239af1ad0] Took 16.69 seconds to build instance. [ 1016.580597] env[61806]: DEBUG nova.network.neutron [req-80d50b1a-908a-40d8-9c33-8b23f3ea7bc9 req-a8255949-5019-475a-bf9c-592897441883 service nova] [instance: fd9e520a-2e7b-43d4-b84d-8933c32cc6a6] Updated VIF entry in instance network info cache for port 670f75e3-58f9-493b-9148-416e2e9cea5d. {{(pid=61806) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1016.580975] env[61806]: DEBUG nova.network.neutron [req-80d50b1a-908a-40d8-9c33-8b23f3ea7bc9 req-a8255949-5019-475a-bf9c-592897441883 service nova] [instance: fd9e520a-2e7b-43d4-b84d-8933c32cc6a6] Updating instance_info_cache with network_info: [{"id": "670f75e3-58f9-493b-9148-416e2e9cea5d", "address": "fa:16:3e:76:be:e1", "network": {"id": "a9d993bc-0817-4ee6-ab6e-1ef9432a110f", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-939246736-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.226", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "bbbb09a49203421d807f733616c854df", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ad4c015b-4a8b-46ca-9556-74bad8db9fb3", "external-id": "nsx-vlan-transportzone-246", "segmentation_id": 246, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap670f75e3-58", "ovs_interfaceid": "670f75e3-58f9-493b-9148-416e2e9cea5d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61806) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1016.668330] env[61806]: DEBUG oslo_concurrency.lockutils [None req-8755c097-96dd-4051-93b1-330ffc947718 tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Lock "71999b04-bf6e-4e4c-87c9-d5ec1932a80f" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 18.171s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1016.737263] env[61806]: DEBUG oslo_concurrency.lockutils [None req-8b3f01e9-87f4-442a-a746-46f4f6533d96 tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.749s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1016.740298] env[61806]: DEBUG oslo_concurrency.lockutils [None req-666a872d-3efd-4dd5-b0c9-c25d09581b9c tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" :: waited 1.908s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1016.759025] env[61806]: INFO nova.scheduler.client.report [None req-8b3f01e9-87f4-442a-a746-46f4f6533d96 tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] Deleted allocations for instance fe1d4ace-7de3-423b-b423-9d7397aa2413 [ 1016.926653] env[61806]: DEBUG oslo_vmware.api [-] Task: {'id': task-1295104, 'name': CreateVM_Task} progress is 99%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1016.974417] env[61806]: INFO nova.compute.manager [None req-2bebc059-5179-46cc-9f09-746fe48209df tempest-ServersNegativeTestJSON-891856253 tempest-ServersNegativeTestJSON-891856253-project-member] [instance: bdbaef31-7d37-40cb-ac3e-ad152b37e2d2] Shelve offloading [ 1016.976497] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [None req-2bebc059-5179-46cc-9f09-746fe48209df tempest-ServersNegativeTestJSON-891856253 tempest-ServersNegativeTestJSON-891856253-project-member] [instance: bdbaef31-7d37-40cb-ac3e-ad152b37e2d2] Powering off the VM {{(pid=61806) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1016.976791] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-46b663d1-9c4d-43f2-850c-b25a69ba350f {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1016.985681] env[61806]: DEBUG oslo_vmware.api [None req-2bebc059-5179-46cc-9f09-746fe48209df tempest-ServersNegativeTestJSON-891856253 tempest-ServersNegativeTestJSON-891856253-project-member] Waiting for the task: (returnval){ [ 1016.985681] env[61806]: value = "task-1295106" [ 1016.985681] env[61806]: _type = "Task" [ 1016.985681] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1016.996529] env[61806]: DEBUG oslo_concurrency.lockutils [None req-4b7244a6-0e60-4b9b-ac2e-5c7b0ff2b8c7 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Lock "d0669232-6f33-4b30-97c2-2e4239af1ad0" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 18.199s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1016.996846] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [None req-2bebc059-5179-46cc-9f09-746fe48209df tempest-ServersNegativeTestJSON-891856253 tempest-ServersNegativeTestJSON-891856253-project-member] [instance: bdbaef31-7d37-40cb-ac3e-ad152b37e2d2] VM already powered off {{(pid=61806) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1509}} [ 1016.997105] env[61806]: DEBUG nova.compute.manager [None req-2bebc059-5179-46cc-9f09-746fe48209df tempest-ServersNegativeTestJSON-891856253 tempest-ServersNegativeTestJSON-891856253-project-member] [instance: bdbaef31-7d37-40cb-ac3e-ad152b37e2d2] Checking state {{(pid=61806) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1016.997938] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fd7fb657-88b2-450f-8d9e-e335936c7795 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1017.004835] env[61806]: DEBUG oslo_concurrency.lockutils [None req-2bebc059-5179-46cc-9f09-746fe48209df tempest-ServersNegativeTestJSON-891856253 tempest-ServersNegativeTestJSON-891856253-project-member] Acquiring lock "refresh_cache-bdbaef31-7d37-40cb-ac3e-ad152b37e2d2" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1017.005024] env[61806]: DEBUG oslo_concurrency.lockutils [None req-2bebc059-5179-46cc-9f09-746fe48209df tempest-ServersNegativeTestJSON-891856253 tempest-ServersNegativeTestJSON-891856253-project-member] Acquired lock "refresh_cache-bdbaef31-7d37-40cb-ac3e-ad152b37e2d2" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1017.005209] env[61806]: DEBUG nova.network.neutron [None req-2bebc059-5179-46cc-9f09-746fe48209df tempest-ServersNegativeTestJSON-891856253 tempest-ServersNegativeTestJSON-891856253-project-member] [instance: bdbaef31-7d37-40cb-ac3e-ad152b37e2d2] Building network info cache for instance {{(pid=61806) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1017.024790] env[61806]: DEBUG oslo_concurrency.lockutils [None req-1a0d0139-1a16-409a-b523-f987ba54d59c tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Acquiring lock "67d1bdd8-cf8f-4fb5-bd3c-42008e4d60e3" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1017.025070] env[61806]: DEBUG oslo_concurrency.lockutils [None req-1a0d0139-1a16-409a-b523-f987ba54d59c tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Lock "67d1bdd8-cf8f-4fb5-bd3c-42008e4d60e3" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1017.025295] env[61806]: DEBUG oslo_concurrency.lockutils [None req-1a0d0139-1a16-409a-b523-f987ba54d59c tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Acquiring lock "67d1bdd8-cf8f-4fb5-bd3c-42008e4d60e3-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1017.025485] env[61806]: DEBUG oslo_concurrency.lockutils [None req-1a0d0139-1a16-409a-b523-f987ba54d59c tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Lock "67d1bdd8-cf8f-4fb5-bd3c-42008e4d60e3-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1017.025659] env[61806]: DEBUG oslo_concurrency.lockutils [None req-1a0d0139-1a16-409a-b523-f987ba54d59c tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Lock "67d1bdd8-cf8f-4fb5-bd3c-42008e4d60e3-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1017.027663] env[61806]: INFO nova.compute.manager [None req-1a0d0139-1a16-409a-b523-f987ba54d59c tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] [instance: 67d1bdd8-cf8f-4fb5-bd3c-42008e4d60e3] Terminating instance [ 1017.029381] env[61806]: DEBUG nova.compute.manager [None req-1a0d0139-1a16-409a-b523-f987ba54d59c tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] [instance: 67d1bdd8-cf8f-4fb5-bd3c-42008e4d60e3] Start destroying the instance on the hypervisor. {{(pid=61806) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1017.029585] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-1a0d0139-1a16-409a-b523-f987ba54d59c tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] [instance: 67d1bdd8-cf8f-4fb5-bd3c-42008e4d60e3] Destroying instance {{(pid=61806) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1017.030618] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-25cb26f0-fe78-428b-9e61-4b95018919b1 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1017.040083] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [None req-1a0d0139-1a16-409a-b523-f987ba54d59c tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] [instance: 67d1bdd8-cf8f-4fb5-bd3c-42008e4d60e3] Powering off the VM {{(pid=61806) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1017.040083] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-1f95cd1a-f8c2-4f2f-b3cd-5a50b683d2da {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1017.047457] env[61806]: DEBUG oslo_vmware.api [None req-1a0d0139-1a16-409a-b523-f987ba54d59c tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Waiting for the task: (returnval){ [ 1017.047457] env[61806]: value = "task-1295107" [ 1017.047457] env[61806]: _type = "Task" [ 1017.047457] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1017.058508] env[61806]: DEBUG oslo_vmware.api [None req-1a0d0139-1a16-409a-b523-f987ba54d59c tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Task: {'id': task-1295107, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1017.084092] env[61806]: DEBUG oslo_concurrency.lockutils [req-80d50b1a-908a-40d8-9c33-8b23f3ea7bc9 req-a8255949-5019-475a-bf9c-592897441883 service nova] Releasing lock "refresh_cache-fd9e520a-2e7b-43d4-b84d-8933c32cc6a6" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1017.225536] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2850c8e8-0e0f-46b6-b609-bd1d11b90717 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1017.233323] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-fb581627-64fc-4a53-be25-73413ad0fac5 tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] [instance: 71999b04-bf6e-4e4c-87c9-d5ec1932a80f] Suspending the VM {{(pid=61806) suspend /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1163}} [ 1017.233619] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.SuspendVM_Task with opID=oslo.vmware-502a31d3-ec70-40a1-9024-b9d6fbcfad73 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1017.242460] env[61806]: DEBUG oslo_vmware.api [None req-fb581627-64fc-4a53-be25-73413ad0fac5 tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Waiting for the task: (returnval){ [ 1017.242460] env[61806]: value = "task-1295108" [ 1017.242460] env[61806]: _type = "Task" [ 1017.242460] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1017.255402] env[61806]: DEBUG oslo_vmware.api [None req-fb581627-64fc-4a53-be25-73413ad0fac5 tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Task: {'id': task-1295108, 'name': SuspendVM_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1017.269579] env[61806]: DEBUG oslo_concurrency.lockutils [None req-8b3f01e9-87f4-442a-a746-46f4f6533d96 tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] Lock "fe1d4ace-7de3-423b-b423-9d7397aa2413" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 6.443s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1017.384821] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d5b56f54-0001-4fb0-acda-12f967a2d913 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1017.394287] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-83e9d613-4a69-46ff-889f-9a5a78332f05 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1017.439567] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2bed91ef-f4b3-4e48-8f72-0d442188582a {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1017.452687] env[61806]: DEBUG oslo_vmware.api [-] Task: {'id': task-1295104, 'name': CreateVM_Task, 'duration_secs': 1.438972} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1017.454038] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d2034639-31c7-40c6-85d0-3287da36b9d0 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1017.458271] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: fd9e520a-2e7b-43d4-b84d-8933c32cc6a6] Created VM on the ESX host {{(pid=61806) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 1017.459197] env[61806]: DEBUG oslo_concurrency.lockutils [None req-de9a979c-45ed-463a-871e-b7dda8c366a2 tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/b08c43d9-cfd5-4a1a-b99d-0aebb718b592" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1017.459450] env[61806]: DEBUG oslo_concurrency.lockutils [None req-de9a979c-45ed-463a-871e-b7dda8c366a2 tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] Acquired lock "[datastore2] devstack-image-cache_base/b08c43d9-cfd5-4a1a-b99d-0aebb718b592" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1017.459853] env[61806]: DEBUG oslo_concurrency.lockutils [None req-de9a979c-45ed-463a-871e-b7dda8c366a2 tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/b08c43d9-cfd5-4a1a-b99d-0aebb718b592" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1017.460758] env[61806]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-5776cba2-5068-4c34-825a-f0f58851446b {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1017.472580] env[61806]: DEBUG nova.compute.provider_tree [None req-666a872d-3efd-4dd5-b0c9-c25d09581b9c tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Inventory has not changed in ProviderTree for provider: a2858be1-fd22-4e08-979e-87ad25293407 {{(pid=61806) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1017.475467] env[61806]: DEBUG oslo_vmware.api [None req-de9a979c-45ed-463a-871e-b7dda8c366a2 tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] Waiting for the task: (returnval){ [ 1017.475467] env[61806]: value = "session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]5201ec54-7628-13c5-6659-e6fa1bfb3c8c" [ 1017.475467] env[61806]: _type = "Task" [ 1017.475467] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1017.486614] env[61806]: DEBUG oslo_concurrency.lockutils [None req-de9a979c-45ed-463a-871e-b7dda8c366a2 tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] Releasing lock "[datastore2] devstack-image-cache_base/b08c43d9-cfd5-4a1a-b99d-0aebb718b592" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1017.487445] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-de9a979c-45ed-463a-871e-b7dda8c366a2 tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] [instance: fd9e520a-2e7b-43d4-b84d-8933c32cc6a6] Processing image b08c43d9-cfd5-4a1a-b99d-0aebb718b592 {{(pid=61806) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1017.487445] env[61806]: DEBUG oslo_concurrency.lockutils [None req-de9a979c-45ed-463a-871e-b7dda8c366a2 tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/b08c43d9-cfd5-4a1a-b99d-0aebb718b592/b08c43d9-cfd5-4a1a-b99d-0aebb718b592.vmdk" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1017.489344] env[61806]: DEBUG oslo_concurrency.lockutils [None req-de9a979c-45ed-463a-871e-b7dda8c366a2 tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] Acquired lock "[datastore2] devstack-image-cache_base/b08c43d9-cfd5-4a1a-b99d-0aebb718b592/b08c43d9-cfd5-4a1a-b99d-0aebb718b592.vmdk" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1017.489344] env[61806]: DEBUG nova.virt.vmwareapi.ds_util [None req-de9a979c-45ed-463a-871e-b7dda8c366a2 tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61806) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1017.490835] env[61806]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-5038f4a5-692d-4807-9a90-c45b85216194 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1017.497624] env[61806]: DEBUG nova.objects.instance [None req-a22e273c-c0ad-47bf-9acf-b3a1dd13247a tempest-AttachVolumeNegativeTest-1683033704 tempest-AttachVolumeNegativeTest-1683033704-project-member] Lazy-loading 'flavor' on Instance uuid e4638faa-7be1-4909-a595-c437837fa314 {{(pid=61806) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1017.517158] env[61806]: DEBUG nova.virt.vmwareapi.ds_util [None req-de9a979c-45ed-463a-871e-b7dda8c366a2 tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61806) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1017.517411] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-de9a979c-45ed-463a-871e-b7dda8c366a2 tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61806) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 1017.519182] env[61806]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-5e37d6a1-e014-4c38-a92b-b3052bf0e0ad {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1017.529362] env[61806]: DEBUG oslo_vmware.api [None req-de9a979c-45ed-463a-871e-b7dda8c366a2 tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] Waiting for the task: (returnval){ [ 1017.529362] env[61806]: value = "session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]52aaa419-be30-3051-bab4-31083a38a94e" [ 1017.529362] env[61806]: _type = "Task" [ 1017.529362] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1017.539648] env[61806]: DEBUG oslo_vmware.api [None req-de9a979c-45ed-463a-871e-b7dda8c366a2 tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] Task: {'id': session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]52aaa419-be30-3051-bab4-31083a38a94e, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1017.559469] env[61806]: DEBUG oslo_vmware.api [None req-1a0d0139-1a16-409a-b523-f987ba54d59c tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Task: {'id': task-1295107, 'name': PowerOffVM_Task} progress is 100%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1017.618804] env[61806]: DEBUG nova.compute.manager [req-6afcb937-5f6c-4572-b949-13c96537fc1f req-a911fe2f-cf40-4b2b-83ef-903d4a0c9595 service nova] [instance: d0669232-6f33-4b30-97c2-2e4239af1ad0] Received event network-changed-f958e9a3-f8d6-4665-a59f-876aab66edfb {{(pid=61806) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1017.619013] env[61806]: DEBUG nova.compute.manager [req-6afcb937-5f6c-4572-b949-13c96537fc1f req-a911fe2f-cf40-4b2b-83ef-903d4a0c9595 service nova] [instance: d0669232-6f33-4b30-97c2-2e4239af1ad0] Refreshing instance network info cache due to event network-changed-f958e9a3-f8d6-4665-a59f-876aab66edfb. {{(pid=61806) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 1017.619438] env[61806]: DEBUG oslo_concurrency.lockutils [req-6afcb937-5f6c-4572-b949-13c96537fc1f req-a911fe2f-cf40-4b2b-83ef-903d4a0c9595 service nova] Acquiring lock "refresh_cache-d0669232-6f33-4b30-97c2-2e4239af1ad0" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1017.619495] env[61806]: DEBUG oslo_concurrency.lockutils [req-6afcb937-5f6c-4572-b949-13c96537fc1f req-a911fe2f-cf40-4b2b-83ef-903d4a0c9595 service nova] Acquired lock "refresh_cache-d0669232-6f33-4b30-97c2-2e4239af1ad0" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1017.619632] env[61806]: DEBUG nova.network.neutron [req-6afcb937-5f6c-4572-b949-13c96537fc1f req-a911fe2f-cf40-4b2b-83ef-903d4a0c9595 service nova] [instance: d0669232-6f33-4b30-97c2-2e4239af1ad0] Refreshing network info cache for port f958e9a3-f8d6-4665-a59f-876aab66edfb {{(pid=61806) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1017.753420] env[61806]: DEBUG oslo_vmware.api [None req-fb581627-64fc-4a53-be25-73413ad0fac5 tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Task: {'id': task-1295108, 'name': SuspendVM_Task} progress is 58%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1017.812213] env[61806]: DEBUG nova.network.neutron [None req-2bebc059-5179-46cc-9f09-746fe48209df tempest-ServersNegativeTestJSON-891856253 tempest-ServersNegativeTestJSON-891856253-project-member] [instance: bdbaef31-7d37-40cb-ac3e-ad152b37e2d2] Updating instance_info_cache with network_info: [{"id": "b8b714a4-b17c-4b3d-889e-407bec6c30ae", "address": "fa:16:3e:33:74:bb", "network": {"id": "fd35ea93-8914-4b54-a4ea-f45fc750982b", "bridge": "br-int", "label": "tempest-ServersNegativeTestJSON-2117068882-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d2e2b7eec47e4b10898215123408692a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "2d94740a-bce8-4103-8ecf-230d02ec0a44", "external-id": "nsx-vlan-transportzone-149", "segmentation_id": 149, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb8b714a4-b1", "ovs_interfaceid": "b8b714a4-b17c-4b3d-889e-407bec6c30ae", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61806) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1017.977812] env[61806]: DEBUG nova.scheduler.client.report [None req-666a872d-3efd-4dd5-b0c9-c25d09581b9c tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Inventory has not changed for provider a2858be1-fd22-4e08-979e-87ad25293407 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 138, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61806) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1018.002790] env[61806]: DEBUG oslo_concurrency.lockutils [None req-a22e273c-c0ad-47bf-9acf-b3a1dd13247a tempest-AttachVolumeNegativeTest-1683033704 tempest-AttachVolumeNegativeTest-1683033704-project-member] Lock "e4638faa-7be1-4909-a595-c437837fa314" "released" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: held 7.343s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1018.040149] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-de9a979c-45ed-463a-871e-b7dda8c366a2 tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] [instance: fd9e520a-2e7b-43d4-b84d-8933c32cc6a6] Preparing fetch location {{(pid=61806) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:633}} [ 1018.040549] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-de9a979c-45ed-463a-871e-b7dda8c366a2 tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] [instance: fd9e520a-2e7b-43d4-b84d-8933c32cc6a6] Fetch image to [datastore2] OSTACK_IMG_db03ccaf-05b7-4eb9-8bd9-fe317c49b0ef/OSTACK_IMG_db03ccaf-05b7-4eb9-8bd9-fe317c49b0ef.vmdk {{(pid=61806) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:635}} [ 1018.040793] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-de9a979c-45ed-463a-871e-b7dda8c366a2 tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] [instance: fd9e520a-2e7b-43d4-b84d-8933c32cc6a6] Downloading stream optimized image b08c43d9-cfd5-4a1a-b99d-0aebb718b592 to [datastore2] OSTACK_IMG_db03ccaf-05b7-4eb9-8bd9-fe317c49b0ef/OSTACK_IMG_db03ccaf-05b7-4eb9-8bd9-fe317c49b0ef.vmdk on the data store datastore2 as vApp {{(pid=61806) _fetch_image_as_vapp /opt/stack/nova/nova/virt/vmwareapi/vmops.py:437}} [ 1018.041076] env[61806]: DEBUG nova.virt.vmwareapi.images [None req-de9a979c-45ed-463a-871e-b7dda8c366a2 tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] [instance: fd9e520a-2e7b-43d4-b84d-8933c32cc6a6] Downloading image file data b08c43d9-cfd5-4a1a-b99d-0aebb718b592 to the ESX as VM named 'OSTACK_IMG_db03ccaf-05b7-4eb9-8bd9-fe317c49b0ef' {{(pid=61806) fetch_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:336}} [ 1018.050242] env[61806]: DEBUG oslo_concurrency.lockutils [None req-d6fbd9e3-c911-4268-af0d-195628dc566f tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] Acquiring lock "713d4e6e-6f0e-4844-9b55-cc010030eda6" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1018.050494] env[61806]: DEBUG oslo_concurrency.lockutils [None req-d6fbd9e3-c911-4268-af0d-195628dc566f tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] Lock "713d4e6e-6f0e-4844-9b55-cc010030eda6" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1018.062397] env[61806]: DEBUG oslo_vmware.api [None req-1a0d0139-1a16-409a-b523-f987ba54d59c tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Task: {'id': task-1295107, 'name': PowerOffVM_Task} progress is 100%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1018.125220] env[61806]: DEBUG oslo_vmware.rw_handles [None req-de9a979c-45ed-463a-871e-b7dda8c366a2 tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] Creating HttpNfcLease lease for vApp import into resource pool: (val){ [ 1018.125220] env[61806]: value = "resgroup-9" [ 1018.125220] env[61806]: _type = "ResourcePool" [ 1018.125220] env[61806]: }. {{(pid=61806) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:453}} [ 1018.125693] env[61806]: DEBUG oslo_vmware.service [-] Invoking ResourcePool.ImportVApp with opID=oslo.vmware-3b769140-48ab-4803-9e75-967f0564d590 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1018.150860] env[61806]: DEBUG oslo_vmware.rw_handles [None req-de9a979c-45ed-463a-871e-b7dda8c366a2 tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] Lease: (returnval){ [ 1018.150860] env[61806]: value = "session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]5277526b-72f7-1c6b-b267-0036f1ddd96b" [ 1018.150860] env[61806]: _type = "HttpNfcLease" [ 1018.150860] env[61806]: } obtained for vApp import into resource pool (val){ [ 1018.150860] env[61806]: value = "resgroup-9" [ 1018.150860] env[61806]: _type = "ResourcePool" [ 1018.150860] env[61806]: }. {{(pid=61806) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:461}} [ 1018.151222] env[61806]: DEBUG oslo_vmware.api [None req-de9a979c-45ed-463a-871e-b7dda8c366a2 tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] Waiting for the lease: (returnval){ [ 1018.151222] env[61806]: value = "session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]5277526b-72f7-1c6b-b267-0036f1ddd96b" [ 1018.151222] env[61806]: _type = "HttpNfcLease" [ 1018.151222] env[61806]: } to be ready. {{(pid=61806) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 1018.158941] env[61806]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1018.158941] env[61806]: value = "session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]5277526b-72f7-1c6b-b267-0036f1ddd96b" [ 1018.158941] env[61806]: _type = "HttpNfcLease" [ 1018.158941] env[61806]: } is initializing. {{(pid=61806) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 1018.257886] env[61806]: DEBUG oslo_vmware.api [None req-fb581627-64fc-4a53-be25-73413ad0fac5 tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Task: {'id': task-1295108, 'name': SuspendVM_Task} progress is 100%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1018.266566] env[61806]: DEBUG oslo_concurrency.lockutils [None req-d0b88989-5e9c-4d68-be85-24933492125d tempest-AttachVolumeNegativeTest-1683033704 tempest-AttachVolumeNegativeTest-1683033704-project-member] Acquiring lock "e4638faa-7be1-4909-a595-c437837fa314" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1018.266837] env[61806]: DEBUG oslo_concurrency.lockutils [None req-d0b88989-5e9c-4d68-be85-24933492125d tempest-AttachVolumeNegativeTest-1683033704 tempest-AttachVolumeNegativeTest-1683033704-project-member] Lock "e4638faa-7be1-4909-a595-c437837fa314" acquired by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: waited 0.000s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1018.316851] env[61806]: DEBUG oslo_concurrency.lockutils [None req-2bebc059-5179-46cc-9f09-746fe48209df tempest-ServersNegativeTestJSON-891856253 tempest-ServersNegativeTestJSON-891856253-project-member] Releasing lock "refresh_cache-bdbaef31-7d37-40cb-ac3e-ad152b37e2d2" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1018.422556] env[61806]: DEBUG nova.network.neutron [req-6afcb937-5f6c-4572-b949-13c96537fc1f req-a911fe2f-cf40-4b2b-83ef-903d4a0c9595 service nova] [instance: d0669232-6f33-4b30-97c2-2e4239af1ad0] Updated VIF entry in instance network info cache for port f958e9a3-f8d6-4665-a59f-876aab66edfb. {{(pid=61806) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1018.423065] env[61806]: DEBUG nova.network.neutron [req-6afcb937-5f6c-4572-b949-13c96537fc1f req-a911fe2f-cf40-4b2b-83ef-903d4a0c9595 service nova] [instance: d0669232-6f33-4b30-97c2-2e4239af1ad0] Updating instance_info_cache with network_info: [{"id": "f958e9a3-f8d6-4665-a59f-876aab66edfb", "address": "fa:16:3e:fc:93:23", "network": {"id": "33bf5b21-82b0-432e-ae3d-9c2ca75161d0", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-1294159336-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.187", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "1c55cc0455584fbb8cda88d33b8f8b7e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "5b0fa7a2-ebd9-4788-8904-7bf250ce466c", "external-id": "nsx-vlan-transportzone-669", "segmentation_id": 669, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf958e9a3-f8", "ovs_interfaceid": "f958e9a3-f8d6-4665-a59f-876aab66edfb", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61806) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1018.555950] env[61806]: DEBUG nova.compute.manager [None req-d6fbd9e3-c911-4268-af0d-195628dc566f tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] [instance: 713d4e6e-6f0e-4844-9b55-cc010030eda6] Starting instance... {{(pid=61806) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1018.562899] env[61806]: DEBUG oslo_vmware.api [None req-1a0d0139-1a16-409a-b523-f987ba54d59c tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Task: {'id': task-1295107, 'name': PowerOffVM_Task, 'duration_secs': 1.215854} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1018.563708] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [None req-1a0d0139-1a16-409a-b523-f987ba54d59c tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] [instance: 67d1bdd8-cf8f-4fb5-bd3c-42008e4d60e3] Powered off the VM {{(pid=61806) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1018.563991] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-1a0d0139-1a16-409a-b523-f987ba54d59c tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] [instance: 67d1bdd8-cf8f-4fb5-bd3c-42008e4d60e3] Unregistering the VM {{(pid=61806) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1018.564336] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-86805513-a792-4d70-8ad4-014b292b3cc2 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1018.629897] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-2bebc059-5179-46cc-9f09-746fe48209df tempest-ServersNegativeTestJSON-891856253 tempest-ServersNegativeTestJSON-891856253-project-member] [instance: bdbaef31-7d37-40cb-ac3e-ad152b37e2d2] Destroying instance {{(pid=61806) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1018.630847] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8632fb8b-de4c-4f56-80c5-0d2b6bf45e03 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1018.639197] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-2bebc059-5179-46cc-9f09-746fe48209df tempest-ServersNegativeTestJSON-891856253 tempest-ServersNegativeTestJSON-891856253-project-member] [instance: bdbaef31-7d37-40cb-ac3e-ad152b37e2d2] Unregistering the VM {{(pid=61806) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1018.639479] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-fb37f557-fded-4ac2-a31a-9785f18c25e9 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1018.662431] env[61806]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1018.662431] env[61806]: value = "session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]5277526b-72f7-1c6b-b267-0036f1ddd96b" [ 1018.662431] env[61806]: _type = "HttpNfcLease" [ 1018.662431] env[61806]: } is initializing. {{(pid=61806) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 1018.677938] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-1a0d0139-1a16-409a-b523-f987ba54d59c tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] [instance: 67d1bdd8-cf8f-4fb5-bd3c-42008e4d60e3] Unregistered the VM {{(pid=61806) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1018.678286] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-1a0d0139-1a16-409a-b523-f987ba54d59c tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] [instance: 67d1bdd8-cf8f-4fb5-bd3c-42008e4d60e3] Deleting contents of the VM from datastore datastore1 {{(pid=61806) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1018.678601] env[61806]: DEBUG nova.virt.vmwareapi.ds_util [None req-1a0d0139-1a16-409a-b523-f987ba54d59c tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Deleting the datastore file [datastore1] 67d1bdd8-cf8f-4fb5-bd3c-42008e4d60e3 {{(pid=61806) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1018.678944] env[61806]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-80eca74c-9d20-47f9-87c7-7c2e159e9930 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1018.686694] env[61806]: DEBUG oslo_vmware.api [None req-1a0d0139-1a16-409a-b523-f987ba54d59c tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Waiting for the task: (returnval){ [ 1018.686694] env[61806]: value = "task-1295112" [ 1018.686694] env[61806]: _type = "Task" [ 1018.686694] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1018.695159] env[61806]: DEBUG oslo_vmware.api [None req-1a0d0139-1a16-409a-b523-f987ba54d59c tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Task: {'id': task-1295112, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1018.730806] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-2bebc059-5179-46cc-9f09-746fe48209df tempest-ServersNegativeTestJSON-891856253 tempest-ServersNegativeTestJSON-891856253-project-member] [instance: bdbaef31-7d37-40cb-ac3e-ad152b37e2d2] Unregistered the VM {{(pid=61806) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1018.731133] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-2bebc059-5179-46cc-9f09-746fe48209df tempest-ServersNegativeTestJSON-891856253 tempest-ServersNegativeTestJSON-891856253-project-member] [instance: bdbaef31-7d37-40cb-ac3e-ad152b37e2d2] Deleting contents of the VM from datastore datastore1 {{(pid=61806) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1018.731509] env[61806]: DEBUG nova.virt.vmwareapi.ds_util [None req-2bebc059-5179-46cc-9f09-746fe48209df tempest-ServersNegativeTestJSON-891856253 tempest-ServersNegativeTestJSON-891856253-project-member] Deleting the datastore file [datastore1] bdbaef31-7d37-40cb-ac3e-ad152b37e2d2 {{(pid=61806) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1018.731874] env[61806]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-25031a70-e9fc-406e-afe1-59d93b891b3b {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1018.740951] env[61806]: DEBUG oslo_vmware.api [None req-2bebc059-5179-46cc-9f09-746fe48209df tempest-ServersNegativeTestJSON-891856253 tempest-ServersNegativeTestJSON-891856253-project-member] Waiting for the task: (returnval){ [ 1018.740951] env[61806]: value = "task-1295113" [ 1018.740951] env[61806]: _type = "Task" [ 1018.740951] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1018.751365] env[61806]: DEBUG oslo_vmware.api [None req-2bebc059-5179-46cc-9f09-746fe48209df tempest-ServersNegativeTestJSON-891856253 tempest-ServersNegativeTestJSON-891856253-project-member] Task: {'id': task-1295113, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1018.757605] env[61806]: DEBUG oslo_vmware.api [None req-fb581627-64fc-4a53-be25-73413ad0fac5 tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Task: {'id': task-1295108, 'name': SuspendVM_Task, 'duration_secs': 1.028492} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1018.757897] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-fb581627-64fc-4a53-be25-73413ad0fac5 tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] [instance: 71999b04-bf6e-4e4c-87c9-d5ec1932a80f] Suspended the VM {{(pid=61806) suspend /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1167}} [ 1018.758118] env[61806]: DEBUG nova.compute.manager [None req-fb581627-64fc-4a53-be25-73413ad0fac5 tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] [instance: 71999b04-bf6e-4e4c-87c9-d5ec1932a80f] Checking state {{(pid=61806) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1018.758883] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-90bafbad-b49e-4a80-9934-c2db3288f6cb {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1018.769672] env[61806]: INFO nova.compute.manager [None req-d0b88989-5e9c-4d68-be85-24933492125d tempest-AttachVolumeNegativeTest-1683033704 tempest-AttachVolumeNegativeTest-1683033704-project-member] [instance: e4638faa-7be1-4909-a595-c437837fa314] Detaching volume 1492823b-b43f-42fd-bc67-94fe82531df9 [ 1018.807013] env[61806]: INFO nova.virt.block_device [None req-d0b88989-5e9c-4d68-be85-24933492125d tempest-AttachVolumeNegativeTest-1683033704 tempest-AttachVolumeNegativeTest-1683033704-project-member] [instance: e4638faa-7be1-4909-a595-c437837fa314] Attempting to driver detach volume 1492823b-b43f-42fd-bc67-94fe82531df9 from mountpoint /dev/sdb [ 1018.807368] env[61806]: DEBUG nova.virt.vmwareapi.volumeops [None req-d0b88989-5e9c-4d68-be85-24933492125d tempest-AttachVolumeNegativeTest-1683033704 tempest-AttachVolumeNegativeTest-1683033704-project-member] [instance: e4638faa-7be1-4909-a595-c437837fa314] Volume detach. Driver type: vmdk {{(pid=61806) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 1018.807632] env[61806]: DEBUG nova.virt.vmwareapi.volumeops [None req-d0b88989-5e9c-4d68-be85-24933492125d tempest-AttachVolumeNegativeTest-1683033704 tempest-AttachVolumeNegativeTest-1683033704-project-member] [instance: e4638faa-7be1-4909-a595-c437837fa314] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-277745', 'volume_id': '1492823b-b43f-42fd-bc67-94fe82531df9', 'name': 'volume-1492823b-b43f-42fd-bc67-94fe82531df9', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'e4638faa-7be1-4909-a595-c437837fa314', 'attached_at': '', 'detached_at': '', 'volume_id': '1492823b-b43f-42fd-bc67-94fe82531df9', 'serial': '1492823b-b43f-42fd-bc67-94fe82531df9'} {{(pid=61806) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 1018.808842] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7280f03e-0ae7-45b9-aef8-21bbf500d8f2 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1018.832909] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2344158e-868e-4eef-a3c2-816ea3b104fe {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1018.841538] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1eb6f2b6-998d-4cc4-8d67-59394dab0906 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1018.863968] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a27cfcec-efa1-4864-b067-c37e4f72b540 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1018.880768] env[61806]: DEBUG nova.virt.vmwareapi.volumeops [None req-d0b88989-5e9c-4d68-be85-24933492125d tempest-AttachVolumeNegativeTest-1683033704 tempest-AttachVolumeNegativeTest-1683033704-project-member] The volume has not been displaced from its original location: [datastore2] volume-1492823b-b43f-42fd-bc67-94fe82531df9/volume-1492823b-b43f-42fd-bc67-94fe82531df9.vmdk. No consolidation needed. {{(pid=61806) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 1018.886520] env[61806]: DEBUG nova.virt.vmwareapi.volumeops [None req-d0b88989-5e9c-4d68-be85-24933492125d tempest-AttachVolumeNegativeTest-1683033704 tempest-AttachVolumeNegativeTest-1683033704-project-member] [instance: e4638faa-7be1-4909-a595-c437837fa314] Reconfiguring VM instance instance-0000005d to detach disk 2001 {{(pid=61806) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1018.887024] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-076d374b-f945-4365-952c-94dfd4274675 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1018.907009] env[61806]: DEBUG oslo_vmware.api [None req-d0b88989-5e9c-4d68-be85-24933492125d tempest-AttachVolumeNegativeTest-1683033704 tempest-AttachVolumeNegativeTest-1683033704-project-member] Waiting for the task: (returnval){ [ 1018.907009] env[61806]: value = "task-1295114" [ 1018.907009] env[61806]: _type = "Task" [ 1018.907009] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1018.915568] env[61806]: DEBUG oslo_vmware.api [None req-d0b88989-5e9c-4d68-be85-24933492125d tempest-AttachVolumeNegativeTest-1683033704 tempest-AttachVolumeNegativeTest-1683033704-project-member] Task: {'id': task-1295114, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1018.926233] env[61806]: DEBUG oslo_concurrency.lockutils [req-6afcb937-5f6c-4572-b949-13c96537fc1f req-a911fe2f-cf40-4b2b-83ef-903d4a0c9595 service nova] Releasing lock "refresh_cache-d0669232-6f33-4b30-97c2-2e4239af1ad0" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1018.989563] env[61806]: DEBUG oslo_concurrency.lockutils [None req-666a872d-3efd-4dd5-b0c9-c25d09581b9c tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" :: held 2.249s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1019.086749] env[61806]: DEBUG oslo_concurrency.lockutils [None req-d6fbd9e3-c911-4268-af0d-195628dc566f tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1019.087120] env[61806]: DEBUG oslo_concurrency.lockutils [None req-d6fbd9e3-c911-4268-af0d-195628dc566f tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.001s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1019.088859] env[61806]: INFO nova.compute.claims [None req-d6fbd9e3-c911-4268-af0d-195628dc566f tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] [instance: 713d4e6e-6f0e-4844-9b55-cc010030eda6] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1019.161666] env[61806]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1019.161666] env[61806]: value = "session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]5277526b-72f7-1c6b-b267-0036f1ddd96b" [ 1019.161666] env[61806]: _type = "HttpNfcLease" [ 1019.161666] env[61806]: } is ready. {{(pid=61806) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 1019.161998] env[61806]: DEBUG oslo_vmware.rw_handles [None req-de9a979c-45ed-463a-871e-b7dda8c366a2 tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 1019.161998] env[61806]: value = "session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]5277526b-72f7-1c6b-b267-0036f1ddd96b" [ 1019.161998] env[61806]: _type = "HttpNfcLease" [ 1019.161998] env[61806]: }. {{(pid=61806) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:467}} [ 1019.162863] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e1c4856f-f353-42a9-8443-772ac1d8b1b5 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1019.172016] env[61806]: DEBUG oslo_vmware.rw_handles [None req-de9a979c-45ed-463a-871e-b7dda8c366a2 tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] Found VMDK URL: https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52534545-3613-8aec-5fa6-95bfdb3934f8/disk-0.vmdk from lease info. {{(pid=61806) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 1019.172315] env[61806]: DEBUG oslo_vmware.rw_handles [None req-de9a979c-45ed-463a-871e-b7dda8c366a2 tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] Creating HTTP connection to write to file with size = 31668736 and URL = https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52534545-3613-8aec-5fa6-95bfdb3934f8/disk-0.vmdk. {{(pid=61806) _create_write_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:122}} [ 1019.237941] env[61806]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-e019142d-4af1-4796-a28c-32b5f0299746 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1019.243286] env[61806]: DEBUG oslo_vmware.api [None req-1a0d0139-1a16-409a-b523-f987ba54d59c tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Task: {'id': task-1295112, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.249884} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1019.246533] env[61806]: DEBUG nova.virt.vmwareapi.ds_util [None req-1a0d0139-1a16-409a-b523-f987ba54d59c tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Deleted the datastore file {{(pid=61806) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1019.246742] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-1a0d0139-1a16-409a-b523-f987ba54d59c tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] [instance: 67d1bdd8-cf8f-4fb5-bd3c-42008e4d60e3] Deleted contents of the VM from datastore datastore1 {{(pid=61806) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1019.246928] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-1a0d0139-1a16-409a-b523-f987ba54d59c tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] [instance: 67d1bdd8-cf8f-4fb5-bd3c-42008e4d60e3] Instance destroyed {{(pid=61806) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1019.247137] env[61806]: INFO nova.compute.manager [None req-1a0d0139-1a16-409a-b523-f987ba54d59c tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] [instance: 67d1bdd8-cf8f-4fb5-bd3c-42008e4d60e3] Took 2.22 seconds to destroy the instance on the hypervisor. [ 1019.247401] env[61806]: DEBUG oslo.service.loopingcall [None req-1a0d0139-1a16-409a-b523-f987ba54d59c tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61806) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1019.248426] env[61806]: DEBUG nova.compute.manager [-] [instance: 67d1bdd8-cf8f-4fb5-bd3c-42008e4d60e3] Deallocating network for instance {{(pid=61806) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1019.248529] env[61806]: DEBUG nova.network.neutron [-] [instance: 67d1bdd8-cf8f-4fb5-bd3c-42008e4d60e3] deallocate_for_instance() {{(pid=61806) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1019.255044] env[61806]: DEBUG oslo_vmware.api [None req-2bebc059-5179-46cc-9f09-746fe48209df tempest-ServersNegativeTestJSON-891856253 tempest-ServersNegativeTestJSON-891856253-project-member] Task: {'id': task-1295113, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.224933} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1019.255276] env[61806]: DEBUG nova.virt.vmwareapi.ds_util [None req-2bebc059-5179-46cc-9f09-746fe48209df tempest-ServersNegativeTestJSON-891856253 tempest-ServersNegativeTestJSON-891856253-project-member] Deleted the datastore file {{(pid=61806) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1019.255458] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-2bebc059-5179-46cc-9f09-746fe48209df tempest-ServersNegativeTestJSON-891856253 tempest-ServersNegativeTestJSON-891856253-project-member] [instance: bdbaef31-7d37-40cb-ac3e-ad152b37e2d2] Deleted contents of the VM from datastore datastore1 {{(pid=61806) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1019.255638] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-2bebc059-5179-46cc-9f09-746fe48209df tempest-ServersNegativeTestJSON-891856253 tempest-ServersNegativeTestJSON-891856253-project-member] [instance: bdbaef31-7d37-40cb-ac3e-ad152b37e2d2] Instance destroyed {{(pid=61806) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1019.276772] env[61806]: INFO nova.scheduler.client.report [None req-2bebc059-5179-46cc-9f09-746fe48209df tempest-ServersNegativeTestJSON-891856253 tempest-ServersNegativeTestJSON-891856253-project-member] Deleted allocations for instance bdbaef31-7d37-40cb-ac3e-ad152b37e2d2 [ 1019.418558] env[61806]: DEBUG oslo_vmware.api [None req-d0b88989-5e9c-4d68-be85-24933492125d tempest-AttachVolumeNegativeTest-1683033704 tempest-AttachVolumeNegativeTest-1683033704-project-member] Task: {'id': task-1295114, 'name': ReconfigVM_Task, 'duration_secs': 0.258578} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1019.418872] env[61806]: DEBUG nova.virt.vmwareapi.volumeops [None req-d0b88989-5e9c-4d68-be85-24933492125d tempest-AttachVolumeNegativeTest-1683033704 tempest-AttachVolumeNegativeTest-1683033704-project-member] [instance: e4638faa-7be1-4909-a595-c437837fa314] Reconfigured VM instance instance-0000005d to detach disk 2001 {{(pid=61806) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1019.425255] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-bfaec779-0b42-4f71-bbdb-afeda63dd005 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1019.442304] env[61806]: DEBUG oslo_vmware.api [None req-d0b88989-5e9c-4d68-be85-24933492125d tempest-AttachVolumeNegativeTest-1683033704 tempest-AttachVolumeNegativeTest-1683033704-project-member] Waiting for the task: (returnval){ [ 1019.442304] env[61806]: value = "task-1295115" [ 1019.442304] env[61806]: _type = "Task" [ 1019.442304] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1019.455028] env[61806]: DEBUG oslo_vmware.api [None req-d0b88989-5e9c-4d68-be85-24933492125d tempest-AttachVolumeNegativeTest-1683033704 tempest-AttachVolumeNegativeTest-1683033704-project-member] Task: {'id': task-1295115, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1019.549457] env[61806]: INFO nova.scheduler.client.report [None req-666a872d-3efd-4dd5-b0c9-c25d09581b9c tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Deleted allocation for migration ce240126-0844-456c-b7bf-b9d90e170466 [ 1019.639027] env[61806]: DEBUG oslo_concurrency.lockutils [None req-d16494e9-ced1-4dfc-b5e8-4d317df5ec0e tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Acquiring lock "71999b04-bf6e-4e4c-87c9-d5ec1932a80f" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1019.639355] env[61806]: DEBUG oslo_concurrency.lockutils [None req-d16494e9-ced1-4dfc-b5e8-4d317df5ec0e tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Lock "71999b04-bf6e-4e4c-87c9-d5ec1932a80f" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1019.639611] env[61806]: DEBUG oslo_concurrency.lockutils [None req-d16494e9-ced1-4dfc-b5e8-4d317df5ec0e tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Acquiring lock "71999b04-bf6e-4e4c-87c9-d5ec1932a80f-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1019.639867] env[61806]: DEBUG oslo_concurrency.lockutils [None req-d16494e9-ced1-4dfc-b5e8-4d317df5ec0e tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Lock "71999b04-bf6e-4e4c-87c9-d5ec1932a80f-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1019.640182] env[61806]: DEBUG oslo_concurrency.lockutils [None req-d16494e9-ced1-4dfc-b5e8-4d317df5ec0e tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Lock "71999b04-bf6e-4e4c-87c9-d5ec1932a80f-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1019.644389] env[61806]: INFO nova.compute.manager [None req-d16494e9-ced1-4dfc-b5e8-4d317df5ec0e tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] [instance: 71999b04-bf6e-4e4c-87c9-d5ec1932a80f] Terminating instance [ 1019.646984] env[61806]: DEBUG nova.compute.manager [None req-d16494e9-ced1-4dfc-b5e8-4d317df5ec0e tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] [instance: 71999b04-bf6e-4e4c-87c9-d5ec1932a80f] Start destroying the instance on the hypervisor. {{(pid=61806) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1019.647222] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-d16494e9-ced1-4dfc-b5e8-4d317df5ec0e tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] [instance: 71999b04-bf6e-4e4c-87c9-d5ec1932a80f] Destroying instance {{(pid=61806) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1019.648321] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c98be64e-3c88-4e33-8093-9d598d394369 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1019.657010] env[61806]: DEBUG nova.compute.manager [req-daa52ad5-0afa-4dd6-9ac7-5b8208ac384d req-2aa74e9f-eca1-4f4b-b797-fd776e75b155 service nova] [instance: bdbaef31-7d37-40cb-ac3e-ad152b37e2d2] Received event network-vif-unplugged-b8b714a4-b17c-4b3d-889e-407bec6c30ae {{(pid=61806) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1019.657946] env[61806]: DEBUG oslo_concurrency.lockutils [req-daa52ad5-0afa-4dd6-9ac7-5b8208ac384d req-2aa74e9f-eca1-4f4b-b797-fd776e75b155 service nova] Acquiring lock "bdbaef31-7d37-40cb-ac3e-ad152b37e2d2-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1019.657946] env[61806]: DEBUG oslo_concurrency.lockutils [req-daa52ad5-0afa-4dd6-9ac7-5b8208ac384d req-2aa74e9f-eca1-4f4b-b797-fd776e75b155 service nova] Lock "bdbaef31-7d37-40cb-ac3e-ad152b37e2d2-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1019.657946] env[61806]: DEBUG oslo_concurrency.lockutils [req-daa52ad5-0afa-4dd6-9ac7-5b8208ac384d req-2aa74e9f-eca1-4f4b-b797-fd776e75b155 service nova] Lock "bdbaef31-7d37-40cb-ac3e-ad152b37e2d2-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1019.658170] env[61806]: DEBUG nova.compute.manager [req-daa52ad5-0afa-4dd6-9ac7-5b8208ac384d req-2aa74e9f-eca1-4f4b-b797-fd776e75b155 service nova] [instance: bdbaef31-7d37-40cb-ac3e-ad152b37e2d2] No waiting events found dispatching network-vif-unplugged-b8b714a4-b17c-4b3d-889e-407bec6c30ae {{(pid=61806) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1019.658401] env[61806]: WARNING nova.compute.manager [req-daa52ad5-0afa-4dd6-9ac7-5b8208ac384d req-2aa74e9f-eca1-4f4b-b797-fd776e75b155 service nova] [instance: bdbaef31-7d37-40cb-ac3e-ad152b37e2d2] Received unexpected event network-vif-unplugged-b8b714a4-b17c-4b3d-889e-407bec6c30ae for instance with vm_state shelved_offloaded and task_state None. [ 1019.658608] env[61806]: DEBUG nova.compute.manager [req-daa52ad5-0afa-4dd6-9ac7-5b8208ac384d req-2aa74e9f-eca1-4f4b-b797-fd776e75b155 service nova] [instance: bdbaef31-7d37-40cb-ac3e-ad152b37e2d2] Received event network-changed-b8b714a4-b17c-4b3d-889e-407bec6c30ae {{(pid=61806) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1019.658779] env[61806]: DEBUG nova.compute.manager [req-daa52ad5-0afa-4dd6-9ac7-5b8208ac384d req-2aa74e9f-eca1-4f4b-b797-fd776e75b155 service nova] [instance: bdbaef31-7d37-40cb-ac3e-ad152b37e2d2] Refreshing instance network info cache due to event network-changed-b8b714a4-b17c-4b3d-889e-407bec6c30ae. {{(pid=61806) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 1019.659025] env[61806]: DEBUG oslo_concurrency.lockutils [req-daa52ad5-0afa-4dd6-9ac7-5b8208ac384d req-2aa74e9f-eca1-4f4b-b797-fd776e75b155 service nova] Acquiring lock "refresh_cache-bdbaef31-7d37-40cb-ac3e-ad152b37e2d2" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1019.659229] env[61806]: DEBUG oslo_concurrency.lockutils [req-daa52ad5-0afa-4dd6-9ac7-5b8208ac384d req-2aa74e9f-eca1-4f4b-b797-fd776e75b155 service nova] Acquired lock "refresh_cache-bdbaef31-7d37-40cb-ac3e-ad152b37e2d2" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1019.659436] env[61806]: DEBUG nova.network.neutron [req-daa52ad5-0afa-4dd6-9ac7-5b8208ac384d req-2aa74e9f-eca1-4f4b-b797-fd776e75b155 service nova] [instance: bdbaef31-7d37-40cb-ac3e-ad152b37e2d2] Refreshing network info cache for port b8b714a4-b17c-4b3d-889e-407bec6c30ae {{(pid=61806) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1019.671770] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-d16494e9-ced1-4dfc-b5e8-4d317df5ec0e tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] [instance: 71999b04-bf6e-4e4c-87c9-d5ec1932a80f] Unregistering the VM {{(pid=61806) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1019.672104] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-1f4a8ab6-1714-43d9-8654-7369567eac44 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1019.760918] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-d16494e9-ced1-4dfc-b5e8-4d317df5ec0e tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] [instance: 71999b04-bf6e-4e4c-87c9-d5ec1932a80f] Unregistered the VM {{(pid=61806) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1019.761214] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-d16494e9-ced1-4dfc-b5e8-4d317df5ec0e tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] [instance: 71999b04-bf6e-4e4c-87c9-d5ec1932a80f] Deleting contents of the VM from datastore datastore2 {{(pid=61806) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1019.761444] env[61806]: DEBUG nova.virt.vmwareapi.ds_util [None req-d16494e9-ced1-4dfc-b5e8-4d317df5ec0e tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Deleting the datastore file [datastore2] 71999b04-bf6e-4e4c-87c9-d5ec1932a80f {{(pid=61806) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1019.761742] env[61806]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-10659860-93d0-4293-a858-0c6028a5223b {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1019.770266] env[61806]: DEBUG oslo_vmware.api [None req-d16494e9-ced1-4dfc-b5e8-4d317df5ec0e tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Waiting for the task: (returnval){ [ 1019.770266] env[61806]: value = "task-1295117" [ 1019.770266] env[61806]: _type = "Task" [ 1019.770266] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1019.781926] env[61806]: DEBUG oslo_vmware.api [None req-d16494e9-ced1-4dfc-b5e8-4d317df5ec0e tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Task: {'id': task-1295117, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1019.784344] env[61806]: DEBUG oslo_concurrency.lockutils [None req-2bebc059-5179-46cc-9f09-746fe48209df tempest-ServersNegativeTestJSON-891856253 tempest-ServersNegativeTestJSON-891856253-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1019.963907] env[61806]: DEBUG oslo_vmware.api [None req-d0b88989-5e9c-4d68-be85-24933492125d tempest-AttachVolumeNegativeTest-1683033704 tempest-AttachVolumeNegativeTest-1683033704-project-member] Task: {'id': task-1295115, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1020.059029] env[61806]: DEBUG oslo_concurrency.lockutils [None req-666a872d-3efd-4dd5-b0c9-c25d09581b9c tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Lock "9357e61f-1628-43bd-ab46-de13c1529f51" "released" by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" :: held 8.676s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1020.226829] env[61806]: DEBUG nova.network.neutron [-] [instance: 67d1bdd8-cf8f-4fb5-bd3c-42008e4d60e3] Updating instance_info_cache with network_info: [] {{(pid=61806) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1020.256718] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-004a8ada-2383-4c98-b4d8-6b0996ed4ecc {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1020.276168] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-74bf77bb-4644-4b40-bb63-f296d4709494 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1020.318104] env[61806]: DEBUG oslo_vmware.api [None req-d16494e9-ced1-4dfc-b5e8-4d317df5ec0e tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Task: {'id': task-1295117, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.201575} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1020.325531] env[61806]: DEBUG nova.virt.vmwareapi.ds_util [None req-d16494e9-ced1-4dfc-b5e8-4d317df5ec0e tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Deleted the datastore file {{(pid=61806) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1020.325750] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-d16494e9-ced1-4dfc-b5e8-4d317df5ec0e tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] [instance: 71999b04-bf6e-4e4c-87c9-d5ec1932a80f] Deleted contents of the VM from datastore datastore2 {{(pid=61806) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1020.325998] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-d16494e9-ced1-4dfc-b5e8-4d317df5ec0e tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] [instance: 71999b04-bf6e-4e4c-87c9-d5ec1932a80f] Instance destroyed {{(pid=61806) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1020.326273] env[61806]: INFO nova.compute.manager [None req-d16494e9-ced1-4dfc-b5e8-4d317df5ec0e tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] [instance: 71999b04-bf6e-4e4c-87c9-d5ec1932a80f] Took 0.68 seconds to destroy the instance on the hypervisor. [ 1020.326436] env[61806]: DEBUG oslo.service.loopingcall [None req-d16494e9-ced1-4dfc-b5e8-4d317df5ec0e tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61806) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1020.327275] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cfd308ba-4226-42e9-8c9d-e932d559387b {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1020.332036] env[61806]: DEBUG nova.compute.manager [-] [instance: 71999b04-bf6e-4e4c-87c9-d5ec1932a80f] Deallocating network for instance {{(pid=61806) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1020.332168] env[61806]: DEBUG nova.network.neutron [-] [instance: 71999b04-bf6e-4e4c-87c9-d5ec1932a80f] deallocate_for_instance() {{(pid=61806) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1020.341035] env[61806]: DEBUG oslo_vmware.rw_handles [None req-de9a979c-45ed-463a-871e-b7dda8c366a2 tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] Completed reading data from the image iterator. {{(pid=61806) read /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:765}} [ 1020.341251] env[61806]: DEBUG oslo_vmware.rw_handles [None req-de9a979c-45ed-463a-871e-b7dda8c366a2 tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] Getting lease state for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52534545-3613-8aec-5fa6-95bfdb3934f8/disk-0.vmdk. {{(pid=61806) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 1020.342571] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0e8b485a-d28d-49f2-89b7-3530cac0cd9f {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1020.347218] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a661e720-0aaa-4ac2-9f98-f978d21ec933 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1020.353897] env[61806]: DEBUG oslo_vmware.rw_handles [None req-de9a979c-45ed-463a-871e-b7dda8c366a2 tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] Lease for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52534545-3613-8aec-5fa6-95bfdb3934f8/disk-0.vmdk is in state: ready. {{(pid=61806) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 1020.354079] env[61806]: DEBUG oslo_vmware.rw_handles [None req-de9a979c-45ed-463a-871e-b7dda8c366a2 tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] Releasing lease for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52534545-3613-8aec-5fa6-95bfdb3934f8/disk-0.vmdk. {{(pid=61806) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:440}} [ 1020.363282] env[61806]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseComplete with opID=oslo.vmware-285407fd-d74a-413a-b5a2-4f461d53157a {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1020.366838] env[61806]: DEBUG nova.compute.provider_tree [None req-d6fbd9e3-c911-4268-af0d-195628dc566f tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] Inventory has not changed in ProviderTree for provider: a2858be1-fd22-4e08-979e-87ad25293407 {{(pid=61806) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1020.457141] env[61806]: DEBUG oslo_vmware.api [None req-d0b88989-5e9c-4d68-be85-24933492125d tempest-AttachVolumeNegativeTest-1683033704 tempest-AttachVolumeNegativeTest-1683033704-project-member] Task: {'id': task-1295115, 'name': ReconfigVM_Task, 'duration_secs': 0.822254} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1020.457528] env[61806]: DEBUG nova.virt.vmwareapi.volumeops [None req-d0b88989-5e9c-4d68-be85-24933492125d tempest-AttachVolumeNegativeTest-1683033704 tempest-AttachVolumeNegativeTest-1683033704-project-member] [instance: e4638faa-7be1-4909-a595-c437837fa314] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-277745', 'volume_id': '1492823b-b43f-42fd-bc67-94fe82531df9', 'name': 'volume-1492823b-b43f-42fd-bc67-94fe82531df9', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'e4638faa-7be1-4909-a595-c437837fa314', 'attached_at': '', 'detached_at': '', 'volume_id': '1492823b-b43f-42fd-bc67-94fe82531df9', 'serial': '1492823b-b43f-42fd-bc67-94fe82531df9'} {{(pid=61806) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 1020.567474] env[61806]: DEBUG oslo_vmware.rw_handles [None req-de9a979c-45ed-463a-871e-b7dda8c366a2 tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] Closed VMDK write handle for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52534545-3613-8aec-5fa6-95bfdb3934f8/disk-0.vmdk. {{(pid=61806) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:646}} [ 1020.567706] env[61806]: INFO nova.virt.vmwareapi.images [None req-de9a979c-45ed-463a-871e-b7dda8c366a2 tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] [instance: fd9e520a-2e7b-43d4-b84d-8933c32cc6a6] Downloaded image file data b08c43d9-cfd5-4a1a-b99d-0aebb718b592 [ 1020.568634] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d69c5b0f-2dca-4c29-9b40-422bffb83327 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1020.587085] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-4b6452ef-db56-4f23-9e68-846bacca8ca3 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1020.615038] env[61806]: INFO nova.virt.vmwareapi.images [None req-de9a979c-45ed-463a-871e-b7dda8c366a2 tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] [instance: fd9e520a-2e7b-43d4-b84d-8933c32cc6a6] The imported VM was unregistered [ 1020.617522] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-de9a979c-45ed-463a-871e-b7dda8c366a2 tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] [instance: fd9e520a-2e7b-43d4-b84d-8933c32cc6a6] Caching image {{(pid=61806) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:638}} [ 1020.617783] env[61806]: DEBUG nova.virt.vmwareapi.ds_util [None req-de9a979c-45ed-463a-871e-b7dda8c366a2 tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] Creating directory with path [datastore2] devstack-image-cache_base/b08c43d9-cfd5-4a1a-b99d-0aebb718b592 {{(pid=61806) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1020.618758] env[61806]: DEBUG nova.network.neutron [req-daa52ad5-0afa-4dd6-9ac7-5b8208ac384d req-2aa74e9f-eca1-4f4b-b797-fd776e75b155 service nova] [instance: bdbaef31-7d37-40cb-ac3e-ad152b37e2d2] Updated VIF entry in instance network info cache for port b8b714a4-b17c-4b3d-889e-407bec6c30ae. {{(pid=61806) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1020.619145] env[61806]: DEBUG nova.network.neutron [req-daa52ad5-0afa-4dd6-9ac7-5b8208ac384d req-2aa74e9f-eca1-4f4b-b797-fd776e75b155 service nova] [instance: bdbaef31-7d37-40cb-ac3e-ad152b37e2d2] Updating instance_info_cache with network_info: [{"id": "b8b714a4-b17c-4b3d-889e-407bec6c30ae", "address": "fa:16:3e:33:74:bb", "network": {"id": "fd35ea93-8914-4b54-a4ea-f45fc750982b", "bridge": null, "label": "tempest-ServersNegativeTestJSON-2117068882-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d2e2b7eec47e4b10898215123408692a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "unbound", "details": {}, "devname": "tapb8b714a4-b1", "ovs_interfaceid": null, "qbh_params": null, "qbg_params": null, "active": false, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61806) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1020.620467] env[61806]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-596b28a2-0256-47ea-9c2d-fe1f3c55f072 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1020.632782] env[61806]: DEBUG nova.virt.vmwareapi.ds_util [None req-de9a979c-45ed-463a-871e-b7dda8c366a2 tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] Created directory with path [datastore2] devstack-image-cache_base/b08c43d9-cfd5-4a1a-b99d-0aebb718b592 {{(pid=61806) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1020.632782] env[61806]: DEBUG nova.virt.vmwareapi.ds_util [None req-de9a979c-45ed-463a-871e-b7dda8c366a2 tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] Moving virtual disk from [datastore2] OSTACK_IMG_db03ccaf-05b7-4eb9-8bd9-fe317c49b0ef/OSTACK_IMG_db03ccaf-05b7-4eb9-8bd9-fe317c49b0ef.vmdk to [datastore2] devstack-image-cache_base/b08c43d9-cfd5-4a1a-b99d-0aebb718b592/b08c43d9-cfd5-4a1a-b99d-0aebb718b592.vmdk. {{(pid=61806) disk_move /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:263}} [ 1020.632959] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.MoveVirtualDisk_Task with opID=oslo.vmware-ead879db-7253-49ad-92e1-17ed9605c415 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1020.642267] env[61806]: DEBUG oslo_vmware.api [None req-de9a979c-45ed-463a-871e-b7dda8c366a2 tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] Waiting for the task: (returnval){ [ 1020.642267] env[61806]: value = "task-1295119" [ 1020.642267] env[61806]: _type = "Task" [ 1020.642267] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1020.651783] env[61806]: DEBUG oslo_vmware.api [None req-de9a979c-45ed-463a-871e-b7dda8c366a2 tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] Task: {'id': task-1295119, 'name': MoveVirtualDisk_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1020.733368] env[61806]: INFO nova.compute.manager [-] [instance: 67d1bdd8-cf8f-4fb5-bd3c-42008e4d60e3] Took 1.48 seconds to deallocate network for instance. [ 1020.870871] env[61806]: DEBUG nova.scheduler.client.report [None req-d6fbd9e3-c911-4268-af0d-195628dc566f tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] Inventory has not changed for provider a2858be1-fd22-4e08-979e-87ad25293407 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 138, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61806) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1021.013461] env[61806]: DEBUG nova.objects.instance [None req-d0b88989-5e9c-4d68-be85-24933492125d tempest-AttachVolumeNegativeTest-1683033704 tempest-AttachVolumeNegativeTest-1683033704-project-member] Lazy-loading 'flavor' on Instance uuid e4638faa-7be1-4909-a595-c437837fa314 {{(pid=61806) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1021.124155] env[61806]: DEBUG oslo_concurrency.lockutils [req-daa52ad5-0afa-4dd6-9ac7-5b8208ac384d req-2aa74e9f-eca1-4f4b-b797-fd776e75b155 service nova] Releasing lock "refresh_cache-bdbaef31-7d37-40cb-ac3e-ad152b37e2d2" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1021.157660] env[61806]: DEBUG oslo_vmware.api [None req-de9a979c-45ed-463a-871e-b7dda8c366a2 tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] Task: {'id': task-1295119, 'name': MoveVirtualDisk_Task} progress is 21%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1021.195249] env[61806]: DEBUG oslo_concurrency.lockutils [None req-6a1c6218-d12d-4e5e-8616-7dd5071f5819 tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Acquiring lock "9357e61f-1628-43bd-ab46-de13c1529f51" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1021.195635] env[61806]: DEBUG oslo_concurrency.lockutils [None req-6a1c6218-d12d-4e5e-8616-7dd5071f5819 tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Lock "9357e61f-1628-43bd-ab46-de13c1529f51" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1021.195897] env[61806]: DEBUG oslo_concurrency.lockutils [None req-6a1c6218-d12d-4e5e-8616-7dd5071f5819 tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Acquiring lock "9357e61f-1628-43bd-ab46-de13c1529f51-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1021.196139] env[61806]: DEBUG oslo_concurrency.lockutils [None req-6a1c6218-d12d-4e5e-8616-7dd5071f5819 tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Lock "9357e61f-1628-43bd-ab46-de13c1529f51-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1021.196388] env[61806]: DEBUG oslo_concurrency.lockutils [None req-6a1c6218-d12d-4e5e-8616-7dd5071f5819 tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Lock "9357e61f-1628-43bd-ab46-de13c1529f51-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1021.198771] env[61806]: INFO nova.compute.manager [None req-6a1c6218-d12d-4e5e-8616-7dd5071f5819 tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] [instance: 9357e61f-1628-43bd-ab46-de13c1529f51] Terminating instance [ 1021.201469] env[61806]: DEBUG nova.compute.manager [None req-6a1c6218-d12d-4e5e-8616-7dd5071f5819 tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] [instance: 9357e61f-1628-43bd-ab46-de13c1529f51] Start destroying the instance on the hypervisor. {{(pid=61806) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1021.201701] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-6a1c6218-d12d-4e5e-8616-7dd5071f5819 tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] [instance: 9357e61f-1628-43bd-ab46-de13c1529f51] Destroying instance {{(pid=61806) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1021.202702] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-88806606-1e5e-4752-a16e-5fd33f6c35b0 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1021.214659] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [None req-6a1c6218-d12d-4e5e-8616-7dd5071f5819 tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] [instance: 9357e61f-1628-43bd-ab46-de13c1529f51] Powering off the VM {{(pid=61806) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1021.215011] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-57baa4e1-de76-481f-a393-8943997a84dd {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1021.221284] env[61806]: DEBUG nova.network.neutron [-] [instance: 71999b04-bf6e-4e4c-87c9-d5ec1932a80f] Updating instance_info_cache with network_info: [] {{(pid=61806) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1021.226094] env[61806]: DEBUG oslo_vmware.api [None req-6a1c6218-d12d-4e5e-8616-7dd5071f5819 tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Waiting for the task: (returnval){ [ 1021.226094] env[61806]: value = "task-1295120" [ 1021.226094] env[61806]: _type = "Task" [ 1021.226094] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1021.239221] env[61806]: DEBUG oslo_vmware.api [None req-6a1c6218-d12d-4e5e-8616-7dd5071f5819 tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Task: {'id': task-1295120, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1021.242467] env[61806]: DEBUG oslo_concurrency.lockutils [None req-1a0d0139-1a16-409a-b523-f987ba54d59c tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1021.376460] env[61806]: DEBUG oslo_concurrency.lockutils [None req-d6fbd9e3-c911-4268-af0d-195628dc566f tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.289s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1021.377142] env[61806]: DEBUG nova.compute.manager [None req-d6fbd9e3-c911-4268-af0d-195628dc566f tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] [instance: 713d4e6e-6f0e-4844-9b55-cc010030eda6] Start building networks asynchronously for instance. {{(pid=61806) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1021.379981] env[61806]: DEBUG oslo_concurrency.lockutils [None req-2bebc059-5179-46cc-9f09-746fe48209df tempest-ServersNegativeTestJSON-891856253 tempest-ServersNegativeTestJSON-891856253-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 1.596s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1021.380286] env[61806]: DEBUG nova.objects.instance [None req-2bebc059-5179-46cc-9f09-746fe48209df tempest-ServersNegativeTestJSON-891856253 tempest-ServersNegativeTestJSON-891856253-project-member] Lazy-loading 'resources' on Instance uuid bdbaef31-7d37-40cb-ac3e-ad152b37e2d2 {{(pid=61806) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1021.659020] env[61806]: DEBUG oslo_vmware.api [None req-de9a979c-45ed-463a-871e-b7dda8c366a2 tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] Task: {'id': task-1295119, 'name': MoveVirtualDisk_Task} progress is 43%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1021.681477] env[61806]: DEBUG nova.compute.manager [req-96ea789e-d26c-44c8-a79d-de5368ac2b57 req-a155bb30-f4d3-4ba5-a017-22b339bdb469 service nova] [instance: 67d1bdd8-cf8f-4fb5-bd3c-42008e4d60e3] Received event network-vif-deleted-859fb723-4a1c-4ea8-bd7c-0ed308f3a9e3 {{(pid=61806) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1021.681705] env[61806]: DEBUG nova.compute.manager [req-96ea789e-d26c-44c8-a79d-de5368ac2b57 req-a155bb30-f4d3-4ba5-a017-22b339bdb469 service nova] [instance: 71999b04-bf6e-4e4c-87c9-d5ec1932a80f] Received event network-vif-deleted-927351ab-9ed5-41c7-beea-77033dc7601b {{(pid=61806) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1021.724120] env[61806]: INFO nova.compute.manager [-] [instance: 71999b04-bf6e-4e4c-87c9-d5ec1932a80f] Took 1.39 seconds to deallocate network for instance. [ 1021.741860] env[61806]: DEBUG oslo_vmware.api [None req-6a1c6218-d12d-4e5e-8616-7dd5071f5819 tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Task: {'id': task-1295120, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1021.883493] env[61806]: DEBUG nova.compute.utils [None req-d6fbd9e3-c911-4268-af0d-195628dc566f tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] Using /dev/sd instead of None {{(pid=61806) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1021.885732] env[61806]: DEBUG nova.objects.instance [None req-2bebc059-5179-46cc-9f09-746fe48209df tempest-ServersNegativeTestJSON-891856253 tempest-ServersNegativeTestJSON-891856253-project-member] Lazy-loading 'numa_topology' on Instance uuid bdbaef31-7d37-40cb-ac3e-ad152b37e2d2 {{(pid=61806) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1021.887078] env[61806]: DEBUG nova.compute.manager [None req-d6fbd9e3-c911-4268-af0d-195628dc566f tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] [instance: 713d4e6e-6f0e-4844-9b55-cc010030eda6] Allocating IP information in the background. {{(pid=61806) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1021.887266] env[61806]: DEBUG nova.network.neutron [None req-d6fbd9e3-c911-4268-af0d-195628dc566f tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] [instance: 713d4e6e-6f0e-4844-9b55-cc010030eda6] allocate_for_instance() {{(pid=61806) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 1021.928218] env[61806]: DEBUG nova.policy [None req-d6fbd9e3-c911-4268-af0d-195628dc566f tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '4c5d4c6e1ad844a585c08c4644a60bbf', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '73ebcc4d79d248efb653a084a5e44302', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61806) authorize /opt/stack/nova/nova/policy.py:201}} [ 1022.021101] env[61806]: DEBUG oslo_concurrency.lockutils [None req-d0b88989-5e9c-4d68-be85-24933492125d tempest-AttachVolumeNegativeTest-1683033704 tempest-AttachVolumeNegativeTest-1683033704-project-member] Lock "e4638faa-7be1-4909-a595-c437837fa314" "released" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: held 3.754s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1022.085363] env[61806]: DEBUG oslo_concurrency.lockutils [None req-a5c61942-d486-43f3-834d-7dfb87ef5b63 tempest-ServersNegativeTestJSON-891856253 tempest-ServersNegativeTestJSON-891856253-project-member] Acquiring lock "bdbaef31-7d37-40cb-ac3e-ad152b37e2d2" by "nova.compute.manager.ComputeManager.unshelve_instance..do_unshelve_instance" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1022.160234] env[61806]: DEBUG oslo_vmware.api [None req-de9a979c-45ed-463a-871e-b7dda8c366a2 tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] Task: {'id': task-1295119, 'name': MoveVirtualDisk_Task} progress is 63%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1022.234054] env[61806]: DEBUG oslo_concurrency.lockutils [None req-d16494e9-ced1-4dfc-b5e8-4d317df5ec0e tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1022.242397] env[61806]: DEBUG oslo_vmware.api [None req-6a1c6218-d12d-4e5e-8616-7dd5071f5819 tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Task: {'id': task-1295120, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1022.253517] env[61806]: DEBUG nova.network.neutron [None req-d6fbd9e3-c911-4268-af0d-195628dc566f tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] [instance: 713d4e6e-6f0e-4844-9b55-cc010030eda6] Successfully created port: 7828f677-3fd0-441f-ac22-2b837a4c5d54 {{(pid=61806) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1022.391024] env[61806]: DEBUG nova.compute.manager [None req-d6fbd9e3-c911-4268-af0d-195628dc566f tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] [instance: 713d4e6e-6f0e-4844-9b55-cc010030eda6] Start building block device mappings for instance. {{(pid=61806) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1022.391640] env[61806]: DEBUG nova.objects.base [None req-2bebc059-5179-46cc-9f09-746fe48209df tempest-ServersNegativeTestJSON-891856253 tempest-ServersNegativeTestJSON-891856253-project-member] Object Instance lazy-loaded attributes: resources,numa_topology {{(pid=61806) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 1022.551594] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-43f8cd2b-0f89-491d-b926-2daf8c669a3e {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1022.563670] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b3d08582-28ea-4569-97fa-790af20dca83 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1022.599546] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ace561e3-88bf-4a01-9e89-6fb415f0d528 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1022.609968] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e46d4b51-c210-4c18-bdb1-6d36bd31b967 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1022.626669] env[61806]: DEBUG nova.compute.provider_tree [None req-2bebc059-5179-46cc-9f09-746fe48209df tempest-ServersNegativeTestJSON-891856253 tempest-ServersNegativeTestJSON-891856253-project-member] Inventory has not changed in ProviderTree for provider: a2858be1-fd22-4e08-979e-87ad25293407 {{(pid=61806) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1022.659462] env[61806]: DEBUG oslo_vmware.api [None req-de9a979c-45ed-463a-871e-b7dda8c366a2 tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] Task: {'id': task-1295119, 'name': MoveVirtualDisk_Task} progress is 85%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1022.698169] env[61806]: DEBUG oslo_concurrency.lockutils [None req-439a5433-afde-41f1-8525-78229f3cf29f tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Acquiring lock "a61abfd9-3cc1-4a2b-b68b-a446094daf87" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1022.698411] env[61806]: DEBUG oslo_concurrency.lockutils [None req-439a5433-afde-41f1-8525-78229f3cf29f tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Lock "a61abfd9-3cc1-4a2b-b68b-a446094daf87" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1022.741760] env[61806]: DEBUG oslo_vmware.api [None req-6a1c6218-d12d-4e5e-8616-7dd5071f5819 tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Task: {'id': task-1295120, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1022.832574] env[61806]: DEBUG oslo_concurrency.lockutils [None req-78fe7ba4-1e10-47a3-bab9-f40fd2226008 tempest-AttachVolumeNegativeTest-1683033704 tempest-AttachVolumeNegativeTest-1683033704-project-member] Acquiring lock "e4638faa-7be1-4909-a595-c437837fa314" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1022.832987] env[61806]: DEBUG oslo_concurrency.lockutils [None req-78fe7ba4-1e10-47a3-bab9-f40fd2226008 tempest-AttachVolumeNegativeTest-1683033704 tempest-AttachVolumeNegativeTest-1683033704-project-member] Lock "e4638faa-7be1-4909-a595-c437837fa314" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1022.833254] env[61806]: DEBUG oslo_concurrency.lockutils [None req-78fe7ba4-1e10-47a3-bab9-f40fd2226008 tempest-AttachVolumeNegativeTest-1683033704 tempest-AttachVolumeNegativeTest-1683033704-project-member] Acquiring lock "e4638faa-7be1-4909-a595-c437837fa314-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1022.833462] env[61806]: DEBUG oslo_concurrency.lockutils [None req-78fe7ba4-1e10-47a3-bab9-f40fd2226008 tempest-AttachVolumeNegativeTest-1683033704 tempest-AttachVolumeNegativeTest-1683033704-project-member] Lock "e4638faa-7be1-4909-a595-c437837fa314-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1022.833665] env[61806]: DEBUG oslo_concurrency.lockutils [None req-78fe7ba4-1e10-47a3-bab9-f40fd2226008 tempest-AttachVolumeNegativeTest-1683033704 tempest-AttachVolumeNegativeTest-1683033704-project-member] Lock "e4638faa-7be1-4909-a595-c437837fa314-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1022.838021] env[61806]: INFO nova.compute.manager [None req-78fe7ba4-1e10-47a3-bab9-f40fd2226008 tempest-AttachVolumeNegativeTest-1683033704 tempest-AttachVolumeNegativeTest-1683033704-project-member] [instance: e4638faa-7be1-4909-a595-c437837fa314] Terminating instance [ 1022.839573] env[61806]: DEBUG nova.compute.manager [None req-78fe7ba4-1e10-47a3-bab9-f40fd2226008 tempest-AttachVolumeNegativeTest-1683033704 tempest-AttachVolumeNegativeTest-1683033704-project-member] [instance: e4638faa-7be1-4909-a595-c437837fa314] Start destroying the instance on the hypervisor. {{(pid=61806) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1022.839791] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-78fe7ba4-1e10-47a3-bab9-f40fd2226008 tempest-AttachVolumeNegativeTest-1683033704 tempest-AttachVolumeNegativeTest-1683033704-project-member] [instance: e4638faa-7be1-4909-a595-c437837fa314] Destroying instance {{(pid=61806) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1022.840920] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8f961cf5-2bb8-4736-b4d7-f82fe087cd4e {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1022.852141] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [None req-78fe7ba4-1e10-47a3-bab9-f40fd2226008 tempest-AttachVolumeNegativeTest-1683033704 tempest-AttachVolumeNegativeTest-1683033704-project-member] [instance: e4638faa-7be1-4909-a595-c437837fa314] Powering off the VM {{(pid=61806) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1022.852383] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-6586fef2-64ec-45b6-9b05-af50e3c034ee {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1022.863160] env[61806]: DEBUG oslo_vmware.api [None req-78fe7ba4-1e10-47a3-bab9-f40fd2226008 tempest-AttachVolumeNegativeTest-1683033704 tempest-AttachVolumeNegativeTest-1683033704-project-member] Waiting for the task: (returnval){ [ 1022.863160] env[61806]: value = "task-1295121" [ 1022.863160] env[61806]: _type = "Task" [ 1022.863160] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1022.875876] env[61806]: DEBUG oslo_vmware.api [None req-78fe7ba4-1e10-47a3-bab9-f40fd2226008 tempest-AttachVolumeNegativeTest-1683033704 tempest-AttachVolumeNegativeTest-1683033704-project-member] Task: {'id': task-1295121, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1023.129959] env[61806]: DEBUG nova.scheduler.client.report [None req-2bebc059-5179-46cc-9f09-746fe48209df tempest-ServersNegativeTestJSON-891856253 tempest-ServersNegativeTestJSON-891856253-project-member] Inventory has not changed for provider a2858be1-fd22-4e08-979e-87ad25293407 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 138, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61806) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1023.157909] env[61806]: DEBUG oslo_vmware.api [None req-de9a979c-45ed-463a-871e-b7dda8c366a2 tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] Task: {'id': task-1295119, 'name': MoveVirtualDisk_Task, 'duration_secs': 2.457904} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1023.158200] env[61806]: INFO nova.virt.vmwareapi.ds_util [None req-de9a979c-45ed-463a-871e-b7dda8c366a2 tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] Moved virtual disk from [datastore2] OSTACK_IMG_db03ccaf-05b7-4eb9-8bd9-fe317c49b0ef/OSTACK_IMG_db03ccaf-05b7-4eb9-8bd9-fe317c49b0ef.vmdk to [datastore2] devstack-image-cache_base/b08c43d9-cfd5-4a1a-b99d-0aebb718b592/b08c43d9-cfd5-4a1a-b99d-0aebb718b592.vmdk. [ 1023.158396] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-de9a979c-45ed-463a-871e-b7dda8c366a2 tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] [instance: fd9e520a-2e7b-43d4-b84d-8933c32cc6a6] Cleaning up location [datastore2] OSTACK_IMG_db03ccaf-05b7-4eb9-8bd9-fe317c49b0ef {{(pid=61806) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:640}} [ 1023.158571] env[61806]: DEBUG nova.virt.vmwareapi.ds_util [None req-de9a979c-45ed-463a-871e-b7dda8c366a2 tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] Deleting the datastore file [datastore2] OSTACK_IMG_db03ccaf-05b7-4eb9-8bd9-fe317c49b0ef {{(pid=61806) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1023.158834] env[61806]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-3ff2436a-a30f-418c-9347-848f7438c8b8 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1023.165697] env[61806]: DEBUG oslo_vmware.api [None req-de9a979c-45ed-463a-871e-b7dda8c366a2 tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] Waiting for the task: (returnval){ [ 1023.165697] env[61806]: value = "task-1295122" [ 1023.165697] env[61806]: _type = "Task" [ 1023.165697] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1023.173450] env[61806]: DEBUG oslo_vmware.api [None req-de9a979c-45ed-463a-871e-b7dda8c366a2 tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] Task: {'id': task-1295122, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1023.200621] env[61806]: DEBUG nova.compute.manager [None req-439a5433-afde-41f1-8525-78229f3cf29f tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] [instance: a61abfd9-3cc1-4a2b-b68b-a446094daf87] Starting instance... {{(pid=61806) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1023.241254] env[61806]: DEBUG oslo_vmware.api [None req-6a1c6218-d12d-4e5e-8616-7dd5071f5819 tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Task: {'id': task-1295120, 'name': PowerOffVM_Task, 'duration_secs': 1.900047} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1023.241525] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [None req-6a1c6218-d12d-4e5e-8616-7dd5071f5819 tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] [instance: 9357e61f-1628-43bd-ab46-de13c1529f51] Powered off the VM {{(pid=61806) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1023.241702] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-6a1c6218-d12d-4e5e-8616-7dd5071f5819 tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] [instance: 9357e61f-1628-43bd-ab46-de13c1529f51] Unregistering the VM {{(pid=61806) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1023.241958] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-5d11ee54-d680-41e7-b265-d5716cd4744d {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1023.331508] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-6a1c6218-d12d-4e5e-8616-7dd5071f5819 tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] [instance: 9357e61f-1628-43bd-ab46-de13c1529f51] Unregistered the VM {{(pid=61806) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1023.331831] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-6a1c6218-d12d-4e5e-8616-7dd5071f5819 tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] [instance: 9357e61f-1628-43bd-ab46-de13c1529f51] Deleting contents of the VM from datastore datastore1 {{(pid=61806) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1023.332607] env[61806]: DEBUG nova.virt.vmwareapi.ds_util [None req-6a1c6218-d12d-4e5e-8616-7dd5071f5819 tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Deleting the datastore file [datastore1] 9357e61f-1628-43bd-ab46-de13c1529f51 {{(pid=61806) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1023.333015] env[61806]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-523e4ed4-7de2-4c55-b679-48f36e4616ac {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1023.342205] env[61806]: DEBUG oslo_vmware.api [None req-6a1c6218-d12d-4e5e-8616-7dd5071f5819 tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Waiting for the task: (returnval){ [ 1023.342205] env[61806]: value = "task-1295124" [ 1023.342205] env[61806]: _type = "Task" [ 1023.342205] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1023.350813] env[61806]: DEBUG oslo_vmware.api [None req-6a1c6218-d12d-4e5e-8616-7dd5071f5819 tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Task: {'id': task-1295124, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1023.373743] env[61806]: DEBUG oslo_vmware.api [None req-78fe7ba4-1e10-47a3-bab9-f40fd2226008 tempest-AttachVolumeNegativeTest-1683033704 tempest-AttachVolumeNegativeTest-1683033704-project-member] Task: {'id': task-1295121, 'name': PowerOffVM_Task, 'duration_secs': 0.283965} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1023.374096] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [None req-78fe7ba4-1e10-47a3-bab9-f40fd2226008 tempest-AttachVolumeNegativeTest-1683033704 tempest-AttachVolumeNegativeTest-1683033704-project-member] [instance: e4638faa-7be1-4909-a595-c437837fa314] Powered off the VM {{(pid=61806) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1023.374286] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-78fe7ba4-1e10-47a3-bab9-f40fd2226008 tempest-AttachVolumeNegativeTest-1683033704 tempest-AttachVolumeNegativeTest-1683033704-project-member] [instance: e4638faa-7be1-4909-a595-c437837fa314] Unregistering the VM {{(pid=61806) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1023.374549] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-a3c60d3e-7ca1-4190-b05c-296b26d2736e {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1023.400868] env[61806]: DEBUG nova.compute.manager [None req-d6fbd9e3-c911-4268-af0d-195628dc566f tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] [instance: 713d4e6e-6f0e-4844-9b55-cc010030eda6] Start spawning the instance on the hypervisor. {{(pid=61806) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1023.428615] env[61806]: DEBUG nova.virt.hardware [None req-d6fbd9e3-c911-4268-af0d-195628dc566f tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-15T18:05:10Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-15T18:04:52Z,direct_url=,disk_format='vmdk',id=ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='84ba83fce2c349a6988173c5d6fc3b07',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-15T18:04:53Z,virtual_size=,visibility=), allow threads: False {{(pid=61806) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1023.428911] env[61806]: DEBUG nova.virt.hardware [None req-d6fbd9e3-c911-4268-af0d-195628dc566f tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] Flavor limits 0:0:0 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1023.429094] env[61806]: DEBUG nova.virt.hardware [None req-d6fbd9e3-c911-4268-af0d-195628dc566f tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] Image limits 0:0:0 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1023.429336] env[61806]: DEBUG nova.virt.hardware [None req-d6fbd9e3-c911-4268-af0d-195628dc566f tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] Flavor pref 0:0:0 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1023.429503] env[61806]: DEBUG nova.virt.hardware [None req-d6fbd9e3-c911-4268-af0d-195628dc566f tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] Image pref 0:0:0 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1023.429662] env[61806]: DEBUG nova.virt.hardware [None req-d6fbd9e3-c911-4268-af0d-195628dc566f tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1023.429882] env[61806]: DEBUG nova.virt.hardware [None req-d6fbd9e3-c911-4268-af0d-195628dc566f tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61806) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1023.430070] env[61806]: DEBUG nova.virt.hardware [None req-d6fbd9e3-c911-4268-af0d-195628dc566f tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61806) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1023.430257] env[61806]: DEBUG nova.virt.hardware [None req-d6fbd9e3-c911-4268-af0d-195628dc566f tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] Got 1 possible topologies {{(pid=61806) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1023.430430] env[61806]: DEBUG nova.virt.hardware [None req-d6fbd9e3-c911-4268-af0d-195628dc566f tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61806) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1023.430609] env[61806]: DEBUG nova.virt.hardware [None req-d6fbd9e3-c911-4268-af0d-195628dc566f tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61806) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1023.431617] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d383d5e7-0bd5-4d22-9fb6-05a05dcc2502 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1023.441487] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e33f2af7-d235-47ef-aff2-7dac087af6f8 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1023.446817] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-78fe7ba4-1e10-47a3-bab9-f40fd2226008 tempest-AttachVolumeNegativeTest-1683033704 tempest-AttachVolumeNegativeTest-1683033704-project-member] [instance: e4638faa-7be1-4909-a595-c437837fa314] Unregistered the VM {{(pid=61806) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1023.447113] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-78fe7ba4-1e10-47a3-bab9-f40fd2226008 tempest-AttachVolumeNegativeTest-1683033704 tempest-AttachVolumeNegativeTest-1683033704-project-member] [instance: e4638faa-7be1-4909-a595-c437837fa314] Deleting contents of the VM from datastore datastore2 {{(pid=61806) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1023.447350] env[61806]: DEBUG nova.virt.vmwareapi.ds_util [None req-78fe7ba4-1e10-47a3-bab9-f40fd2226008 tempest-AttachVolumeNegativeTest-1683033704 tempest-AttachVolumeNegativeTest-1683033704-project-member] Deleting the datastore file [datastore2] e4638faa-7be1-4909-a595-c437837fa314 {{(pid=61806) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1023.448103] env[61806]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-7e0c51a1-76cb-417d-93bf-097be5004dc2 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1023.464457] env[61806]: DEBUG oslo_vmware.api [None req-78fe7ba4-1e10-47a3-bab9-f40fd2226008 tempest-AttachVolumeNegativeTest-1683033704 tempest-AttachVolumeNegativeTest-1683033704-project-member] Waiting for the task: (returnval){ [ 1023.464457] env[61806]: value = "task-1295126" [ 1023.464457] env[61806]: _type = "Task" [ 1023.464457] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1023.472816] env[61806]: DEBUG oslo_vmware.api [None req-78fe7ba4-1e10-47a3-bab9-f40fd2226008 tempest-AttachVolumeNegativeTest-1683033704 tempest-AttachVolumeNegativeTest-1683033704-project-member] Task: {'id': task-1295126, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1023.635107] env[61806]: DEBUG oslo_concurrency.lockutils [None req-2bebc059-5179-46cc-9f09-746fe48209df tempest-ServersNegativeTestJSON-891856253 tempest-ServersNegativeTestJSON-891856253-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.255s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1023.638824] env[61806]: DEBUG oslo_concurrency.lockutils [None req-1a0d0139-1a16-409a-b523-f987ba54d59c tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 2.396s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1023.638824] env[61806]: DEBUG oslo_concurrency.lockutils [None req-1a0d0139-1a16-409a-b523-f987ba54d59c tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1023.641017] env[61806]: DEBUG oslo_concurrency.lockutils [None req-d16494e9-ced1-4dfc-b5e8-4d317df5ec0e tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 1.407s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1023.641282] env[61806]: DEBUG nova.objects.instance [None req-d16494e9-ced1-4dfc-b5e8-4d317df5ec0e tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Lazy-loading 'resources' on Instance uuid 71999b04-bf6e-4e4c-87c9-d5ec1932a80f {{(pid=61806) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1023.663531] env[61806]: INFO nova.scheduler.client.report [None req-1a0d0139-1a16-409a-b523-f987ba54d59c tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Deleted allocations for instance 67d1bdd8-cf8f-4fb5-bd3c-42008e4d60e3 [ 1023.677542] env[61806]: DEBUG oslo_vmware.api [None req-de9a979c-45ed-463a-871e-b7dda8c366a2 tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] Task: {'id': task-1295122, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.046499} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1023.678408] env[61806]: DEBUG nova.virt.vmwareapi.ds_util [None req-de9a979c-45ed-463a-871e-b7dda8c366a2 tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] Deleted the datastore file {{(pid=61806) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1023.678624] env[61806]: DEBUG oslo_concurrency.lockutils [None req-de9a979c-45ed-463a-871e-b7dda8c366a2 tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] Releasing lock "[datastore2] devstack-image-cache_base/b08c43d9-cfd5-4a1a-b99d-0aebb718b592/b08c43d9-cfd5-4a1a-b99d-0aebb718b592.vmdk" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1023.678842] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [None req-de9a979c-45ed-463a-871e-b7dda8c366a2 tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/b08c43d9-cfd5-4a1a-b99d-0aebb718b592/b08c43d9-cfd5-4a1a-b99d-0aebb718b592.vmdk to [datastore2] fd9e520a-2e7b-43d4-b84d-8933c32cc6a6/fd9e520a-2e7b-43d4-b84d-8933c32cc6a6.vmdk {{(pid=61806) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 1023.679437] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-483a410d-8a8d-4538-b5f9-832d8b4a053b {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1023.687399] env[61806]: DEBUG oslo_vmware.api [None req-de9a979c-45ed-463a-871e-b7dda8c366a2 tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] Waiting for the task: (returnval){ [ 1023.687399] env[61806]: value = "task-1295127" [ 1023.687399] env[61806]: _type = "Task" [ 1023.687399] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1023.696903] env[61806]: DEBUG oslo_vmware.api [None req-de9a979c-45ed-463a-871e-b7dda8c366a2 tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] Task: {'id': task-1295127, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1023.713550] env[61806]: DEBUG nova.compute.manager [req-15e78512-8ab0-4ed9-a486-477a0e2f338b req-5bf2b658-69e4-4fe2-a7ba-3bd6fbdbc626 service nova] [instance: 713d4e6e-6f0e-4844-9b55-cc010030eda6] Received event network-vif-plugged-7828f677-3fd0-441f-ac22-2b837a4c5d54 {{(pid=61806) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1023.713696] env[61806]: DEBUG oslo_concurrency.lockutils [req-15e78512-8ab0-4ed9-a486-477a0e2f338b req-5bf2b658-69e4-4fe2-a7ba-3bd6fbdbc626 service nova] Acquiring lock "713d4e6e-6f0e-4844-9b55-cc010030eda6-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1023.713906] env[61806]: DEBUG oslo_concurrency.lockutils [req-15e78512-8ab0-4ed9-a486-477a0e2f338b req-5bf2b658-69e4-4fe2-a7ba-3bd6fbdbc626 service nova] Lock "713d4e6e-6f0e-4844-9b55-cc010030eda6-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1023.714179] env[61806]: DEBUG oslo_concurrency.lockutils [req-15e78512-8ab0-4ed9-a486-477a0e2f338b req-5bf2b658-69e4-4fe2-a7ba-3bd6fbdbc626 service nova] Lock "713d4e6e-6f0e-4844-9b55-cc010030eda6-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1023.714305] env[61806]: DEBUG nova.compute.manager [req-15e78512-8ab0-4ed9-a486-477a0e2f338b req-5bf2b658-69e4-4fe2-a7ba-3bd6fbdbc626 service nova] [instance: 713d4e6e-6f0e-4844-9b55-cc010030eda6] No waiting events found dispatching network-vif-plugged-7828f677-3fd0-441f-ac22-2b837a4c5d54 {{(pid=61806) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1023.714497] env[61806]: WARNING nova.compute.manager [req-15e78512-8ab0-4ed9-a486-477a0e2f338b req-5bf2b658-69e4-4fe2-a7ba-3bd6fbdbc626 service nova] [instance: 713d4e6e-6f0e-4844-9b55-cc010030eda6] Received unexpected event network-vif-plugged-7828f677-3fd0-441f-ac22-2b837a4c5d54 for instance with vm_state building and task_state spawning. [ 1023.721806] env[61806]: DEBUG oslo_concurrency.lockutils [None req-439a5433-afde-41f1-8525-78229f3cf29f tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1023.809279] env[61806]: DEBUG nova.network.neutron [None req-d6fbd9e3-c911-4268-af0d-195628dc566f tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] [instance: 713d4e6e-6f0e-4844-9b55-cc010030eda6] Successfully updated port: 7828f677-3fd0-441f-ac22-2b837a4c5d54 {{(pid=61806) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1023.852744] env[61806]: DEBUG oslo_vmware.api [None req-6a1c6218-d12d-4e5e-8616-7dd5071f5819 tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Task: {'id': task-1295124, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.175169} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1023.853057] env[61806]: DEBUG nova.virt.vmwareapi.ds_util [None req-6a1c6218-d12d-4e5e-8616-7dd5071f5819 tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Deleted the datastore file {{(pid=61806) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1023.853267] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-6a1c6218-d12d-4e5e-8616-7dd5071f5819 tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] [instance: 9357e61f-1628-43bd-ab46-de13c1529f51] Deleted contents of the VM from datastore datastore1 {{(pid=61806) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1023.853451] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-6a1c6218-d12d-4e5e-8616-7dd5071f5819 tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] [instance: 9357e61f-1628-43bd-ab46-de13c1529f51] Instance destroyed {{(pid=61806) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1023.853627] env[61806]: INFO nova.compute.manager [None req-6a1c6218-d12d-4e5e-8616-7dd5071f5819 tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] [instance: 9357e61f-1628-43bd-ab46-de13c1529f51] Took 2.65 seconds to destroy the instance on the hypervisor. [ 1023.853868] env[61806]: DEBUG oslo.service.loopingcall [None req-6a1c6218-d12d-4e5e-8616-7dd5071f5819 tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61806) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1023.854076] env[61806]: DEBUG nova.compute.manager [-] [instance: 9357e61f-1628-43bd-ab46-de13c1529f51] Deallocating network for instance {{(pid=61806) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1023.854172] env[61806]: DEBUG nova.network.neutron [-] [instance: 9357e61f-1628-43bd-ab46-de13c1529f51] deallocate_for_instance() {{(pid=61806) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1023.975591] env[61806]: DEBUG oslo_vmware.api [None req-78fe7ba4-1e10-47a3-bab9-f40fd2226008 tempest-AttachVolumeNegativeTest-1683033704 tempest-AttachVolumeNegativeTest-1683033704-project-member] Task: {'id': task-1295126, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1024.152226] env[61806]: DEBUG oslo_concurrency.lockutils [None req-2bebc059-5179-46cc-9f09-746fe48209df tempest-ServersNegativeTestJSON-891856253 tempest-ServersNegativeTestJSON-891856253-project-member] Lock "bdbaef31-7d37-40cb-ac3e-ad152b37e2d2" "released" by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" :: held 26.131s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1024.153646] env[61806]: DEBUG oslo_concurrency.lockutils [None req-a5c61942-d486-43f3-834d-7dfb87ef5b63 tempest-ServersNegativeTestJSON-891856253 tempest-ServersNegativeTestJSON-891856253-project-member] Lock "bdbaef31-7d37-40cb-ac3e-ad152b37e2d2" acquired by "nova.compute.manager.ComputeManager.unshelve_instance..do_unshelve_instance" :: waited 2.068s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1024.153978] env[61806]: INFO nova.compute.manager [None req-a5c61942-d486-43f3-834d-7dfb87ef5b63 tempest-ServersNegativeTestJSON-891856253 tempest-ServersNegativeTestJSON-891856253-project-member] [instance: bdbaef31-7d37-40cb-ac3e-ad152b37e2d2] Unshelving [ 1024.174071] env[61806]: DEBUG oslo_concurrency.lockutils [None req-1a0d0139-1a16-409a-b523-f987ba54d59c tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Lock "67d1bdd8-cf8f-4fb5-bd3c-42008e4d60e3" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 7.149s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1024.200901] env[61806]: DEBUG oslo_vmware.api [None req-de9a979c-45ed-463a-871e-b7dda8c366a2 tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] Task: {'id': task-1295127, 'name': CopyVirtualDisk_Task} progress is 9%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1024.275466] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-447f54c3-7eff-4dd9-8c8b-b91d9aad55d7 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1024.283906] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-26318740-08f7-4120-aeda-0f2a98e20e69 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1024.314196] env[61806]: DEBUG oslo_concurrency.lockutils [None req-d6fbd9e3-c911-4268-af0d-195628dc566f tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] Acquiring lock "refresh_cache-713d4e6e-6f0e-4844-9b55-cc010030eda6" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1024.314360] env[61806]: DEBUG oslo_concurrency.lockutils [None req-d6fbd9e3-c911-4268-af0d-195628dc566f tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] Acquired lock "refresh_cache-713d4e6e-6f0e-4844-9b55-cc010030eda6" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1024.314513] env[61806]: DEBUG nova.network.neutron [None req-d6fbd9e3-c911-4268-af0d-195628dc566f tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] [instance: 713d4e6e-6f0e-4844-9b55-cc010030eda6] Building network info cache for instance {{(pid=61806) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1024.316543] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5c53f631-ad5d-444c-a401-8b45448127cf {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1024.326482] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d7d7c6f1-0128-46e4-aa54-db7d3fd731a8 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1024.345356] env[61806]: DEBUG nova.compute.provider_tree [None req-d16494e9-ced1-4dfc-b5e8-4d317df5ec0e tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Inventory has not changed in ProviderTree for provider: a2858be1-fd22-4e08-979e-87ad25293407 {{(pid=61806) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1024.477195] env[61806]: DEBUG oslo_vmware.api [None req-78fe7ba4-1e10-47a3-bab9-f40fd2226008 tempest-AttachVolumeNegativeTest-1683033704 tempest-AttachVolumeNegativeTest-1683033704-project-member] Task: {'id': task-1295126, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.52235} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1024.477563] env[61806]: DEBUG nova.virt.vmwareapi.ds_util [None req-78fe7ba4-1e10-47a3-bab9-f40fd2226008 tempest-AttachVolumeNegativeTest-1683033704 tempest-AttachVolumeNegativeTest-1683033704-project-member] Deleted the datastore file {{(pid=61806) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1024.477702] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-78fe7ba4-1e10-47a3-bab9-f40fd2226008 tempest-AttachVolumeNegativeTest-1683033704 tempest-AttachVolumeNegativeTest-1683033704-project-member] [instance: e4638faa-7be1-4909-a595-c437837fa314] Deleted contents of the VM from datastore datastore2 {{(pid=61806) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1024.477951] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-78fe7ba4-1e10-47a3-bab9-f40fd2226008 tempest-AttachVolumeNegativeTest-1683033704 tempest-AttachVolumeNegativeTest-1683033704-project-member] [instance: e4638faa-7be1-4909-a595-c437837fa314] Instance destroyed {{(pid=61806) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1024.478453] env[61806]: INFO nova.compute.manager [None req-78fe7ba4-1e10-47a3-bab9-f40fd2226008 tempest-AttachVolumeNegativeTest-1683033704 tempest-AttachVolumeNegativeTest-1683033704-project-member] [instance: e4638faa-7be1-4909-a595-c437837fa314] Took 1.64 seconds to destroy the instance on the hypervisor. [ 1024.478753] env[61806]: DEBUG oslo.service.loopingcall [None req-78fe7ba4-1e10-47a3-bab9-f40fd2226008 tempest-AttachVolumeNegativeTest-1683033704 tempest-AttachVolumeNegativeTest-1683033704-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61806) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1024.478991] env[61806]: DEBUG nova.compute.manager [-] [instance: e4638faa-7be1-4909-a595-c437837fa314] Deallocating network for instance {{(pid=61806) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1024.479116] env[61806]: DEBUG nova.network.neutron [-] [instance: e4638faa-7be1-4909-a595-c437837fa314] deallocate_for_instance() {{(pid=61806) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1024.605107] env[61806]: DEBUG nova.network.neutron [-] [instance: 9357e61f-1628-43bd-ab46-de13c1529f51] Updating instance_info_cache with network_info: [] {{(pid=61806) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1024.703088] env[61806]: DEBUG oslo_vmware.api [None req-de9a979c-45ed-463a-871e-b7dda8c366a2 tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] Task: {'id': task-1295127, 'name': CopyVirtualDisk_Task} progress is 26%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1024.852129] env[61806]: DEBUG nova.scheduler.client.report [None req-d16494e9-ced1-4dfc-b5e8-4d317df5ec0e tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Inventory has not changed for provider a2858be1-fd22-4e08-979e-87ad25293407 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 138, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61806) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1024.864537] env[61806]: DEBUG nova.network.neutron [None req-d6fbd9e3-c911-4268-af0d-195628dc566f tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] [instance: 713d4e6e-6f0e-4844-9b55-cc010030eda6] Instance cache missing network info. {{(pid=61806) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 1025.105931] env[61806]: DEBUG nova.network.neutron [None req-d6fbd9e3-c911-4268-af0d-195628dc566f tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] [instance: 713d4e6e-6f0e-4844-9b55-cc010030eda6] Updating instance_info_cache with network_info: [{"id": "7828f677-3fd0-441f-ac22-2b837a4c5d54", "address": "fa:16:3e:fd:d7:4e", "network": {"id": "f280a01b-7000-4e54-ab9a-20eab156a410", "bridge": "br-int", "label": "tempest-ServersTestJSON-2080718146-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "73ebcc4d79d248efb653a084a5e44302", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "43798f54-0c07-4417-a23f-58bb6b7e204b", "external-id": "nsx-vlan-transportzone-571", "segmentation_id": 571, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7828f677-3f", "ovs_interfaceid": "7828f677-3fd0-441f-ac22-2b837a4c5d54", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61806) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1025.108888] env[61806]: INFO nova.compute.manager [-] [instance: 9357e61f-1628-43bd-ab46-de13c1529f51] Took 1.25 seconds to deallocate network for instance. [ 1025.184441] env[61806]: DEBUG oslo_concurrency.lockutils [None req-a5c61942-d486-43f3-834d-7dfb87ef5b63 tempest-ServersNegativeTestJSON-891856253 tempest-ServersNegativeTestJSON-891856253-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1025.203326] env[61806]: DEBUG oslo_vmware.api [None req-de9a979c-45ed-463a-871e-b7dda8c366a2 tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] Task: {'id': task-1295127, 'name': CopyVirtualDisk_Task} progress is 46%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1025.356169] env[61806]: DEBUG oslo_concurrency.lockutils [None req-d16494e9-ced1-4dfc-b5e8-4d317df5ec0e tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.715s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1025.358497] env[61806]: DEBUG oslo_concurrency.lockutils [None req-439a5433-afde-41f1-8525-78229f3cf29f tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 1.637s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1025.360073] env[61806]: INFO nova.compute.claims [None req-439a5433-afde-41f1-8525-78229f3cf29f tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] [instance: a61abfd9-3cc1-4a2b-b68b-a446094daf87] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1025.611689] env[61806]: DEBUG oslo_concurrency.lockutils [None req-d6fbd9e3-c911-4268-af0d-195628dc566f tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] Releasing lock "refresh_cache-713d4e6e-6f0e-4844-9b55-cc010030eda6" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1025.612106] env[61806]: DEBUG nova.compute.manager [None req-d6fbd9e3-c911-4268-af0d-195628dc566f tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] [instance: 713d4e6e-6f0e-4844-9b55-cc010030eda6] Instance network_info: |[{"id": "7828f677-3fd0-441f-ac22-2b837a4c5d54", "address": "fa:16:3e:fd:d7:4e", "network": {"id": "f280a01b-7000-4e54-ab9a-20eab156a410", "bridge": "br-int", "label": "tempest-ServersTestJSON-2080718146-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "73ebcc4d79d248efb653a084a5e44302", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "43798f54-0c07-4417-a23f-58bb6b7e204b", "external-id": "nsx-vlan-transportzone-571", "segmentation_id": 571, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7828f677-3f", "ovs_interfaceid": "7828f677-3fd0-441f-ac22-2b837a4c5d54", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61806) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1025.612451] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-d6fbd9e3-c911-4268-af0d-195628dc566f tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] [instance: 713d4e6e-6f0e-4844-9b55-cc010030eda6] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:fd:d7:4e', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '43798f54-0c07-4417-a23f-58bb6b7e204b', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '7828f677-3fd0-441f-ac22-2b837a4c5d54', 'vif_model': 'vmxnet3'}] {{(pid=61806) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1025.620228] env[61806]: DEBUG oslo.service.loopingcall [None req-d6fbd9e3-c911-4268-af0d-195628dc566f tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61806) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1025.621085] env[61806]: DEBUG oslo_concurrency.lockutils [None req-6a1c6218-d12d-4e5e-8616-7dd5071f5819 tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1025.621344] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 713d4e6e-6f0e-4844-9b55-cc010030eda6] Creating VM on the ESX host {{(pid=61806) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 1025.621578] env[61806]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-ba0756be-2662-4fd0-95d7-630293a54d38 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1025.642776] env[61806]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1025.642776] env[61806]: value = "task-1295128" [ 1025.642776] env[61806]: _type = "Task" [ 1025.642776] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1025.651164] env[61806]: DEBUG oslo_vmware.api [-] Task: {'id': task-1295128, 'name': CreateVM_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1025.698803] env[61806]: DEBUG nova.network.neutron [-] [instance: e4638faa-7be1-4909-a595-c437837fa314] Updating instance_info_cache with network_info: [] {{(pid=61806) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1025.700759] env[61806]: DEBUG oslo_vmware.api [None req-de9a979c-45ed-463a-871e-b7dda8c366a2 tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] Task: {'id': task-1295127, 'name': CopyVirtualDisk_Task} progress is 54%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1025.746242] env[61806]: DEBUG nova.compute.manager [req-df43e91e-d58d-43db-a96c-4b959bd70b6c req-66306544-147a-4a79-910d-e579a20a9c04 service nova] [instance: 713d4e6e-6f0e-4844-9b55-cc010030eda6] Received event network-changed-7828f677-3fd0-441f-ac22-2b837a4c5d54 {{(pid=61806) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1025.746456] env[61806]: DEBUG nova.compute.manager [req-df43e91e-d58d-43db-a96c-4b959bd70b6c req-66306544-147a-4a79-910d-e579a20a9c04 service nova] [instance: 713d4e6e-6f0e-4844-9b55-cc010030eda6] Refreshing instance network info cache due to event network-changed-7828f677-3fd0-441f-ac22-2b837a4c5d54. {{(pid=61806) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 1025.746675] env[61806]: DEBUG oslo_concurrency.lockutils [req-df43e91e-d58d-43db-a96c-4b959bd70b6c req-66306544-147a-4a79-910d-e579a20a9c04 service nova] Acquiring lock "refresh_cache-713d4e6e-6f0e-4844-9b55-cc010030eda6" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1025.746828] env[61806]: DEBUG oslo_concurrency.lockutils [req-df43e91e-d58d-43db-a96c-4b959bd70b6c req-66306544-147a-4a79-910d-e579a20a9c04 service nova] Acquired lock "refresh_cache-713d4e6e-6f0e-4844-9b55-cc010030eda6" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1025.746997] env[61806]: DEBUG nova.network.neutron [req-df43e91e-d58d-43db-a96c-4b959bd70b6c req-66306544-147a-4a79-910d-e579a20a9c04 service nova] [instance: 713d4e6e-6f0e-4844-9b55-cc010030eda6] Refreshing network info cache for port 7828f677-3fd0-441f-ac22-2b837a4c5d54 {{(pid=61806) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1026.155059] env[61806]: DEBUG oslo_vmware.api [-] Task: {'id': task-1295128, 'name': CreateVM_Task, 'duration_secs': 0.429685} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1026.155237] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 713d4e6e-6f0e-4844-9b55-cc010030eda6] Created VM on the ESX host {{(pid=61806) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 1026.155918] env[61806]: DEBUG oslo_concurrency.lockutils [None req-d6fbd9e3-c911-4268-af0d-195628dc566f tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1026.156114] env[61806]: DEBUG oslo_concurrency.lockutils [None req-d6fbd9e3-c911-4268-af0d-195628dc566f tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] Acquired lock "[datastore1] devstack-image-cache_base/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1026.156448] env[61806]: DEBUG oslo_concurrency.lockutils [None req-d6fbd9e3-c911-4268-af0d-195628dc566f tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1026.156704] env[61806]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-6d6f761d-dbd2-40f3-858d-36c10d231043 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1026.161192] env[61806]: DEBUG oslo_vmware.api [None req-d6fbd9e3-c911-4268-af0d-195628dc566f tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] Waiting for the task: (returnval){ [ 1026.161192] env[61806]: value = "session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]5223e18b-857c-7f73-f624-eea961696953" [ 1026.161192] env[61806]: _type = "Task" [ 1026.161192] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1026.168510] env[61806]: DEBUG oslo_vmware.api [None req-d6fbd9e3-c911-4268-af0d-195628dc566f tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] Task: {'id': session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]5223e18b-857c-7f73-f624-eea961696953, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1026.197240] env[61806]: DEBUG oslo_vmware.api [None req-de9a979c-45ed-463a-871e-b7dda8c366a2 tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] Task: {'id': task-1295127, 'name': CopyVirtualDisk_Task} progress is 54%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1026.432875] env[61806]: DEBUG nova.network.neutron [req-df43e91e-d58d-43db-a96c-4b959bd70b6c req-66306544-147a-4a79-910d-e579a20a9c04 service nova] [instance: 713d4e6e-6f0e-4844-9b55-cc010030eda6] Updated VIF entry in instance network info cache for port 7828f677-3fd0-441f-ac22-2b837a4c5d54. {{(pid=61806) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1026.433329] env[61806]: DEBUG nova.network.neutron [req-df43e91e-d58d-43db-a96c-4b959bd70b6c req-66306544-147a-4a79-910d-e579a20a9c04 service nova] [instance: 713d4e6e-6f0e-4844-9b55-cc010030eda6] Updating instance_info_cache with network_info: [{"id": "7828f677-3fd0-441f-ac22-2b837a4c5d54", "address": "fa:16:3e:fd:d7:4e", "network": {"id": "f280a01b-7000-4e54-ab9a-20eab156a410", "bridge": "br-int", "label": "tempest-ServersTestJSON-2080718146-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "73ebcc4d79d248efb653a084a5e44302", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "43798f54-0c07-4417-a23f-58bb6b7e204b", "external-id": "nsx-vlan-transportzone-571", "segmentation_id": 571, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7828f677-3f", "ovs_interfaceid": "7828f677-3fd0-441f-ac22-2b837a4c5d54", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61806) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1026.673033] env[61806]: DEBUG oslo_vmware.api [None req-d6fbd9e3-c911-4268-af0d-195628dc566f tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] Task: {'id': session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]5223e18b-857c-7f73-f624-eea961696953, 'name': SearchDatastore_Task, 'duration_secs': 0.011942} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1026.673033] env[61806]: DEBUG oslo_concurrency.lockutils [None req-d6fbd9e3-c911-4268-af0d-195628dc566f tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] Releasing lock "[datastore1] devstack-image-cache_base/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1026.673033] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-d6fbd9e3-c911-4268-af0d-195628dc566f tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] [instance: 713d4e6e-6f0e-4844-9b55-cc010030eda6] Processing image ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa {{(pid=61806) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1026.673432] env[61806]: DEBUG oslo_concurrency.lockutils [None req-d6fbd9e3-c911-4268-af0d-195628dc566f tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa.vmdk" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1026.673432] env[61806]: DEBUG oslo_concurrency.lockutils [None req-d6fbd9e3-c911-4268-af0d-195628dc566f tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] Acquired lock "[datastore1] devstack-image-cache_base/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa.vmdk" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1026.673554] env[61806]: DEBUG nova.virt.vmwareapi.ds_util [None req-d6fbd9e3-c911-4268-af0d-195628dc566f tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61806) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1026.673821] env[61806]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-dd648a97-06c6-4404-a6aa-74348b11d7c9 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1026.698826] env[61806]: DEBUG oslo_vmware.api [None req-de9a979c-45ed-463a-871e-b7dda8c366a2 tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] Task: {'id': task-1295127, 'name': CopyVirtualDisk_Task} progress is 54%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1026.936361] env[61806]: DEBUG oslo_concurrency.lockutils [req-df43e91e-d58d-43db-a96c-4b959bd70b6c req-66306544-147a-4a79-910d-e579a20a9c04 service nova] Releasing lock "refresh_cache-713d4e6e-6f0e-4844-9b55-cc010030eda6" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1026.936615] env[61806]: DEBUG nova.compute.manager [req-df43e91e-d58d-43db-a96c-4b959bd70b6c req-66306544-147a-4a79-910d-e579a20a9c04 service nova] [instance: 9357e61f-1628-43bd-ab46-de13c1529f51] Received event network-vif-deleted-a4ca7c56-b6a1-4bd2-bd47-02b0ed4aa097 {{(pid=61806) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1026.936816] env[61806]: DEBUG nova.compute.manager [req-df43e91e-d58d-43db-a96c-4b959bd70b6c req-66306544-147a-4a79-910d-e579a20a9c04 service nova] [instance: e4638faa-7be1-4909-a595-c437837fa314] Received event network-vif-deleted-bfb37474-8e7a-4820-9045-e1b388cfb87d {{(pid=61806) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1027.200544] env[61806]: DEBUG oslo_vmware.api [None req-de9a979c-45ed-463a-871e-b7dda8c366a2 tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] Task: {'id': task-1295127, 'name': CopyVirtualDisk_Task} progress is 54%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1027.576844] env[61806]: INFO nova.scheduler.client.report [None req-d16494e9-ced1-4dfc-b5e8-4d317df5ec0e tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Deleted allocations for instance 71999b04-bf6e-4e4c-87c9-d5ec1932a80f [ 1027.651847] env[61806]: DEBUG oslo_concurrency.lockutils [None req-138f28bc-d88a-4f34-babf-797a70b177f1 tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Acquiring lock "06658134-4c6d-4911-9a55-b6805f555c78" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1027.652125] env[61806]: DEBUG oslo_concurrency.lockutils [None req-138f28bc-d88a-4f34-babf-797a70b177f1 tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Lock "06658134-4c6d-4911-9a55-b6805f555c78" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1027.702326] env[61806]: DEBUG oslo_vmware.api [None req-de9a979c-45ed-463a-871e-b7dda8c366a2 tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] Task: {'id': task-1295127, 'name': CopyVirtualDisk_Task} progress is 63%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1027.703915] env[61806]: INFO nova.compute.manager [-] [instance: e4638faa-7be1-4909-a595-c437837fa314] Took 3.22 seconds to deallocate network for instance. [ 1028.088460] env[61806]: DEBUG oslo_concurrency.lockutils [None req-d16494e9-ced1-4dfc-b5e8-4d317df5ec0e tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Lock "71999b04-bf6e-4e4c-87c9-d5ec1932a80f" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 8.449s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1028.154981] env[61806]: DEBUG nova.compute.manager [None req-138f28bc-d88a-4f34-babf-797a70b177f1 tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] [instance: 06658134-4c6d-4911-9a55-b6805f555c78] Starting instance... {{(pid=61806) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1028.202832] env[61806]: DEBUG oslo_vmware.api [None req-de9a979c-45ed-463a-871e-b7dda8c366a2 tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] Task: {'id': task-1295127, 'name': CopyVirtualDisk_Task} progress is 83%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1028.210287] env[61806]: DEBUG oslo_concurrency.lockutils [None req-78fe7ba4-1e10-47a3-bab9-f40fd2226008 tempest-AttachVolumeNegativeTest-1683033704 tempest-AttachVolumeNegativeTest-1683033704-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1028.512527] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8165142c-743a-43e9-89e4-18b3762faf8d {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1028.522505] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3632e99d-7e57-4c58-bb8d-ae8427f703a4 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1028.557557] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-411b2d3f-df95-4db9-977e-58ef98f2ec24 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1028.567011] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fff69157-11e6-4d1e-8b06-bb3e7ba48799 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1028.582358] env[61806]: DEBUG nova.compute.provider_tree [None req-439a5433-afde-41f1-8525-78229f3cf29f tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Inventory has not changed in ProviderTree for provider: a2858be1-fd22-4e08-979e-87ad25293407 {{(pid=61806) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1028.675453] env[61806]: DEBUG oslo_concurrency.lockutils [None req-138f28bc-d88a-4f34-babf-797a70b177f1 tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1028.703306] env[61806]: DEBUG oslo_vmware.api [None req-de9a979c-45ed-463a-871e-b7dda8c366a2 tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] Task: {'id': task-1295127, 'name': CopyVirtualDisk_Task} progress is 100%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1028.916757] env[61806]: DEBUG nova.virt.vmwareapi.ds_util [None req-d6fbd9e3-c911-4268-af0d-195628dc566f tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61806) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1028.916958] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-d6fbd9e3-c911-4268-af0d-195628dc566f tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61806) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 1028.917753] env[61806]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-583289af-5f99-4654-a1bd-df384a4040eb {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1028.923975] env[61806]: DEBUG oslo_vmware.api [None req-d6fbd9e3-c911-4268-af0d-195628dc566f tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] Waiting for the task: (returnval){ [ 1028.923975] env[61806]: value = "session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]5244c7a0-139d-f589-7ef7-f2f22d6cfb5d" [ 1028.923975] env[61806]: _type = "Task" [ 1028.923975] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1028.932467] env[61806]: DEBUG oslo_vmware.api [None req-d6fbd9e3-c911-4268-af0d-195628dc566f tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] Task: {'id': session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]5244c7a0-139d-f589-7ef7-f2f22d6cfb5d, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1028.970522] env[61806]: DEBUG oslo_concurrency.lockutils [None req-54021779-f9e2-42db-b76c-c250eacb9121 tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Acquiring lock "7445d2c5-26bd-4f8d-8653-51c721ea801f" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1028.970810] env[61806]: DEBUG oslo_concurrency.lockutils [None req-54021779-f9e2-42db-b76c-c250eacb9121 tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Lock "7445d2c5-26bd-4f8d-8653-51c721ea801f" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1029.086624] env[61806]: DEBUG nova.scheduler.client.report [None req-439a5433-afde-41f1-8525-78229f3cf29f tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Inventory has not changed for provider a2858be1-fd22-4e08-979e-87ad25293407 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 138, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61806) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1029.203674] env[61806]: DEBUG oslo_vmware.api [None req-de9a979c-45ed-463a-871e-b7dda8c366a2 tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] Task: {'id': task-1295127, 'name': CopyVirtualDisk_Task, 'duration_secs': 5.055052} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1029.203960] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [None req-de9a979c-45ed-463a-871e-b7dda8c366a2 tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/b08c43d9-cfd5-4a1a-b99d-0aebb718b592/b08c43d9-cfd5-4a1a-b99d-0aebb718b592.vmdk to [datastore2] fd9e520a-2e7b-43d4-b84d-8933c32cc6a6/fd9e520a-2e7b-43d4-b84d-8933c32cc6a6.vmdk {{(pid=61806) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 1029.204754] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c65a7ce9-42e3-4f0e-ab5d-1e9b15e717db {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1029.226233] env[61806]: DEBUG nova.virt.vmwareapi.volumeops [None req-de9a979c-45ed-463a-871e-b7dda8c366a2 tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] [instance: fd9e520a-2e7b-43d4-b84d-8933c32cc6a6] Reconfiguring VM instance instance-00000057 to attach disk [datastore2] fd9e520a-2e7b-43d4-b84d-8933c32cc6a6/fd9e520a-2e7b-43d4-b84d-8933c32cc6a6.vmdk or device None with type streamOptimized {{(pid=61806) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1029.226504] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-3b3cc958-eb9b-4482-a36b-c57971052711 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1029.247066] env[61806]: DEBUG oslo_vmware.api [None req-de9a979c-45ed-463a-871e-b7dda8c366a2 tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] Waiting for the task: (returnval){ [ 1029.247066] env[61806]: value = "task-1295129" [ 1029.247066] env[61806]: _type = "Task" [ 1029.247066] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1029.255490] env[61806]: DEBUG oslo_vmware.api [None req-de9a979c-45ed-463a-871e-b7dda8c366a2 tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] Task: {'id': task-1295129, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1029.435587] env[61806]: DEBUG oslo_vmware.api [None req-d6fbd9e3-c911-4268-af0d-195628dc566f tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] Task: {'id': session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]5244c7a0-139d-f589-7ef7-f2f22d6cfb5d, 'name': SearchDatastore_Task, 'duration_secs': 0.150906} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1029.436582] env[61806]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d7385c75-ff7a-4b2c-ba2d-dbe8b8511c94 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1029.442214] env[61806]: DEBUG oslo_vmware.api [None req-d6fbd9e3-c911-4268-af0d-195628dc566f tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] Waiting for the task: (returnval){ [ 1029.442214] env[61806]: value = "session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]529032c7-28e9-189e-1fa8-a7729a34d0a7" [ 1029.442214] env[61806]: _type = "Task" [ 1029.442214] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1029.450263] env[61806]: DEBUG oslo_vmware.api [None req-d6fbd9e3-c911-4268-af0d-195628dc566f tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] Task: {'id': session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]529032c7-28e9-189e-1fa8-a7729a34d0a7, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1029.472827] env[61806]: DEBUG nova.compute.manager [None req-54021779-f9e2-42db-b76c-c250eacb9121 tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] [instance: 7445d2c5-26bd-4f8d-8653-51c721ea801f] Starting instance... {{(pid=61806) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1029.591767] env[61806]: DEBUG oslo_concurrency.lockutils [None req-439a5433-afde-41f1-8525-78229f3cf29f tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 4.233s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1029.592367] env[61806]: DEBUG nova.compute.manager [None req-439a5433-afde-41f1-8525-78229f3cf29f tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] [instance: a61abfd9-3cc1-4a2b-b68b-a446094daf87] Start building networks asynchronously for instance. {{(pid=61806) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1029.595378] env[61806]: DEBUG oslo_concurrency.lockutils [None req-a5c61942-d486-43f3-834d-7dfb87ef5b63 tempest-ServersNegativeTestJSON-891856253 tempest-ServersNegativeTestJSON-891856253-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 4.411s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1029.595599] env[61806]: DEBUG nova.objects.instance [None req-a5c61942-d486-43f3-834d-7dfb87ef5b63 tempest-ServersNegativeTestJSON-891856253 tempest-ServersNegativeTestJSON-891856253-project-member] Lazy-loading 'pci_requests' on Instance uuid bdbaef31-7d37-40cb-ac3e-ad152b37e2d2 {{(pid=61806) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1029.757309] env[61806]: DEBUG oslo_vmware.api [None req-de9a979c-45ed-463a-871e-b7dda8c366a2 tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] Task: {'id': task-1295129, 'name': ReconfigVM_Task, 'duration_secs': 0.339277} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1029.757653] env[61806]: DEBUG nova.virt.vmwareapi.volumeops [None req-de9a979c-45ed-463a-871e-b7dda8c366a2 tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] [instance: fd9e520a-2e7b-43d4-b84d-8933c32cc6a6] Reconfigured VM instance instance-00000057 to attach disk [datastore2] fd9e520a-2e7b-43d4-b84d-8933c32cc6a6/fd9e520a-2e7b-43d4-b84d-8933c32cc6a6.vmdk or device None with type streamOptimized {{(pid=61806) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1029.759017] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-de9a979c-45ed-463a-871e-b7dda8c366a2 tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] [instance: fd9e520a-2e7b-43d4-b84d-8933c32cc6a6] Block device information present: {'root_device_name': '/dev/sda', 'image': [{'encryption_format': None, 'device_type': 'disk', 'guest_format': None, 'encryption_options': None, 'device_name': '/dev/sda', 'encryption_secret_uuid': None, 'size': 0, 'boot_index': 0, 'encrypted': False, 'disk_bus': None, 'image_id': 'ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa'}], 'ephemerals': [], 'block_device_mapping': [{'delete_on_termination': False, 'device_type': None, 'attachment_id': '60011335-d621-48ac-bdd9-7da611ca0e30', 'connection_info': {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-277740', 'volume_id': 'dd35bef0-37b9-4821-9a9d-7f5c7789746c', 'name': 'volume-dd35bef0-37b9-4821-9a9d-7f5c7789746c', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'attached', 'instance': 'fd9e520a-2e7b-43d4-b84d-8933c32cc6a6', 'attached_at': '', 'detached_at': '', 'volume_id': 'dd35bef0-37b9-4821-9a9d-7f5c7789746c', 'serial': 'dd35bef0-37b9-4821-9a9d-7f5c7789746c'}, 'guest_format': None, 'mount_device': '/dev/sdb', 'boot_index': None, 'disk_bus': None, 'volume_type': None}], 'swap': None} {{(pid=61806) spawn /opt/stack/nova/nova/virt/vmwareapi/vmops.py:799}} [ 1029.759242] env[61806]: DEBUG nova.virt.vmwareapi.volumeops [None req-de9a979c-45ed-463a-871e-b7dda8c366a2 tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] [instance: fd9e520a-2e7b-43d4-b84d-8933c32cc6a6] Volume attach. Driver type: vmdk {{(pid=61806) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 1029.759441] env[61806]: DEBUG nova.virt.vmwareapi.volumeops [None req-de9a979c-45ed-463a-871e-b7dda8c366a2 tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] [instance: fd9e520a-2e7b-43d4-b84d-8933c32cc6a6] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-277740', 'volume_id': 'dd35bef0-37b9-4821-9a9d-7f5c7789746c', 'name': 'volume-dd35bef0-37b9-4821-9a9d-7f5c7789746c', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'attached', 'instance': 'fd9e520a-2e7b-43d4-b84d-8933c32cc6a6', 'attached_at': '', 'detached_at': '', 'volume_id': 'dd35bef0-37b9-4821-9a9d-7f5c7789746c', 'serial': 'dd35bef0-37b9-4821-9a9d-7f5c7789746c'} {{(pid=61806) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 1029.760255] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-852de546-c049-4230-b34c-bca5a11cf7ff {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1029.777450] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-462ade60-8463-4d1f-9c78-c5143f7913b1 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1029.805125] env[61806]: DEBUG nova.virt.vmwareapi.volumeops [None req-de9a979c-45ed-463a-871e-b7dda8c366a2 tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] [instance: fd9e520a-2e7b-43d4-b84d-8933c32cc6a6] Reconfiguring VM instance instance-00000057 to attach disk [datastore2] volume-dd35bef0-37b9-4821-9a9d-7f5c7789746c/volume-dd35bef0-37b9-4821-9a9d-7f5c7789746c.vmdk or device None with type thin {{(pid=61806) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1029.805487] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-74ca874d-dbaf-493d-ac28-859db4734a57 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1029.824566] env[61806]: DEBUG oslo_vmware.api [None req-de9a979c-45ed-463a-871e-b7dda8c366a2 tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] Waiting for the task: (returnval){ [ 1029.824566] env[61806]: value = "task-1295130" [ 1029.824566] env[61806]: _type = "Task" [ 1029.824566] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1029.833747] env[61806]: DEBUG oslo_vmware.api [None req-de9a979c-45ed-463a-871e-b7dda8c366a2 tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] Task: {'id': task-1295130, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1029.953926] env[61806]: DEBUG oslo_vmware.api [None req-d6fbd9e3-c911-4268-af0d-195628dc566f tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] Task: {'id': session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]529032c7-28e9-189e-1fa8-a7729a34d0a7, 'name': SearchDatastore_Task, 'duration_secs': 0.010329} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1029.954310] env[61806]: DEBUG oslo_concurrency.lockutils [None req-d6fbd9e3-c911-4268-af0d-195628dc566f tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] Releasing lock "[datastore1] devstack-image-cache_base/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa.vmdk" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1029.954598] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [None req-d6fbd9e3-c911-4268-af0d-195628dc566f tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa.vmdk to [datastore1] 713d4e6e-6f0e-4844-9b55-cc010030eda6/713d4e6e-6f0e-4844-9b55-cc010030eda6.vmdk {{(pid=61806) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 1029.954886] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-184e62ad-0b20-4a43-bfd1-39e27e0c2e87 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1029.963876] env[61806]: DEBUG oslo_vmware.api [None req-d6fbd9e3-c911-4268-af0d-195628dc566f tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] Waiting for the task: (returnval){ [ 1029.963876] env[61806]: value = "task-1295131" [ 1029.963876] env[61806]: _type = "Task" [ 1029.963876] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1029.973021] env[61806]: DEBUG oslo_vmware.api [None req-d6fbd9e3-c911-4268-af0d-195628dc566f tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] Task: {'id': task-1295131, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1030.003872] env[61806]: DEBUG oslo_concurrency.lockutils [None req-54021779-f9e2-42db-b76c-c250eacb9121 tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1030.099303] env[61806]: DEBUG nova.compute.utils [None req-439a5433-afde-41f1-8525-78229f3cf29f tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Using /dev/sd instead of None {{(pid=61806) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1030.102441] env[61806]: DEBUG nova.objects.instance [None req-a5c61942-d486-43f3-834d-7dfb87ef5b63 tempest-ServersNegativeTestJSON-891856253 tempest-ServersNegativeTestJSON-891856253-project-member] Lazy-loading 'numa_topology' on Instance uuid bdbaef31-7d37-40cb-ac3e-ad152b37e2d2 {{(pid=61806) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1030.103806] env[61806]: DEBUG nova.compute.manager [None req-439a5433-afde-41f1-8525-78229f3cf29f tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] [instance: a61abfd9-3cc1-4a2b-b68b-a446094daf87] Allocating IP information in the background. {{(pid=61806) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1030.103979] env[61806]: DEBUG nova.network.neutron [None req-439a5433-afde-41f1-8525-78229f3cf29f tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] [instance: a61abfd9-3cc1-4a2b-b68b-a446094daf87] allocate_for_instance() {{(pid=61806) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 1030.147021] env[61806]: DEBUG nova.policy [None req-439a5433-afde-41f1-8525-78229f3cf29f tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'b2631ce1f6dd40a4bfc5da7f90cde260', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '4b9d94c3d684474b99e305fb6cd9b95f', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61806) authorize /opt/stack/nova/nova/policy.py:201}} [ 1030.335838] env[61806]: DEBUG oslo_vmware.api [None req-de9a979c-45ed-463a-871e-b7dda8c366a2 tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] Task: {'id': task-1295130, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1030.403849] env[61806]: DEBUG nova.network.neutron [None req-439a5433-afde-41f1-8525-78229f3cf29f tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] [instance: a61abfd9-3cc1-4a2b-b68b-a446094daf87] Successfully created port: 2950cc6d-7a09-4217-abf5-8a5a940e6653 {{(pid=61806) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1030.479343] env[61806]: DEBUG oslo_vmware.api [None req-d6fbd9e3-c911-4268-af0d-195628dc566f tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] Task: {'id': task-1295131, 'name': CopyVirtualDisk_Task} progress is 4%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1030.608029] env[61806]: DEBUG nova.compute.manager [None req-439a5433-afde-41f1-8525-78229f3cf29f tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] [instance: a61abfd9-3cc1-4a2b-b68b-a446094daf87] Start building block device mappings for instance. {{(pid=61806) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1030.608248] env[61806]: INFO nova.compute.claims [None req-a5c61942-d486-43f3-834d-7dfb87ef5b63 tempest-ServersNegativeTestJSON-891856253 tempest-ServersNegativeTestJSON-891856253-project-member] [instance: bdbaef31-7d37-40cb-ac3e-ad152b37e2d2] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1030.835039] env[61806]: DEBUG oslo_vmware.api [None req-de9a979c-45ed-463a-871e-b7dda8c366a2 tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] Task: {'id': task-1295130, 'name': ReconfigVM_Task, 'duration_secs': 0.715051} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1030.835466] env[61806]: DEBUG nova.virt.vmwareapi.volumeops [None req-de9a979c-45ed-463a-871e-b7dda8c366a2 tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] [instance: fd9e520a-2e7b-43d4-b84d-8933c32cc6a6] Reconfigured VM instance instance-00000057 to attach disk [datastore2] volume-dd35bef0-37b9-4821-9a9d-7f5c7789746c/volume-dd35bef0-37b9-4821-9a9d-7f5c7789746c.vmdk or device None with type thin {{(pid=61806) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1030.840409] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-dba2be17-0c03-471e-a9ac-86290d92e411 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1030.858526] env[61806]: DEBUG oslo_vmware.api [None req-de9a979c-45ed-463a-871e-b7dda8c366a2 tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] Waiting for the task: (returnval){ [ 1030.858526] env[61806]: value = "task-1295132" [ 1030.858526] env[61806]: _type = "Task" [ 1030.858526] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1030.870357] env[61806]: DEBUG oslo_vmware.api [None req-de9a979c-45ed-463a-871e-b7dda8c366a2 tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] Task: {'id': task-1295132, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1030.977231] env[61806]: DEBUG oslo_vmware.api [None req-d6fbd9e3-c911-4268-af0d-195628dc566f tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] Task: {'id': task-1295131, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.837821} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1030.977745] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [None req-d6fbd9e3-c911-4268-af0d-195628dc566f tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa.vmdk to [datastore1] 713d4e6e-6f0e-4844-9b55-cc010030eda6/713d4e6e-6f0e-4844-9b55-cc010030eda6.vmdk {{(pid=61806) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 1030.978170] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-d6fbd9e3-c911-4268-af0d-195628dc566f tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] [instance: 713d4e6e-6f0e-4844-9b55-cc010030eda6] Extending root virtual disk to 1048576 {{(pid=61806) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1030.978622] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-3e20f236-eca1-4f67-8480-72caaa36eb86 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1030.988947] env[61806]: DEBUG oslo_vmware.api [None req-d6fbd9e3-c911-4268-af0d-195628dc566f tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] Waiting for the task: (returnval){ [ 1030.988947] env[61806]: value = "task-1295133" [ 1030.988947] env[61806]: _type = "Task" [ 1030.988947] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1031.006191] env[61806]: DEBUG oslo_vmware.api [None req-d6fbd9e3-c911-4268-af0d-195628dc566f tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] Task: {'id': task-1295133, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1031.370415] env[61806]: DEBUG oslo_vmware.api [None req-de9a979c-45ed-463a-871e-b7dda8c366a2 tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] Task: {'id': task-1295132, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1031.506838] env[61806]: DEBUG oslo_vmware.api [None req-d6fbd9e3-c911-4268-af0d-195628dc566f tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] Task: {'id': task-1295133, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.221249} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1031.506838] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-d6fbd9e3-c911-4268-af0d-195628dc566f tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] [instance: 713d4e6e-6f0e-4844-9b55-cc010030eda6] Extended root virtual disk {{(pid=61806) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1031.507209] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-555247e2-42c4-4045-8811-0a83a5634067 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1031.529235] env[61806]: DEBUG nova.virt.vmwareapi.volumeops [None req-d6fbd9e3-c911-4268-af0d-195628dc566f tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] [instance: 713d4e6e-6f0e-4844-9b55-cc010030eda6] Reconfiguring VM instance instance-00000064 to attach disk [datastore1] 713d4e6e-6f0e-4844-9b55-cc010030eda6/713d4e6e-6f0e-4844-9b55-cc010030eda6.vmdk or device None with type sparse {{(pid=61806) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1031.529499] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-6479a012-3ae4-431e-9e39-58e002d09243 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1031.549224] env[61806]: DEBUG oslo_vmware.api [None req-d6fbd9e3-c911-4268-af0d-195628dc566f tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] Waiting for the task: (returnval){ [ 1031.549224] env[61806]: value = "task-1295134" [ 1031.549224] env[61806]: _type = "Task" [ 1031.549224] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1031.557300] env[61806]: DEBUG oslo_vmware.api [None req-d6fbd9e3-c911-4268-af0d-195628dc566f tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] Task: {'id': task-1295134, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1031.622454] env[61806]: DEBUG nova.compute.manager [None req-439a5433-afde-41f1-8525-78229f3cf29f tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] [instance: a61abfd9-3cc1-4a2b-b68b-a446094daf87] Start spawning the instance on the hypervisor. {{(pid=61806) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1031.648305] env[61806]: DEBUG nova.virt.hardware [None req-439a5433-afde-41f1-8525-78229f3cf29f tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-15T18:05:10Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-15T18:04:52Z,direct_url=,disk_format='vmdk',id=ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='84ba83fce2c349a6988173c5d6fc3b07',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-15T18:04:53Z,virtual_size=,visibility=), allow threads: False {{(pid=61806) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1031.648586] env[61806]: DEBUG nova.virt.hardware [None req-439a5433-afde-41f1-8525-78229f3cf29f tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Flavor limits 0:0:0 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1031.648776] env[61806]: DEBUG nova.virt.hardware [None req-439a5433-afde-41f1-8525-78229f3cf29f tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Image limits 0:0:0 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1031.648939] env[61806]: DEBUG nova.virt.hardware [None req-439a5433-afde-41f1-8525-78229f3cf29f tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Flavor pref 0:0:0 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1031.649108] env[61806]: DEBUG nova.virt.hardware [None req-439a5433-afde-41f1-8525-78229f3cf29f tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Image pref 0:0:0 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1031.649292] env[61806]: DEBUG nova.virt.hardware [None req-439a5433-afde-41f1-8525-78229f3cf29f tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1031.649546] env[61806]: DEBUG nova.virt.hardware [None req-439a5433-afde-41f1-8525-78229f3cf29f tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61806) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1031.649721] env[61806]: DEBUG nova.virt.hardware [None req-439a5433-afde-41f1-8525-78229f3cf29f tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61806) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1031.649901] env[61806]: DEBUG nova.virt.hardware [None req-439a5433-afde-41f1-8525-78229f3cf29f tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Got 1 possible topologies {{(pid=61806) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1031.650094] env[61806]: DEBUG nova.virt.hardware [None req-439a5433-afde-41f1-8525-78229f3cf29f tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61806) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1031.650284] env[61806]: DEBUG nova.virt.hardware [None req-439a5433-afde-41f1-8525-78229f3cf29f tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61806) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1031.651211] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-447dcb1d-8731-4b21-9df2-ef1db2759f95 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1031.660135] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-08caa735-04b1-4be8-9908-5446125d4d26 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1031.767041] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8d38def2-c9be-4377-8335-40d396f72696 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1031.775892] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ee91a984-cbb1-40a8-982b-3331643ebd9a {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1031.809396] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-81c7b8b9-8390-4747-ba06-90f9205727e3 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1031.817699] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8687d559-2b45-48b9-b24c-31495eb0e7c5 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1031.832699] env[61806]: DEBUG nova.compute.provider_tree [None req-a5c61942-d486-43f3-834d-7dfb87ef5b63 tempest-ServersNegativeTestJSON-891856253 tempest-ServersNegativeTestJSON-891856253-project-member] Inventory has not changed in ProviderTree for provider: a2858be1-fd22-4e08-979e-87ad25293407 {{(pid=61806) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1031.898706] env[61806]: DEBUG oslo_vmware.api [None req-de9a979c-45ed-463a-871e-b7dda8c366a2 tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] Task: {'id': task-1295132, 'name': ReconfigVM_Task, 'duration_secs': 0.549451} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1031.898706] env[61806]: DEBUG nova.virt.vmwareapi.volumeops [None req-de9a979c-45ed-463a-871e-b7dda8c366a2 tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] [instance: fd9e520a-2e7b-43d4-b84d-8933c32cc6a6] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-277740', 'volume_id': 'dd35bef0-37b9-4821-9a9d-7f5c7789746c', 'name': 'volume-dd35bef0-37b9-4821-9a9d-7f5c7789746c', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'attached', 'instance': 'fd9e520a-2e7b-43d4-b84d-8933c32cc6a6', 'attached_at': '', 'detached_at': '', 'volume_id': 'dd35bef0-37b9-4821-9a9d-7f5c7789746c', 'serial': 'dd35bef0-37b9-4821-9a9d-7f5c7789746c'} {{(pid=61806) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 1031.898706] env[61806]: DEBUG nova.compute.manager [req-20f0748f-e3ff-4221-a568-e12e3c62c4b8 req-3b85f99a-7b46-4c60-904b-023348d1254d service nova] [instance: a61abfd9-3cc1-4a2b-b68b-a446094daf87] Received event network-vif-plugged-2950cc6d-7a09-4217-abf5-8a5a940e6653 {{(pid=61806) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1031.898706] env[61806]: DEBUG oslo_concurrency.lockutils [req-20f0748f-e3ff-4221-a568-e12e3c62c4b8 req-3b85f99a-7b46-4c60-904b-023348d1254d service nova] Acquiring lock "a61abfd9-3cc1-4a2b-b68b-a446094daf87-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1031.898706] env[61806]: DEBUG oslo_concurrency.lockutils [req-20f0748f-e3ff-4221-a568-e12e3c62c4b8 req-3b85f99a-7b46-4c60-904b-023348d1254d service nova] Lock "a61abfd9-3cc1-4a2b-b68b-a446094daf87-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1031.898706] env[61806]: DEBUG oslo_concurrency.lockutils [req-20f0748f-e3ff-4221-a568-e12e3c62c4b8 req-3b85f99a-7b46-4c60-904b-023348d1254d service nova] Lock "a61abfd9-3cc1-4a2b-b68b-a446094daf87-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1031.898706] env[61806]: DEBUG nova.compute.manager [req-20f0748f-e3ff-4221-a568-e12e3c62c4b8 req-3b85f99a-7b46-4c60-904b-023348d1254d service nova] [instance: a61abfd9-3cc1-4a2b-b68b-a446094daf87] No waiting events found dispatching network-vif-plugged-2950cc6d-7a09-4217-abf5-8a5a940e6653 {{(pid=61806) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1031.898706] env[61806]: WARNING nova.compute.manager [req-20f0748f-e3ff-4221-a568-e12e3c62c4b8 req-3b85f99a-7b46-4c60-904b-023348d1254d service nova] [instance: a61abfd9-3cc1-4a2b-b68b-a446094daf87] Received unexpected event network-vif-plugged-2950cc6d-7a09-4217-abf5-8a5a940e6653 for instance with vm_state building and task_state spawning. [ 1031.898706] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-e0349b5a-4325-4cff-99e3-49b8aae59e07 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1031.898706] env[61806]: DEBUG oslo_vmware.api [None req-de9a979c-45ed-463a-871e-b7dda8c366a2 tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] Waiting for the task: (returnval){ [ 1031.898706] env[61806]: value = "task-1295135" [ 1031.898706] env[61806]: _type = "Task" [ 1031.898706] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1031.898706] env[61806]: DEBUG oslo_vmware.api [None req-de9a979c-45ed-463a-871e-b7dda8c366a2 tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] Task: {'id': task-1295135, 'name': Rename_Task} progress is 5%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1031.909445] env[61806]: DEBUG oslo_service.periodic_task [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=61806) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1031.909540] env[61806]: DEBUG oslo_service.periodic_task [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=61806) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1031.909639] env[61806]: DEBUG nova.compute.manager [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Starting heal instance info cache {{(pid=61806) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9926}} [ 1031.960343] env[61806]: DEBUG nova.network.neutron [None req-439a5433-afde-41f1-8525-78229f3cf29f tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] [instance: a61abfd9-3cc1-4a2b-b68b-a446094daf87] Successfully updated port: 2950cc6d-7a09-4217-abf5-8a5a940e6653 {{(pid=61806) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1032.061867] env[61806]: DEBUG oslo_vmware.api [None req-d6fbd9e3-c911-4268-af0d-195628dc566f tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] Task: {'id': task-1295134, 'name': ReconfigVM_Task} progress is 99%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1032.336269] env[61806]: DEBUG nova.scheduler.client.report [None req-a5c61942-d486-43f3-834d-7dfb87ef5b63 tempest-ServersNegativeTestJSON-891856253 tempest-ServersNegativeTestJSON-891856253-project-member] Inventory has not changed for provider a2858be1-fd22-4e08-979e-87ad25293407 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 138, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61806) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1032.392324] env[61806]: DEBUG oslo_vmware.api [None req-de9a979c-45ed-463a-871e-b7dda8c366a2 tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] Task: {'id': task-1295135, 'name': Rename_Task} progress is 99%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1032.415155] env[61806]: DEBUG nova.compute.manager [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Didn't find any instances for network info cache update. {{(pid=61806) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10012}} [ 1032.415365] env[61806]: DEBUG oslo_service.periodic_task [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=61806) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1032.415543] env[61806]: DEBUG oslo_service.periodic_task [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=61806) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1032.415695] env[61806]: DEBUG oslo_service.periodic_task [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=61806) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1032.415848] env[61806]: DEBUG oslo_service.periodic_task [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=61806) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1032.415996] env[61806]: DEBUG oslo_service.periodic_task [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=61806) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1032.416163] env[61806]: DEBUG oslo_service.periodic_task [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=61806) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1032.416294] env[61806]: DEBUG nova.compute.manager [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=61806) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10545}} [ 1032.416440] env[61806]: DEBUG oslo_service.periodic_task [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Running periodic task ComputeManager.update_available_resource {{(pid=61806) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1032.462740] env[61806]: DEBUG oslo_concurrency.lockutils [None req-439a5433-afde-41f1-8525-78229f3cf29f tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Acquiring lock "refresh_cache-a61abfd9-3cc1-4a2b-b68b-a446094daf87" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1032.462874] env[61806]: DEBUG oslo_concurrency.lockutils [None req-439a5433-afde-41f1-8525-78229f3cf29f tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Acquired lock "refresh_cache-a61abfd9-3cc1-4a2b-b68b-a446094daf87" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1032.463036] env[61806]: DEBUG nova.network.neutron [None req-439a5433-afde-41f1-8525-78229f3cf29f tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] [instance: a61abfd9-3cc1-4a2b-b68b-a446094daf87] Building network info cache for instance {{(pid=61806) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1032.560755] env[61806]: DEBUG oslo_vmware.api [None req-d6fbd9e3-c911-4268-af0d-195628dc566f tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] Task: {'id': task-1295134, 'name': ReconfigVM_Task, 'duration_secs': 0.575514} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1032.561051] env[61806]: DEBUG nova.virt.vmwareapi.volumeops [None req-d6fbd9e3-c911-4268-af0d-195628dc566f tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] [instance: 713d4e6e-6f0e-4844-9b55-cc010030eda6] Reconfigured VM instance instance-00000064 to attach disk [datastore1] 713d4e6e-6f0e-4844-9b55-cc010030eda6/713d4e6e-6f0e-4844-9b55-cc010030eda6.vmdk or device None with type sparse {{(pid=61806) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1032.561712] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-11fe1046-368d-4a0a-8bda-1e04c26f6c9a {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1032.568579] env[61806]: DEBUG oslo_vmware.api [None req-d6fbd9e3-c911-4268-af0d-195628dc566f tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] Waiting for the task: (returnval){ [ 1032.568579] env[61806]: value = "task-1295136" [ 1032.568579] env[61806]: _type = "Task" [ 1032.568579] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1032.576572] env[61806]: DEBUG oslo_vmware.api [None req-d6fbd9e3-c911-4268-af0d-195628dc566f tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] Task: {'id': task-1295136, 'name': Rename_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1032.841079] env[61806]: DEBUG oslo_concurrency.lockutils [None req-a5c61942-d486-43f3-834d-7dfb87ef5b63 tempest-ServersNegativeTestJSON-891856253 tempest-ServersNegativeTestJSON-891856253-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 3.246s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1032.843671] env[61806]: DEBUG oslo_concurrency.lockutils [None req-6a1c6218-d12d-4e5e-8616-7dd5071f5819 tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 7.223s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1032.843839] env[61806]: DEBUG oslo_concurrency.lockutils [None req-6a1c6218-d12d-4e5e-8616-7dd5071f5819 tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1032.845772] env[61806]: DEBUG oslo_concurrency.lockutils [None req-78fe7ba4-1e10-47a3-bab9-f40fd2226008 tempest-AttachVolumeNegativeTest-1683033704 tempest-AttachVolumeNegativeTest-1683033704-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 4.636s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1032.846111] env[61806]: DEBUG nova.objects.instance [None req-78fe7ba4-1e10-47a3-bab9-f40fd2226008 tempest-AttachVolumeNegativeTest-1683033704 tempest-AttachVolumeNegativeTest-1683033704-project-member] Lazy-loading 'resources' on Instance uuid e4638faa-7be1-4909-a595-c437837fa314 {{(pid=61806) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1032.867997] env[61806]: INFO nova.scheduler.client.report [None req-6a1c6218-d12d-4e5e-8616-7dd5071f5819 tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Deleted allocations for instance 9357e61f-1628-43bd-ab46-de13c1529f51 [ 1032.879550] env[61806]: INFO nova.network.neutron [None req-a5c61942-d486-43f3-834d-7dfb87ef5b63 tempest-ServersNegativeTestJSON-891856253 tempest-ServersNegativeTestJSON-891856253-project-member] [instance: bdbaef31-7d37-40cb-ac3e-ad152b37e2d2] Updating port b8b714a4-b17c-4b3d-889e-407bec6c30ae with attributes {'binding:host_id': 'cpu-1', 'device_owner': 'compute:nova'} [ 1032.892784] env[61806]: DEBUG oslo_vmware.api [None req-de9a979c-45ed-463a-871e-b7dda8c366a2 tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] Task: {'id': task-1295135, 'name': Rename_Task} progress is 99%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1032.918856] env[61806]: DEBUG oslo_concurrency.lockutils [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1033.004442] env[61806]: DEBUG nova.network.neutron [None req-439a5433-afde-41f1-8525-78229f3cf29f tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] [instance: a61abfd9-3cc1-4a2b-b68b-a446094daf87] Instance cache missing network info. {{(pid=61806) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 1033.079233] env[61806]: DEBUG oslo_vmware.api [None req-d6fbd9e3-c911-4268-af0d-195628dc566f tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] Task: {'id': task-1295136, 'name': Rename_Task, 'duration_secs': 0.140447} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1033.081656] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [None req-d6fbd9e3-c911-4268-af0d-195628dc566f tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] [instance: 713d4e6e-6f0e-4844-9b55-cc010030eda6] Powering on the VM {{(pid=61806) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1033.081935] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-90d8391a-3dd8-40e8-a62a-c51b9c03fec8 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1033.089272] env[61806]: DEBUG oslo_vmware.api [None req-d6fbd9e3-c911-4268-af0d-195628dc566f tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] Waiting for the task: (returnval){ [ 1033.089272] env[61806]: value = "task-1295137" [ 1033.089272] env[61806]: _type = "Task" [ 1033.089272] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1033.097452] env[61806]: DEBUG oslo_vmware.api [None req-d6fbd9e3-c911-4268-af0d-195628dc566f tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] Task: {'id': task-1295137, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1033.293658] env[61806]: DEBUG nova.network.neutron [None req-439a5433-afde-41f1-8525-78229f3cf29f tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] [instance: a61abfd9-3cc1-4a2b-b68b-a446094daf87] Updating instance_info_cache with network_info: [{"id": "2950cc6d-7a09-4217-abf5-8a5a940e6653", "address": "fa:16:3e:31:5f:a7", "network": {"id": "46515a4e-610f-42fb-a424-205a03df0640", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-1133623547-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4b9d94c3d684474b99e305fb6cd9b95f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "82ca17df-257e-40e6-9ec9-310ed6f05ccb", "external-id": "nsx-vlan-transportzone-778", "segmentation_id": 778, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2950cc6d-7a", "ovs_interfaceid": "2950cc6d-7a09-4217-abf5-8a5a940e6653", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61806) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1033.375665] env[61806]: DEBUG oslo_concurrency.lockutils [None req-6a1c6218-d12d-4e5e-8616-7dd5071f5819 tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Lock "9357e61f-1628-43bd-ab46-de13c1529f51" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 12.180s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1033.393415] env[61806]: DEBUG oslo_vmware.api [None req-de9a979c-45ed-463a-871e-b7dda8c366a2 tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] Task: {'id': task-1295135, 'name': Rename_Task, 'duration_secs': 1.186871} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1033.394281] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [None req-de9a979c-45ed-463a-871e-b7dda8c366a2 tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] [instance: fd9e520a-2e7b-43d4-b84d-8933c32cc6a6] Powering on the VM {{(pid=61806) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1033.394549] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-d5ce2dfe-71d0-4372-adc9-21e838997772 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1033.403759] env[61806]: DEBUG oslo_vmware.api [None req-de9a979c-45ed-463a-871e-b7dda8c366a2 tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] Waiting for the task: (returnval){ [ 1033.403759] env[61806]: value = "task-1295138" [ 1033.403759] env[61806]: _type = "Task" [ 1033.403759] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1033.414548] env[61806]: DEBUG oslo_vmware.api [None req-de9a979c-45ed-463a-871e-b7dda8c366a2 tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] Task: {'id': task-1295138, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1033.476069] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-53cffa25-b957-437b-ab98-06aa6b8c0840 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1033.483852] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-517cfa05-5325-4cb6-976b-4f7af98b5847 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1033.514593] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a64025a5-2c8a-4573-92da-134c1d6a4da5 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1033.523229] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aec9cdb1-2cde-4863-bdbf-6223f34b245a {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1033.537913] env[61806]: DEBUG nova.compute.provider_tree [None req-78fe7ba4-1e10-47a3-bab9-f40fd2226008 tempest-AttachVolumeNegativeTest-1683033704 tempest-AttachVolumeNegativeTest-1683033704-project-member] Inventory has not changed in ProviderTree for provider: a2858be1-fd22-4e08-979e-87ad25293407 {{(pid=61806) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1033.601206] env[61806]: DEBUG oslo_vmware.api [None req-d6fbd9e3-c911-4268-af0d-195628dc566f tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] Task: {'id': task-1295137, 'name': PowerOnVM_Task, 'duration_secs': 0.465685} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1033.601487] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [None req-d6fbd9e3-c911-4268-af0d-195628dc566f tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] [instance: 713d4e6e-6f0e-4844-9b55-cc010030eda6] Powered on the VM {{(pid=61806) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1033.601693] env[61806]: INFO nova.compute.manager [None req-d6fbd9e3-c911-4268-af0d-195628dc566f tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] [instance: 713d4e6e-6f0e-4844-9b55-cc010030eda6] Took 10.20 seconds to spawn the instance on the hypervisor. [ 1033.601877] env[61806]: DEBUG nova.compute.manager [None req-d6fbd9e3-c911-4268-af0d-195628dc566f tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] [instance: 713d4e6e-6f0e-4844-9b55-cc010030eda6] Checking state {{(pid=61806) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1033.602677] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cb48dd87-afe2-48d9-89eb-7810e34f503e {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1033.798050] env[61806]: DEBUG oslo_concurrency.lockutils [None req-439a5433-afde-41f1-8525-78229f3cf29f tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Releasing lock "refresh_cache-a61abfd9-3cc1-4a2b-b68b-a446094daf87" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1033.798050] env[61806]: DEBUG nova.compute.manager [None req-439a5433-afde-41f1-8525-78229f3cf29f tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] [instance: a61abfd9-3cc1-4a2b-b68b-a446094daf87] Instance network_info: |[{"id": "2950cc6d-7a09-4217-abf5-8a5a940e6653", "address": "fa:16:3e:31:5f:a7", "network": {"id": "46515a4e-610f-42fb-a424-205a03df0640", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-1133623547-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4b9d94c3d684474b99e305fb6cd9b95f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "82ca17df-257e-40e6-9ec9-310ed6f05ccb", "external-id": "nsx-vlan-transportzone-778", "segmentation_id": 778, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2950cc6d-7a", "ovs_interfaceid": "2950cc6d-7a09-4217-abf5-8a5a940e6653", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61806) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1033.798292] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-439a5433-afde-41f1-8525-78229f3cf29f tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] [instance: a61abfd9-3cc1-4a2b-b68b-a446094daf87] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:31:5f:a7', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '82ca17df-257e-40e6-9ec9-310ed6f05ccb', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '2950cc6d-7a09-4217-abf5-8a5a940e6653', 'vif_model': 'vmxnet3'}] {{(pid=61806) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1033.806746] env[61806]: DEBUG oslo.service.loopingcall [None req-439a5433-afde-41f1-8525-78229f3cf29f tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61806) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1033.807018] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: a61abfd9-3cc1-4a2b-b68b-a446094daf87] Creating VM on the ESX host {{(pid=61806) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 1033.807260] env[61806]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-69c4c2ed-00bb-484b-97e9-da102d6918ee {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1033.828854] env[61806]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1033.828854] env[61806]: value = "task-1295139" [ 1033.828854] env[61806]: _type = "Task" [ 1033.828854] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1033.836689] env[61806]: DEBUG oslo_vmware.api [-] Task: {'id': task-1295139, 'name': CreateVM_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1033.895141] env[61806]: DEBUG nova.compute.manager [req-a451dd13-e9bd-47f8-8a28-5df6512efcfc req-43e08733-9b79-4557-ac1a-91e1fe019eae service nova] [instance: a61abfd9-3cc1-4a2b-b68b-a446094daf87] Received event network-changed-2950cc6d-7a09-4217-abf5-8a5a940e6653 {{(pid=61806) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1033.895306] env[61806]: DEBUG nova.compute.manager [req-a451dd13-e9bd-47f8-8a28-5df6512efcfc req-43e08733-9b79-4557-ac1a-91e1fe019eae service nova] [instance: a61abfd9-3cc1-4a2b-b68b-a446094daf87] Refreshing instance network info cache due to event network-changed-2950cc6d-7a09-4217-abf5-8a5a940e6653. {{(pid=61806) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 1033.895605] env[61806]: DEBUG oslo_concurrency.lockutils [req-a451dd13-e9bd-47f8-8a28-5df6512efcfc req-43e08733-9b79-4557-ac1a-91e1fe019eae service nova] Acquiring lock "refresh_cache-a61abfd9-3cc1-4a2b-b68b-a446094daf87" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1033.895771] env[61806]: DEBUG oslo_concurrency.lockutils [req-a451dd13-e9bd-47f8-8a28-5df6512efcfc req-43e08733-9b79-4557-ac1a-91e1fe019eae service nova] Acquired lock "refresh_cache-a61abfd9-3cc1-4a2b-b68b-a446094daf87" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1033.895948] env[61806]: DEBUG nova.network.neutron [req-a451dd13-e9bd-47f8-8a28-5df6512efcfc req-43e08733-9b79-4557-ac1a-91e1fe019eae service nova] [instance: a61abfd9-3cc1-4a2b-b68b-a446094daf87] Refreshing network info cache for port 2950cc6d-7a09-4217-abf5-8a5a940e6653 {{(pid=61806) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1033.919749] env[61806]: DEBUG oslo_vmware.api [None req-de9a979c-45ed-463a-871e-b7dda8c366a2 tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] Task: {'id': task-1295138, 'name': PowerOnVM_Task} progress is 89%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1034.041661] env[61806]: DEBUG nova.scheduler.client.report [None req-78fe7ba4-1e10-47a3-bab9-f40fd2226008 tempest-AttachVolumeNegativeTest-1683033704 tempest-AttachVolumeNegativeTest-1683033704-project-member] Inventory has not changed for provider a2858be1-fd22-4e08-979e-87ad25293407 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 138, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61806) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1034.118274] env[61806]: INFO nova.compute.manager [None req-d6fbd9e3-c911-4268-af0d-195628dc566f tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] [instance: 713d4e6e-6f0e-4844-9b55-cc010030eda6] Took 15.05 seconds to build instance. [ 1034.340455] env[61806]: DEBUG oslo_vmware.api [-] Task: {'id': task-1295139, 'name': CreateVM_Task, 'duration_secs': 0.441523} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1034.340627] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: a61abfd9-3cc1-4a2b-b68b-a446094daf87] Created VM on the ESX host {{(pid=61806) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 1034.341321] env[61806]: DEBUG oslo_concurrency.lockutils [None req-439a5433-afde-41f1-8525-78229f3cf29f tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1034.341501] env[61806]: DEBUG oslo_concurrency.lockutils [None req-439a5433-afde-41f1-8525-78229f3cf29f tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Acquired lock "[datastore1] devstack-image-cache_base/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1034.341892] env[61806]: DEBUG oslo_concurrency.lockutils [None req-439a5433-afde-41f1-8525-78229f3cf29f tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1034.342169] env[61806]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-10bceb0c-1dd1-4d1d-883d-2aee3d23b20b {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1034.347175] env[61806]: DEBUG oslo_vmware.api [None req-439a5433-afde-41f1-8525-78229f3cf29f tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Waiting for the task: (returnval){ [ 1034.347175] env[61806]: value = "session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]527b96d7-2159-87dc-3b60-cab86d8adbc8" [ 1034.347175] env[61806]: _type = "Task" [ 1034.347175] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1034.356253] env[61806]: DEBUG oslo_vmware.api [None req-439a5433-afde-41f1-8525-78229f3cf29f tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Task: {'id': session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]527b96d7-2159-87dc-3b60-cab86d8adbc8, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1034.421068] env[61806]: DEBUG oslo_vmware.api [None req-de9a979c-45ed-463a-871e-b7dda8c366a2 tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] Task: {'id': task-1295138, 'name': PowerOnVM_Task, 'duration_secs': 0.60119} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1034.421391] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [None req-de9a979c-45ed-463a-871e-b7dda8c366a2 tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] [instance: fd9e520a-2e7b-43d4-b84d-8933c32cc6a6] Powered on the VM {{(pid=61806) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1034.524747] env[61806]: DEBUG nova.compute.manager [None req-de9a979c-45ed-463a-871e-b7dda8c366a2 tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] [instance: fd9e520a-2e7b-43d4-b84d-8933c32cc6a6] Checking state {{(pid=61806) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1034.525956] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bde64616-96f3-4281-a774-20423fefd37f {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1034.547786] env[61806]: DEBUG oslo_concurrency.lockutils [None req-78fe7ba4-1e10-47a3-bab9-f40fd2226008 tempest-AttachVolumeNegativeTest-1683033704 tempest-AttachVolumeNegativeTest-1683033704-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.702s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1034.555177] env[61806]: DEBUG oslo_concurrency.lockutils [None req-138f28bc-d88a-4f34-babf-797a70b177f1 tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 5.878s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1034.556394] env[61806]: INFO nova.compute.claims [None req-138f28bc-d88a-4f34-babf-797a70b177f1 tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] [instance: 06658134-4c6d-4911-9a55-b6805f555c78] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1034.591422] env[61806]: INFO nova.scheduler.client.report [None req-78fe7ba4-1e10-47a3-bab9-f40fd2226008 tempest-AttachVolumeNegativeTest-1683033704 tempest-AttachVolumeNegativeTest-1683033704-project-member] Deleted allocations for instance e4638faa-7be1-4909-a595-c437837fa314 [ 1034.620684] env[61806]: DEBUG oslo_concurrency.lockutils [None req-d6fbd9e3-c911-4268-af0d-195628dc566f tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] Lock "713d4e6e-6f0e-4844-9b55-cc010030eda6" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 16.570s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1034.775026] env[61806]: DEBUG oslo_concurrency.lockutils [None req-a5c61942-d486-43f3-834d-7dfb87ef5b63 tempest-ServersNegativeTestJSON-891856253 tempest-ServersNegativeTestJSON-891856253-project-member] Acquiring lock "refresh_cache-bdbaef31-7d37-40cb-ac3e-ad152b37e2d2" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1034.775274] env[61806]: DEBUG oslo_concurrency.lockutils [None req-a5c61942-d486-43f3-834d-7dfb87ef5b63 tempest-ServersNegativeTestJSON-891856253 tempest-ServersNegativeTestJSON-891856253-project-member] Acquired lock "refresh_cache-bdbaef31-7d37-40cb-ac3e-ad152b37e2d2" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1034.775466] env[61806]: DEBUG nova.network.neutron [None req-a5c61942-d486-43f3-834d-7dfb87ef5b63 tempest-ServersNegativeTestJSON-891856253 tempest-ServersNegativeTestJSON-891856253-project-member] [instance: bdbaef31-7d37-40cb-ac3e-ad152b37e2d2] Building network info cache for instance {{(pid=61806) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1034.803239] env[61806]: DEBUG nova.network.neutron [req-a451dd13-e9bd-47f8-8a28-5df6512efcfc req-43e08733-9b79-4557-ac1a-91e1fe019eae service nova] [instance: a61abfd9-3cc1-4a2b-b68b-a446094daf87] Updated VIF entry in instance network info cache for port 2950cc6d-7a09-4217-abf5-8a5a940e6653. {{(pid=61806) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1034.803608] env[61806]: DEBUG nova.network.neutron [req-a451dd13-e9bd-47f8-8a28-5df6512efcfc req-43e08733-9b79-4557-ac1a-91e1fe019eae service nova] [instance: a61abfd9-3cc1-4a2b-b68b-a446094daf87] Updating instance_info_cache with network_info: [{"id": "2950cc6d-7a09-4217-abf5-8a5a940e6653", "address": "fa:16:3e:31:5f:a7", "network": {"id": "46515a4e-610f-42fb-a424-205a03df0640", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-1133623547-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4b9d94c3d684474b99e305fb6cd9b95f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "82ca17df-257e-40e6-9ec9-310ed6f05ccb", "external-id": "nsx-vlan-transportzone-778", "segmentation_id": 778, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2950cc6d-7a", "ovs_interfaceid": "2950cc6d-7a09-4217-abf5-8a5a940e6653", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61806) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1034.859430] env[61806]: DEBUG oslo_vmware.api [None req-439a5433-afde-41f1-8525-78229f3cf29f tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Task: {'id': session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]527b96d7-2159-87dc-3b60-cab86d8adbc8, 'name': SearchDatastore_Task, 'duration_secs': 0.009913} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1034.859766] env[61806]: DEBUG oslo_concurrency.lockutils [None req-439a5433-afde-41f1-8525-78229f3cf29f tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Releasing lock "[datastore1] devstack-image-cache_base/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1034.860026] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-439a5433-afde-41f1-8525-78229f3cf29f tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] [instance: a61abfd9-3cc1-4a2b-b68b-a446094daf87] Processing image ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa {{(pid=61806) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1034.860318] env[61806]: DEBUG oslo_concurrency.lockutils [None req-439a5433-afde-41f1-8525-78229f3cf29f tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa.vmdk" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1034.860481] env[61806]: DEBUG oslo_concurrency.lockutils [None req-439a5433-afde-41f1-8525-78229f3cf29f tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Acquired lock "[datastore1] devstack-image-cache_base/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa.vmdk" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1034.860670] env[61806]: DEBUG nova.virt.vmwareapi.ds_util [None req-439a5433-afde-41f1-8525-78229f3cf29f tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61806) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1034.860947] env[61806]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-fbcac39b-6acc-4e4d-b223-e8081410b55e {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1034.870825] env[61806]: DEBUG nova.virt.vmwareapi.ds_util [None req-439a5433-afde-41f1-8525-78229f3cf29f tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61806) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1034.871027] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-439a5433-afde-41f1-8525-78229f3cf29f tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61806) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 1034.871759] env[61806]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-05e7146e-7efb-472d-aa78-6d83fec839d3 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1034.877774] env[61806]: DEBUG oslo_vmware.api [None req-439a5433-afde-41f1-8525-78229f3cf29f tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Waiting for the task: (returnval){ [ 1034.877774] env[61806]: value = "session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]52c76b61-c6ad-c73e-e43b-cf76ea03d3c5" [ 1034.877774] env[61806]: _type = "Task" [ 1034.877774] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1034.885691] env[61806]: DEBUG oslo_vmware.api [None req-439a5433-afde-41f1-8525-78229f3cf29f tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Task: {'id': session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]52c76b61-c6ad-c73e-e43b-cf76ea03d3c5, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1035.043975] env[61806]: DEBUG oslo_concurrency.lockutils [None req-de9a979c-45ed-463a-871e-b7dda8c366a2 tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] Lock "fd9e520a-2e7b-43d4-b84d-8933c32cc6a6" "released" by "nova.compute.manager.ComputeManager.unshelve_instance..do_unshelve_instance" :: held 32.993s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1035.106564] env[61806]: DEBUG oslo_concurrency.lockutils [None req-78fe7ba4-1e10-47a3-bab9-f40fd2226008 tempest-AttachVolumeNegativeTest-1683033704 tempest-AttachVolumeNegativeTest-1683033704-project-member] Lock "e4638faa-7be1-4909-a595-c437837fa314" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 12.273s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1035.251324] env[61806]: DEBUG oslo_concurrency.lockutils [None req-42bedba7-0c5b-4f33-8173-c0388dea4ce8 tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] Acquiring lock "713d4e6e-6f0e-4844-9b55-cc010030eda6" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1035.251581] env[61806]: DEBUG oslo_concurrency.lockutils [None req-42bedba7-0c5b-4f33-8173-c0388dea4ce8 tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] Lock "713d4e6e-6f0e-4844-9b55-cc010030eda6" acquired by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: waited 0.001s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1035.251763] env[61806]: DEBUG nova.compute.manager [None req-42bedba7-0c5b-4f33-8173-c0388dea4ce8 tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] [instance: 713d4e6e-6f0e-4844-9b55-cc010030eda6] Checking state {{(pid=61806) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1035.252680] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-733e4df5-4b9e-4fde-af7b-434e2a302e4d {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1035.259772] env[61806]: DEBUG nova.compute.manager [None req-42bedba7-0c5b-4f33-8173-c0388dea4ce8 tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] [instance: 713d4e6e-6f0e-4844-9b55-cc010030eda6] Stopping instance; current vm_state: active, current task_state: powering-off, current DB power_state: 1, current VM power_state: 1 {{(pid=61806) do_stop_instance /opt/stack/nova/nova/compute/manager.py:3368}} [ 1035.260368] env[61806]: DEBUG nova.objects.instance [None req-42bedba7-0c5b-4f33-8173-c0388dea4ce8 tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] Lazy-loading 'flavor' on Instance uuid 713d4e6e-6f0e-4844-9b55-cc010030eda6 {{(pid=61806) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1035.307929] env[61806]: DEBUG oslo_concurrency.lockutils [req-a451dd13-e9bd-47f8-8a28-5df6512efcfc req-43e08733-9b79-4557-ac1a-91e1fe019eae service nova] Releasing lock "refresh_cache-a61abfd9-3cc1-4a2b-b68b-a446094daf87" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1035.388861] env[61806]: DEBUG oslo_vmware.api [None req-439a5433-afde-41f1-8525-78229f3cf29f tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Task: {'id': session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]52c76b61-c6ad-c73e-e43b-cf76ea03d3c5, 'name': SearchDatastore_Task, 'duration_secs': 0.008977} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1035.390054] env[61806]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-48e415f6-ebc9-4db8-8e81-7d0c01bb1fe6 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1035.395960] env[61806]: DEBUG oslo_vmware.api [None req-439a5433-afde-41f1-8525-78229f3cf29f tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Waiting for the task: (returnval){ [ 1035.395960] env[61806]: value = "session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]524b092c-8444-5180-e35c-a258c66b0f88" [ 1035.395960] env[61806]: _type = "Task" [ 1035.395960] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1035.406407] env[61806]: DEBUG oslo_vmware.api [None req-439a5433-afde-41f1-8525-78229f3cf29f tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Task: {'id': session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]524b092c-8444-5180-e35c-a258c66b0f88, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1035.477434] env[61806]: DEBUG nova.network.neutron [None req-a5c61942-d486-43f3-834d-7dfb87ef5b63 tempest-ServersNegativeTestJSON-891856253 tempest-ServersNegativeTestJSON-891856253-project-member] [instance: bdbaef31-7d37-40cb-ac3e-ad152b37e2d2] Updating instance_info_cache with network_info: [{"id": "b8b714a4-b17c-4b3d-889e-407bec6c30ae", "address": "fa:16:3e:33:74:bb", "network": {"id": "fd35ea93-8914-4b54-a4ea-f45fc750982b", "bridge": "br-int", "label": "tempest-ServersNegativeTestJSON-2117068882-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d2e2b7eec47e4b10898215123408692a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "2d94740a-bce8-4103-8ecf-230d02ec0a44", "external-id": "nsx-vlan-transportzone-149", "segmentation_id": 149, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb8b714a4-b1", "ovs_interfaceid": "b8b714a4-b17c-4b3d-889e-407bec6c30ae", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61806) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1035.692169] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-405e09d9-10ed-4d99-b05e-795f61d8737e {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1035.700765] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-43813904-055e-47b2-b1b2-d0c956ec9470 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1035.735807] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bcbdec9f-9014-4b24-9465-4a9a791b301b {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1035.741132] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5042f421-c246-4467-83f3-37be1692b82c {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1035.755216] env[61806]: DEBUG nova.compute.provider_tree [None req-138f28bc-d88a-4f34-babf-797a70b177f1 tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Inventory has not changed in ProviderTree for provider: a2858be1-fd22-4e08-979e-87ad25293407 {{(pid=61806) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1035.764493] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [None req-42bedba7-0c5b-4f33-8173-c0388dea4ce8 tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] [instance: 713d4e6e-6f0e-4844-9b55-cc010030eda6] Powering off the VM {{(pid=61806) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1035.764715] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-db17c877-46e7-47e1-b885-db01adedb51b {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1035.772095] env[61806]: DEBUG oslo_vmware.api [None req-42bedba7-0c5b-4f33-8173-c0388dea4ce8 tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] Waiting for the task: (returnval){ [ 1035.772095] env[61806]: value = "task-1295140" [ 1035.772095] env[61806]: _type = "Task" [ 1035.772095] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1035.781727] env[61806]: DEBUG oslo_vmware.api [None req-42bedba7-0c5b-4f33-8173-c0388dea4ce8 tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] Task: {'id': task-1295140, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1035.906987] env[61806]: DEBUG oslo_vmware.api [None req-439a5433-afde-41f1-8525-78229f3cf29f tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Task: {'id': session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]524b092c-8444-5180-e35c-a258c66b0f88, 'name': SearchDatastore_Task, 'duration_secs': 0.012249} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1035.907394] env[61806]: DEBUG oslo_concurrency.lockutils [None req-439a5433-afde-41f1-8525-78229f3cf29f tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Releasing lock "[datastore1] devstack-image-cache_base/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa.vmdk" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1035.907721] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [None req-439a5433-afde-41f1-8525-78229f3cf29f tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa.vmdk to [datastore1] a61abfd9-3cc1-4a2b-b68b-a446094daf87/a61abfd9-3cc1-4a2b-b68b-a446094daf87.vmdk {{(pid=61806) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 1035.908045] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-9d2b02f1-d41a-49a3-93fd-5709a90dbe0f {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1035.918039] env[61806]: DEBUG oslo_vmware.api [None req-439a5433-afde-41f1-8525-78229f3cf29f tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Waiting for the task: (returnval){ [ 1035.918039] env[61806]: value = "task-1295141" [ 1035.918039] env[61806]: _type = "Task" [ 1035.918039] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1035.924917] env[61806]: DEBUG nova.compute.manager [req-6db4f388-ded0-41c4-933a-70c5211bc730 req-0f45f9d8-b4c0-4f15-8079-e220ca3175e6 service nova] [instance: bdbaef31-7d37-40cb-ac3e-ad152b37e2d2] Received event network-vif-plugged-b8b714a4-b17c-4b3d-889e-407bec6c30ae {{(pid=61806) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1035.926499] env[61806]: DEBUG oslo_concurrency.lockutils [req-6db4f388-ded0-41c4-933a-70c5211bc730 req-0f45f9d8-b4c0-4f15-8079-e220ca3175e6 service nova] Acquiring lock "bdbaef31-7d37-40cb-ac3e-ad152b37e2d2-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1035.926843] env[61806]: DEBUG oslo_concurrency.lockutils [req-6db4f388-ded0-41c4-933a-70c5211bc730 req-0f45f9d8-b4c0-4f15-8079-e220ca3175e6 service nova] Lock "bdbaef31-7d37-40cb-ac3e-ad152b37e2d2-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.002s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1035.927089] env[61806]: DEBUG oslo_concurrency.lockutils [req-6db4f388-ded0-41c4-933a-70c5211bc730 req-0f45f9d8-b4c0-4f15-8079-e220ca3175e6 service nova] Lock "bdbaef31-7d37-40cb-ac3e-ad152b37e2d2-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1035.927332] env[61806]: DEBUG nova.compute.manager [req-6db4f388-ded0-41c4-933a-70c5211bc730 req-0f45f9d8-b4c0-4f15-8079-e220ca3175e6 service nova] [instance: bdbaef31-7d37-40cb-ac3e-ad152b37e2d2] No waiting events found dispatching network-vif-plugged-b8b714a4-b17c-4b3d-889e-407bec6c30ae {{(pid=61806) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1035.927538] env[61806]: WARNING nova.compute.manager [req-6db4f388-ded0-41c4-933a-70c5211bc730 req-0f45f9d8-b4c0-4f15-8079-e220ca3175e6 service nova] [instance: bdbaef31-7d37-40cb-ac3e-ad152b37e2d2] Received unexpected event network-vif-plugged-b8b714a4-b17c-4b3d-889e-407bec6c30ae for instance with vm_state shelved_offloaded and task_state spawning. [ 1035.927745] env[61806]: DEBUG nova.compute.manager [req-6db4f388-ded0-41c4-933a-70c5211bc730 req-0f45f9d8-b4c0-4f15-8079-e220ca3175e6 service nova] [instance: bdbaef31-7d37-40cb-ac3e-ad152b37e2d2] Received event network-changed-b8b714a4-b17c-4b3d-889e-407bec6c30ae {{(pid=61806) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1035.927954] env[61806]: DEBUG nova.compute.manager [req-6db4f388-ded0-41c4-933a-70c5211bc730 req-0f45f9d8-b4c0-4f15-8079-e220ca3175e6 service nova] [instance: bdbaef31-7d37-40cb-ac3e-ad152b37e2d2] Refreshing instance network info cache due to event network-changed-b8b714a4-b17c-4b3d-889e-407bec6c30ae. {{(pid=61806) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 1035.928209] env[61806]: DEBUG oslo_concurrency.lockutils [req-6db4f388-ded0-41c4-933a-70c5211bc730 req-0f45f9d8-b4c0-4f15-8079-e220ca3175e6 service nova] Acquiring lock "refresh_cache-bdbaef31-7d37-40cb-ac3e-ad152b37e2d2" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1035.932296] env[61806]: DEBUG oslo_vmware.api [None req-439a5433-afde-41f1-8525-78229f3cf29f tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Task: {'id': task-1295141, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1035.980163] env[61806]: DEBUG oslo_concurrency.lockutils [None req-a5c61942-d486-43f3-834d-7dfb87ef5b63 tempest-ServersNegativeTestJSON-891856253 tempest-ServersNegativeTestJSON-891856253-project-member] Releasing lock "refresh_cache-bdbaef31-7d37-40cb-ac3e-ad152b37e2d2" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1035.983058] env[61806]: DEBUG oslo_concurrency.lockutils [req-6db4f388-ded0-41c4-933a-70c5211bc730 req-0f45f9d8-b4c0-4f15-8079-e220ca3175e6 service nova] Acquired lock "refresh_cache-bdbaef31-7d37-40cb-ac3e-ad152b37e2d2" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1035.983322] env[61806]: DEBUG nova.network.neutron [req-6db4f388-ded0-41c4-933a-70c5211bc730 req-0f45f9d8-b4c0-4f15-8079-e220ca3175e6 service nova] [instance: bdbaef31-7d37-40cb-ac3e-ad152b37e2d2] Refreshing network info cache for port b8b714a4-b17c-4b3d-889e-407bec6c30ae {{(pid=61806) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1036.011935] env[61806]: DEBUG nova.virt.hardware [None req-a5c61942-d486-43f3-834d-7dfb87ef5b63 tempest-ServersNegativeTestJSON-891856253 tempest-ServersNegativeTestJSON-891856253-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-15T18:05:10Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='7e05a4fe97c65f65d95918efbcf9fc73',container_format='bare',created_at=2024-10-15T18:14:58Z,direct_url=,disk_format='vmdk',id=20c8ea6e-0db5-4788-9e34-92c1d63bb7b8,min_disk=1,min_ram=0,name='tempest-ServersNegativeTestJSON-server-1314647859-shelved',owner='d2e2b7eec47e4b10898215123408692a',properties=ImageMetaProps,protected=,size=31667200,status='active',tags=,updated_at=2024-10-15T18:15:15Z,virtual_size=,visibility=), allow threads: False {{(pid=61806) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1036.011935] env[61806]: DEBUG nova.virt.hardware [None req-a5c61942-d486-43f3-834d-7dfb87ef5b63 tempest-ServersNegativeTestJSON-891856253 tempest-ServersNegativeTestJSON-891856253-project-member] Flavor limits 0:0:0 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1036.011935] env[61806]: DEBUG nova.virt.hardware [None req-a5c61942-d486-43f3-834d-7dfb87ef5b63 tempest-ServersNegativeTestJSON-891856253 tempest-ServersNegativeTestJSON-891856253-project-member] Image limits 0:0:0 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1036.011935] env[61806]: DEBUG nova.virt.hardware [None req-a5c61942-d486-43f3-834d-7dfb87ef5b63 tempest-ServersNegativeTestJSON-891856253 tempest-ServersNegativeTestJSON-891856253-project-member] Flavor pref 0:0:0 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1036.011935] env[61806]: DEBUG nova.virt.hardware [None req-a5c61942-d486-43f3-834d-7dfb87ef5b63 tempest-ServersNegativeTestJSON-891856253 tempest-ServersNegativeTestJSON-891856253-project-member] Image pref 0:0:0 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1036.011935] env[61806]: DEBUG nova.virt.hardware [None req-a5c61942-d486-43f3-834d-7dfb87ef5b63 tempest-ServersNegativeTestJSON-891856253 tempest-ServersNegativeTestJSON-891856253-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1036.011935] env[61806]: DEBUG nova.virt.hardware [None req-a5c61942-d486-43f3-834d-7dfb87ef5b63 tempest-ServersNegativeTestJSON-891856253 tempest-ServersNegativeTestJSON-891856253-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61806) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1036.015782] env[61806]: DEBUG nova.virt.hardware [None req-a5c61942-d486-43f3-834d-7dfb87ef5b63 tempest-ServersNegativeTestJSON-891856253 tempest-ServersNegativeTestJSON-891856253-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61806) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1036.015782] env[61806]: DEBUG nova.virt.hardware [None req-a5c61942-d486-43f3-834d-7dfb87ef5b63 tempest-ServersNegativeTestJSON-891856253 tempest-ServersNegativeTestJSON-891856253-project-member] Got 1 possible topologies {{(pid=61806) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1036.015782] env[61806]: DEBUG nova.virt.hardware [None req-a5c61942-d486-43f3-834d-7dfb87ef5b63 tempest-ServersNegativeTestJSON-891856253 tempest-ServersNegativeTestJSON-891856253-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61806) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1036.015782] env[61806]: DEBUG nova.virt.hardware [None req-a5c61942-d486-43f3-834d-7dfb87ef5b63 tempest-ServersNegativeTestJSON-891856253 tempest-ServersNegativeTestJSON-891856253-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61806) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1036.016655] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-87cf9979-7557-4803-9f5f-298f53b78d11 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1036.028423] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-29b796aa-ec88-4111-a1b5-908f30b96438 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1036.049963] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-a5c61942-d486-43f3-834d-7dfb87ef5b63 tempest-ServersNegativeTestJSON-891856253 tempest-ServersNegativeTestJSON-891856253-project-member] [instance: bdbaef31-7d37-40cb-ac3e-ad152b37e2d2] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:33:74:bb', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '2d94740a-bce8-4103-8ecf-230d02ec0a44', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'b8b714a4-b17c-4b3d-889e-407bec6c30ae', 'vif_model': 'vmxnet3'}] {{(pid=61806) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1036.062982] env[61806]: DEBUG oslo.service.loopingcall [None req-a5c61942-d486-43f3-834d-7dfb87ef5b63 tempest-ServersNegativeTestJSON-891856253 tempest-ServersNegativeTestJSON-891856253-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61806) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1036.063889] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: bdbaef31-7d37-40cb-ac3e-ad152b37e2d2] Creating VM on the ESX host {{(pid=61806) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 1036.064253] env[61806]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-02995fd5-57e9-45b0-8e8d-56ebe5f99443 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1036.090498] env[61806]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1036.090498] env[61806]: value = "task-1295142" [ 1036.090498] env[61806]: _type = "Task" [ 1036.090498] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1036.099606] env[61806]: DEBUG oslo_vmware.api [-] Task: {'id': task-1295142, 'name': CreateVM_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1036.258706] env[61806]: DEBUG nova.scheduler.client.report [None req-138f28bc-d88a-4f34-babf-797a70b177f1 tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Inventory has not changed for provider a2858be1-fd22-4e08-979e-87ad25293407 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 138, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61806) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1036.292426] env[61806]: DEBUG oslo_vmware.api [None req-42bedba7-0c5b-4f33-8173-c0388dea4ce8 tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] Task: {'id': task-1295140, 'name': PowerOffVM_Task, 'duration_secs': 0.187871} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1036.292784] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [None req-42bedba7-0c5b-4f33-8173-c0388dea4ce8 tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] [instance: 713d4e6e-6f0e-4844-9b55-cc010030eda6] Powered off the VM {{(pid=61806) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1036.293059] env[61806]: DEBUG nova.compute.manager [None req-42bedba7-0c5b-4f33-8173-c0388dea4ce8 tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] [instance: 713d4e6e-6f0e-4844-9b55-cc010030eda6] Checking state {{(pid=61806) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1036.294236] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e45eb7d9-ac0b-4a73-8041-3df51c955bbd {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1036.428358] env[61806]: DEBUG oslo_vmware.api [None req-439a5433-afde-41f1-8525-78229f3cf29f tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Task: {'id': task-1295141, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.506263} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1036.428725] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [None req-439a5433-afde-41f1-8525-78229f3cf29f tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa.vmdk to [datastore1] a61abfd9-3cc1-4a2b-b68b-a446094daf87/a61abfd9-3cc1-4a2b-b68b-a446094daf87.vmdk {{(pid=61806) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 1036.429031] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-439a5433-afde-41f1-8525-78229f3cf29f tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] [instance: a61abfd9-3cc1-4a2b-b68b-a446094daf87] Extending root virtual disk to 1048576 {{(pid=61806) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1036.429280] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-854ae121-dcd6-47ee-a508-d50ba49c0f58 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1036.439210] env[61806]: DEBUG oslo_vmware.api [None req-439a5433-afde-41f1-8525-78229f3cf29f tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Waiting for the task: (returnval){ [ 1036.439210] env[61806]: value = "task-1295143" [ 1036.439210] env[61806]: _type = "Task" [ 1036.439210] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1036.447236] env[61806]: DEBUG oslo_vmware.api [None req-439a5433-afde-41f1-8525-78229f3cf29f tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Task: {'id': task-1295143, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1036.601916] env[61806]: DEBUG oslo_vmware.api [-] Task: {'id': task-1295142, 'name': CreateVM_Task, 'duration_secs': 0.47632} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1036.602174] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: bdbaef31-7d37-40cb-ac3e-ad152b37e2d2] Created VM on the ESX host {{(pid=61806) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 1036.604930] env[61806]: DEBUG oslo_concurrency.lockutils [None req-a5c61942-d486-43f3-834d-7dfb87ef5b63 tempest-ServersNegativeTestJSON-891856253 tempest-ServersNegativeTestJSON-891856253-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/20c8ea6e-0db5-4788-9e34-92c1d63bb7b8" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1036.605143] env[61806]: DEBUG oslo_concurrency.lockutils [None req-a5c61942-d486-43f3-834d-7dfb87ef5b63 tempest-ServersNegativeTestJSON-891856253 tempest-ServersNegativeTestJSON-891856253-project-member] Acquired lock "[datastore1] devstack-image-cache_base/20c8ea6e-0db5-4788-9e34-92c1d63bb7b8" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1036.605629] env[61806]: DEBUG oslo_concurrency.lockutils [None req-a5c61942-d486-43f3-834d-7dfb87ef5b63 tempest-ServersNegativeTestJSON-891856253 tempest-ServersNegativeTestJSON-891856253-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/20c8ea6e-0db5-4788-9e34-92c1d63bb7b8" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1036.605893] env[61806]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-83682236-59d3-4813-a20d-642399ef9492 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1036.611080] env[61806]: DEBUG oslo_vmware.api [None req-a5c61942-d486-43f3-834d-7dfb87ef5b63 tempest-ServersNegativeTestJSON-891856253 tempest-ServersNegativeTestJSON-891856253-project-member] Waiting for the task: (returnval){ [ 1036.611080] env[61806]: value = "session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]5205bbb1-5b06-6620-a6e6-1cd7c79921d6" [ 1036.611080] env[61806]: _type = "Task" [ 1036.611080] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1036.620290] env[61806]: DEBUG oslo_vmware.api [None req-a5c61942-d486-43f3-834d-7dfb87ef5b63 tempest-ServersNegativeTestJSON-891856253 tempest-ServersNegativeTestJSON-891856253-project-member] Task: {'id': session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]5205bbb1-5b06-6620-a6e6-1cd7c79921d6, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1036.765231] env[61806]: DEBUG oslo_concurrency.lockutils [None req-138f28bc-d88a-4f34-babf-797a70b177f1 tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.212s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1036.765874] env[61806]: DEBUG nova.compute.manager [None req-138f28bc-d88a-4f34-babf-797a70b177f1 tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] [instance: 06658134-4c6d-4911-9a55-b6805f555c78] Start building networks asynchronously for instance. {{(pid=61806) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1036.769084] env[61806]: DEBUG oslo_concurrency.lockutils [None req-54021779-f9e2-42db-b76c-c250eacb9121 tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 6.765s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1036.771049] env[61806]: INFO nova.compute.claims [None req-54021779-f9e2-42db-b76c-c250eacb9121 tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] [instance: 7445d2c5-26bd-4f8d-8653-51c721ea801f] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1036.809581] env[61806]: DEBUG oslo_concurrency.lockutils [None req-42bedba7-0c5b-4f33-8173-c0388dea4ce8 tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] Lock "713d4e6e-6f0e-4844-9b55-cc010030eda6" "released" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: held 1.558s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1036.824761] env[61806]: DEBUG nova.network.neutron [req-6db4f388-ded0-41c4-933a-70c5211bc730 req-0f45f9d8-b4c0-4f15-8079-e220ca3175e6 service nova] [instance: bdbaef31-7d37-40cb-ac3e-ad152b37e2d2] Updated VIF entry in instance network info cache for port b8b714a4-b17c-4b3d-889e-407bec6c30ae. {{(pid=61806) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1036.825278] env[61806]: DEBUG nova.network.neutron [req-6db4f388-ded0-41c4-933a-70c5211bc730 req-0f45f9d8-b4c0-4f15-8079-e220ca3175e6 service nova] [instance: bdbaef31-7d37-40cb-ac3e-ad152b37e2d2] Updating instance_info_cache with network_info: [{"id": "b8b714a4-b17c-4b3d-889e-407bec6c30ae", "address": "fa:16:3e:33:74:bb", "network": {"id": "fd35ea93-8914-4b54-a4ea-f45fc750982b", "bridge": "br-int", "label": "tempest-ServersNegativeTestJSON-2117068882-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d2e2b7eec47e4b10898215123408692a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "2d94740a-bce8-4103-8ecf-230d02ec0a44", "external-id": "nsx-vlan-transportzone-149", "segmentation_id": 149, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb8b714a4-b1", "ovs_interfaceid": "b8b714a4-b17c-4b3d-889e-407bec6c30ae", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61806) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1036.949344] env[61806]: DEBUG oslo_vmware.api [None req-439a5433-afde-41f1-8525-78229f3cf29f tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Task: {'id': task-1295143, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.085259} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1036.949714] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-439a5433-afde-41f1-8525-78229f3cf29f tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] [instance: a61abfd9-3cc1-4a2b-b68b-a446094daf87] Extended root virtual disk {{(pid=61806) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1036.950584] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-81213edd-7ef9-48ab-b738-7c51b108451c {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1036.974512] env[61806]: DEBUG nova.virt.vmwareapi.volumeops [None req-439a5433-afde-41f1-8525-78229f3cf29f tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] [instance: a61abfd9-3cc1-4a2b-b68b-a446094daf87] Reconfiguring VM instance instance-00000065 to attach disk [datastore1] a61abfd9-3cc1-4a2b-b68b-a446094daf87/a61abfd9-3cc1-4a2b-b68b-a446094daf87.vmdk or device None with type sparse {{(pid=61806) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1036.974808] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-2d9b9e19-9484-4e8c-a011-e8ae25dd227a {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1036.995956] env[61806]: DEBUG oslo_vmware.api [None req-439a5433-afde-41f1-8525-78229f3cf29f tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Waiting for the task: (returnval){ [ 1036.995956] env[61806]: value = "task-1295144" [ 1036.995956] env[61806]: _type = "Task" [ 1036.995956] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1037.004112] env[61806]: DEBUG oslo_vmware.api [None req-439a5433-afde-41f1-8525-78229f3cf29f tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Task: {'id': task-1295144, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1037.121822] env[61806]: DEBUG oslo_concurrency.lockutils [None req-a5c61942-d486-43f3-834d-7dfb87ef5b63 tempest-ServersNegativeTestJSON-891856253 tempest-ServersNegativeTestJSON-891856253-project-member] Releasing lock "[datastore1] devstack-image-cache_base/20c8ea6e-0db5-4788-9e34-92c1d63bb7b8" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1037.122115] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-a5c61942-d486-43f3-834d-7dfb87ef5b63 tempest-ServersNegativeTestJSON-891856253 tempest-ServersNegativeTestJSON-891856253-project-member] [instance: bdbaef31-7d37-40cb-ac3e-ad152b37e2d2] Processing image 20c8ea6e-0db5-4788-9e34-92c1d63bb7b8 {{(pid=61806) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1037.122626] env[61806]: DEBUG oslo_concurrency.lockutils [None req-a5c61942-d486-43f3-834d-7dfb87ef5b63 tempest-ServersNegativeTestJSON-891856253 tempest-ServersNegativeTestJSON-891856253-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/20c8ea6e-0db5-4788-9e34-92c1d63bb7b8/20c8ea6e-0db5-4788-9e34-92c1d63bb7b8.vmdk" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1037.122626] env[61806]: DEBUG oslo_concurrency.lockutils [None req-a5c61942-d486-43f3-834d-7dfb87ef5b63 tempest-ServersNegativeTestJSON-891856253 tempest-ServersNegativeTestJSON-891856253-project-member] Acquired lock "[datastore1] devstack-image-cache_base/20c8ea6e-0db5-4788-9e34-92c1d63bb7b8/20c8ea6e-0db5-4788-9e34-92c1d63bb7b8.vmdk" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1037.122734] env[61806]: DEBUG nova.virt.vmwareapi.ds_util [None req-a5c61942-d486-43f3-834d-7dfb87ef5b63 tempest-ServersNegativeTestJSON-891856253 tempest-ServersNegativeTestJSON-891856253-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61806) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1037.122958] env[61806]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-20a92dc9-a159-455f-a42d-961058836649 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1037.131807] env[61806]: DEBUG nova.virt.vmwareapi.ds_util [None req-a5c61942-d486-43f3-834d-7dfb87ef5b63 tempest-ServersNegativeTestJSON-891856253 tempest-ServersNegativeTestJSON-891856253-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61806) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1037.131995] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-a5c61942-d486-43f3-834d-7dfb87ef5b63 tempest-ServersNegativeTestJSON-891856253 tempest-ServersNegativeTestJSON-891856253-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61806) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 1037.132735] env[61806]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-8f8ffec3-9850-4a14-8ac0-2a1be78ad353 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1037.138329] env[61806]: DEBUG oslo_vmware.api [None req-a5c61942-d486-43f3-834d-7dfb87ef5b63 tempest-ServersNegativeTestJSON-891856253 tempest-ServersNegativeTestJSON-891856253-project-member] Waiting for the task: (returnval){ [ 1037.138329] env[61806]: value = "session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]528e354a-9727-a5a4-3c92-9970f807e7e9" [ 1037.138329] env[61806]: _type = "Task" [ 1037.138329] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1037.147984] env[61806]: DEBUG oslo_vmware.api [None req-a5c61942-d486-43f3-834d-7dfb87ef5b63 tempest-ServersNegativeTestJSON-891856253 tempest-ServersNegativeTestJSON-891856253-project-member] Task: {'id': session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]528e354a-9727-a5a4-3c92-9970f807e7e9, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1037.276469] env[61806]: DEBUG nova.compute.utils [None req-138f28bc-d88a-4f34-babf-797a70b177f1 tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Using /dev/sd instead of None {{(pid=61806) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1037.278589] env[61806]: DEBUG nova.compute.manager [None req-138f28bc-d88a-4f34-babf-797a70b177f1 tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] [instance: 06658134-4c6d-4911-9a55-b6805f555c78] Allocating IP information in the background. {{(pid=61806) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1037.278791] env[61806]: DEBUG nova.network.neutron [None req-138f28bc-d88a-4f34-babf-797a70b177f1 tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] [instance: 06658134-4c6d-4911-9a55-b6805f555c78] allocate_for_instance() {{(pid=61806) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 1037.321707] env[61806]: DEBUG nova.policy [None req-138f28bc-d88a-4f34-babf-797a70b177f1 tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '921aeb4e90e64f89888a9c65f873d66b', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'c1f2d161c50f4b70987b90e6364c3779', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61806) authorize /opt/stack/nova/nova/policy.py:201}} [ 1037.327270] env[61806]: DEBUG oslo_concurrency.lockutils [req-6db4f388-ded0-41c4-933a-70c5211bc730 req-0f45f9d8-b4c0-4f15-8079-e220ca3175e6 service nova] Releasing lock "refresh_cache-bdbaef31-7d37-40cb-ac3e-ad152b37e2d2" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1037.505745] env[61806]: DEBUG oslo_vmware.api [None req-439a5433-afde-41f1-8525-78229f3cf29f tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Task: {'id': task-1295144, 'name': ReconfigVM_Task, 'duration_secs': 0.31533} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1037.506172] env[61806]: DEBUG nova.virt.vmwareapi.volumeops [None req-439a5433-afde-41f1-8525-78229f3cf29f tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] [instance: a61abfd9-3cc1-4a2b-b68b-a446094daf87] Reconfigured VM instance instance-00000065 to attach disk [datastore1] a61abfd9-3cc1-4a2b-b68b-a446094daf87/a61abfd9-3cc1-4a2b-b68b-a446094daf87.vmdk or device None with type sparse {{(pid=61806) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1037.508063] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-6297f9a1-e19a-401d-b51f-a3432c7b1744 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1037.512305] env[61806]: DEBUG oslo_concurrency.lockutils [None req-c257ef4d-1b9a-4377-aada-40cc9a4843fd tempest-AttachVolumeNegativeTest-1683033704 tempest-AttachVolumeNegativeTest-1683033704-project-member] Acquiring lock "ce2ad097-4c3f-4b55-b018-ac788aa5b662" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1037.512499] env[61806]: DEBUG oslo_concurrency.lockutils [None req-c257ef4d-1b9a-4377-aada-40cc9a4843fd tempest-AttachVolumeNegativeTest-1683033704 tempest-AttachVolumeNegativeTest-1683033704-project-member] Lock "ce2ad097-4c3f-4b55-b018-ac788aa5b662" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1037.521349] env[61806]: DEBUG oslo_vmware.api [None req-439a5433-afde-41f1-8525-78229f3cf29f tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Waiting for the task: (returnval){ [ 1037.521349] env[61806]: value = "task-1295145" [ 1037.521349] env[61806]: _type = "Task" [ 1037.521349] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1037.530873] env[61806]: DEBUG oslo_vmware.api [None req-439a5433-afde-41f1-8525-78229f3cf29f tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Task: {'id': task-1295145, 'name': Rename_Task} progress is 5%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1037.633820] env[61806]: DEBUG nova.network.neutron [None req-138f28bc-d88a-4f34-babf-797a70b177f1 tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] [instance: 06658134-4c6d-4911-9a55-b6805f555c78] Successfully created port: 783dae1f-d1c0-4958-b039-ed57b202b64a {{(pid=61806) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1037.649449] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-a5c61942-d486-43f3-834d-7dfb87ef5b63 tempest-ServersNegativeTestJSON-891856253 tempest-ServersNegativeTestJSON-891856253-project-member] [instance: bdbaef31-7d37-40cb-ac3e-ad152b37e2d2] Preparing fetch location {{(pid=61806) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:633}} [ 1037.649574] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-a5c61942-d486-43f3-834d-7dfb87ef5b63 tempest-ServersNegativeTestJSON-891856253 tempest-ServersNegativeTestJSON-891856253-project-member] [instance: bdbaef31-7d37-40cb-ac3e-ad152b37e2d2] Fetch image to [datastore1] OSTACK_IMG_b3afb79f-96af-441f-a8dc-c28cd8d6a22d/OSTACK_IMG_b3afb79f-96af-441f-a8dc-c28cd8d6a22d.vmdk {{(pid=61806) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:635}} [ 1037.649755] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-a5c61942-d486-43f3-834d-7dfb87ef5b63 tempest-ServersNegativeTestJSON-891856253 tempest-ServersNegativeTestJSON-891856253-project-member] [instance: bdbaef31-7d37-40cb-ac3e-ad152b37e2d2] Downloading stream optimized image 20c8ea6e-0db5-4788-9e34-92c1d63bb7b8 to [datastore1] OSTACK_IMG_b3afb79f-96af-441f-a8dc-c28cd8d6a22d/OSTACK_IMG_b3afb79f-96af-441f-a8dc-c28cd8d6a22d.vmdk on the data store datastore1 as vApp {{(pid=61806) _fetch_image_as_vapp /opt/stack/nova/nova/virt/vmwareapi/vmops.py:437}} [ 1037.649927] env[61806]: DEBUG nova.virt.vmwareapi.images [None req-a5c61942-d486-43f3-834d-7dfb87ef5b63 tempest-ServersNegativeTestJSON-891856253 tempest-ServersNegativeTestJSON-891856253-project-member] [instance: bdbaef31-7d37-40cb-ac3e-ad152b37e2d2] Downloading image file data 20c8ea6e-0db5-4788-9e34-92c1d63bb7b8 to the ESX as VM named 'OSTACK_IMG_b3afb79f-96af-441f-a8dc-c28cd8d6a22d' {{(pid=61806) fetch_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:336}} [ 1037.737897] env[61806]: DEBUG oslo_vmware.rw_handles [None req-a5c61942-d486-43f3-834d-7dfb87ef5b63 tempest-ServersNegativeTestJSON-891856253 tempest-ServersNegativeTestJSON-891856253-project-member] Creating HttpNfcLease lease for vApp import into resource pool: (val){ [ 1037.737897] env[61806]: value = "resgroup-9" [ 1037.737897] env[61806]: _type = "ResourcePool" [ 1037.737897] env[61806]: }. {{(pid=61806) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:453}} [ 1037.738221] env[61806]: DEBUG oslo_vmware.service [-] Invoking ResourcePool.ImportVApp with opID=oslo.vmware-3797fae0-0017-47c2-a4e5-ff2b218f5fea {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1037.761438] env[61806]: DEBUG oslo_vmware.rw_handles [None req-a5c61942-d486-43f3-834d-7dfb87ef5b63 tempest-ServersNegativeTestJSON-891856253 tempest-ServersNegativeTestJSON-891856253-project-member] Lease: (returnval){ [ 1037.761438] env[61806]: value = "session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]52ca8d24-e431-b01a-384b-74121f2819a3" [ 1037.761438] env[61806]: _type = "HttpNfcLease" [ 1037.761438] env[61806]: } obtained for vApp import into resource pool (val){ [ 1037.761438] env[61806]: value = "resgroup-9" [ 1037.761438] env[61806]: _type = "ResourcePool" [ 1037.761438] env[61806]: }. {{(pid=61806) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:461}} [ 1037.761694] env[61806]: DEBUG oslo_vmware.api [None req-a5c61942-d486-43f3-834d-7dfb87ef5b63 tempest-ServersNegativeTestJSON-891856253 tempest-ServersNegativeTestJSON-891856253-project-member] Waiting for the lease: (returnval){ [ 1037.761694] env[61806]: value = "session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]52ca8d24-e431-b01a-384b-74121f2819a3" [ 1037.761694] env[61806]: _type = "HttpNfcLease" [ 1037.761694] env[61806]: } to be ready. {{(pid=61806) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 1037.768479] env[61806]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1037.768479] env[61806]: value = "session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]52ca8d24-e431-b01a-384b-74121f2819a3" [ 1037.768479] env[61806]: _type = "HttpNfcLease" [ 1037.768479] env[61806]: } is initializing. {{(pid=61806) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 1037.784114] env[61806]: DEBUG nova.compute.manager [None req-138f28bc-d88a-4f34-babf-797a70b177f1 tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] [instance: 06658134-4c6d-4911-9a55-b6805f555c78] Start building block device mappings for instance. {{(pid=61806) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1037.917589] env[61806]: DEBUG oslo_concurrency.lockutils [None req-5eac2f24-7650-441c-94d5-4e1470c88591 tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] Acquiring lock "713d4e6e-6f0e-4844-9b55-cc010030eda6" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1037.917805] env[61806]: DEBUG oslo_concurrency.lockutils [None req-5eac2f24-7650-441c-94d5-4e1470c88591 tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] Lock "713d4e6e-6f0e-4844-9b55-cc010030eda6" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1037.917907] env[61806]: DEBUG oslo_concurrency.lockutils [None req-5eac2f24-7650-441c-94d5-4e1470c88591 tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] Acquiring lock "713d4e6e-6f0e-4844-9b55-cc010030eda6-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1037.918135] env[61806]: DEBUG oslo_concurrency.lockutils [None req-5eac2f24-7650-441c-94d5-4e1470c88591 tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] Lock "713d4e6e-6f0e-4844-9b55-cc010030eda6-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1037.918316] env[61806]: DEBUG oslo_concurrency.lockutils [None req-5eac2f24-7650-441c-94d5-4e1470c88591 tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] Lock "713d4e6e-6f0e-4844-9b55-cc010030eda6-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1037.921557] env[61806]: INFO nova.compute.manager [None req-5eac2f24-7650-441c-94d5-4e1470c88591 tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] [instance: 713d4e6e-6f0e-4844-9b55-cc010030eda6] Terminating instance [ 1037.931742] env[61806]: DEBUG nova.compute.manager [None req-5eac2f24-7650-441c-94d5-4e1470c88591 tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] [instance: 713d4e6e-6f0e-4844-9b55-cc010030eda6] Start destroying the instance on the hypervisor. {{(pid=61806) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1037.932012] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-5eac2f24-7650-441c-94d5-4e1470c88591 tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] [instance: 713d4e6e-6f0e-4844-9b55-cc010030eda6] Destroying instance {{(pid=61806) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1037.932923] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cab42c1b-43d4-4eb1-946b-d54cd1936480 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1037.938157] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-000252e2-f39f-47d6-84d7-f34ae229828c {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1037.947026] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-5eac2f24-7650-441c-94d5-4e1470c88591 tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] [instance: 713d4e6e-6f0e-4844-9b55-cc010030eda6] Unregistering the VM {{(pid=61806) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1037.949855] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-b5a343a8-0fdb-4a79-a4da-1f5f7836230c {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1037.953729] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eedfe022-4c61-4619-bd43-22367cbef5d3 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1037.990752] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d7031fc3-b307-4cfb-995e-8c0f1bd6569d {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1037.998509] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e4b430dc-6898-46ff-99fb-8266cf4a0663 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1038.012880] env[61806]: DEBUG nova.compute.provider_tree [None req-54021779-f9e2-42db-b76c-c250eacb9121 tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Inventory has not changed in ProviderTree for provider: a2858be1-fd22-4e08-979e-87ad25293407 {{(pid=61806) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1038.014869] env[61806]: DEBUG nova.compute.manager [None req-c257ef4d-1b9a-4377-aada-40cc9a4843fd tempest-AttachVolumeNegativeTest-1683033704 tempest-AttachVolumeNegativeTest-1683033704-project-member] [instance: ce2ad097-4c3f-4b55-b018-ac788aa5b662] Starting instance... {{(pid=61806) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1038.027975] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-5eac2f24-7650-441c-94d5-4e1470c88591 tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] [instance: 713d4e6e-6f0e-4844-9b55-cc010030eda6] Unregistered the VM {{(pid=61806) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1038.028447] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-5eac2f24-7650-441c-94d5-4e1470c88591 tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] [instance: 713d4e6e-6f0e-4844-9b55-cc010030eda6] Deleting contents of the VM from datastore datastore1 {{(pid=61806) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1038.028735] env[61806]: DEBUG nova.virt.vmwareapi.ds_util [None req-5eac2f24-7650-441c-94d5-4e1470c88591 tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] Deleting the datastore file [datastore1] 713d4e6e-6f0e-4844-9b55-cc010030eda6 {{(pid=61806) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1038.029504] env[61806]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-667c162b-2d22-4a86-9d16-89bdd88d5c10 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1038.036115] env[61806]: DEBUG oslo_vmware.api [None req-439a5433-afde-41f1-8525-78229f3cf29f tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Task: {'id': task-1295145, 'name': Rename_Task, 'duration_secs': 0.14857} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1038.036442] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [None req-439a5433-afde-41f1-8525-78229f3cf29f tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] [instance: a61abfd9-3cc1-4a2b-b68b-a446094daf87] Powering on the VM {{(pid=61806) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1038.036763] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-1b636b83-02e0-4210-b633-c5c3a9850922 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1038.040610] env[61806]: DEBUG oslo_vmware.api [None req-5eac2f24-7650-441c-94d5-4e1470c88591 tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] Waiting for the task: (returnval){ [ 1038.040610] env[61806]: value = "task-1295148" [ 1038.040610] env[61806]: _type = "Task" [ 1038.040610] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1038.045492] env[61806]: DEBUG oslo_vmware.api [None req-439a5433-afde-41f1-8525-78229f3cf29f tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Waiting for the task: (returnval){ [ 1038.045492] env[61806]: value = "task-1295149" [ 1038.045492] env[61806]: _type = "Task" [ 1038.045492] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1038.052040] env[61806]: DEBUG oslo_vmware.api [None req-5eac2f24-7650-441c-94d5-4e1470c88591 tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] Task: {'id': task-1295148, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1038.057974] env[61806]: DEBUG oslo_vmware.api [None req-439a5433-afde-41f1-8525-78229f3cf29f tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Task: {'id': task-1295149, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1038.270726] env[61806]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1038.270726] env[61806]: value = "session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]52ca8d24-e431-b01a-384b-74121f2819a3" [ 1038.270726] env[61806]: _type = "HttpNfcLease" [ 1038.270726] env[61806]: } is initializing. {{(pid=61806) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 1038.515834] env[61806]: DEBUG nova.scheduler.client.report [None req-54021779-f9e2-42db-b76c-c250eacb9121 tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Inventory has not changed for provider a2858be1-fd22-4e08-979e-87ad25293407 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 138, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61806) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1038.536500] env[61806]: DEBUG oslo_concurrency.lockutils [None req-c257ef4d-1b9a-4377-aada-40cc9a4843fd tempest-AttachVolumeNegativeTest-1683033704 tempest-AttachVolumeNegativeTest-1683033704-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1038.553305] env[61806]: DEBUG oslo_vmware.api [None req-5eac2f24-7650-441c-94d5-4e1470c88591 tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] Task: {'id': task-1295148, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.186703} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1038.553988] env[61806]: DEBUG nova.virt.vmwareapi.ds_util [None req-5eac2f24-7650-441c-94d5-4e1470c88591 tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] Deleted the datastore file {{(pid=61806) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1038.554209] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-5eac2f24-7650-441c-94d5-4e1470c88591 tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] [instance: 713d4e6e-6f0e-4844-9b55-cc010030eda6] Deleted contents of the VM from datastore datastore1 {{(pid=61806) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1038.554404] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-5eac2f24-7650-441c-94d5-4e1470c88591 tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] [instance: 713d4e6e-6f0e-4844-9b55-cc010030eda6] Instance destroyed {{(pid=61806) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1038.554580] env[61806]: INFO nova.compute.manager [None req-5eac2f24-7650-441c-94d5-4e1470c88591 tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] [instance: 713d4e6e-6f0e-4844-9b55-cc010030eda6] Took 0.62 seconds to destroy the instance on the hypervisor. [ 1038.554839] env[61806]: DEBUG oslo.service.loopingcall [None req-5eac2f24-7650-441c-94d5-4e1470c88591 tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61806) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1038.555065] env[61806]: DEBUG nova.compute.manager [-] [instance: 713d4e6e-6f0e-4844-9b55-cc010030eda6] Deallocating network for instance {{(pid=61806) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1038.555196] env[61806]: DEBUG nova.network.neutron [-] [instance: 713d4e6e-6f0e-4844-9b55-cc010030eda6] deallocate_for_instance() {{(pid=61806) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1038.559682] env[61806]: DEBUG oslo_vmware.api [None req-439a5433-afde-41f1-8525-78229f3cf29f tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Task: {'id': task-1295149, 'name': PowerOnVM_Task, 'duration_secs': 0.504559} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1038.560222] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [None req-439a5433-afde-41f1-8525-78229f3cf29f tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] [instance: a61abfd9-3cc1-4a2b-b68b-a446094daf87] Powered on the VM {{(pid=61806) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1038.560436] env[61806]: INFO nova.compute.manager [None req-439a5433-afde-41f1-8525-78229f3cf29f tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] [instance: a61abfd9-3cc1-4a2b-b68b-a446094daf87] Took 6.94 seconds to spawn the instance on the hypervisor. [ 1038.560624] env[61806]: DEBUG nova.compute.manager [None req-439a5433-afde-41f1-8525-78229f3cf29f tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] [instance: a61abfd9-3cc1-4a2b-b68b-a446094daf87] Checking state {{(pid=61806) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1038.561404] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-beca55e5-4050-414e-8131-cbf5387ef818 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1038.773435] env[61806]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1038.773435] env[61806]: value = "session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]52ca8d24-e431-b01a-384b-74121f2819a3" [ 1038.773435] env[61806]: _type = "HttpNfcLease" [ 1038.773435] env[61806]: } is ready. {{(pid=61806) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 1038.773856] env[61806]: DEBUG oslo_vmware.rw_handles [None req-a5c61942-d486-43f3-834d-7dfb87ef5b63 tempest-ServersNegativeTestJSON-891856253 tempest-ServersNegativeTestJSON-891856253-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 1038.773856] env[61806]: value = "session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]52ca8d24-e431-b01a-384b-74121f2819a3" [ 1038.773856] env[61806]: _type = "HttpNfcLease" [ 1038.773856] env[61806]: }. {{(pid=61806) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:467}} [ 1038.774906] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f427775b-9368-44b6-a146-aa6b0918a80b {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1038.781197] env[61806]: DEBUG nova.compute.manager [req-a24f1d67-86bc-4444-8587-9b2e80806ed3 req-f1dc9b73-684d-415a-aef1-453a8e3aee68 service nova] [instance: 713d4e6e-6f0e-4844-9b55-cc010030eda6] Received event network-vif-deleted-7828f677-3fd0-441f-ac22-2b837a4c5d54 {{(pid=61806) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1038.781414] env[61806]: INFO nova.compute.manager [req-a24f1d67-86bc-4444-8587-9b2e80806ed3 req-f1dc9b73-684d-415a-aef1-453a8e3aee68 service nova] [instance: 713d4e6e-6f0e-4844-9b55-cc010030eda6] Neutron deleted interface 7828f677-3fd0-441f-ac22-2b837a4c5d54; detaching it from the instance and deleting it from the info cache [ 1038.781604] env[61806]: DEBUG nova.network.neutron [req-a24f1d67-86bc-4444-8587-9b2e80806ed3 req-f1dc9b73-684d-415a-aef1-453a8e3aee68 service nova] [instance: 713d4e6e-6f0e-4844-9b55-cc010030eda6] Updating instance_info_cache with network_info: [] {{(pid=61806) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1038.786134] env[61806]: DEBUG oslo_vmware.rw_handles [None req-a5c61942-d486-43f3-834d-7dfb87ef5b63 tempest-ServersNegativeTestJSON-891856253 tempest-ServersNegativeTestJSON-891856253-project-member] Found VMDK URL: https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52a2032f-a228-1c10-99ea-4211bd18d59d/disk-0.vmdk from lease info. {{(pid=61806) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 1038.786459] env[61806]: DEBUG oslo_vmware.rw_handles [None req-a5c61942-d486-43f3-834d-7dfb87ef5b63 tempest-ServersNegativeTestJSON-891856253 tempest-ServersNegativeTestJSON-891856253-project-member] Creating HTTP connection to write to file with size = 31667200 and URL = https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52a2032f-a228-1c10-99ea-4211bd18d59d/disk-0.vmdk. {{(pid=61806) _create_write_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:122}} [ 1038.843339] env[61806]: DEBUG nova.compute.manager [None req-138f28bc-d88a-4f34-babf-797a70b177f1 tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] [instance: 06658134-4c6d-4911-9a55-b6805f555c78] Start spawning the instance on the hypervisor. {{(pid=61806) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1038.851284] env[61806]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-baf7948c-8056-460e-aab8-abe073eb9b8f {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1038.869596] env[61806]: DEBUG nova.virt.hardware [None req-138f28bc-d88a-4f34-babf-797a70b177f1 tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-15T18:05:10Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-15T18:04:52Z,direct_url=,disk_format='vmdk',id=ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='84ba83fce2c349a6988173c5d6fc3b07',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-15T18:04:53Z,virtual_size=,visibility=), allow threads: False {{(pid=61806) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1038.869876] env[61806]: DEBUG nova.virt.hardware [None req-138f28bc-d88a-4f34-babf-797a70b177f1 tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Flavor limits 0:0:0 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1038.870060] env[61806]: DEBUG nova.virt.hardware [None req-138f28bc-d88a-4f34-babf-797a70b177f1 tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Image limits 0:0:0 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1038.870259] env[61806]: DEBUG nova.virt.hardware [None req-138f28bc-d88a-4f34-babf-797a70b177f1 tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Flavor pref 0:0:0 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1038.870441] env[61806]: DEBUG nova.virt.hardware [None req-138f28bc-d88a-4f34-babf-797a70b177f1 tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Image pref 0:0:0 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1038.870602] env[61806]: DEBUG nova.virt.hardware [None req-138f28bc-d88a-4f34-babf-797a70b177f1 tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1038.870815] env[61806]: DEBUG nova.virt.hardware [None req-138f28bc-d88a-4f34-babf-797a70b177f1 tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61806) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1038.870995] env[61806]: DEBUG nova.virt.hardware [None req-138f28bc-d88a-4f34-babf-797a70b177f1 tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61806) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1038.871175] env[61806]: DEBUG nova.virt.hardware [None req-138f28bc-d88a-4f34-babf-797a70b177f1 tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Got 1 possible topologies {{(pid=61806) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1038.871345] env[61806]: DEBUG nova.virt.hardware [None req-138f28bc-d88a-4f34-babf-797a70b177f1 tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61806) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1038.871526] env[61806]: DEBUG nova.virt.hardware [None req-138f28bc-d88a-4f34-babf-797a70b177f1 tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61806) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1038.872660] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1d941742-0c22-4536-bfed-badc451e8984 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1038.880975] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-896dd529-c35b-46d2-9ad4-7e1ee5a5d0e0 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1039.020722] env[61806]: DEBUG oslo_concurrency.lockutils [None req-54021779-f9e2-42db-b76c-c250eacb9121 tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.252s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1039.021287] env[61806]: DEBUG nova.compute.manager [None req-54021779-f9e2-42db-b76c-c250eacb9121 tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] [instance: 7445d2c5-26bd-4f8d-8653-51c721ea801f] Start building networks asynchronously for instance. {{(pid=61806) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1039.023842] env[61806]: DEBUG oslo_concurrency.lockutils [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 6.105s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1039.024737] env[61806]: DEBUG oslo_concurrency.lockutils [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1039.024922] env[61806]: DEBUG nova.compute.resource_tracker [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=61806) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 1039.025330] env[61806]: DEBUG oslo_concurrency.lockutils [None req-c257ef4d-1b9a-4377-aada-40cc9a4843fd tempest-AttachVolumeNegativeTest-1683033704 tempest-AttachVolumeNegativeTest-1683033704-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.489s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1039.026914] env[61806]: INFO nova.compute.claims [None req-c257ef4d-1b9a-4377-aada-40cc9a4843fd tempest-AttachVolumeNegativeTest-1683033704 tempest-AttachVolumeNegativeTest-1683033704-project-member] [instance: ce2ad097-4c3f-4b55-b018-ac788aa5b662] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1039.030494] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-45b86a13-b4f6-4899-b337-66d5861aeb1c {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1039.041174] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aba1b685-d8b0-49fd-9c92-1127fa1c6e6b {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1039.057591] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0dcbb54b-0375-43b4-b9b7-4616587103d5 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1039.065733] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e2aecd61-c338-4565-ac58-ebb37b4fe358 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1039.104471] env[61806]: DEBUG nova.compute.resource_tracker [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=180598MB free_disk=138GB free_vcpus=48 pci_devices=None {{(pid=61806) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 1039.104637] env[61806]: DEBUG oslo_concurrency.lockutils [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1039.107130] env[61806]: INFO nova.compute.manager [None req-439a5433-afde-41f1-8525-78229f3cf29f tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] [instance: a61abfd9-3cc1-4a2b-b68b-a446094daf87] Took 15.40 seconds to build instance. [ 1039.257639] env[61806]: DEBUG nova.network.neutron [-] [instance: 713d4e6e-6f0e-4844-9b55-cc010030eda6] Updating instance_info_cache with network_info: [] {{(pid=61806) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1039.290830] env[61806]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-116707af-efa4-424a-85b5-22e57d3a2873 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1039.301163] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-67d8c27b-a4ce-456d-9c1d-2e6cdbbd81c0 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1039.330581] env[61806]: DEBUG nova.compute.manager [req-a24f1d67-86bc-4444-8587-9b2e80806ed3 req-f1dc9b73-684d-415a-aef1-453a8e3aee68 service nova] [instance: 713d4e6e-6f0e-4844-9b55-cc010030eda6] Detach interface failed, port_id=7828f677-3fd0-441f-ac22-2b837a4c5d54, reason: Instance 713d4e6e-6f0e-4844-9b55-cc010030eda6 could not be found. {{(pid=61806) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10963}} [ 1039.385825] env[61806]: DEBUG nova.network.neutron [None req-138f28bc-d88a-4f34-babf-797a70b177f1 tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] [instance: 06658134-4c6d-4911-9a55-b6805f555c78] Successfully updated port: 783dae1f-d1c0-4958-b039-ed57b202b64a {{(pid=61806) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1039.532163] env[61806]: DEBUG nova.compute.utils [None req-54021779-f9e2-42db-b76c-c250eacb9121 tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Using /dev/sd instead of None {{(pid=61806) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1039.534026] env[61806]: DEBUG nova.compute.manager [None req-54021779-f9e2-42db-b76c-c250eacb9121 tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] [instance: 7445d2c5-26bd-4f8d-8653-51c721ea801f] Allocating IP information in the background. {{(pid=61806) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1039.534245] env[61806]: DEBUG nova.network.neutron [None req-54021779-f9e2-42db-b76c-c250eacb9121 tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] [instance: 7445d2c5-26bd-4f8d-8653-51c721ea801f] allocate_for_instance() {{(pid=61806) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 1039.601215] env[61806]: DEBUG nova.policy [None req-54021779-f9e2-42db-b76c-c250eacb9121 tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '2b86d596e89645c2ab31332afdcbc3c2', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '7ab95fb72b5d46c3b6c7bebbccf897cd', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61806) authorize /opt/stack/nova/nova/policy.py:201}} [ 1039.609557] env[61806]: DEBUG oslo_concurrency.lockutils [None req-439a5433-afde-41f1-8525-78229f3cf29f tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Lock "a61abfd9-3cc1-4a2b-b68b-a446094daf87" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 16.911s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1039.760205] env[61806]: INFO nova.compute.manager [-] [instance: 713d4e6e-6f0e-4844-9b55-cc010030eda6] Took 1.20 seconds to deallocate network for instance. [ 1039.886510] env[61806]: DEBUG nova.network.neutron [None req-54021779-f9e2-42db-b76c-c250eacb9121 tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] [instance: 7445d2c5-26bd-4f8d-8653-51c721ea801f] Successfully created port: 773e4470-b54c-4449-b825-5bb680cf8d2f {{(pid=61806) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1039.889469] env[61806]: DEBUG oslo_concurrency.lockutils [None req-138f28bc-d88a-4f34-babf-797a70b177f1 tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Acquiring lock "refresh_cache-06658134-4c6d-4911-9a55-b6805f555c78" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1039.889640] env[61806]: DEBUG oslo_concurrency.lockutils [None req-138f28bc-d88a-4f34-babf-797a70b177f1 tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Acquired lock "refresh_cache-06658134-4c6d-4911-9a55-b6805f555c78" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1039.889791] env[61806]: DEBUG nova.network.neutron [None req-138f28bc-d88a-4f34-babf-797a70b177f1 tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] [instance: 06658134-4c6d-4911-9a55-b6805f555c78] Building network info cache for instance {{(pid=61806) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1040.037575] env[61806]: DEBUG nova.compute.manager [None req-54021779-f9e2-42db-b76c-c250eacb9121 tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] [instance: 7445d2c5-26bd-4f8d-8653-51c721ea801f] Start building block device mappings for instance. {{(pid=61806) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1040.187280] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b086dc59-5e60-434e-831e-e98e148d4967 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1040.199955] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e1bf69c4-dbc6-4716-9185-6d7a660aba87 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1040.239704] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-92a25eaa-eaa8-4011-8134-edfeb1f660e9 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1040.251992] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ae4e8ef4-0dcd-4c57-8ae6-9bc4504b7bdc {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1040.269358] env[61806]: DEBUG oslo_concurrency.lockutils [None req-5eac2f24-7650-441c-94d5-4e1470c88591 tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1040.269896] env[61806]: DEBUG nova.compute.provider_tree [None req-c257ef4d-1b9a-4377-aada-40cc9a4843fd tempest-AttachVolumeNegativeTest-1683033704 tempest-AttachVolumeNegativeTest-1683033704-project-member] Inventory has not changed in ProviderTree for provider: a2858be1-fd22-4e08-979e-87ad25293407 {{(pid=61806) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1040.306705] env[61806]: DEBUG oslo_concurrency.lockutils [None req-4da0f2fc-feb0-4b02-97f3-bd412b867242 tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Acquiring lock "a61abfd9-3cc1-4a2b-b68b-a446094daf87" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1040.307251] env[61806]: DEBUG oslo_concurrency.lockutils [None req-4da0f2fc-feb0-4b02-97f3-bd412b867242 tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Lock "a61abfd9-3cc1-4a2b-b68b-a446094daf87" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1040.307309] env[61806]: DEBUG oslo_concurrency.lockutils [None req-4da0f2fc-feb0-4b02-97f3-bd412b867242 tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Acquiring lock "a61abfd9-3cc1-4a2b-b68b-a446094daf87-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1040.307549] env[61806]: DEBUG oslo_concurrency.lockutils [None req-4da0f2fc-feb0-4b02-97f3-bd412b867242 tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Lock "a61abfd9-3cc1-4a2b-b68b-a446094daf87-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1040.307738] env[61806]: DEBUG oslo_concurrency.lockutils [None req-4da0f2fc-feb0-4b02-97f3-bd412b867242 tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Lock "a61abfd9-3cc1-4a2b-b68b-a446094daf87-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1040.310040] env[61806]: INFO nova.compute.manager [None req-4da0f2fc-feb0-4b02-97f3-bd412b867242 tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] [instance: a61abfd9-3cc1-4a2b-b68b-a446094daf87] Terminating instance [ 1040.312068] env[61806]: DEBUG nova.compute.manager [None req-4da0f2fc-feb0-4b02-97f3-bd412b867242 tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] [instance: a61abfd9-3cc1-4a2b-b68b-a446094daf87] Start destroying the instance on the hypervisor. {{(pid=61806) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1040.312322] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-4da0f2fc-feb0-4b02-97f3-bd412b867242 tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] [instance: a61abfd9-3cc1-4a2b-b68b-a446094daf87] Destroying instance {{(pid=61806) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1040.313344] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cc8e0e18-d615-4432-9d69-b059914a4502 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1040.324774] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [None req-4da0f2fc-feb0-4b02-97f3-bd412b867242 tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] [instance: a61abfd9-3cc1-4a2b-b68b-a446094daf87] Powering off the VM {{(pid=61806) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1040.326800] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-df7f3eda-c662-42c7-b931-f862e8551f0b {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1040.334480] env[61806]: DEBUG oslo_vmware.api [None req-4da0f2fc-feb0-4b02-97f3-bd412b867242 tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Waiting for the task: (returnval){ [ 1040.334480] env[61806]: value = "task-1295150" [ 1040.334480] env[61806]: _type = "Task" [ 1040.334480] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1040.345804] env[61806]: DEBUG oslo_vmware.api [None req-4da0f2fc-feb0-4b02-97f3-bd412b867242 tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Task: {'id': task-1295150, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1040.429147] env[61806]: DEBUG nova.network.neutron [None req-138f28bc-d88a-4f34-babf-797a70b177f1 tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] [instance: 06658134-4c6d-4911-9a55-b6805f555c78] Instance cache missing network info. {{(pid=61806) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 1040.597055] env[61806]: DEBUG nova.network.neutron [None req-138f28bc-d88a-4f34-babf-797a70b177f1 tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] [instance: 06658134-4c6d-4911-9a55-b6805f555c78] Updating instance_info_cache with network_info: [{"id": "783dae1f-d1c0-4958-b039-ed57b202b64a", "address": "fa:16:3e:89:24:35", "network": {"id": "746a03b6-c1a6-4d27-98b9-0bee6376634f", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-1971382587-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c1f2d161c50f4b70987b90e6364c3779", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6a31b224-77d7-48a4-af87-312758611b8e", "external-id": "nsx-vlan-transportzone-761", "segmentation_id": 761, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap783dae1f-d1", "ovs_interfaceid": "783dae1f-d1c0-4958-b039-ed57b202b64a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61806) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1040.667191] env[61806]: DEBUG oslo_vmware.rw_handles [None req-a5c61942-d486-43f3-834d-7dfb87ef5b63 tempest-ServersNegativeTestJSON-891856253 tempest-ServersNegativeTestJSON-891856253-project-member] Completed reading data from the image iterator. {{(pid=61806) read /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:765}} [ 1040.667463] env[61806]: DEBUG oslo_vmware.rw_handles [None req-a5c61942-d486-43f3-834d-7dfb87ef5b63 tempest-ServersNegativeTestJSON-891856253 tempest-ServersNegativeTestJSON-891856253-project-member] Getting lease state for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52a2032f-a228-1c10-99ea-4211bd18d59d/disk-0.vmdk. {{(pid=61806) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 1040.668441] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a2c3a8ab-aa77-4197-9643-9fc46373107c {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1040.674888] env[61806]: DEBUG oslo_vmware.rw_handles [None req-a5c61942-d486-43f3-834d-7dfb87ef5b63 tempest-ServersNegativeTestJSON-891856253 tempest-ServersNegativeTestJSON-891856253-project-member] Lease for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52a2032f-a228-1c10-99ea-4211bd18d59d/disk-0.vmdk is in state: ready. {{(pid=61806) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 1040.675139] env[61806]: DEBUG oslo_vmware.rw_handles [None req-a5c61942-d486-43f3-834d-7dfb87ef5b63 tempest-ServersNegativeTestJSON-891856253 tempest-ServersNegativeTestJSON-891856253-project-member] Releasing lease for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52a2032f-a228-1c10-99ea-4211bd18d59d/disk-0.vmdk. {{(pid=61806) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:440}} [ 1040.675352] env[61806]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseComplete with opID=oslo.vmware-aad68d2c-f62d-4957-b1b8-cf80ed8ac9c9 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1040.776142] env[61806]: DEBUG nova.scheduler.client.report [None req-c257ef4d-1b9a-4377-aada-40cc9a4843fd tempest-AttachVolumeNegativeTest-1683033704 tempest-AttachVolumeNegativeTest-1683033704-project-member] Inventory has not changed for provider a2858be1-fd22-4e08-979e-87ad25293407 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 138, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61806) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1040.806272] env[61806]: DEBUG nova.compute.manager [req-ae44934b-7a0e-4a63-af7c-8831b2c897b1 req-ca2c6373-de8d-4112-a681-fd3295fa03a2 service nova] [instance: 06658134-4c6d-4911-9a55-b6805f555c78] Received event network-vif-plugged-783dae1f-d1c0-4958-b039-ed57b202b64a {{(pid=61806) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1040.806512] env[61806]: DEBUG oslo_concurrency.lockutils [req-ae44934b-7a0e-4a63-af7c-8831b2c897b1 req-ca2c6373-de8d-4112-a681-fd3295fa03a2 service nova] Acquiring lock "06658134-4c6d-4911-9a55-b6805f555c78-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1040.806724] env[61806]: DEBUG oslo_concurrency.lockutils [req-ae44934b-7a0e-4a63-af7c-8831b2c897b1 req-ca2c6373-de8d-4112-a681-fd3295fa03a2 service nova] Lock "06658134-4c6d-4911-9a55-b6805f555c78-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1040.806899] env[61806]: DEBUG oslo_concurrency.lockutils [req-ae44934b-7a0e-4a63-af7c-8831b2c897b1 req-ca2c6373-de8d-4112-a681-fd3295fa03a2 service nova] Lock "06658134-4c6d-4911-9a55-b6805f555c78-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1040.807186] env[61806]: DEBUG nova.compute.manager [req-ae44934b-7a0e-4a63-af7c-8831b2c897b1 req-ca2c6373-de8d-4112-a681-fd3295fa03a2 service nova] [instance: 06658134-4c6d-4911-9a55-b6805f555c78] No waiting events found dispatching network-vif-plugged-783dae1f-d1c0-4958-b039-ed57b202b64a {{(pid=61806) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1040.807325] env[61806]: WARNING nova.compute.manager [req-ae44934b-7a0e-4a63-af7c-8831b2c897b1 req-ca2c6373-de8d-4112-a681-fd3295fa03a2 service nova] [instance: 06658134-4c6d-4911-9a55-b6805f555c78] Received unexpected event network-vif-plugged-783dae1f-d1c0-4958-b039-ed57b202b64a for instance with vm_state building and task_state spawning. [ 1040.807526] env[61806]: DEBUG nova.compute.manager [req-ae44934b-7a0e-4a63-af7c-8831b2c897b1 req-ca2c6373-de8d-4112-a681-fd3295fa03a2 service nova] [instance: 06658134-4c6d-4911-9a55-b6805f555c78] Received event network-changed-783dae1f-d1c0-4958-b039-ed57b202b64a {{(pid=61806) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1040.807693] env[61806]: DEBUG nova.compute.manager [req-ae44934b-7a0e-4a63-af7c-8831b2c897b1 req-ca2c6373-de8d-4112-a681-fd3295fa03a2 service nova] [instance: 06658134-4c6d-4911-9a55-b6805f555c78] Refreshing instance network info cache due to event network-changed-783dae1f-d1c0-4958-b039-ed57b202b64a. {{(pid=61806) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 1040.807861] env[61806]: DEBUG oslo_concurrency.lockutils [req-ae44934b-7a0e-4a63-af7c-8831b2c897b1 req-ca2c6373-de8d-4112-a681-fd3295fa03a2 service nova] Acquiring lock "refresh_cache-06658134-4c6d-4911-9a55-b6805f555c78" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1040.842291] env[61806]: DEBUG oslo_vmware.rw_handles [None req-a5c61942-d486-43f3-834d-7dfb87ef5b63 tempest-ServersNegativeTestJSON-891856253 tempest-ServersNegativeTestJSON-891856253-project-member] Closed VMDK write handle for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52a2032f-a228-1c10-99ea-4211bd18d59d/disk-0.vmdk. {{(pid=61806) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:646}} [ 1040.842518] env[61806]: INFO nova.virt.vmwareapi.images [None req-a5c61942-d486-43f3-834d-7dfb87ef5b63 tempest-ServersNegativeTestJSON-891856253 tempest-ServersNegativeTestJSON-891856253-project-member] [instance: bdbaef31-7d37-40cb-ac3e-ad152b37e2d2] Downloaded image file data 20c8ea6e-0db5-4788-9e34-92c1d63bb7b8 [ 1040.843711] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1f00ffdf-5ad9-49bc-ab37-273e4b14aaf5 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1040.849352] env[61806]: DEBUG oslo_vmware.api [None req-4da0f2fc-feb0-4b02-97f3-bd412b867242 tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Task: {'id': task-1295150, 'name': PowerOffVM_Task, 'duration_secs': 0.211643} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1040.849920] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [None req-4da0f2fc-feb0-4b02-97f3-bd412b867242 tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] [instance: a61abfd9-3cc1-4a2b-b68b-a446094daf87] Powered off the VM {{(pid=61806) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1040.850141] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-4da0f2fc-feb0-4b02-97f3-bd412b867242 tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] [instance: a61abfd9-3cc1-4a2b-b68b-a446094daf87] Unregistering the VM {{(pid=61806) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1040.850367] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-a98d24e6-4f34-4ce7-8db1-a94222e28170 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1040.862264] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-5dd0ac77-94c4-4bf5-bb45-2f579a23f637 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1040.897995] env[61806]: INFO nova.virt.vmwareapi.images [None req-a5c61942-d486-43f3-834d-7dfb87ef5b63 tempest-ServersNegativeTestJSON-891856253 tempest-ServersNegativeTestJSON-891856253-project-member] [instance: bdbaef31-7d37-40cb-ac3e-ad152b37e2d2] The imported VM was unregistered [ 1040.900687] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-a5c61942-d486-43f3-834d-7dfb87ef5b63 tempest-ServersNegativeTestJSON-891856253 tempest-ServersNegativeTestJSON-891856253-project-member] [instance: bdbaef31-7d37-40cb-ac3e-ad152b37e2d2] Caching image {{(pid=61806) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:638}} [ 1040.900935] env[61806]: DEBUG nova.virt.vmwareapi.ds_util [None req-a5c61942-d486-43f3-834d-7dfb87ef5b63 tempest-ServersNegativeTestJSON-891856253 tempest-ServersNegativeTestJSON-891856253-project-member] Creating directory with path [datastore1] devstack-image-cache_base/20c8ea6e-0db5-4788-9e34-92c1d63bb7b8 {{(pid=61806) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1040.901219] env[61806]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-f8a50401-9f58-43e6-9e51-bfd7f89edde3 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1040.911177] env[61806]: DEBUG nova.virt.vmwareapi.ds_util [None req-a5c61942-d486-43f3-834d-7dfb87ef5b63 tempest-ServersNegativeTestJSON-891856253 tempest-ServersNegativeTestJSON-891856253-project-member] Created directory with path [datastore1] devstack-image-cache_base/20c8ea6e-0db5-4788-9e34-92c1d63bb7b8 {{(pid=61806) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1040.911374] env[61806]: DEBUG nova.virt.vmwareapi.ds_util [None req-a5c61942-d486-43f3-834d-7dfb87ef5b63 tempest-ServersNegativeTestJSON-891856253 tempest-ServersNegativeTestJSON-891856253-project-member] Moving virtual disk from [datastore1] OSTACK_IMG_b3afb79f-96af-441f-a8dc-c28cd8d6a22d/OSTACK_IMG_b3afb79f-96af-441f-a8dc-c28cd8d6a22d.vmdk to [datastore1] devstack-image-cache_base/20c8ea6e-0db5-4788-9e34-92c1d63bb7b8/20c8ea6e-0db5-4788-9e34-92c1d63bb7b8.vmdk. {{(pid=61806) disk_move /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:263}} [ 1040.911621] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.MoveVirtualDisk_Task with opID=oslo.vmware-990de6d0-898f-488c-b99f-a5733df9e927 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1040.918359] env[61806]: DEBUG oslo_vmware.api [None req-a5c61942-d486-43f3-834d-7dfb87ef5b63 tempest-ServersNegativeTestJSON-891856253 tempest-ServersNegativeTestJSON-891856253-project-member] Waiting for the task: (returnval){ [ 1040.918359] env[61806]: value = "task-1295153" [ 1040.918359] env[61806]: _type = "Task" [ 1040.918359] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1040.926760] env[61806]: DEBUG oslo_vmware.api [None req-a5c61942-d486-43f3-834d-7dfb87ef5b63 tempest-ServersNegativeTestJSON-891856253 tempest-ServersNegativeTestJSON-891856253-project-member] Task: {'id': task-1295153, 'name': MoveVirtualDisk_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1040.941561] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-4da0f2fc-feb0-4b02-97f3-bd412b867242 tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] [instance: a61abfd9-3cc1-4a2b-b68b-a446094daf87] Unregistered the VM {{(pid=61806) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1040.942243] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-4da0f2fc-feb0-4b02-97f3-bd412b867242 tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] [instance: a61abfd9-3cc1-4a2b-b68b-a446094daf87] Deleting contents of the VM from datastore datastore1 {{(pid=61806) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1040.942243] env[61806]: DEBUG nova.virt.vmwareapi.ds_util [None req-4da0f2fc-feb0-4b02-97f3-bd412b867242 tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Deleting the datastore file [datastore1] a61abfd9-3cc1-4a2b-b68b-a446094daf87 {{(pid=61806) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1040.942372] env[61806]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-0999c525-e648-4b71-966c-493e5d39efd7 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1040.948608] env[61806]: DEBUG oslo_vmware.api [None req-4da0f2fc-feb0-4b02-97f3-bd412b867242 tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Waiting for the task: (returnval){ [ 1040.948608] env[61806]: value = "task-1295154" [ 1040.948608] env[61806]: _type = "Task" [ 1040.948608] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1040.956138] env[61806]: DEBUG oslo_vmware.api [None req-4da0f2fc-feb0-4b02-97f3-bd412b867242 tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Task: {'id': task-1295154, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1041.055430] env[61806]: DEBUG nova.compute.manager [None req-54021779-f9e2-42db-b76c-c250eacb9121 tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] [instance: 7445d2c5-26bd-4f8d-8653-51c721ea801f] Start spawning the instance on the hypervisor. {{(pid=61806) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1041.081176] env[61806]: DEBUG nova.virt.hardware [None req-54021779-f9e2-42db-b76c-c250eacb9121 tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-15T18:05:10Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-15T18:04:52Z,direct_url=,disk_format='vmdk',id=ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='84ba83fce2c349a6988173c5d6fc3b07',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-15T18:04:53Z,virtual_size=,visibility=), allow threads: False {{(pid=61806) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1041.081442] env[61806]: DEBUG nova.virt.hardware [None req-54021779-f9e2-42db-b76c-c250eacb9121 tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Flavor limits 0:0:0 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1041.081654] env[61806]: DEBUG nova.virt.hardware [None req-54021779-f9e2-42db-b76c-c250eacb9121 tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Image limits 0:0:0 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1041.081888] env[61806]: DEBUG nova.virt.hardware [None req-54021779-f9e2-42db-b76c-c250eacb9121 tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Flavor pref 0:0:0 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1041.082106] env[61806]: DEBUG nova.virt.hardware [None req-54021779-f9e2-42db-b76c-c250eacb9121 tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Image pref 0:0:0 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1041.082319] env[61806]: DEBUG nova.virt.hardware [None req-54021779-f9e2-42db-b76c-c250eacb9121 tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1041.082591] env[61806]: DEBUG nova.virt.hardware [None req-54021779-f9e2-42db-b76c-c250eacb9121 tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61806) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1041.082804] env[61806]: DEBUG nova.virt.hardware [None req-54021779-f9e2-42db-b76c-c250eacb9121 tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61806) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1041.083060] env[61806]: DEBUG nova.virt.hardware [None req-54021779-f9e2-42db-b76c-c250eacb9121 tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Got 1 possible topologies {{(pid=61806) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1041.083305] env[61806]: DEBUG nova.virt.hardware [None req-54021779-f9e2-42db-b76c-c250eacb9121 tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61806) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1041.083529] env[61806]: DEBUG nova.virt.hardware [None req-54021779-f9e2-42db-b76c-c250eacb9121 tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61806) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1041.084451] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dcd59053-5e63-4790-99f1-106b1c0b4e12 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1041.092905] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fd159b03-05e2-49ae-bc8b-706d9332bdc9 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1041.106085] env[61806]: DEBUG oslo_concurrency.lockutils [None req-138f28bc-d88a-4f34-babf-797a70b177f1 tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Releasing lock "refresh_cache-06658134-4c6d-4911-9a55-b6805f555c78" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1041.106436] env[61806]: DEBUG nova.compute.manager [None req-138f28bc-d88a-4f34-babf-797a70b177f1 tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] [instance: 06658134-4c6d-4911-9a55-b6805f555c78] Instance network_info: |[{"id": "783dae1f-d1c0-4958-b039-ed57b202b64a", "address": "fa:16:3e:89:24:35", "network": {"id": "746a03b6-c1a6-4d27-98b9-0bee6376634f", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-1971382587-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c1f2d161c50f4b70987b90e6364c3779", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6a31b224-77d7-48a4-af87-312758611b8e", "external-id": "nsx-vlan-transportzone-761", "segmentation_id": 761, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap783dae1f-d1", "ovs_interfaceid": "783dae1f-d1c0-4958-b039-ed57b202b64a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61806) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1041.106973] env[61806]: DEBUG oslo_concurrency.lockutils [req-ae44934b-7a0e-4a63-af7c-8831b2c897b1 req-ca2c6373-de8d-4112-a681-fd3295fa03a2 service nova] Acquired lock "refresh_cache-06658134-4c6d-4911-9a55-b6805f555c78" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1041.107177] env[61806]: DEBUG nova.network.neutron [req-ae44934b-7a0e-4a63-af7c-8831b2c897b1 req-ca2c6373-de8d-4112-a681-fd3295fa03a2 service nova] [instance: 06658134-4c6d-4911-9a55-b6805f555c78] Refreshing network info cache for port 783dae1f-d1c0-4958-b039-ed57b202b64a {{(pid=61806) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1041.108468] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-138f28bc-d88a-4f34-babf-797a70b177f1 tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] [instance: 06658134-4c6d-4911-9a55-b6805f555c78] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:89:24:35', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '6a31b224-77d7-48a4-af87-312758611b8e', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '783dae1f-d1c0-4958-b039-ed57b202b64a', 'vif_model': 'vmxnet3'}] {{(pid=61806) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1041.116000] env[61806]: DEBUG oslo.service.loopingcall [None req-138f28bc-d88a-4f34-babf-797a70b177f1 tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61806) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1041.117082] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 06658134-4c6d-4911-9a55-b6805f555c78] Creating VM on the ESX host {{(pid=61806) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 1041.117331] env[61806]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-896eea10-d386-4ad6-9a2a-e89f1a54982b {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1041.138679] env[61806]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1041.138679] env[61806]: value = "task-1295155" [ 1041.138679] env[61806]: _type = "Task" [ 1041.138679] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1041.149213] env[61806]: DEBUG oslo_vmware.api [-] Task: {'id': task-1295155, 'name': CreateVM_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1041.281354] env[61806]: DEBUG oslo_concurrency.lockutils [None req-c257ef4d-1b9a-4377-aada-40cc9a4843fd tempest-AttachVolumeNegativeTest-1683033704 tempest-AttachVolumeNegativeTest-1683033704-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.256s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1041.281924] env[61806]: DEBUG nova.compute.manager [None req-c257ef4d-1b9a-4377-aada-40cc9a4843fd tempest-AttachVolumeNegativeTest-1683033704 tempest-AttachVolumeNegativeTest-1683033704-project-member] [instance: ce2ad097-4c3f-4b55-b018-ac788aa5b662] Start building networks asynchronously for instance. {{(pid=61806) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1041.285014] env[61806]: DEBUG oslo_concurrency.lockutils [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 2.180s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1041.430773] env[61806]: DEBUG oslo_vmware.api [None req-a5c61942-d486-43f3-834d-7dfb87ef5b63 tempest-ServersNegativeTestJSON-891856253 tempest-ServersNegativeTestJSON-891856253-project-member] Task: {'id': task-1295153, 'name': MoveVirtualDisk_Task} progress is 21%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1041.461320] env[61806]: DEBUG oslo_vmware.api [None req-4da0f2fc-feb0-4b02-97f3-bd412b867242 tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Task: {'id': task-1295154, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.191777} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1041.461320] env[61806]: DEBUG nova.virt.vmwareapi.ds_util [None req-4da0f2fc-feb0-4b02-97f3-bd412b867242 tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Deleted the datastore file {{(pid=61806) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1041.461320] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-4da0f2fc-feb0-4b02-97f3-bd412b867242 tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] [instance: a61abfd9-3cc1-4a2b-b68b-a446094daf87] Deleted contents of the VM from datastore datastore1 {{(pid=61806) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1041.461320] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-4da0f2fc-feb0-4b02-97f3-bd412b867242 tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] [instance: a61abfd9-3cc1-4a2b-b68b-a446094daf87] Instance destroyed {{(pid=61806) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1041.461320] env[61806]: INFO nova.compute.manager [None req-4da0f2fc-feb0-4b02-97f3-bd412b867242 tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] [instance: a61abfd9-3cc1-4a2b-b68b-a446094daf87] Took 1.15 seconds to destroy the instance on the hypervisor. [ 1041.461664] env[61806]: DEBUG oslo.service.loopingcall [None req-4da0f2fc-feb0-4b02-97f3-bd412b867242 tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61806) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1041.462933] env[61806]: DEBUG nova.compute.manager [-] [instance: a61abfd9-3cc1-4a2b-b68b-a446094daf87] Deallocating network for instance {{(pid=61806) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1041.462933] env[61806]: DEBUG nova.network.neutron [-] [instance: a61abfd9-3cc1-4a2b-b68b-a446094daf87] deallocate_for_instance() {{(pid=61806) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1041.650885] env[61806]: DEBUG oslo_vmware.api [-] Task: {'id': task-1295155, 'name': CreateVM_Task} progress is 99%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1041.674337] env[61806]: DEBUG nova.network.neutron [None req-54021779-f9e2-42db-b76c-c250eacb9121 tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] [instance: 7445d2c5-26bd-4f8d-8653-51c721ea801f] Successfully updated port: 773e4470-b54c-4449-b825-5bb680cf8d2f {{(pid=61806) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1041.790998] env[61806]: DEBUG nova.compute.utils [None req-c257ef4d-1b9a-4377-aada-40cc9a4843fd tempest-AttachVolumeNegativeTest-1683033704 tempest-AttachVolumeNegativeTest-1683033704-project-member] Using /dev/sd instead of None {{(pid=61806) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1041.799935] env[61806]: DEBUG nova.compute.manager [None req-c257ef4d-1b9a-4377-aada-40cc9a4843fd tempest-AttachVolumeNegativeTest-1683033704 tempest-AttachVolumeNegativeTest-1683033704-project-member] [instance: ce2ad097-4c3f-4b55-b018-ac788aa5b662] Allocating IP information in the background. {{(pid=61806) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1041.799935] env[61806]: DEBUG nova.network.neutron [None req-c257ef4d-1b9a-4377-aada-40cc9a4843fd tempest-AttachVolumeNegativeTest-1683033704 tempest-AttachVolumeNegativeTest-1683033704-project-member] [instance: ce2ad097-4c3f-4b55-b018-ac788aa5b662] allocate_for_instance() {{(pid=61806) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 1041.853792] env[61806]: DEBUG nova.policy [None req-c257ef4d-1b9a-4377-aada-40cc9a4843fd tempest-AttachVolumeNegativeTest-1683033704 tempest-AttachVolumeNegativeTest-1683033704-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '62b5e014cfeb4403a1edf89108c55e11', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '22e47882c2c44ce492dff6b1c5d782e2', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61806) authorize /opt/stack/nova/nova/policy.py:201}} [ 1041.887796] env[61806]: DEBUG nova.network.neutron [req-ae44934b-7a0e-4a63-af7c-8831b2c897b1 req-ca2c6373-de8d-4112-a681-fd3295fa03a2 service nova] [instance: 06658134-4c6d-4911-9a55-b6805f555c78] Updated VIF entry in instance network info cache for port 783dae1f-d1c0-4958-b039-ed57b202b64a. {{(pid=61806) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1041.888248] env[61806]: DEBUG nova.network.neutron [req-ae44934b-7a0e-4a63-af7c-8831b2c897b1 req-ca2c6373-de8d-4112-a681-fd3295fa03a2 service nova] [instance: 06658134-4c6d-4911-9a55-b6805f555c78] Updating instance_info_cache with network_info: [{"id": "783dae1f-d1c0-4958-b039-ed57b202b64a", "address": "fa:16:3e:89:24:35", "network": {"id": "746a03b6-c1a6-4d27-98b9-0bee6376634f", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-1971382587-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c1f2d161c50f4b70987b90e6364c3779", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6a31b224-77d7-48a4-af87-312758611b8e", "external-id": "nsx-vlan-transportzone-761", "segmentation_id": 761, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap783dae1f-d1", "ovs_interfaceid": "783dae1f-d1c0-4958-b039-ed57b202b64a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61806) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1041.931667] env[61806]: DEBUG oslo_vmware.api [None req-a5c61942-d486-43f3-834d-7dfb87ef5b63 tempest-ServersNegativeTestJSON-891856253 tempest-ServersNegativeTestJSON-891856253-project-member] Task: {'id': task-1295153, 'name': MoveVirtualDisk_Task} progress is 40%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1042.152745] env[61806]: DEBUG oslo_vmware.api [-] Task: {'id': task-1295155, 'name': CreateVM_Task, 'duration_secs': 0.562441} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1042.152992] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 06658134-4c6d-4911-9a55-b6805f555c78] Created VM on the ESX host {{(pid=61806) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 1042.153686] env[61806]: DEBUG oslo_concurrency.lockutils [None req-138f28bc-d88a-4f34-babf-797a70b177f1 tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1042.153865] env[61806]: DEBUG oslo_concurrency.lockutils [None req-138f28bc-d88a-4f34-babf-797a70b177f1 tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Acquired lock "[datastore2] devstack-image-cache_base/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1042.154240] env[61806]: DEBUG oslo_concurrency.lockutils [None req-138f28bc-d88a-4f34-babf-797a70b177f1 tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1042.154516] env[61806]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d05e1cf9-fbdf-45a0-83bb-da634197cca3 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1042.161963] env[61806]: DEBUG oslo_vmware.api [None req-138f28bc-d88a-4f34-babf-797a70b177f1 tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Waiting for the task: (returnval){ [ 1042.161963] env[61806]: value = "session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]522742d1-8d03-571b-116a-1af5f04a2607" [ 1042.161963] env[61806]: _type = "Task" [ 1042.161963] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1042.166528] env[61806]: DEBUG nova.network.neutron [None req-c257ef4d-1b9a-4377-aada-40cc9a4843fd tempest-AttachVolumeNegativeTest-1683033704 tempest-AttachVolumeNegativeTest-1683033704-project-member] [instance: ce2ad097-4c3f-4b55-b018-ac788aa5b662] Successfully created port: 6792e7fd-4056-4d2f-86a5-45ea41172f3f {{(pid=61806) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1042.175457] env[61806]: DEBUG oslo_vmware.api [None req-138f28bc-d88a-4f34-babf-797a70b177f1 tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Task: {'id': session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]522742d1-8d03-571b-116a-1af5f04a2607, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1042.177172] env[61806]: DEBUG oslo_concurrency.lockutils [None req-54021779-f9e2-42db-b76c-c250eacb9121 tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Acquiring lock "refresh_cache-7445d2c5-26bd-4f8d-8653-51c721ea801f" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1042.177439] env[61806]: DEBUG oslo_concurrency.lockutils [None req-54021779-f9e2-42db-b76c-c250eacb9121 tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Acquired lock "refresh_cache-7445d2c5-26bd-4f8d-8653-51c721ea801f" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1042.177621] env[61806]: DEBUG nova.network.neutron [None req-54021779-f9e2-42db-b76c-c250eacb9121 tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] [instance: 7445d2c5-26bd-4f8d-8653-51c721ea801f] Building network info cache for instance {{(pid=61806) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1042.297280] env[61806]: DEBUG nova.compute.manager [None req-c257ef4d-1b9a-4377-aada-40cc9a4843fd tempest-AttachVolumeNegativeTest-1683033704 tempest-AttachVolumeNegativeTest-1683033704-project-member] [instance: ce2ad097-4c3f-4b55-b018-ac788aa5b662] Start building block device mappings for instance. {{(pid=61806) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1042.303744] env[61806]: DEBUG nova.network.neutron [-] [instance: a61abfd9-3cc1-4a2b-b68b-a446094daf87] Updating instance_info_cache with network_info: [] {{(pid=61806) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1042.327283] env[61806]: DEBUG nova.compute.resource_tracker [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Instance d1802da4-a670-427d-84d7-0fcb4717e18f actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61806) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1042.327724] env[61806]: DEBUG nova.compute.resource_tracker [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Instance fd9e520a-2e7b-43d4-b84d-8933c32cc6a6 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61806) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1042.327724] env[61806]: DEBUG nova.compute.resource_tracker [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Instance d0669232-6f33-4b30-97c2-2e4239af1ad0 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61806) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1042.327724] env[61806]: WARNING nova.compute.resource_tracker [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Instance 713d4e6e-6f0e-4844-9b55-cc010030eda6 is not being actively managed by this compute host but has allocations referencing this compute host: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. Skipping heal of allocation because we do not know what to do. [ 1042.327832] env[61806]: DEBUG nova.compute.resource_tracker [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Instance bdbaef31-7d37-40cb-ac3e-ad152b37e2d2 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61806) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1042.327917] env[61806]: DEBUG nova.compute.resource_tracker [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Instance a61abfd9-3cc1-4a2b-b68b-a446094daf87 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61806) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1042.328039] env[61806]: DEBUG nova.compute.resource_tracker [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Instance 06658134-4c6d-4911-9a55-b6805f555c78 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61806) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1042.328428] env[61806]: DEBUG nova.compute.resource_tracker [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Instance 7445d2c5-26bd-4f8d-8653-51c721ea801f actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61806) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1042.329177] env[61806]: DEBUG nova.compute.resource_tracker [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Instance ce2ad097-4c3f-4b55-b018-ac788aa5b662 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61806) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1042.329436] env[61806]: DEBUG nova.compute.resource_tracker [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Total usable vcpus: 48, total allocated vcpus: 8 {{(pid=61806) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 1042.329591] env[61806]: DEBUG nova.compute.resource_tracker [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=2048MB phys_disk=200GB used_disk=8GB total_vcpus=48 used_vcpus=8 pci_stats=[] {{(pid=61806) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 1042.391051] env[61806]: DEBUG oslo_concurrency.lockutils [req-ae44934b-7a0e-4a63-af7c-8831b2c897b1 req-ca2c6373-de8d-4112-a681-fd3295fa03a2 service nova] Releasing lock "refresh_cache-06658134-4c6d-4911-9a55-b6805f555c78" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1042.435434] env[61806]: DEBUG oslo_vmware.api [None req-a5c61942-d486-43f3-834d-7dfb87ef5b63 tempest-ServersNegativeTestJSON-891856253 tempest-ServersNegativeTestJSON-891856253-project-member] Task: {'id': task-1295153, 'name': MoveVirtualDisk_Task} progress is 63%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1042.511152] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d6dce469-d154-4884-85e4-9869eae92f18 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1042.520114] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-be4f1a8e-51d0-4c1b-9e50-df871c3bfb16 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1042.555971] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3bbfdf64-fdd5-4265-9ea0-203fecb50e1a {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1042.566600] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2916a6a2-6dcb-43ae-8750-cf0b8eaee4d1 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1042.584460] env[61806]: DEBUG nova.compute.provider_tree [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Inventory has not changed in ProviderTree for provider: a2858be1-fd22-4e08-979e-87ad25293407 {{(pid=61806) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1042.672641] env[61806]: DEBUG oslo_vmware.api [None req-138f28bc-d88a-4f34-babf-797a70b177f1 tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Task: {'id': session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]522742d1-8d03-571b-116a-1af5f04a2607, 'name': SearchDatastore_Task, 'duration_secs': 0.022319} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1042.673192] env[61806]: DEBUG oslo_concurrency.lockutils [None req-138f28bc-d88a-4f34-babf-797a70b177f1 tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Releasing lock "[datastore2] devstack-image-cache_base/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1042.673370] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-138f28bc-d88a-4f34-babf-797a70b177f1 tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] [instance: 06658134-4c6d-4911-9a55-b6805f555c78] Processing image ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa {{(pid=61806) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1042.673682] env[61806]: DEBUG oslo_concurrency.lockutils [None req-138f28bc-d88a-4f34-babf-797a70b177f1 tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa.vmdk" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1042.673940] env[61806]: DEBUG oslo_concurrency.lockutils [None req-138f28bc-d88a-4f34-babf-797a70b177f1 tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Acquired lock "[datastore2] devstack-image-cache_base/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa.vmdk" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1042.674197] env[61806]: DEBUG nova.virt.vmwareapi.ds_util [None req-138f28bc-d88a-4f34-babf-797a70b177f1 tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61806) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1042.674485] env[61806]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-188496cd-7b01-4530-808e-2f155d75e717 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1042.687519] env[61806]: DEBUG nova.virt.vmwareapi.ds_util [None req-138f28bc-d88a-4f34-babf-797a70b177f1 tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61806) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1042.687519] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-138f28bc-d88a-4f34-babf-797a70b177f1 tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61806) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 1042.688235] env[61806]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-2a1d3987-7751-4c86-9287-1d754ec4ca41 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1042.698029] env[61806]: DEBUG oslo_vmware.api [None req-138f28bc-d88a-4f34-babf-797a70b177f1 tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Waiting for the task: (returnval){ [ 1042.698029] env[61806]: value = "session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]528d5d3b-4107-f67d-c8ea-d642071de443" [ 1042.698029] env[61806]: _type = "Task" [ 1042.698029] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1042.708413] env[61806]: DEBUG oslo_vmware.api [None req-138f28bc-d88a-4f34-babf-797a70b177f1 tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Task: {'id': session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]528d5d3b-4107-f67d-c8ea-d642071de443, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1042.722278] env[61806]: DEBUG nova.network.neutron [None req-54021779-f9e2-42db-b76c-c250eacb9121 tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] [instance: 7445d2c5-26bd-4f8d-8653-51c721ea801f] Instance cache missing network info. {{(pid=61806) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 1042.811637] env[61806]: INFO nova.compute.manager [-] [instance: a61abfd9-3cc1-4a2b-b68b-a446094daf87] Took 1.35 seconds to deallocate network for instance. [ 1042.839982] env[61806]: DEBUG nova.compute.manager [req-a7425dce-615e-4503-b23f-02554902280f req-744a4108-be48-4b81-bcf5-428bfeb2f44b service nova] [instance: 7445d2c5-26bd-4f8d-8653-51c721ea801f] Received event network-vif-plugged-773e4470-b54c-4449-b825-5bb680cf8d2f {{(pid=61806) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1042.841073] env[61806]: DEBUG oslo_concurrency.lockutils [req-a7425dce-615e-4503-b23f-02554902280f req-744a4108-be48-4b81-bcf5-428bfeb2f44b service nova] Acquiring lock "7445d2c5-26bd-4f8d-8653-51c721ea801f-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1042.841073] env[61806]: DEBUG oslo_concurrency.lockutils [req-a7425dce-615e-4503-b23f-02554902280f req-744a4108-be48-4b81-bcf5-428bfeb2f44b service nova] Lock "7445d2c5-26bd-4f8d-8653-51c721ea801f-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1042.841073] env[61806]: DEBUG oslo_concurrency.lockutils [req-a7425dce-615e-4503-b23f-02554902280f req-744a4108-be48-4b81-bcf5-428bfeb2f44b service nova] Lock "7445d2c5-26bd-4f8d-8653-51c721ea801f-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1042.841073] env[61806]: DEBUG nova.compute.manager [req-a7425dce-615e-4503-b23f-02554902280f req-744a4108-be48-4b81-bcf5-428bfeb2f44b service nova] [instance: 7445d2c5-26bd-4f8d-8653-51c721ea801f] No waiting events found dispatching network-vif-plugged-773e4470-b54c-4449-b825-5bb680cf8d2f {{(pid=61806) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1042.841073] env[61806]: WARNING nova.compute.manager [req-a7425dce-615e-4503-b23f-02554902280f req-744a4108-be48-4b81-bcf5-428bfeb2f44b service nova] [instance: 7445d2c5-26bd-4f8d-8653-51c721ea801f] Received unexpected event network-vif-plugged-773e4470-b54c-4449-b825-5bb680cf8d2f for instance with vm_state building and task_state spawning. [ 1042.841433] env[61806]: DEBUG nova.compute.manager [req-a7425dce-615e-4503-b23f-02554902280f req-744a4108-be48-4b81-bcf5-428bfeb2f44b service nova] [instance: 7445d2c5-26bd-4f8d-8653-51c721ea801f] Received event network-changed-773e4470-b54c-4449-b825-5bb680cf8d2f {{(pid=61806) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1042.841433] env[61806]: DEBUG nova.compute.manager [req-a7425dce-615e-4503-b23f-02554902280f req-744a4108-be48-4b81-bcf5-428bfeb2f44b service nova] [instance: 7445d2c5-26bd-4f8d-8653-51c721ea801f] Refreshing instance network info cache due to event network-changed-773e4470-b54c-4449-b825-5bb680cf8d2f. {{(pid=61806) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 1042.841433] env[61806]: DEBUG oslo_concurrency.lockutils [req-a7425dce-615e-4503-b23f-02554902280f req-744a4108-be48-4b81-bcf5-428bfeb2f44b service nova] Acquiring lock "refresh_cache-7445d2c5-26bd-4f8d-8653-51c721ea801f" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1042.872989] env[61806]: DEBUG nova.network.neutron [None req-54021779-f9e2-42db-b76c-c250eacb9121 tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] [instance: 7445d2c5-26bd-4f8d-8653-51c721ea801f] Updating instance_info_cache with network_info: [{"id": "773e4470-b54c-4449-b825-5bb680cf8d2f", "address": "fa:16:3e:09:73:90", "network": {"id": "b7e78559-d885-4b59-9c7f-9d966b424d04", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-1456721354-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7ab95fb72b5d46c3b6c7bebbccf897cd", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d891a662-7da0-4be5-8d0d-01b4ff70552a", "external-id": "nsx-vlan-transportzone-815", "segmentation_id": 815, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap773e4470-b5", "ovs_interfaceid": "773e4470-b54c-4449-b825-5bb680cf8d2f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61806) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1042.929672] env[61806]: DEBUG oslo_vmware.api [None req-a5c61942-d486-43f3-834d-7dfb87ef5b63 tempest-ServersNegativeTestJSON-891856253 tempest-ServersNegativeTestJSON-891856253-project-member] Task: {'id': task-1295153, 'name': MoveVirtualDisk_Task} progress is 85%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1043.087771] env[61806]: DEBUG nova.scheduler.client.report [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Inventory has not changed for provider a2858be1-fd22-4e08-979e-87ad25293407 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 138, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61806) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1043.209911] env[61806]: DEBUG oslo_vmware.api [None req-138f28bc-d88a-4f34-babf-797a70b177f1 tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Task: {'id': session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]528d5d3b-4107-f67d-c8ea-d642071de443, 'name': SearchDatastore_Task, 'duration_secs': 0.015075} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1043.210815] env[61806]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b8da4155-f670-4d8c-b1ae-c424460910b4 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1043.217626] env[61806]: DEBUG oslo_vmware.api [None req-138f28bc-d88a-4f34-babf-797a70b177f1 tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Waiting for the task: (returnval){ [ 1043.217626] env[61806]: value = "session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]528eba6c-26a6-6bb9-a4fd-df19daae9a51" [ 1043.217626] env[61806]: _type = "Task" [ 1043.217626] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1043.227242] env[61806]: DEBUG oslo_vmware.api [None req-138f28bc-d88a-4f34-babf-797a70b177f1 tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Task: {'id': session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]528eba6c-26a6-6bb9-a4fd-df19daae9a51, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1043.312515] env[61806]: DEBUG nova.compute.manager [None req-c257ef4d-1b9a-4377-aada-40cc9a4843fd tempest-AttachVolumeNegativeTest-1683033704 tempest-AttachVolumeNegativeTest-1683033704-project-member] [instance: ce2ad097-4c3f-4b55-b018-ac788aa5b662] Start spawning the instance on the hypervisor. {{(pid=61806) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1043.315353] env[61806]: DEBUG oslo_concurrency.lockutils [None req-4da0f2fc-feb0-4b02-97f3-bd412b867242 tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1043.339311] env[61806]: DEBUG nova.virt.hardware [None req-c257ef4d-1b9a-4377-aada-40cc9a4843fd tempest-AttachVolumeNegativeTest-1683033704 tempest-AttachVolumeNegativeTest-1683033704-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-15T18:05:10Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-15T18:04:52Z,direct_url=,disk_format='vmdk',id=ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='84ba83fce2c349a6988173c5d6fc3b07',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-15T18:04:53Z,virtual_size=,visibility=), allow threads: False {{(pid=61806) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1043.339580] env[61806]: DEBUG nova.virt.hardware [None req-c257ef4d-1b9a-4377-aada-40cc9a4843fd tempest-AttachVolumeNegativeTest-1683033704 tempest-AttachVolumeNegativeTest-1683033704-project-member] Flavor limits 0:0:0 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1043.339744] env[61806]: DEBUG nova.virt.hardware [None req-c257ef4d-1b9a-4377-aada-40cc9a4843fd tempest-AttachVolumeNegativeTest-1683033704 tempest-AttachVolumeNegativeTest-1683033704-project-member] Image limits 0:0:0 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1043.339985] env[61806]: DEBUG nova.virt.hardware [None req-c257ef4d-1b9a-4377-aada-40cc9a4843fd tempest-AttachVolumeNegativeTest-1683033704 tempest-AttachVolumeNegativeTest-1683033704-project-member] Flavor pref 0:0:0 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1043.340222] env[61806]: DEBUG nova.virt.hardware [None req-c257ef4d-1b9a-4377-aada-40cc9a4843fd tempest-AttachVolumeNegativeTest-1683033704 tempest-AttachVolumeNegativeTest-1683033704-project-member] Image pref 0:0:0 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1043.340416] env[61806]: DEBUG nova.virt.hardware [None req-c257ef4d-1b9a-4377-aada-40cc9a4843fd tempest-AttachVolumeNegativeTest-1683033704 tempest-AttachVolumeNegativeTest-1683033704-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1043.340642] env[61806]: DEBUG nova.virt.hardware [None req-c257ef4d-1b9a-4377-aada-40cc9a4843fd tempest-AttachVolumeNegativeTest-1683033704 tempest-AttachVolumeNegativeTest-1683033704-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61806) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1043.340807] env[61806]: DEBUG nova.virt.hardware [None req-c257ef4d-1b9a-4377-aada-40cc9a4843fd tempest-AttachVolumeNegativeTest-1683033704 tempest-AttachVolumeNegativeTest-1683033704-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61806) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1043.340980] env[61806]: DEBUG nova.virt.hardware [None req-c257ef4d-1b9a-4377-aada-40cc9a4843fd tempest-AttachVolumeNegativeTest-1683033704 tempest-AttachVolumeNegativeTest-1683033704-project-member] Got 1 possible topologies {{(pid=61806) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1043.341189] env[61806]: DEBUG nova.virt.hardware [None req-c257ef4d-1b9a-4377-aada-40cc9a4843fd tempest-AttachVolumeNegativeTest-1683033704 tempest-AttachVolumeNegativeTest-1683033704-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61806) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1043.341430] env[61806]: DEBUG nova.virt.hardware [None req-c257ef4d-1b9a-4377-aada-40cc9a4843fd tempest-AttachVolumeNegativeTest-1683033704 tempest-AttachVolumeNegativeTest-1683033704-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61806) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1043.342316] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1ec3cfe7-6ba2-4f50-8231-03458b7790b6 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1043.349895] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-763d8cee-32f0-4e79-b781-a9859dba95ad {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1043.375720] env[61806]: DEBUG oslo_concurrency.lockutils [None req-54021779-f9e2-42db-b76c-c250eacb9121 tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Releasing lock "refresh_cache-7445d2c5-26bd-4f8d-8653-51c721ea801f" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1043.376019] env[61806]: DEBUG nova.compute.manager [None req-54021779-f9e2-42db-b76c-c250eacb9121 tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] [instance: 7445d2c5-26bd-4f8d-8653-51c721ea801f] Instance network_info: |[{"id": "773e4470-b54c-4449-b825-5bb680cf8d2f", "address": "fa:16:3e:09:73:90", "network": {"id": "b7e78559-d885-4b59-9c7f-9d966b424d04", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-1456721354-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7ab95fb72b5d46c3b6c7bebbccf897cd", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d891a662-7da0-4be5-8d0d-01b4ff70552a", "external-id": "nsx-vlan-transportzone-815", "segmentation_id": 815, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap773e4470-b5", "ovs_interfaceid": "773e4470-b54c-4449-b825-5bb680cf8d2f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61806) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1043.376325] env[61806]: DEBUG oslo_concurrency.lockutils [req-a7425dce-615e-4503-b23f-02554902280f req-744a4108-be48-4b81-bcf5-428bfeb2f44b service nova] Acquired lock "refresh_cache-7445d2c5-26bd-4f8d-8653-51c721ea801f" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1043.376515] env[61806]: DEBUG nova.network.neutron [req-a7425dce-615e-4503-b23f-02554902280f req-744a4108-be48-4b81-bcf5-428bfeb2f44b service nova] [instance: 7445d2c5-26bd-4f8d-8653-51c721ea801f] Refreshing network info cache for port 773e4470-b54c-4449-b825-5bb680cf8d2f {{(pid=61806) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1043.377628] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-54021779-f9e2-42db-b76c-c250eacb9121 tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] [instance: 7445d2c5-26bd-4f8d-8653-51c721ea801f] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:09:73:90', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'd891a662-7da0-4be5-8d0d-01b4ff70552a', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '773e4470-b54c-4449-b825-5bb680cf8d2f', 'vif_model': 'vmxnet3'}] {{(pid=61806) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1043.384859] env[61806]: DEBUG oslo.service.loopingcall [None req-54021779-f9e2-42db-b76c-c250eacb9121 tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61806) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1043.385748] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 7445d2c5-26bd-4f8d-8653-51c721ea801f] Creating VM on the ESX host {{(pid=61806) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 1043.385967] env[61806]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-e045e6ec-f843-499e-bfd8-760ab84a48ec {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1043.405468] env[61806]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1043.405468] env[61806]: value = "task-1295156" [ 1043.405468] env[61806]: _type = "Task" [ 1043.405468] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1043.413138] env[61806]: DEBUG oslo_vmware.api [-] Task: {'id': task-1295156, 'name': CreateVM_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1043.428200] env[61806]: DEBUG oslo_vmware.api [None req-a5c61942-d486-43f3-834d-7dfb87ef5b63 tempest-ServersNegativeTestJSON-891856253 tempest-ServersNegativeTestJSON-891856253-project-member] Task: {'id': task-1295153, 'name': MoveVirtualDisk_Task, 'duration_secs': 2.410193} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1043.428500] env[61806]: INFO nova.virt.vmwareapi.ds_util [None req-a5c61942-d486-43f3-834d-7dfb87ef5b63 tempest-ServersNegativeTestJSON-891856253 tempest-ServersNegativeTestJSON-891856253-project-member] Moved virtual disk from [datastore1] OSTACK_IMG_b3afb79f-96af-441f-a8dc-c28cd8d6a22d/OSTACK_IMG_b3afb79f-96af-441f-a8dc-c28cd8d6a22d.vmdk to [datastore1] devstack-image-cache_base/20c8ea6e-0db5-4788-9e34-92c1d63bb7b8/20c8ea6e-0db5-4788-9e34-92c1d63bb7b8.vmdk. [ 1043.428726] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-a5c61942-d486-43f3-834d-7dfb87ef5b63 tempest-ServersNegativeTestJSON-891856253 tempest-ServersNegativeTestJSON-891856253-project-member] [instance: bdbaef31-7d37-40cb-ac3e-ad152b37e2d2] Cleaning up location [datastore1] OSTACK_IMG_b3afb79f-96af-441f-a8dc-c28cd8d6a22d {{(pid=61806) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:640}} [ 1043.428929] env[61806]: DEBUG nova.virt.vmwareapi.ds_util [None req-a5c61942-d486-43f3-834d-7dfb87ef5b63 tempest-ServersNegativeTestJSON-891856253 tempest-ServersNegativeTestJSON-891856253-project-member] Deleting the datastore file [datastore1] OSTACK_IMG_b3afb79f-96af-441f-a8dc-c28cd8d6a22d {{(pid=61806) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1043.429229] env[61806]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-16c6b8ca-22b7-4a42-ae84-5594ff958c17 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1043.436019] env[61806]: DEBUG oslo_vmware.api [None req-a5c61942-d486-43f3-834d-7dfb87ef5b63 tempest-ServersNegativeTestJSON-891856253 tempest-ServersNegativeTestJSON-891856253-project-member] Waiting for the task: (returnval){ [ 1043.436019] env[61806]: value = "task-1295157" [ 1043.436019] env[61806]: _type = "Task" [ 1043.436019] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1043.443621] env[61806]: DEBUG oslo_vmware.api [None req-a5c61942-d486-43f3-834d-7dfb87ef5b63 tempest-ServersNegativeTestJSON-891856253 tempest-ServersNegativeTestJSON-891856253-project-member] Task: {'id': task-1295157, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1043.592895] env[61806]: DEBUG nova.compute.resource_tracker [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=61806) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 1043.593145] env[61806]: DEBUG oslo_concurrency.lockutils [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 2.308s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1043.593469] env[61806]: DEBUG oslo_concurrency.lockutils [None req-5eac2f24-7650-441c-94d5-4e1470c88591 tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 3.324s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1043.593690] env[61806]: DEBUG oslo_concurrency.lockutils [None req-5eac2f24-7650-441c-94d5-4e1470c88591 tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1043.596031] env[61806]: DEBUG oslo_concurrency.lockutils [None req-4da0f2fc-feb0-4b02-97f3-bd412b867242 tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.281s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1043.596317] env[61806]: DEBUG nova.objects.instance [None req-4da0f2fc-feb0-4b02-97f3-bd412b867242 tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Lazy-loading 'resources' on Instance uuid a61abfd9-3cc1-4a2b-b68b-a446094daf87 {{(pid=61806) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1043.619498] env[61806]: INFO nova.scheduler.client.report [None req-5eac2f24-7650-441c-94d5-4e1470c88591 tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] Deleted allocations for instance 713d4e6e-6f0e-4844-9b55-cc010030eda6 [ 1043.728950] env[61806]: DEBUG oslo_vmware.api [None req-138f28bc-d88a-4f34-babf-797a70b177f1 tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Task: {'id': session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]528eba6c-26a6-6bb9-a4fd-df19daae9a51, 'name': SearchDatastore_Task, 'duration_secs': 0.017498} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1043.729251] env[61806]: DEBUG oslo_concurrency.lockutils [None req-138f28bc-d88a-4f34-babf-797a70b177f1 tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Releasing lock "[datastore2] devstack-image-cache_base/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa.vmdk" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1043.729511] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [None req-138f28bc-d88a-4f34-babf-797a70b177f1 tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa.vmdk to [datastore2] 06658134-4c6d-4911-9a55-b6805f555c78/06658134-4c6d-4911-9a55-b6805f555c78.vmdk {{(pid=61806) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 1043.729782] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-09d1efb3-3866-4ae9-8c86-807e91239659 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1043.737326] env[61806]: DEBUG oslo_vmware.api [None req-138f28bc-d88a-4f34-babf-797a70b177f1 tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Waiting for the task: (returnval){ [ 1043.737326] env[61806]: value = "task-1295158" [ 1043.737326] env[61806]: _type = "Task" [ 1043.737326] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1043.744739] env[61806]: DEBUG oslo_vmware.api [None req-138f28bc-d88a-4f34-babf-797a70b177f1 tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Task: {'id': task-1295158, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1043.890613] env[61806]: DEBUG nova.network.neutron [None req-c257ef4d-1b9a-4377-aada-40cc9a4843fd tempest-AttachVolumeNegativeTest-1683033704 tempest-AttachVolumeNegativeTest-1683033704-project-member] [instance: ce2ad097-4c3f-4b55-b018-ac788aa5b662] Successfully updated port: 6792e7fd-4056-4d2f-86a5-45ea41172f3f {{(pid=61806) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1043.915464] env[61806]: DEBUG oslo_vmware.api [-] Task: {'id': task-1295156, 'name': CreateVM_Task, 'duration_secs': 0.417302} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1043.916778] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 7445d2c5-26bd-4f8d-8653-51c721ea801f] Created VM on the ESX host {{(pid=61806) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 1043.916921] env[61806]: DEBUG oslo_concurrency.lockutils [None req-54021779-f9e2-42db-b76c-c250eacb9121 tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1043.917122] env[61806]: DEBUG oslo_concurrency.lockutils [None req-54021779-f9e2-42db-b76c-c250eacb9121 tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Acquired lock "[datastore2] devstack-image-cache_base/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1043.917483] env[61806]: DEBUG oslo_concurrency.lockutils [None req-54021779-f9e2-42db-b76c-c250eacb9121 tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1043.920169] env[61806]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-1296d03a-870d-4296-a1ab-ebb0b89cfa40 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1043.925792] env[61806]: DEBUG oslo_vmware.api [None req-54021779-f9e2-42db-b76c-c250eacb9121 tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Waiting for the task: (returnval){ [ 1043.925792] env[61806]: value = "session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]52067950-eb9d-2dd9-fc63-706698999935" [ 1043.925792] env[61806]: _type = "Task" [ 1043.925792] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1043.937164] env[61806]: DEBUG oslo_vmware.api [None req-54021779-f9e2-42db-b76c-c250eacb9121 tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Task: {'id': session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]52067950-eb9d-2dd9-fc63-706698999935, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1043.945488] env[61806]: DEBUG oslo_vmware.api [None req-a5c61942-d486-43f3-834d-7dfb87ef5b63 tempest-ServersNegativeTestJSON-891856253 tempest-ServersNegativeTestJSON-891856253-project-member] Task: {'id': task-1295157, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.034723} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1043.945748] env[61806]: DEBUG nova.virt.vmwareapi.ds_util [None req-a5c61942-d486-43f3-834d-7dfb87ef5b63 tempest-ServersNegativeTestJSON-891856253 tempest-ServersNegativeTestJSON-891856253-project-member] Deleted the datastore file {{(pid=61806) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1043.945922] env[61806]: DEBUG oslo_concurrency.lockutils [None req-a5c61942-d486-43f3-834d-7dfb87ef5b63 tempest-ServersNegativeTestJSON-891856253 tempest-ServersNegativeTestJSON-891856253-project-member] Releasing lock "[datastore1] devstack-image-cache_base/20c8ea6e-0db5-4788-9e34-92c1d63bb7b8/20c8ea6e-0db5-4788-9e34-92c1d63bb7b8.vmdk" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1043.946211] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [None req-a5c61942-d486-43f3-834d-7dfb87ef5b63 tempest-ServersNegativeTestJSON-891856253 tempest-ServersNegativeTestJSON-891856253-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/20c8ea6e-0db5-4788-9e34-92c1d63bb7b8/20c8ea6e-0db5-4788-9e34-92c1d63bb7b8.vmdk to [datastore1] bdbaef31-7d37-40cb-ac3e-ad152b37e2d2/bdbaef31-7d37-40cb-ac3e-ad152b37e2d2.vmdk {{(pid=61806) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 1043.946464] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-7a52d6d1-75b8-4919-beab-ef985461a218 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1043.953347] env[61806]: DEBUG oslo_vmware.api [None req-a5c61942-d486-43f3-834d-7dfb87ef5b63 tempest-ServersNegativeTestJSON-891856253 tempest-ServersNegativeTestJSON-891856253-project-member] Waiting for the task: (returnval){ [ 1043.953347] env[61806]: value = "task-1295159" [ 1043.953347] env[61806]: _type = "Task" [ 1043.953347] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1043.961514] env[61806]: DEBUG oslo_vmware.api [None req-a5c61942-d486-43f3-834d-7dfb87ef5b63 tempest-ServersNegativeTestJSON-891856253 tempest-ServersNegativeTestJSON-891856253-project-member] Task: {'id': task-1295159, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1044.106353] env[61806]: DEBUG nova.network.neutron [req-a7425dce-615e-4503-b23f-02554902280f req-744a4108-be48-4b81-bcf5-428bfeb2f44b service nova] [instance: 7445d2c5-26bd-4f8d-8653-51c721ea801f] Updated VIF entry in instance network info cache for port 773e4470-b54c-4449-b825-5bb680cf8d2f. {{(pid=61806) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1044.106738] env[61806]: DEBUG nova.network.neutron [req-a7425dce-615e-4503-b23f-02554902280f req-744a4108-be48-4b81-bcf5-428bfeb2f44b service nova] [instance: 7445d2c5-26bd-4f8d-8653-51c721ea801f] Updating instance_info_cache with network_info: [{"id": "773e4470-b54c-4449-b825-5bb680cf8d2f", "address": "fa:16:3e:09:73:90", "network": {"id": "b7e78559-d885-4b59-9c7f-9d966b424d04", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-1456721354-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7ab95fb72b5d46c3b6c7bebbccf897cd", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d891a662-7da0-4be5-8d0d-01b4ff70552a", "external-id": "nsx-vlan-transportzone-815", "segmentation_id": 815, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap773e4470-b5", "ovs_interfaceid": "773e4470-b54c-4449-b825-5bb680cf8d2f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61806) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1044.128895] env[61806]: DEBUG oslo_concurrency.lockutils [None req-5eac2f24-7650-441c-94d5-4e1470c88591 tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] Lock "713d4e6e-6f0e-4844-9b55-cc010030eda6" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 6.211s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1044.218083] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-96ad2c8b-ef5b-4201-964f-47ab29d83c13 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1044.225532] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4d5e7d44-7e90-4e10-90d2-e563b54632e7 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1044.258294] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-38d79dd7-c226-46a9-96c2-634e65783eff {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1044.266354] env[61806]: DEBUG oslo_vmware.api [None req-138f28bc-d88a-4f34-babf-797a70b177f1 tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Task: {'id': task-1295158, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.479824} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1044.268481] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [None req-138f28bc-d88a-4f34-babf-797a70b177f1 tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa.vmdk to [datastore2] 06658134-4c6d-4911-9a55-b6805f555c78/06658134-4c6d-4911-9a55-b6805f555c78.vmdk {{(pid=61806) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 1044.268718] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-138f28bc-d88a-4f34-babf-797a70b177f1 tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] [instance: 06658134-4c6d-4911-9a55-b6805f555c78] Extending root virtual disk to 1048576 {{(pid=61806) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1044.269013] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-3112d194-f099-4e43-b934-693bef7923b0 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1044.271724] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ea616a18-e68a-47d6-8539-04008750d465 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1044.279810] env[61806]: DEBUG oslo_vmware.api [None req-138f28bc-d88a-4f34-babf-797a70b177f1 tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Waiting for the task: (returnval){ [ 1044.279810] env[61806]: value = "task-1295160" [ 1044.279810] env[61806]: _type = "Task" [ 1044.279810] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1044.289334] env[61806]: DEBUG nova.compute.provider_tree [None req-4da0f2fc-feb0-4b02-97f3-bd412b867242 tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Inventory has not changed in ProviderTree for provider: a2858be1-fd22-4e08-979e-87ad25293407 {{(pid=61806) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1044.394087] env[61806]: DEBUG oslo_concurrency.lockutils [None req-c257ef4d-1b9a-4377-aada-40cc9a4843fd tempest-AttachVolumeNegativeTest-1683033704 tempest-AttachVolumeNegativeTest-1683033704-project-member] Acquiring lock "refresh_cache-ce2ad097-4c3f-4b55-b018-ac788aa5b662" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1044.394087] env[61806]: DEBUG oslo_concurrency.lockutils [None req-c257ef4d-1b9a-4377-aada-40cc9a4843fd tempest-AttachVolumeNegativeTest-1683033704 tempest-AttachVolumeNegativeTest-1683033704-project-member] Acquired lock "refresh_cache-ce2ad097-4c3f-4b55-b018-ac788aa5b662" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1044.394087] env[61806]: DEBUG nova.network.neutron [None req-c257ef4d-1b9a-4377-aada-40cc9a4843fd tempest-AttachVolumeNegativeTest-1683033704 tempest-AttachVolumeNegativeTest-1683033704-project-member] [instance: ce2ad097-4c3f-4b55-b018-ac788aa5b662] Building network info cache for instance {{(pid=61806) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1044.439473] env[61806]: DEBUG oslo_vmware.api [None req-54021779-f9e2-42db-b76c-c250eacb9121 tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Task: {'id': session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]52067950-eb9d-2dd9-fc63-706698999935, 'name': SearchDatastore_Task, 'duration_secs': 0.060484} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1044.439790] env[61806]: DEBUG oslo_concurrency.lockutils [None req-54021779-f9e2-42db-b76c-c250eacb9121 tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Releasing lock "[datastore2] devstack-image-cache_base/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1044.439826] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-54021779-f9e2-42db-b76c-c250eacb9121 tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] [instance: 7445d2c5-26bd-4f8d-8653-51c721ea801f] Processing image ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa {{(pid=61806) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1044.440104] env[61806]: DEBUG oslo_concurrency.lockutils [None req-54021779-f9e2-42db-b76c-c250eacb9121 tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa.vmdk" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1044.440266] env[61806]: DEBUG oslo_concurrency.lockutils [None req-54021779-f9e2-42db-b76c-c250eacb9121 tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Acquired lock "[datastore2] devstack-image-cache_base/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa.vmdk" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1044.440456] env[61806]: DEBUG nova.virt.vmwareapi.ds_util [None req-54021779-f9e2-42db-b76c-c250eacb9121 tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61806) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1044.440778] env[61806]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-549fa47d-f485-439b-89ca-788e5f3d6fae {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1044.449852] env[61806]: DEBUG nova.virt.vmwareapi.ds_util [None req-54021779-f9e2-42db-b76c-c250eacb9121 tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61806) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1044.450101] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-54021779-f9e2-42db-b76c-c250eacb9121 tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61806) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 1044.450904] env[61806]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-238b1c72-7218-4c43-9659-0274390cb9c0 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1044.462190] env[61806]: DEBUG oslo_vmware.api [None req-a5c61942-d486-43f3-834d-7dfb87ef5b63 tempest-ServersNegativeTestJSON-891856253 tempest-ServersNegativeTestJSON-891856253-project-member] Task: {'id': task-1295159, 'name': CopyVirtualDisk_Task} progress is 12%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1044.463921] env[61806]: DEBUG oslo_vmware.api [None req-54021779-f9e2-42db-b76c-c250eacb9121 tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Waiting for the task: (returnval){ [ 1044.463921] env[61806]: value = "session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]52bea25d-9d3d-1cea-e07a-3b12be220c24" [ 1044.463921] env[61806]: _type = "Task" [ 1044.463921] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1044.474718] env[61806]: DEBUG oslo_vmware.api [None req-54021779-f9e2-42db-b76c-c250eacb9121 tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Task: {'id': session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]52bea25d-9d3d-1cea-e07a-3b12be220c24, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1044.609975] env[61806]: DEBUG oslo_concurrency.lockutils [req-a7425dce-615e-4503-b23f-02554902280f req-744a4108-be48-4b81-bcf5-428bfeb2f44b service nova] Releasing lock "refresh_cache-7445d2c5-26bd-4f8d-8653-51c721ea801f" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1044.610131] env[61806]: DEBUG nova.compute.manager [req-a7425dce-615e-4503-b23f-02554902280f req-744a4108-be48-4b81-bcf5-428bfeb2f44b service nova] [instance: a61abfd9-3cc1-4a2b-b68b-a446094daf87] Received event network-vif-deleted-2950cc6d-7a09-4217-abf5-8a5a940e6653 {{(pid=61806) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1044.795127] env[61806]: DEBUG nova.scheduler.client.report [None req-4da0f2fc-feb0-4b02-97f3-bd412b867242 tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Inventory has not changed for provider a2858be1-fd22-4e08-979e-87ad25293407 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 138, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61806) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1044.803240] env[61806]: DEBUG oslo_vmware.api [None req-138f28bc-d88a-4f34-babf-797a70b177f1 tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Task: {'id': task-1295160, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.135316} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1044.803752] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-138f28bc-d88a-4f34-babf-797a70b177f1 tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] [instance: 06658134-4c6d-4911-9a55-b6805f555c78] Extended root virtual disk {{(pid=61806) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1044.804592] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-94fc096f-f6a3-42cd-914e-058e7702602b {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1044.831843] env[61806]: DEBUG nova.virt.vmwareapi.volumeops [None req-138f28bc-d88a-4f34-babf-797a70b177f1 tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] [instance: 06658134-4c6d-4911-9a55-b6805f555c78] Reconfiguring VM instance instance-00000066 to attach disk [datastore2] 06658134-4c6d-4911-9a55-b6805f555c78/06658134-4c6d-4911-9a55-b6805f555c78.vmdk or device None with type sparse {{(pid=61806) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1044.832834] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-84b7d2a4-b1ea-45be-adaa-8d55db87a4b2 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1044.848376] env[61806]: DEBUG oslo_concurrency.lockutils [None req-8942a881-ca25-4ecf-bd1a-43e4a7d53138 tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] Acquiring lock "d1802da4-a670-427d-84d7-0fcb4717e18f" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1044.848638] env[61806]: DEBUG oslo_concurrency.lockutils [None req-8942a881-ca25-4ecf-bd1a-43e4a7d53138 tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] Lock "d1802da4-a670-427d-84d7-0fcb4717e18f" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1044.848896] env[61806]: DEBUG oslo_concurrency.lockutils [None req-8942a881-ca25-4ecf-bd1a-43e4a7d53138 tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] Acquiring lock "d1802da4-a670-427d-84d7-0fcb4717e18f-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1044.849131] env[61806]: DEBUG oslo_concurrency.lockutils [None req-8942a881-ca25-4ecf-bd1a-43e4a7d53138 tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] Lock "d1802da4-a670-427d-84d7-0fcb4717e18f-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1044.849317] env[61806]: DEBUG oslo_concurrency.lockutils [None req-8942a881-ca25-4ecf-bd1a-43e4a7d53138 tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] Lock "d1802da4-a670-427d-84d7-0fcb4717e18f-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1044.851659] env[61806]: INFO nova.compute.manager [None req-8942a881-ca25-4ecf-bd1a-43e4a7d53138 tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] [instance: d1802da4-a670-427d-84d7-0fcb4717e18f] Terminating instance [ 1044.853931] env[61806]: DEBUG nova.compute.manager [None req-8942a881-ca25-4ecf-bd1a-43e4a7d53138 tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] [instance: d1802da4-a670-427d-84d7-0fcb4717e18f] Start destroying the instance on the hypervisor. {{(pid=61806) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1044.854183] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-8942a881-ca25-4ecf-bd1a-43e4a7d53138 tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] [instance: d1802da4-a670-427d-84d7-0fcb4717e18f] Destroying instance {{(pid=61806) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1044.855171] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-919f5d77-e98a-493d-88bd-aefe477f0b0b {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1044.861203] env[61806]: DEBUG oslo_vmware.api [None req-138f28bc-d88a-4f34-babf-797a70b177f1 tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Waiting for the task: (returnval){ [ 1044.861203] env[61806]: value = "task-1295161" [ 1044.861203] env[61806]: _type = "Task" [ 1044.861203] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1044.863610] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [None req-8942a881-ca25-4ecf-bd1a-43e4a7d53138 tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] [instance: d1802da4-a670-427d-84d7-0fcb4717e18f] Powering off the VM {{(pid=61806) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1044.867399] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-2932de83-9ed2-496b-b47f-7491382a9646 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1044.871443] env[61806]: DEBUG nova.compute.manager [req-1e28f8b4-662a-4750-a679-63add0fd58a3 req-852d2410-b221-4aad-bd38-6c8e9d2cbaec service nova] [instance: ce2ad097-4c3f-4b55-b018-ac788aa5b662] Received event network-vif-plugged-6792e7fd-4056-4d2f-86a5-45ea41172f3f {{(pid=61806) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1044.871711] env[61806]: DEBUG oslo_concurrency.lockutils [req-1e28f8b4-662a-4750-a679-63add0fd58a3 req-852d2410-b221-4aad-bd38-6c8e9d2cbaec service nova] Acquiring lock "ce2ad097-4c3f-4b55-b018-ac788aa5b662-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1044.871963] env[61806]: DEBUG oslo_concurrency.lockutils [req-1e28f8b4-662a-4750-a679-63add0fd58a3 req-852d2410-b221-4aad-bd38-6c8e9d2cbaec service nova] Lock "ce2ad097-4c3f-4b55-b018-ac788aa5b662-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1044.872283] env[61806]: DEBUG oslo_concurrency.lockutils [req-1e28f8b4-662a-4750-a679-63add0fd58a3 req-852d2410-b221-4aad-bd38-6c8e9d2cbaec service nova] Lock "ce2ad097-4c3f-4b55-b018-ac788aa5b662-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1044.872355] env[61806]: DEBUG nova.compute.manager [req-1e28f8b4-662a-4750-a679-63add0fd58a3 req-852d2410-b221-4aad-bd38-6c8e9d2cbaec service nova] [instance: ce2ad097-4c3f-4b55-b018-ac788aa5b662] No waiting events found dispatching network-vif-plugged-6792e7fd-4056-4d2f-86a5-45ea41172f3f {{(pid=61806) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1044.872515] env[61806]: WARNING nova.compute.manager [req-1e28f8b4-662a-4750-a679-63add0fd58a3 req-852d2410-b221-4aad-bd38-6c8e9d2cbaec service nova] [instance: ce2ad097-4c3f-4b55-b018-ac788aa5b662] Received unexpected event network-vif-plugged-6792e7fd-4056-4d2f-86a5-45ea41172f3f for instance with vm_state building and task_state spawning. [ 1044.872692] env[61806]: DEBUG nova.compute.manager [req-1e28f8b4-662a-4750-a679-63add0fd58a3 req-852d2410-b221-4aad-bd38-6c8e9d2cbaec service nova] [instance: ce2ad097-4c3f-4b55-b018-ac788aa5b662] Received event network-changed-6792e7fd-4056-4d2f-86a5-45ea41172f3f {{(pid=61806) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1044.872902] env[61806]: DEBUG nova.compute.manager [req-1e28f8b4-662a-4750-a679-63add0fd58a3 req-852d2410-b221-4aad-bd38-6c8e9d2cbaec service nova] [instance: ce2ad097-4c3f-4b55-b018-ac788aa5b662] Refreshing instance network info cache due to event network-changed-6792e7fd-4056-4d2f-86a5-45ea41172f3f. {{(pid=61806) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 1044.873151] env[61806]: DEBUG oslo_concurrency.lockutils [req-1e28f8b4-662a-4750-a679-63add0fd58a3 req-852d2410-b221-4aad-bd38-6c8e9d2cbaec service nova] Acquiring lock "refresh_cache-ce2ad097-4c3f-4b55-b018-ac788aa5b662" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1044.878959] env[61806]: DEBUG oslo_vmware.api [None req-138f28bc-d88a-4f34-babf-797a70b177f1 tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Task: {'id': task-1295161, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1044.880573] env[61806]: DEBUG oslo_vmware.api [None req-8942a881-ca25-4ecf-bd1a-43e4a7d53138 tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] Waiting for the task: (returnval){ [ 1044.880573] env[61806]: value = "task-1295162" [ 1044.880573] env[61806]: _type = "Task" [ 1044.880573] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1044.892157] env[61806]: DEBUG oslo_vmware.api [None req-8942a881-ca25-4ecf-bd1a-43e4a7d53138 tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] Task: {'id': task-1295162, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1044.929290] env[61806]: DEBUG nova.network.neutron [None req-c257ef4d-1b9a-4377-aada-40cc9a4843fd tempest-AttachVolumeNegativeTest-1683033704 tempest-AttachVolumeNegativeTest-1683033704-project-member] [instance: ce2ad097-4c3f-4b55-b018-ac788aa5b662] Instance cache missing network info. {{(pid=61806) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 1044.968573] env[61806]: DEBUG oslo_vmware.api [None req-a5c61942-d486-43f3-834d-7dfb87ef5b63 tempest-ServersNegativeTestJSON-891856253 tempest-ServersNegativeTestJSON-891856253-project-member] Task: {'id': task-1295159, 'name': CopyVirtualDisk_Task} progress is 35%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1044.984063] env[61806]: DEBUG oslo_vmware.api [None req-54021779-f9e2-42db-b76c-c250eacb9121 tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Task: {'id': session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]52bea25d-9d3d-1cea-e07a-3b12be220c24, 'name': SearchDatastore_Task, 'duration_secs': 0.014505} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1044.984063] env[61806]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-9627efa6-d07b-4686-8d66-a677bf4430f1 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1044.992198] env[61806]: DEBUG oslo_vmware.api [None req-54021779-f9e2-42db-b76c-c250eacb9121 tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Waiting for the task: (returnval){ [ 1044.992198] env[61806]: value = "session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]523a3183-4364-1474-e453-d3495b50ca5b" [ 1044.992198] env[61806]: _type = "Task" [ 1044.992198] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1045.002922] env[61806]: DEBUG oslo_vmware.api [None req-54021779-f9e2-42db-b76c-c250eacb9121 tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Task: {'id': session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]523a3183-4364-1474-e453-d3495b50ca5b, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1045.084290] env[61806]: DEBUG nova.network.neutron [None req-c257ef4d-1b9a-4377-aada-40cc9a4843fd tempest-AttachVolumeNegativeTest-1683033704 tempest-AttachVolumeNegativeTest-1683033704-project-member] [instance: ce2ad097-4c3f-4b55-b018-ac788aa5b662] Updating instance_info_cache with network_info: [{"id": "6792e7fd-4056-4d2f-86a5-45ea41172f3f", "address": "fa:16:3e:64:1b:81", "network": {"id": "d636ca55-0737-48e4-8d9d-7fe0f797f49e", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-2090334349-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "22e47882c2c44ce492dff6b1c5d782e2", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6f493cd8-1cb4-42a1-8d56-bfa6ac7cf563", "external-id": "nsx-vlan-transportzone-931", "segmentation_id": 931, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap6792e7fd-40", "ovs_interfaceid": "6792e7fd-4056-4d2f-86a5-45ea41172f3f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61806) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1045.305479] env[61806]: DEBUG oslo_concurrency.lockutils [None req-4da0f2fc-feb0-4b02-97f3-bd412b867242 tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.709s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1045.327281] env[61806]: INFO nova.scheduler.client.report [None req-4da0f2fc-feb0-4b02-97f3-bd412b867242 tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Deleted allocations for instance a61abfd9-3cc1-4a2b-b68b-a446094daf87 [ 1045.376585] env[61806]: DEBUG oslo_vmware.api [None req-138f28bc-d88a-4f34-babf-797a70b177f1 tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Task: {'id': task-1295161, 'name': ReconfigVM_Task, 'duration_secs': 0.377495} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1045.376970] env[61806]: DEBUG nova.virt.vmwareapi.volumeops [None req-138f28bc-d88a-4f34-babf-797a70b177f1 tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] [instance: 06658134-4c6d-4911-9a55-b6805f555c78] Reconfigured VM instance instance-00000066 to attach disk [datastore2] 06658134-4c6d-4911-9a55-b6805f555c78/06658134-4c6d-4911-9a55-b6805f555c78.vmdk or device None with type sparse {{(pid=61806) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1045.377713] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-17db1dfc-367c-40e7-a707-0e1082cd8a36 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1045.387113] env[61806]: DEBUG oslo_vmware.api [None req-138f28bc-d88a-4f34-babf-797a70b177f1 tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Waiting for the task: (returnval){ [ 1045.387113] env[61806]: value = "task-1295163" [ 1045.387113] env[61806]: _type = "Task" [ 1045.387113] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1045.394021] env[61806]: DEBUG oslo_vmware.api [None req-8942a881-ca25-4ecf-bd1a-43e4a7d53138 tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] Task: {'id': task-1295162, 'name': PowerOffVM_Task, 'duration_secs': 0.504517} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1045.395013] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [None req-8942a881-ca25-4ecf-bd1a-43e4a7d53138 tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] [instance: d1802da4-a670-427d-84d7-0fcb4717e18f] Powered off the VM {{(pid=61806) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1045.395251] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-8942a881-ca25-4ecf-bd1a-43e4a7d53138 tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] [instance: d1802da4-a670-427d-84d7-0fcb4717e18f] Unregistering the VM {{(pid=61806) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1045.395524] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-4285dc19-f232-4f38-96e3-026793b80023 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1045.401749] env[61806]: DEBUG oslo_vmware.api [None req-138f28bc-d88a-4f34-babf-797a70b177f1 tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Task: {'id': task-1295163, 'name': Rename_Task} progress is 6%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1045.466824] env[61806]: DEBUG oslo_vmware.api [None req-a5c61942-d486-43f3-834d-7dfb87ef5b63 tempest-ServersNegativeTestJSON-891856253 tempest-ServersNegativeTestJSON-891856253-project-member] Task: {'id': task-1295159, 'name': CopyVirtualDisk_Task} progress is 57%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1045.507207] env[61806]: DEBUG oslo_vmware.api [None req-54021779-f9e2-42db-b76c-c250eacb9121 tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Task: {'id': session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]523a3183-4364-1474-e453-d3495b50ca5b, 'name': SearchDatastore_Task, 'duration_secs': 0.01234} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1045.508070] env[61806]: DEBUG oslo_concurrency.lockutils [None req-54021779-f9e2-42db-b76c-c250eacb9121 tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Releasing lock "[datastore2] devstack-image-cache_base/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa.vmdk" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1045.508445] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [None req-54021779-f9e2-42db-b76c-c250eacb9121 tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa.vmdk to [datastore2] 7445d2c5-26bd-4f8d-8653-51c721ea801f/7445d2c5-26bd-4f8d-8653-51c721ea801f.vmdk {{(pid=61806) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 1045.508870] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-a32a2e57-9050-4b1f-bc4d-384f0451f10e {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1045.519068] env[61806]: DEBUG oslo_vmware.api [None req-54021779-f9e2-42db-b76c-c250eacb9121 tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Waiting for the task: (returnval){ [ 1045.519068] env[61806]: value = "task-1295165" [ 1045.519068] env[61806]: _type = "Task" [ 1045.519068] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1045.531121] env[61806]: DEBUG oslo_vmware.api [None req-54021779-f9e2-42db-b76c-c250eacb9121 tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Task: {'id': task-1295165, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1045.550670] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-8942a881-ca25-4ecf-bd1a-43e4a7d53138 tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] [instance: d1802da4-a670-427d-84d7-0fcb4717e18f] Unregistered the VM {{(pid=61806) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1045.551138] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-8942a881-ca25-4ecf-bd1a-43e4a7d53138 tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] [instance: d1802da4-a670-427d-84d7-0fcb4717e18f] Deleting contents of the VM from datastore datastore1 {{(pid=61806) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1045.551484] env[61806]: DEBUG nova.virt.vmwareapi.ds_util [None req-8942a881-ca25-4ecf-bd1a-43e4a7d53138 tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] Deleting the datastore file [datastore1] d1802da4-a670-427d-84d7-0fcb4717e18f {{(pid=61806) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1045.551973] env[61806]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-833ad95c-7a63-43b4-a8bd-55f30df4be40 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1045.560165] env[61806]: DEBUG oslo_vmware.api [None req-8942a881-ca25-4ecf-bd1a-43e4a7d53138 tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] Waiting for the task: (returnval){ [ 1045.560165] env[61806]: value = "task-1295166" [ 1045.560165] env[61806]: _type = "Task" [ 1045.560165] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1045.570635] env[61806]: DEBUG oslo_vmware.api [None req-8942a881-ca25-4ecf-bd1a-43e4a7d53138 tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] Task: {'id': task-1295166, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1045.587431] env[61806]: DEBUG oslo_concurrency.lockutils [None req-c257ef4d-1b9a-4377-aada-40cc9a4843fd tempest-AttachVolumeNegativeTest-1683033704 tempest-AttachVolumeNegativeTest-1683033704-project-member] Releasing lock "refresh_cache-ce2ad097-4c3f-4b55-b018-ac788aa5b662" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1045.588126] env[61806]: DEBUG nova.compute.manager [None req-c257ef4d-1b9a-4377-aada-40cc9a4843fd tempest-AttachVolumeNegativeTest-1683033704 tempest-AttachVolumeNegativeTest-1683033704-project-member] [instance: ce2ad097-4c3f-4b55-b018-ac788aa5b662] Instance network_info: |[{"id": "6792e7fd-4056-4d2f-86a5-45ea41172f3f", "address": "fa:16:3e:64:1b:81", "network": {"id": "d636ca55-0737-48e4-8d9d-7fe0f797f49e", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-2090334349-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "22e47882c2c44ce492dff6b1c5d782e2", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6f493cd8-1cb4-42a1-8d56-bfa6ac7cf563", "external-id": "nsx-vlan-transportzone-931", "segmentation_id": 931, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap6792e7fd-40", "ovs_interfaceid": "6792e7fd-4056-4d2f-86a5-45ea41172f3f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61806) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1045.589160] env[61806]: DEBUG oslo_concurrency.lockutils [req-1e28f8b4-662a-4750-a679-63add0fd58a3 req-852d2410-b221-4aad-bd38-6c8e9d2cbaec service nova] Acquired lock "refresh_cache-ce2ad097-4c3f-4b55-b018-ac788aa5b662" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1045.589160] env[61806]: DEBUG nova.network.neutron [req-1e28f8b4-662a-4750-a679-63add0fd58a3 req-852d2410-b221-4aad-bd38-6c8e9d2cbaec service nova] [instance: ce2ad097-4c3f-4b55-b018-ac788aa5b662] Refreshing network info cache for port 6792e7fd-4056-4d2f-86a5-45ea41172f3f {{(pid=61806) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1045.590267] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-c257ef4d-1b9a-4377-aada-40cc9a4843fd tempest-AttachVolumeNegativeTest-1683033704 tempest-AttachVolumeNegativeTest-1683033704-project-member] [instance: ce2ad097-4c3f-4b55-b018-ac788aa5b662] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:64:1b:81', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '6f493cd8-1cb4-42a1-8d56-bfa6ac7cf563', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '6792e7fd-4056-4d2f-86a5-45ea41172f3f', 'vif_model': 'vmxnet3'}] {{(pid=61806) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1045.600035] env[61806]: DEBUG oslo.service.loopingcall [None req-c257ef4d-1b9a-4377-aada-40cc9a4843fd tempest-AttachVolumeNegativeTest-1683033704 tempest-AttachVolumeNegativeTest-1683033704-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61806) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1045.601223] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: ce2ad097-4c3f-4b55-b018-ac788aa5b662] Creating VM on the ESX host {{(pid=61806) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 1045.601596] env[61806]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-3e7da7e5-1375-4cab-ba30-a7eed17a8958 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1045.624971] env[61806]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1045.624971] env[61806]: value = "task-1295167" [ 1045.624971] env[61806]: _type = "Task" [ 1045.624971] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1045.634066] env[61806]: DEBUG oslo_vmware.api [-] Task: {'id': task-1295167, 'name': CreateVM_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1045.838596] env[61806]: DEBUG oslo_concurrency.lockutils [None req-4da0f2fc-feb0-4b02-97f3-bd412b867242 tempest-ServerDiskConfigTestJSON-130516410 tempest-ServerDiskConfigTestJSON-130516410-project-member] Lock "a61abfd9-3cc1-4a2b-b68b-a446094daf87" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.530s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1045.900573] env[61806]: DEBUG oslo_vmware.api [None req-138f28bc-d88a-4f34-babf-797a70b177f1 tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Task: {'id': task-1295163, 'name': Rename_Task, 'duration_secs': 0.188852} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1045.900912] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [None req-138f28bc-d88a-4f34-babf-797a70b177f1 tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] [instance: 06658134-4c6d-4911-9a55-b6805f555c78] Powering on the VM {{(pid=61806) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1045.901273] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-c5690772-2a8e-4304-96f8-3c5655a61ecc {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1045.909599] env[61806]: DEBUG oslo_vmware.api [None req-138f28bc-d88a-4f34-babf-797a70b177f1 tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Waiting for the task: (returnval){ [ 1045.909599] env[61806]: value = "task-1295168" [ 1045.909599] env[61806]: _type = "Task" [ 1045.909599] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1045.918919] env[61806]: DEBUG oslo_vmware.api [None req-138f28bc-d88a-4f34-babf-797a70b177f1 tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Task: {'id': task-1295168, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1045.966936] env[61806]: DEBUG oslo_vmware.api [None req-a5c61942-d486-43f3-834d-7dfb87ef5b63 tempest-ServersNegativeTestJSON-891856253 tempest-ServersNegativeTestJSON-891856253-project-member] Task: {'id': task-1295159, 'name': CopyVirtualDisk_Task} progress is 80%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1046.029756] env[61806]: DEBUG oslo_vmware.api [None req-54021779-f9e2-42db-b76c-c250eacb9121 tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Task: {'id': task-1295165, 'name': CopyVirtualDisk_Task} progress is 25%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1046.070716] env[61806]: DEBUG oslo_vmware.api [None req-8942a881-ca25-4ecf-bd1a-43e4a7d53138 tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] Task: {'id': task-1295166, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1046.136902] env[61806]: DEBUG oslo_vmware.api [-] Task: {'id': task-1295167, 'name': CreateVM_Task} progress is 25%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1046.365326] env[61806]: DEBUG nova.network.neutron [req-1e28f8b4-662a-4750-a679-63add0fd58a3 req-852d2410-b221-4aad-bd38-6c8e9d2cbaec service nova] [instance: ce2ad097-4c3f-4b55-b018-ac788aa5b662] Updated VIF entry in instance network info cache for port 6792e7fd-4056-4d2f-86a5-45ea41172f3f. {{(pid=61806) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1046.365750] env[61806]: DEBUG nova.network.neutron [req-1e28f8b4-662a-4750-a679-63add0fd58a3 req-852d2410-b221-4aad-bd38-6c8e9d2cbaec service nova] [instance: ce2ad097-4c3f-4b55-b018-ac788aa5b662] Updating instance_info_cache with network_info: [{"id": "6792e7fd-4056-4d2f-86a5-45ea41172f3f", "address": "fa:16:3e:64:1b:81", "network": {"id": "d636ca55-0737-48e4-8d9d-7fe0f797f49e", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-2090334349-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "22e47882c2c44ce492dff6b1c5d782e2", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6f493cd8-1cb4-42a1-8d56-bfa6ac7cf563", "external-id": "nsx-vlan-transportzone-931", "segmentation_id": 931, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap6792e7fd-40", "ovs_interfaceid": "6792e7fd-4056-4d2f-86a5-45ea41172f3f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61806) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1046.421630] env[61806]: DEBUG oslo_vmware.api [None req-138f28bc-d88a-4f34-babf-797a70b177f1 tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Task: {'id': task-1295168, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1046.466875] env[61806]: DEBUG oslo_vmware.api [None req-a5c61942-d486-43f3-834d-7dfb87ef5b63 tempest-ServersNegativeTestJSON-891856253 tempest-ServersNegativeTestJSON-891856253-project-member] Task: {'id': task-1295159, 'name': CopyVirtualDisk_Task, 'duration_secs': 2.491714} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1046.467192] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [None req-a5c61942-d486-43f3-834d-7dfb87ef5b63 tempest-ServersNegativeTestJSON-891856253 tempest-ServersNegativeTestJSON-891856253-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/20c8ea6e-0db5-4788-9e34-92c1d63bb7b8/20c8ea6e-0db5-4788-9e34-92c1d63bb7b8.vmdk to [datastore1] bdbaef31-7d37-40cb-ac3e-ad152b37e2d2/bdbaef31-7d37-40cb-ac3e-ad152b37e2d2.vmdk {{(pid=61806) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 1046.468115] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ae1976de-f37f-4186-8bf9-2bbfa2bb56ba {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1046.490270] env[61806]: DEBUG nova.virt.vmwareapi.volumeops [None req-a5c61942-d486-43f3-834d-7dfb87ef5b63 tempest-ServersNegativeTestJSON-891856253 tempest-ServersNegativeTestJSON-891856253-project-member] [instance: bdbaef31-7d37-40cb-ac3e-ad152b37e2d2] Reconfiguring VM instance instance-0000005a to attach disk [datastore1] bdbaef31-7d37-40cb-ac3e-ad152b37e2d2/bdbaef31-7d37-40cb-ac3e-ad152b37e2d2.vmdk or device None with type streamOptimized {{(pid=61806) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1046.490578] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-60acaa4b-7eb6-4de1-b242-43ae8b22ebb5 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1046.510245] env[61806]: DEBUG oslo_vmware.api [None req-a5c61942-d486-43f3-834d-7dfb87ef5b63 tempest-ServersNegativeTestJSON-891856253 tempest-ServersNegativeTestJSON-891856253-project-member] Waiting for the task: (returnval){ [ 1046.510245] env[61806]: value = "task-1295169" [ 1046.510245] env[61806]: _type = "Task" [ 1046.510245] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1046.519419] env[61806]: DEBUG oslo_vmware.api [None req-a5c61942-d486-43f3-834d-7dfb87ef5b63 tempest-ServersNegativeTestJSON-891856253 tempest-ServersNegativeTestJSON-891856253-project-member] Task: {'id': task-1295169, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1046.529722] env[61806]: DEBUG oslo_vmware.api [None req-54021779-f9e2-42db-b76c-c250eacb9121 tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Task: {'id': task-1295165, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.719493} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1046.530088] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [None req-54021779-f9e2-42db-b76c-c250eacb9121 tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa.vmdk to [datastore2] 7445d2c5-26bd-4f8d-8653-51c721ea801f/7445d2c5-26bd-4f8d-8653-51c721ea801f.vmdk {{(pid=61806) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 1046.530350] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-54021779-f9e2-42db-b76c-c250eacb9121 tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] [instance: 7445d2c5-26bd-4f8d-8653-51c721ea801f] Extending root virtual disk to 1048576 {{(pid=61806) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1046.530691] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-b0d9dc11-335e-43ba-b435-8e68a7e78133 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1046.538827] env[61806]: DEBUG oslo_vmware.api [None req-54021779-f9e2-42db-b76c-c250eacb9121 tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Waiting for the task: (returnval){ [ 1046.538827] env[61806]: value = "task-1295170" [ 1046.538827] env[61806]: _type = "Task" [ 1046.538827] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1046.546709] env[61806]: DEBUG oslo_vmware.api [None req-54021779-f9e2-42db-b76c-c250eacb9121 tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Task: {'id': task-1295170, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1046.571391] env[61806]: DEBUG oslo_vmware.api [None req-8942a881-ca25-4ecf-bd1a-43e4a7d53138 tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] Task: {'id': task-1295166, 'name': DeleteDatastoreFile_Task, 'duration_secs': 1.009587} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1046.571729] env[61806]: DEBUG nova.virt.vmwareapi.ds_util [None req-8942a881-ca25-4ecf-bd1a-43e4a7d53138 tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] Deleted the datastore file {{(pid=61806) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1046.571926] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-8942a881-ca25-4ecf-bd1a-43e4a7d53138 tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] [instance: d1802da4-a670-427d-84d7-0fcb4717e18f] Deleted contents of the VM from datastore datastore1 {{(pid=61806) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1046.572171] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-8942a881-ca25-4ecf-bd1a-43e4a7d53138 tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] [instance: d1802da4-a670-427d-84d7-0fcb4717e18f] Instance destroyed {{(pid=61806) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1046.572391] env[61806]: INFO nova.compute.manager [None req-8942a881-ca25-4ecf-bd1a-43e4a7d53138 tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] [instance: d1802da4-a670-427d-84d7-0fcb4717e18f] Took 1.72 seconds to destroy the instance on the hypervisor. [ 1046.572686] env[61806]: DEBUG oslo.service.loopingcall [None req-8942a881-ca25-4ecf-bd1a-43e4a7d53138 tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61806) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1046.572890] env[61806]: DEBUG nova.compute.manager [-] [instance: d1802da4-a670-427d-84d7-0fcb4717e18f] Deallocating network for instance {{(pid=61806) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1046.573046] env[61806]: DEBUG nova.network.neutron [-] [instance: d1802da4-a670-427d-84d7-0fcb4717e18f] deallocate_for_instance() {{(pid=61806) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1046.639568] env[61806]: DEBUG oslo_vmware.api [-] Task: {'id': task-1295167, 'name': CreateVM_Task} progress is 99%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1046.869093] env[61806]: DEBUG oslo_concurrency.lockutils [req-1e28f8b4-662a-4750-a679-63add0fd58a3 req-852d2410-b221-4aad-bd38-6c8e9d2cbaec service nova] Releasing lock "refresh_cache-ce2ad097-4c3f-4b55-b018-ac788aa5b662" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1046.920749] env[61806]: DEBUG oslo_vmware.api [None req-138f28bc-d88a-4f34-babf-797a70b177f1 tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Task: {'id': task-1295168, 'name': PowerOnVM_Task, 'duration_secs': 0.880786} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1046.921111] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [None req-138f28bc-d88a-4f34-babf-797a70b177f1 tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] [instance: 06658134-4c6d-4911-9a55-b6805f555c78] Powered on the VM {{(pid=61806) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1046.921338] env[61806]: INFO nova.compute.manager [None req-138f28bc-d88a-4f34-babf-797a70b177f1 tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] [instance: 06658134-4c6d-4911-9a55-b6805f555c78] Took 8.08 seconds to spawn the instance on the hypervisor. [ 1046.921523] env[61806]: DEBUG nova.compute.manager [None req-138f28bc-d88a-4f34-babf-797a70b177f1 tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] [instance: 06658134-4c6d-4911-9a55-b6805f555c78] Checking state {{(pid=61806) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1046.922689] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0832dc7b-270d-4eec-961d-935217fb087a {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1047.020926] env[61806]: DEBUG oslo_vmware.api [None req-a5c61942-d486-43f3-834d-7dfb87ef5b63 tempest-ServersNegativeTestJSON-891856253 tempest-ServersNegativeTestJSON-891856253-project-member] Task: {'id': task-1295169, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1047.035596] env[61806]: DEBUG nova.compute.manager [req-398ede7d-0d73-4845-a969-0879c06995ca req-5a4eb26d-7365-4d5e-9312-621eea0b5ab1 service nova] [instance: d1802da4-a670-427d-84d7-0fcb4717e18f] Received event network-vif-deleted-16d93306-6614-4a1d-aa33-e18d3f79ed1e {{(pid=61806) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1047.035800] env[61806]: INFO nova.compute.manager [req-398ede7d-0d73-4845-a969-0879c06995ca req-5a4eb26d-7365-4d5e-9312-621eea0b5ab1 service nova] [instance: d1802da4-a670-427d-84d7-0fcb4717e18f] Neutron deleted interface 16d93306-6614-4a1d-aa33-e18d3f79ed1e; detaching it from the instance and deleting it from the info cache [ 1047.035977] env[61806]: DEBUG nova.network.neutron [req-398ede7d-0d73-4845-a969-0879c06995ca req-5a4eb26d-7365-4d5e-9312-621eea0b5ab1 service nova] [instance: d1802da4-a670-427d-84d7-0fcb4717e18f] Updating instance_info_cache with network_info: [] {{(pid=61806) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1047.049244] env[61806]: DEBUG oslo_vmware.api [None req-54021779-f9e2-42db-b76c-c250eacb9121 tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Task: {'id': task-1295170, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.069363} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1047.050181] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-54021779-f9e2-42db-b76c-c250eacb9121 tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] [instance: 7445d2c5-26bd-4f8d-8653-51c721ea801f] Extended root virtual disk {{(pid=61806) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1047.050951] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-55bbc6fc-5137-44fc-88ba-eee4aa0d0ea6 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1047.073642] env[61806]: DEBUG nova.virt.vmwareapi.volumeops [None req-54021779-f9e2-42db-b76c-c250eacb9121 tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] [instance: 7445d2c5-26bd-4f8d-8653-51c721ea801f] Reconfiguring VM instance instance-00000067 to attach disk [datastore2] 7445d2c5-26bd-4f8d-8653-51c721ea801f/7445d2c5-26bd-4f8d-8653-51c721ea801f.vmdk or device None with type sparse {{(pid=61806) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1047.074417] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-f2f8b622-3e26-46e0-9599-bb9e3e159287 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1047.095400] env[61806]: DEBUG oslo_vmware.api [None req-54021779-f9e2-42db-b76c-c250eacb9121 tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Waiting for the task: (returnval){ [ 1047.095400] env[61806]: value = "task-1295171" [ 1047.095400] env[61806]: _type = "Task" [ 1047.095400] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1047.103551] env[61806]: DEBUG oslo_vmware.api [None req-54021779-f9e2-42db-b76c-c250eacb9121 tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Task: {'id': task-1295171, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1047.135738] env[61806]: DEBUG oslo_vmware.api [-] Task: {'id': task-1295167, 'name': CreateVM_Task, 'duration_secs': 1.085198} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1047.135902] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: ce2ad097-4c3f-4b55-b018-ac788aa5b662] Created VM on the ESX host {{(pid=61806) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 1047.136659] env[61806]: DEBUG oslo_concurrency.lockutils [None req-c257ef4d-1b9a-4377-aada-40cc9a4843fd tempest-AttachVolumeNegativeTest-1683033704 tempest-AttachVolumeNegativeTest-1683033704-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1047.136867] env[61806]: DEBUG oslo_concurrency.lockutils [None req-c257ef4d-1b9a-4377-aada-40cc9a4843fd tempest-AttachVolumeNegativeTest-1683033704 tempest-AttachVolumeNegativeTest-1683033704-project-member] Acquired lock "[datastore1] devstack-image-cache_base/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1047.137270] env[61806]: DEBUG oslo_concurrency.lockutils [None req-c257ef4d-1b9a-4377-aada-40cc9a4843fd tempest-AttachVolumeNegativeTest-1683033704 tempest-AttachVolumeNegativeTest-1683033704-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1047.137552] env[61806]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-6642117b-8798-4e6e-aa11-d89a96f31c19 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1047.142579] env[61806]: DEBUG oslo_vmware.api [None req-c257ef4d-1b9a-4377-aada-40cc9a4843fd tempest-AttachVolumeNegativeTest-1683033704 tempest-AttachVolumeNegativeTest-1683033704-project-member] Waiting for the task: (returnval){ [ 1047.142579] env[61806]: value = "session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]5200c996-b110-d840-f60b-905391cd31fb" [ 1047.142579] env[61806]: _type = "Task" [ 1047.142579] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1047.150280] env[61806]: DEBUG oslo_vmware.api [None req-c257ef4d-1b9a-4377-aada-40cc9a4843fd tempest-AttachVolumeNegativeTest-1683033704 tempest-AttachVolumeNegativeTest-1683033704-project-member] Task: {'id': session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]5200c996-b110-d840-f60b-905391cd31fb, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1047.442628] env[61806]: INFO nova.compute.manager [None req-138f28bc-d88a-4f34-babf-797a70b177f1 tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] [instance: 06658134-4c6d-4911-9a55-b6805f555c78] Took 18.78 seconds to build instance. [ 1047.494680] env[61806]: DEBUG nova.network.neutron [-] [instance: d1802da4-a670-427d-84d7-0fcb4717e18f] Updating instance_info_cache with network_info: [] {{(pid=61806) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1047.521236] env[61806]: DEBUG oslo_vmware.api [None req-a5c61942-d486-43f3-834d-7dfb87ef5b63 tempest-ServersNegativeTestJSON-891856253 tempest-ServersNegativeTestJSON-891856253-project-member] Task: {'id': task-1295169, 'name': ReconfigVM_Task, 'duration_secs': 0.855471} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1047.521607] env[61806]: DEBUG nova.virt.vmwareapi.volumeops [None req-a5c61942-d486-43f3-834d-7dfb87ef5b63 tempest-ServersNegativeTestJSON-891856253 tempest-ServersNegativeTestJSON-891856253-project-member] [instance: bdbaef31-7d37-40cb-ac3e-ad152b37e2d2] Reconfigured VM instance instance-0000005a to attach disk [datastore1] bdbaef31-7d37-40cb-ac3e-ad152b37e2d2/bdbaef31-7d37-40cb-ac3e-ad152b37e2d2.vmdk or device None with type streamOptimized {{(pid=61806) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1047.522315] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-b85ae533-d5c9-4b3a-a23c-7dc1ac912b0c {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1047.528975] env[61806]: DEBUG oslo_vmware.api [None req-a5c61942-d486-43f3-834d-7dfb87ef5b63 tempest-ServersNegativeTestJSON-891856253 tempest-ServersNegativeTestJSON-891856253-project-member] Waiting for the task: (returnval){ [ 1047.528975] env[61806]: value = "task-1295172" [ 1047.528975] env[61806]: _type = "Task" [ 1047.528975] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1047.538954] env[61806]: DEBUG oslo_vmware.api [None req-a5c61942-d486-43f3-834d-7dfb87ef5b63 tempest-ServersNegativeTestJSON-891856253 tempest-ServersNegativeTestJSON-891856253-project-member] Task: {'id': task-1295172, 'name': Rename_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1047.539197] env[61806]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-e5b4b67a-9a47-4109-b1fb-2271be138ca2 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1047.546899] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bceffe52-2829-4d46-8450-cc50b15865f5 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1047.575261] env[61806]: DEBUG nova.compute.manager [req-398ede7d-0d73-4845-a969-0879c06995ca req-5a4eb26d-7365-4d5e-9312-621eea0b5ab1 service nova] [instance: d1802da4-a670-427d-84d7-0fcb4717e18f] Detach interface failed, port_id=16d93306-6614-4a1d-aa33-e18d3f79ed1e, reason: Instance d1802da4-a670-427d-84d7-0fcb4717e18f could not be found. {{(pid=61806) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10963}} [ 1047.604571] env[61806]: DEBUG oslo_vmware.api [None req-54021779-f9e2-42db-b76c-c250eacb9121 tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Task: {'id': task-1295171, 'name': ReconfigVM_Task, 'duration_secs': 0.298922} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1047.604922] env[61806]: DEBUG nova.virt.vmwareapi.volumeops [None req-54021779-f9e2-42db-b76c-c250eacb9121 tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] [instance: 7445d2c5-26bd-4f8d-8653-51c721ea801f] Reconfigured VM instance instance-00000067 to attach disk [datastore2] 7445d2c5-26bd-4f8d-8653-51c721ea801f/7445d2c5-26bd-4f8d-8653-51c721ea801f.vmdk or device None with type sparse {{(pid=61806) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1047.605623] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-00bf1f0a-215d-4ec8-ada4-baa462d5d0ec {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1047.611487] env[61806]: DEBUG oslo_vmware.api [None req-54021779-f9e2-42db-b76c-c250eacb9121 tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Waiting for the task: (returnval){ [ 1047.611487] env[61806]: value = "task-1295173" [ 1047.611487] env[61806]: _type = "Task" [ 1047.611487] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1047.619176] env[61806]: DEBUG oslo_vmware.api [None req-54021779-f9e2-42db-b76c-c250eacb9121 tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Task: {'id': task-1295173, 'name': Rename_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1047.652910] env[61806]: DEBUG oslo_vmware.api [None req-c257ef4d-1b9a-4377-aada-40cc9a4843fd tempest-AttachVolumeNegativeTest-1683033704 tempest-AttachVolumeNegativeTest-1683033704-project-member] Task: {'id': session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]5200c996-b110-d840-f60b-905391cd31fb, 'name': SearchDatastore_Task, 'duration_secs': 0.036321} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1047.653284] env[61806]: DEBUG oslo_concurrency.lockutils [None req-c257ef4d-1b9a-4377-aada-40cc9a4843fd tempest-AttachVolumeNegativeTest-1683033704 tempest-AttachVolumeNegativeTest-1683033704-project-member] Releasing lock "[datastore1] devstack-image-cache_base/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1047.653531] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-c257ef4d-1b9a-4377-aada-40cc9a4843fd tempest-AttachVolumeNegativeTest-1683033704 tempest-AttachVolumeNegativeTest-1683033704-project-member] [instance: ce2ad097-4c3f-4b55-b018-ac788aa5b662] Processing image ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa {{(pid=61806) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1047.653764] env[61806]: DEBUG oslo_concurrency.lockutils [None req-c257ef4d-1b9a-4377-aada-40cc9a4843fd tempest-AttachVolumeNegativeTest-1683033704 tempest-AttachVolumeNegativeTest-1683033704-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa.vmdk" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1047.653917] env[61806]: DEBUG oslo_concurrency.lockutils [None req-c257ef4d-1b9a-4377-aada-40cc9a4843fd tempest-AttachVolumeNegativeTest-1683033704 tempest-AttachVolumeNegativeTest-1683033704-project-member] Acquired lock "[datastore1] devstack-image-cache_base/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa.vmdk" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1047.654116] env[61806]: DEBUG nova.virt.vmwareapi.ds_util [None req-c257ef4d-1b9a-4377-aada-40cc9a4843fd tempest-AttachVolumeNegativeTest-1683033704 tempest-AttachVolumeNegativeTest-1683033704-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61806) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1047.654400] env[61806]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-c70a32f8-47b8-4671-9457-ecfbe7d20b37 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1047.663608] env[61806]: DEBUG nova.virt.vmwareapi.ds_util [None req-c257ef4d-1b9a-4377-aada-40cc9a4843fd tempest-AttachVolumeNegativeTest-1683033704 tempest-AttachVolumeNegativeTest-1683033704-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61806) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1047.663840] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-c257ef4d-1b9a-4377-aada-40cc9a4843fd tempest-AttachVolumeNegativeTest-1683033704 tempest-AttachVolumeNegativeTest-1683033704-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61806) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 1047.664670] env[61806]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f4c5e2f2-976e-42f5-a1e0-0624fdeb88ee {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1047.670480] env[61806]: DEBUG oslo_vmware.api [None req-c257ef4d-1b9a-4377-aada-40cc9a4843fd tempest-AttachVolumeNegativeTest-1683033704 tempest-AttachVolumeNegativeTest-1683033704-project-member] Waiting for the task: (returnval){ [ 1047.670480] env[61806]: value = "session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]52c19143-f38c-6594-3dec-f3771e40b35d" [ 1047.670480] env[61806]: _type = "Task" [ 1047.670480] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1047.679516] env[61806]: DEBUG oslo_vmware.api [None req-c257ef4d-1b9a-4377-aada-40cc9a4843fd tempest-AttachVolumeNegativeTest-1683033704 tempest-AttachVolumeNegativeTest-1683033704-project-member] Task: {'id': session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]52c19143-f38c-6594-3dec-f3771e40b35d, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1047.944207] env[61806]: DEBUG oslo_concurrency.lockutils [None req-138f28bc-d88a-4f34-babf-797a70b177f1 tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Lock "06658134-4c6d-4911-9a55-b6805f555c78" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 20.292s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1048.000264] env[61806]: INFO nova.compute.manager [-] [instance: d1802da4-a670-427d-84d7-0fcb4717e18f] Took 1.43 seconds to deallocate network for instance. [ 1048.041276] env[61806]: DEBUG oslo_vmware.api [None req-a5c61942-d486-43f3-834d-7dfb87ef5b63 tempest-ServersNegativeTestJSON-891856253 tempest-ServersNegativeTestJSON-891856253-project-member] Task: {'id': task-1295172, 'name': Rename_Task, 'duration_secs': 0.142809} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1048.041586] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [None req-a5c61942-d486-43f3-834d-7dfb87ef5b63 tempest-ServersNegativeTestJSON-891856253 tempest-ServersNegativeTestJSON-891856253-project-member] [instance: bdbaef31-7d37-40cb-ac3e-ad152b37e2d2] Powering on the VM {{(pid=61806) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1048.041865] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-94f15eb1-4ce2-4d03-87aa-e32c6300f4e8 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1048.048763] env[61806]: DEBUG oslo_vmware.api [None req-a5c61942-d486-43f3-834d-7dfb87ef5b63 tempest-ServersNegativeTestJSON-891856253 tempest-ServersNegativeTestJSON-891856253-project-member] Waiting for the task: (returnval){ [ 1048.048763] env[61806]: value = "task-1295174" [ 1048.048763] env[61806]: _type = "Task" [ 1048.048763] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1048.056811] env[61806]: DEBUG oslo_vmware.api [None req-a5c61942-d486-43f3-834d-7dfb87ef5b63 tempest-ServersNegativeTestJSON-891856253 tempest-ServersNegativeTestJSON-891856253-project-member] Task: {'id': task-1295174, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1048.121411] env[61806]: DEBUG oslo_vmware.api [None req-54021779-f9e2-42db-b76c-c250eacb9121 tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Task: {'id': task-1295173, 'name': Rename_Task, 'duration_secs': 0.166601} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1048.121740] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [None req-54021779-f9e2-42db-b76c-c250eacb9121 tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] [instance: 7445d2c5-26bd-4f8d-8653-51c721ea801f] Powering on the VM {{(pid=61806) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1048.122043] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-6bac50b2-1897-4df2-8d2a-197eabb7b337 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1048.128616] env[61806]: DEBUG oslo_vmware.api [None req-54021779-f9e2-42db-b76c-c250eacb9121 tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Waiting for the task: (returnval){ [ 1048.128616] env[61806]: value = "task-1295175" [ 1048.128616] env[61806]: _type = "Task" [ 1048.128616] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1048.138711] env[61806]: DEBUG oslo_vmware.api [None req-54021779-f9e2-42db-b76c-c250eacb9121 tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Task: {'id': task-1295175, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1048.182759] env[61806]: DEBUG oslo_vmware.api [None req-c257ef4d-1b9a-4377-aada-40cc9a4843fd tempest-AttachVolumeNegativeTest-1683033704 tempest-AttachVolumeNegativeTest-1683033704-project-member] Task: {'id': session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]52c19143-f38c-6594-3dec-f3771e40b35d, 'name': SearchDatastore_Task, 'duration_secs': 0.010418} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1048.183777] env[61806]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-1244e79e-b2c0-410d-8f99-5c3b5e7cd216 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1048.190593] env[61806]: DEBUG oslo_vmware.api [None req-c257ef4d-1b9a-4377-aada-40cc9a4843fd tempest-AttachVolumeNegativeTest-1683033704 tempest-AttachVolumeNegativeTest-1683033704-project-member] Waiting for the task: (returnval){ [ 1048.190593] env[61806]: value = "session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]5214de15-ccdc-759f-e31d-8fe121ff1fdf" [ 1048.190593] env[61806]: _type = "Task" [ 1048.190593] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1048.198891] env[61806]: DEBUG oslo_vmware.api [None req-c257ef4d-1b9a-4377-aada-40cc9a4843fd tempest-AttachVolumeNegativeTest-1683033704 tempest-AttachVolumeNegativeTest-1683033704-project-member] Task: {'id': session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]5214de15-ccdc-759f-e31d-8fe121ff1fdf, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1048.509676] env[61806]: DEBUG oslo_concurrency.lockutils [None req-8942a881-ca25-4ecf-bd1a-43e4a7d53138 tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1048.509993] env[61806]: DEBUG oslo_concurrency.lockutils [None req-8942a881-ca25-4ecf-bd1a-43e4a7d53138 tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.000s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1048.510246] env[61806]: DEBUG nova.objects.instance [None req-8942a881-ca25-4ecf-bd1a-43e4a7d53138 tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] Lazy-loading 'resources' on Instance uuid d1802da4-a670-427d-84d7-0fcb4717e18f {{(pid=61806) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1048.559817] env[61806]: DEBUG oslo_vmware.api [None req-a5c61942-d486-43f3-834d-7dfb87ef5b63 tempest-ServersNegativeTestJSON-891856253 tempest-ServersNegativeTestJSON-891856253-project-member] Task: {'id': task-1295174, 'name': PowerOnVM_Task} progress is 99%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1048.595974] env[61806]: DEBUG nova.compute.manager [req-89e29e60-dce3-4107-96dd-1dd52ce5a1b2 req-2cd6f095-ea05-457a-9c85-8f8601cd955e service nova] [instance: 06658134-4c6d-4911-9a55-b6805f555c78] Received event network-changed-783dae1f-d1c0-4958-b039-ed57b202b64a {{(pid=61806) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1048.596206] env[61806]: DEBUG nova.compute.manager [req-89e29e60-dce3-4107-96dd-1dd52ce5a1b2 req-2cd6f095-ea05-457a-9c85-8f8601cd955e service nova] [instance: 06658134-4c6d-4911-9a55-b6805f555c78] Refreshing instance network info cache due to event network-changed-783dae1f-d1c0-4958-b039-ed57b202b64a. {{(pid=61806) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 1048.596455] env[61806]: DEBUG oslo_concurrency.lockutils [req-89e29e60-dce3-4107-96dd-1dd52ce5a1b2 req-2cd6f095-ea05-457a-9c85-8f8601cd955e service nova] Acquiring lock "refresh_cache-06658134-4c6d-4911-9a55-b6805f555c78" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1048.596647] env[61806]: DEBUG oslo_concurrency.lockutils [req-89e29e60-dce3-4107-96dd-1dd52ce5a1b2 req-2cd6f095-ea05-457a-9c85-8f8601cd955e service nova] Acquired lock "refresh_cache-06658134-4c6d-4911-9a55-b6805f555c78" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1048.597871] env[61806]: DEBUG nova.network.neutron [req-89e29e60-dce3-4107-96dd-1dd52ce5a1b2 req-2cd6f095-ea05-457a-9c85-8f8601cd955e service nova] [instance: 06658134-4c6d-4911-9a55-b6805f555c78] Refreshing network info cache for port 783dae1f-d1c0-4958-b039-ed57b202b64a {{(pid=61806) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1048.642118] env[61806]: DEBUG oslo_vmware.api [None req-54021779-f9e2-42db-b76c-c250eacb9121 tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Task: {'id': task-1295175, 'name': PowerOnVM_Task} progress is 100%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1048.701667] env[61806]: DEBUG oslo_vmware.api [None req-c257ef4d-1b9a-4377-aada-40cc9a4843fd tempest-AttachVolumeNegativeTest-1683033704 tempest-AttachVolumeNegativeTest-1683033704-project-member] Task: {'id': session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]5214de15-ccdc-759f-e31d-8fe121ff1fdf, 'name': SearchDatastore_Task, 'duration_secs': 0.011777} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1048.701970] env[61806]: DEBUG oslo_concurrency.lockutils [None req-c257ef4d-1b9a-4377-aada-40cc9a4843fd tempest-AttachVolumeNegativeTest-1683033704 tempest-AttachVolumeNegativeTest-1683033704-project-member] Releasing lock "[datastore1] devstack-image-cache_base/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa.vmdk" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1048.702831] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [None req-c257ef4d-1b9a-4377-aada-40cc9a4843fd tempest-AttachVolumeNegativeTest-1683033704 tempest-AttachVolumeNegativeTest-1683033704-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa.vmdk to [datastore1] ce2ad097-4c3f-4b55-b018-ac788aa5b662/ce2ad097-4c3f-4b55-b018-ac788aa5b662.vmdk {{(pid=61806) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 1048.702831] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-663b22af-44f1-44ca-8253-709f189768ea {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1048.709690] env[61806]: DEBUG oslo_vmware.api [None req-c257ef4d-1b9a-4377-aada-40cc9a4843fd tempest-AttachVolumeNegativeTest-1683033704 tempest-AttachVolumeNegativeTest-1683033704-project-member] Waiting for the task: (returnval){ [ 1048.709690] env[61806]: value = "task-1295176" [ 1048.709690] env[61806]: _type = "Task" [ 1048.709690] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1048.718471] env[61806]: DEBUG oslo_vmware.api [None req-c257ef4d-1b9a-4377-aada-40cc9a4843fd tempest-AttachVolumeNegativeTest-1683033704 tempest-AttachVolumeNegativeTest-1683033704-project-member] Task: {'id': task-1295176, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1049.061198] env[61806]: DEBUG oslo_vmware.api [None req-a5c61942-d486-43f3-834d-7dfb87ef5b63 tempest-ServersNegativeTestJSON-891856253 tempest-ServersNegativeTestJSON-891856253-project-member] Task: {'id': task-1295174, 'name': PowerOnVM_Task, 'duration_secs': 0.663379} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1049.061492] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [None req-a5c61942-d486-43f3-834d-7dfb87ef5b63 tempest-ServersNegativeTestJSON-891856253 tempest-ServersNegativeTestJSON-891856253-project-member] [instance: bdbaef31-7d37-40cb-ac3e-ad152b37e2d2] Powered on the VM {{(pid=61806) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1049.114840] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-98129998-18d7-4e22-9468-2d908a6fb658 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1049.123663] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cbf0904b-3e88-40cc-abde-b385c54c35c1 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1049.159069] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4d0a4da6-538d-4535-85b1-fccc8d7f24a8 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1049.164839] env[61806]: DEBUG oslo_vmware.api [None req-54021779-f9e2-42db-b76c-c250eacb9121 tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Task: {'id': task-1295175, 'name': PowerOnVM_Task, 'duration_secs': 0.54369} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1049.165407] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [None req-54021779-f9e2-42db-b76c-c250eacb9121 tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] [instance: 7445d2c5-26bd-4f8d-8653-51c721ea801f] Powered on the VM {{(pid=61806) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1049.165642] env[61806]: INFO nova.compute.manager [None req-54021779-f9e2-42db-b76c-c250eacb9121 tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] [instance: 7445d2c5-26bd-4f8d-8653-51c721ea801f] Took 8.11 seconds to spawn the instance on the hypervisor. [ 1049.165860] env[61806]: DEBUG nova.compute.manager [None req-54021779-f9e2-42db-b76c-c250eacb9121 tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] [instance: 7445d2c5-26bd-4f8d-8653-51c721ea801f] Checking state {{(pid=61806) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1049.166698] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-91b90621-a144-4106-9bd4-f41b71df8fb7 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1049.171662] env[61806]: DEBUG nova.compute.manager [None req-a5c61942-d486-43f3-834d-7dfb87ef5b63 tempest-ServersNegativeTestJSON-891856253 tempest-ServersNegativeTestJSON-891856253-project-member] [instance: bdbaef31-7d37-40cb-ac3e-ad152b37e2d2] Checking state {{(pid=61806) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1049.174205] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-92ffb3b1-8b7d-4ad3-ab7e-38410959097b {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1049.177875] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cd8801a6-d585-43a7-8073-8181bceee4c7 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1049.198537] env[61806]: DEBUG nova.compute.provider_tree [None req-8942a881-ca25-4ecf-bd1a-43e4a7d53138 tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] Inventory has not changed in ProviderTree for provider: a2858be1-fd22-4e08-979e-87ad25293407 {{(pid=61806) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1049.221793] env[61806]: DEBUG oslo_vmware.api [None req-c257ef4d-1b9a-4377-aada-40cc9a4843fd tempest-AttachVolumeNegativeTest-1683033704 tempest-AttachVolumeNegativeTest-1683033704-project-member] Task: {'id': task-1295176, 'name': CopyVirtualDisk_Task} progress is 4%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1049.407293] env[61806]: DEBUG nova.network.neutron [req-89e29e60-dce3-4107-96dd-1dd52ce5a1b2 req-2cd6f095-ea05-457a-9c85-8f8601cd955e service nova] [instance: 06658134-4c6d-4911-9a55-b6805f555c78] Updated VIF entry in instance network info cache for port 783dae1f-d1c0-4958-b039-ed57b202b64a. {{(pid=61806) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1049.407926] env[61806]: DEBUG nova.network.neutron [req-89e29e60-dce3-4107-96dd-1dd52ce5a1b2 req-2cd6f095-ea05-457a-9c85-8f8601cd955e service nova] [instance: 06658134-4c6d-4911-9a55-b6805f555c78] Updating instance_info_cache with network_info: [{"id": "783dae1f-d1c0-4958-b039-ed57b202b64a", "address": "fa:16:3e:89:24:35", "network": {"id": "746a03b6-c1a6-4d27-98b9-0bee6376634f", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-1971382587-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.245", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c1f2d161c50f4b70987b90e6364c3779", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6a31b224-77d7-48a4-af87-312758611b8e", "external-id": "nsx-vlan-transportzone-761", "segmentation_id": 761, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap783dae1f-d1", "ovs_interfaceid": "783dae1f-d1c0-4958-b039-ed57b202b64a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61806) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1049.705620] env[61806]: DEBUG nova.scheduler.client.report [None req-8942a881-ca25-4ecf-bd1a-43e4a7d53138 tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] Inventory has not changed for provider a2858be1-fd22-4e08-979e-87ad25293407 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 138, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61806) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1049.711565] env[61806]: DEBUG oslo_concurrency.lockutils [None req-a5c61942-d486-43f3-834d-7dfb87ef5b63 tempest-ServersNegativeTestJSON-891856253 tempest-ServersNegativeTestJSON-891856253-project-member] Lock "bdbaef31-7d37-40cb-ac3e-ad152b37e2d2" "released" by "nova.compute.manager.ComputeManager.unshelve_instance..do_unshelve_instance" :: held 25.558s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1049.712623] env[61806]: INFO nova.compute.manager [None req-54021779-f9e2-42db-b76c-c250eacb9121 tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] [instance: 7445d2c5-26bd-4f8d-8653-51c721ea801f] Took 19.74 seconds to build instance. [ 1049.722655] env[61806]: DEBUG oslo_vmware.api [None req-c257ef4d-1b9a-4377-aada-40cc9a4843fd tempest-AttachVolumeNegativeTest-1683033704 tempest-AttachVolumeNegativeTest-1683033704-project-member] Task: {'id': task-1295176, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.869916} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1049.722927] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [None req-c257ef4d-1b9a-4377-aada-40cc9a4843fd tempest-AttachVolumeNegativeTest-1683033704 tempest-AttachVolumeNegativeTest-1683033704-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa.vmdk to [datastore1] ce2ad097-4c3f-4b55-b018-ac788aa5b662/ce2ad097-4c3f-4b55-b018-ac788aa5b662.vmdk {{(pid=61806) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 1049.723161] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-c257ef4d-1b9a-4377-aada-40cc9a4843fd tempest-AttachVolumeNegativeTest-1683033704 tempest-AttachVolumeNegativeTest-1683033704-project-member] [instance: ce2ad097-4c3f-4b55-b018-ac788aa5b662] Extending root virtual disk to 1048576 {{(pid=61806) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1049.723417] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-cb18aef1-ea55-48ac-871f-5391095efeb7 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1049.730641] env[61806]: DEBUG oslo_vmware.api [None req-c257ef4d-1b9a-4377-aada-40cc9a4843fd tempest-AttachVolumeNegativeTest-1683033704 tempest-AttachVolumeNegativeTest-1683033704-project-member] Waiting for the task: (returnval){ [ 1049.730641] env[61806]: value = "task-1295177" [ 1049.730641] env[61806]: _type = "Task" [ 1049.730641] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1049.738860] env[61806]: DEBUG oslo_vmware.api [None req-c257ef4d-1b9a-4377-aada-40cc9a4843fd tempest-AttachVolumeNegativeTest-1683033704 tempest-AttachVolumeNegativeTest-1683033704-project-member] Task: {'id': task-1295177, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1049.911069] env[61806]: DEBUG oslo_concurrency.lockutils [req-89e29e60-dce3-4107-96dd-1dd52ce5a1b2 req-2cd6f095-ea05-457a-9c85-8f8601cd955e service nova] Releasing lock "refresh_cache-06658134-4c6d-4911-9a55-b6805f555c78" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1050.214339] env[61806]: DEBUG oslo_concurrency.lockutils [None req-8942a881-ca25-4ecf-bd1a-43e4a7d53138 tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.703s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1050.217894] env[61806]: DEBUG oslo_concurrency.lockutils [None req-54021779-f9e2-42db-b76c-c250eacb9121 tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Lock "7445d2c5-26bd-4f8d-8653-51c721ea801f" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 21.247s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1050.236532] env[61806]: INFO nova.scheduler.client.report [None req-8942a881-ca25-4ecf-bd1a-43e4a7d53138 tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] Deleted allocations for instance d1802da4-a670-427d-84d7-0fcb4717e18f [ 1050.243183] env[61806]: DEBUG oslo_vmware.api [None req-c257ef4d-1b9a-4377-aada-40cc9a4843fd tempest-AttachVolumeNegativeTest-1683033704 tempest-AttachVolumeNegativeTest-1683033704-project-member] Task: {'id': task-1295177, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.062466} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1050.243641] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-c257ef4d-1b9a-4377-aada-40cc9a4843fd tempest-AttachVolumeNegativeTest-1683033704 tempest-AttachVolumeNegativeTest-1683033704-project-member] [instance: ce2ad097-4c3f-4b55-b018-ac788aa5b662] Extended root virtual disk {{(pid=61806) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1050.244669] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4928fd56-e0b4-4b2b-8218-43ee666413bb {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1050.268955] env[61806]: DEBUG nova.virt.vmwareapi.volumeops [None req-c257ef4d-1b9a-4377-aada-40cc9a4843fd tempest-AttachVolumeNegativeTest-1683033704 tempest-AttachVolumeNegativeTest-1683033704-project-member] [instance: ce2ad097-4c3f-4b55-b018-ac788aa5b662] Reconfiguring VM instance instance-00000068 to attach disk [datastore1] ce2ad097-4c3f-4b55-b018-ac788aa5b662/ce2ad097-4c3f-4b55-b018-ac788aa5b662.vmdk or device None with type sparse {{(pid=61806) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1050.269324] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-17c21437-b390-4a70-9f6e-9a9fd048e13b {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1050.290031] env[61806]: DEBUG oslo_vmware.api [None req-c257ef4d-1b9a-4377-aada-40cc9a4843fd tempest-AttachVolumeNegativeTest-1683033704 tempest-AttachVolumeNegativeTest-1683033704-project-member] Waiting for the task: (returnval){ [ 1050.290031] env[61806]: value = "task-1295178" [ 1050.290031] env[61806]: _type = "Task" [ 1050.290031] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1050.301195] env[61806]: DEBUG oslo_vmware.api [None req-c257ef4d-1b9a-4377-aada-40cc9a4843fd tempest-AttachVolumeNegativeTest-1683033704 tempest-AttachVolumeNegativeTest-1683033704-project-member] Task: {'id': task-1295178, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1050.691518] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8e4cf817-97ab-47e5-9c72-10477f6a40e8 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1050.698236] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-9f172294-294c-44f1-a510-08d7a003e46f tempest-ServersNegativeTestJSON-891856253 tempest-ServersNegativeTestJSON-891856253-project-member] [instance: bdbaef31-7d37-40cb-ac3e-ad152b37e2d2] Suspending the VM {{(pid=61806) suspend /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1163}} [ 1050.698464] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.SuspendVM_Task with opID=oslo.vmware-3ab3bc1a-3a3f-4546-aade-ddd42db56586 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1050.705021] env[61806]: DEBUG oslo_vmware.api [None req-9f172294-294c-44f1-a510-08d7a003e46f tempest-ServersNegativeTestJSON-891856253 tempest-ServersNegativeTestJSON-891856253-project-member] Waiting for the task: (returnval){ [ 1050.705021] env[61806]: value = "task-1295179" [ 1050.705021] env[61806]: _type = "Task" [ 1050.705021] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1050.712741] env[61806]: DEBUG oslo_vmware.api [None req-9f172294-294c-44f1-a510-08d7a003e46f tempest-ServersNegativeTestJSON-891856253 tempest-ServersNegativeTestJSON-891856253-project-member] Task: {'id': task-1295179, 'name': SuspendVM_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1050.747641] env[61806]: DEBUG oslo_concurrency.lockutils [None req-8942a881-ca25-4ecf-bd1a-43e4a7d53138 tempest-ServersTestJSON-878116636 tempest-ServersTestJSON-878116636-project-member] Lock "d1802da4-a670-427d-84d7-0fcb4717e18f" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.899s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1050.800672] env[61806]: DEBUG oslo_vmware.api [None req-c257ef4d-1b9a-4377-aada-40cc9a4843fd tempest-AttachVolumeNegativeTest-1683033704 tempest-AttachVolumeNegativeTest-1683033704-project-member] Task: {'id': task-1295178, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1051.215475] env[61806]: DEBUG oslo_vmware.api [None req-9f172294-294c-44f1-a510-08d7a003e46f tempest-ServersNegativeTestJSON-891856253 tempest-ServersNegativeTestJSON-891856253-project-member] Task: {'id': task-1295179, 'name': SuspendVM_Task} progress is 66%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1051.301538] env[61806]: DEBUG oslo_vmware.api [None req-c257ef4d-1b9a-4377-aada-40cc9a4843fd tempest-AttachVolumeNegativeTest-1683033704 tempest-AttachVolumeNegativeTest-1683033704-project-member] Task: {'id': task-1295178, 'name': ReconfigVM_Task, 'duration_secs': 0.666449} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1051.301864] env[61806]: DEBUG nova.virt.vmwareapi.volumeops [None req-c257ef4d-1b9a-4377-aada-40cc9a4843fd tempest-AttachVolumeNegativeTest-1683033704 tempest-AttachVolumeNegativeTest-1683033704-project-member] [instance: ce2ad097-4c3f-4b55-b018-ac788aa5b662] Reconfigured VM instance instance-00000068 to attach disk [datastore1] ce2ad097-4c3f-4b55-b018-ac788aa5b662/ce2ad097-4c3f-4b55-b018-ac788aa5b662.vmdk or device None with type sparse {{(pid=61806) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1051.302540] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-771a830b-0e06-4773-b866-02fd7e70e035 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1051.310741] env[61806]: DEBUG oslo_vmware.api [None req-c257ef4d-1b9a-4377-aada-40cc9a4843fd tempest-AttachVolumeNegativeTest-1683033704 tempest-AttachVolumeNegativeTest-1683033704-project-member] Waiting for the task: (returnval){ [ 1051.310741] env[61806]: value = "task-1295180" [ 1051.310741] env[61806]: _type = "Task" [ 1051.310741] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1051.319409] env[61806]: DEBUG oslo_vmware.api [None req-c257ef4d-1b9a-4377-aada-40cc9a4843fd tempest-AttachVolumeNegativeTest-1683033704 tempest-AttachVolumeNegativeTest-1683033704-project-member] Task: {'id': task-1295180, 'name': Rename_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1051.716602] env[61806]: DEBUG oslo_vmware.api [None req-9f172294-294c-44f1-a510-08d7a003e46f tempest-ServersNegativeTestJSON-891856253 tempest-ServersNegativeTestJSON-891856253-project-member] Task: {'id': task-1295179, 'name': SuspendVM_Task, 'duration_secs': 0.631356} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1051.717089] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-9f172294-294c-44f1-a510-08d7a003e46f tempest-ServersNegativeTestJSON-891856253 tempest-ServersNegativeTestJSON-891856253-project-member] [instance: bdbaef31-7d37-40cb-ac3e-ad152b37e2d2] Suspended the VM {{(pid=61806) suspend /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1167}} [ 1051.717341] env[61806]: DEBUG nova.compute.manager [None req-9f172294-294c-44f1-a510-08d7a003e46f tempest-ServersNegativeTestJSON-891856253 tempest-ServersNegativeTestJSON-891856253-project-member] [instance: bdbaef31-7d37-40cb-ac3e-ad152b37e2d2] Checking state {{(pid=61806) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1051.718354] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dab0c478-a7fb-4416-83c0-bb976fa8011a {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1051.766807] env[61806]: DEBUG nova.compute.manager [None req-5fe913ca-36c2-4202-9d2d-1c8428624cce tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] [instance: 7445d2c5-26bd-4f8d-8653-51c721ea801f] Stashing vm_state: active {{(pid=61806) _prep_resize /opt/stack/nova/nova/compute/manager.py:5624}} [ 1051.821016] env[61806]: DEBUG oslo_vmware.api [None req-c257ef4d-1b9a-4377-aada-40cc9a4843fd tempest-AttachVolumeNegativeTest-1683033704 tempest-AttachVolumeNegativeTest-1683033704-project-member] Task: {'id': task-1295180, 'name': Rename_Task, 'duration_secs': 0.156873} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1051.821345] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [None req-c257ef4d-1b9a-4377-aada-40cc9a4843fd tempest-AttachVolumeNegativeTest-1683033704 tempest-AttachVolumeNegativeTest-1683033704-project-member] [instance: ce2ad097-4c3f-4b55-b018-ac788aa5b662] Powering on the VM {{(pid=61806) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1051.821605] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-efd603e4-37e7-43fb-b154-43c82e84b1d6 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1051.827650] env[61806]: DEBUG oslo_vmware.api [None req-c257ef4d-1b9a-4377-aada-40cc9a4843fd tempest-AttachVolumeNegativeTest-1683033704 tempest-AttachVolumeNegativeTest-1683033704-project-member] Waiting for the task: (returnval){ [ 1051.827650] env[61806]: value = "task-1295181" [ 1051.827650] env[61806]: _type = "Task" [ 1051.827650] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1051.835725] env[61806]: DEBUG oslo_vmware.api [None req-c257ef4d-1b9a-4377-aada-40cc9a4843fd tempest-AttachVolumeNegativeTest-1683033704 tempest-AttachVolumeNegativeTest-1683033704-project-member] Task: {'id': task-1295181, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1052.289427] env[61806]: DEBUG oslo_concurrency.lockutils [None req-5fe913ca-36c2-4202-9d2d-1c8428624cce tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1052.289697] env[61806]: DEBUG oslo_concurrency.lockutils [None req-5fe913ca-36c2-4202-9d2d-1c8428624cce tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: waited 0.000s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1052.339031] env[61806]: DEBUG oslo_vmware.api [None req-c257ef4d-1b9a-4377-aada-40cc9a4843fd tempest-AttachVolumeNegativeTest-1683033704 tempest-AttachVolumeNegativeTest-1683033704-project-member] Task: {'id': task-1295181, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1052.796245] env[61806]: INFO nova.compute.claims [None req-5fe913ca-36c2-4202-9d2d-1c8428624cce tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] [instance: 7445d2c5-26bd-4f8d-8653-51c721ea801f] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1052.843027] env[61806]: DEBUG oslo_vmware.api [None req-c257ef4d-1b9a-4377-aada-40cc9a4843fd tempest-AttachVolumeNegativeTest-1683033704 tempest-AttachVolumeNegativeTest-1683033704-project-member] Task: {'id': task-1295181, 'name': PowerOnVM_Task, 'duration_secs': 0.948627} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1052.843027] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [None req-c257ef4d-1b9a-4377-aada-40cc9a4843fd tempest-AttachVolumeNegativeTest-1683033704 tempest-AttachVolumeNegativeTest-1683033704-project-member] [instance: ce2ad097-4c3f-4b55-b018-ac788aa5b662] Powered on the VM {{(pid=61806) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1052.843027] env[61806]: INFO nova.compute.manager [None req-c257ef4d-1b9a-4377-aada-40cc9a4843fd tempest-AttachVolumeNegativeTest-1683033704 tempest-AttachVolumeNegativeTest-1683033704-project-member] [instance: ce2ad097-4c3f-4b55-b018-ac788aa5b662] Took 9.53 seconds to spawn the instance on the hypervisor. [ 1052.843027] env[61806]: DEBUG nova.compute.manager [None req-c257ef4d-1b9a-4377-aada-40cc9a4843fd tempest-AttachVolumeNegativeTest-1683033704 tempest-AttachVolumeNegativeTest-1683033704-project-member] [instance: ce2ad097-4c3f-4b55-b018-ac788aa5b662] Checking state {{(pid=61806) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1052.843027] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-736ff9e8-e9f1-4909-ae01-70f401042e1e {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1053.104561] env[61806]: INFO nova.compute.manager [None req-0535feeb-cc73-4034-b7b1-f380af26fbb0 tempest-ServersNegativeTestJSON-891856253 tempest-ServersNegativeTestJSON-891856253-project-member] [instance: bdbaef31-7d37-40cb-ac3e-ad152b37e2d2] Resuming [ 1053.105241] env[61806]: DEBUG nova.objects.instance [None req-0535feeb-cc73-4034-b7b1-f380af26fbb0 tempest-ServersNegativeTestJSON-891856253 tempest-ServersNegativeTestJSON-891856253-project-member] Lazy-loading 'flavor' on Instance uuid bdbaef31-7d37-40cb-ac3e-ad152b37e2d2 {{(pid=61806) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1053.303956] env[61806]: INFO nova.compute.resource_tracker [None req-5fe913ca-36c2-4202-9d2d-1c8428624cce tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] [instance: 7445d2c5-26bd-4f8d-8653-51c721ea801f] Updating resource usage from migration 523713f3-d834-4268-aee1-f7185bcad0bb [ 1053.364903] env[61806]: INFO nova.compute.manager [None req-c257ef4d-1b9a-4377-aada-40cc9a4843fd tempest-AttachVolumeNegativeTest-1683033704 tempest-AttachVolumeNegativeTest-1683033704-project-member] [instance: ce2ad097-4c3f-4b55-b018-ac788aa5b662] Took 14.84 seconds to build instance. [ 1053.405765] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d3e1cbde-da7f-486a-afbb-72021942e985 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1053.413016] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-297a7cf9-4856-44bb-aa3b-ce092d2b715c {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1053.443891] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0e01a4d2-d665-4877-b7f2-3ac13ed83274 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1053.451189] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ed6d4611-dbc4-442f-9b50-6970928ee2b0 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1053.463869] env[61806]: DEBUG nova.compute.provider_tree [None req-5fe913ca-36c2-4202-9d2d-1c8428624cce tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Inventory has not changed in ProviderTree for provider: a2858be1-fd22-4e08-979e-87ad25293407 {{(pid=61806) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1053.868159] env[61806]: DEBUG oslo_concurrency.lockutils [None req-c257ef4d-1b9a-4377-aada-40cc9a4843fd tempest-AttachVolumeNegativeTest-1683033704 tempest-AttachVolumeNegativeTest-1683033704-project-member] Lock "ce2ad097-4c3f-4b55-b018-ac788aa5b662" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 16.355s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1053.966967] env[61806]: DEBUG nova.scheduler.client.report [None req-5fe913ca-36c2-4202-9d2d-1c8428624cce tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Inventory has not changed for provider a2858be1-fd22-4e08-979e-87ad25293407 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 138, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61806) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1054.113564] env[61806]: DEBUG oslo_concurrency.lockutils [None req-0535feeb-cc73-4034-b7b1-f380af26fbb0 tempest-ServersNegativeTestJSON-891856253 tempest-ServersNegativeTestJSON-891856253-project-member] Acquiring lock "refresh_cache-bdbaef31-7d37-40cb-ac3e-ad152b37e2d2" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1054.113750] env[61806]: DEBUG oslo_concurrency.lockutils [None req-0535feeb-cc73-4034-b7b1-f380af26fbb0 tempest-ServersNegativeTestJSON-891856253 tempest-ServersNegativeTestJSON-891856253-project-member] Acquired lock "refresh_cache-bdbaef31-7d37-40cb-ac3e-ad152b37e2d2" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1054.113940] env[61806]: DEBUG nova.network.neutron [None req-0535feeb-cc73-4034-b7b1-f380af26fbb0 tempest-ServersNegativeTestJSON-891856253 tempest-ServersNegativeTestJSON-891856253-project-member] [instance: bdbaef31-7d37-40cb-ac3e-ad152b37e2d2] Building network info cache for instance {{(pid=61806) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1054.472121] env[61806]: DEBUG oslo_concurrency.lockutils [None req-5fe913ca-36c2-4202-9d2d-1c8428624cce tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: held 2.182s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1054.472454] env[61806]: INFO nova.compute.manager [None req-5fe913ca-36c2-4202-9d2d-1c8428624cce tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] [instance: 7445d2c5-26bd-4f8d-8653-51c721ea801f] Migrating [ 1054.488188] env[61806]: DEBUG nova.compute.manager [req-ae4c7206-ebe7-4922-809a-8d59dc6aa122 req-b34e6bb0-c724-4abe-94fe-5e24e2d8e487 service nova] [instance: ce2ad097-4c3f-4b55-b018-ac788aa5b662] Received event network-changed-6792e7fd-4056-4d2f-86a5-45ea41172f3f {{(pid=61806) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1054.488413] env[61806]: DEBUG nova.compute.manager [req-ae4c7206-ebe7-4922-809a-8d59dc6aa122 req-b34e6bb0-c724-4abe-94fe-5e24e2d8e487 service nova] [instance: ce2ad097-4c3f-4b55-b018-ac788aa5b662] Refreshing instance network info cache due to event network-changed-6792e7fd-4056-4d2f-86a5-45ea41172f3f. {{(pid=61806) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 1054.488648] env[61806]: DEBUG oslo_concurrency.lockutils [req-ae4c7206-ebe7-4922-809a-8d59dc6aa122 req-b34e6bb0-c724-4abe-94fe-5e24e2d8e487 service nova] Acquiring lock "refresh_cache-ce2ad097-4c3f-4b55-b018-ac788aa5b662" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1054.488799] env[61806]: DEBUG oslo_concurrency.lockutils [req-ae4c7206-ebe7-4922-809a-8d59dc6aa122 req-b34e6bb0-c724-4abe-94fe-5e24e2d8e487 service nova] Acquired lock "refresh_cache-ce2ad097-4c3f-4b55-b018-ac788aa5b662" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1054.488963] env[61806]: DEBUG nova.network.neutron [req-ae4c7206-ebe7-4922-809a-8d59dc6aa122 req-b34e6bb0-c724-4abe-94fe-5e24e2d8e487 service nova] [instance: ce2ad097-4c3f-4b55-b018-ac788aa5b662] Refreshing network info cache for port 6792e7fd-4056-4d2f-86a5-45ea41172f3f {{(pid=61806) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1054.803693] env[61806]: DEBUG nova.network.neutron [None req-0535feeb-cc73-4034-b7b1-f380af26fbb0 tempest-ServersNegativeTestJSON-891856253 tempest-ServersNegativeTestJSON-891856253-project-member] [instance: bdbaef31-7d37-40cb-ac3e-ad152b37e2d2] Updating instance_info_cache with network_info: [{"id": "b8b714a4-b17c-4b3d-889e-407bec6c30ae", "address": "fa:16:3e:33:74:bb", "network": {"id": "fd35ea93-8914-4b54-a4ea-f45fc750982b", "bridge": "br-int", "label": "tempest-ServersNegativeTestJSON-2117068882-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d2e2b7eec47e4b10898215123408692a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "2d94740a-bce8-4103-8ecf-230d02ec0a44", "external-id": "nsx-vlan-transportzone-149", "segmentation_id": 149, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb8b714a4-b1", "ovs_interfaceid": "b8b714a4-b17c-4b3d-889e-407bec6c30ae", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61806) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1054.990884] env[61806]: DEBUG oslo_concurrency.lockutils [None req-5fe913ca-36c2-4202-9d2d-1c8428624cce tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Acquiring lock "refresh_cache-7445d2c5-26bd-4f8d-8653-51c721ea801f" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1054.991132] env[61806]: DEBUG oslo_concurrency.lockutils [None req-5fe913ca-36c2-4202-9d2d-1c8428624cce tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Acquired lock "refresh_cache-7445d2c5-26bd-4f8d-8653-51c721ea801f" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1054.991318] env[61806]: DEBUG nova.network.neutron [None req-5fe913ca-36c2-4202-9d2d-1c8428624cce tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] [instance: 7445d2c5-26bd-4f8d-8653-51c721ea801f] Building network info cache for instance {{(pid=61806) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1055.192831] env[61806]: DEBUG nova.network.neutron [req-ae4c7206-ebe7-4922-809a-8d59dc6aa122 req-b34e6bb0-c724-4abe-94fe-5e24e2d8e487 service nova] [instance: ce2ad097-4c3f-4b55-b018-ac788aa5b662] Updated VIF entry in instance network info cache for port 6792e7fd-4056-4d2f-86a5-45ea41172f3f. {{(pid=61806) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1055.193250] env[61806]: DEBUG nova.network.neutron [req-ae4c7206-ebe7-4922-809a-8d59dc6aa122 req-b34e6bb0-c724-4abe-94fe-5e24e2d8e487 service nova] [instance: ce2ad097-4c3f-4b55-b018-ac788aa5b662] Updating instance_info_cache with network_info: [{"id": "6792e7fd-4056-4d2f-86a5-45ea41172f3f", "address": "fa:16:3e:64:1b:81", "network": {"id": "d636ca55-0737-48e4-8d9d-7fe0f797f49e", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-2090334349-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.192", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "22e47882c2c44ce492dff6b1c5d782e2", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6f493cd8-1cb4-42a1-8d56-bfa6ac7cf563", "external-id": "nsx-vlan-transportzone-931", "segmentation_id": 931, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap6792e7fd-40", "ovs_interfaceid": "6792e7fd-4056-4d2f-86a5-45ea41172f3f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61806) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1055.306864] env[61806]: DEBUG oslo_concurrency.lockutils [None req-0535feeb-cc73-4034-b7b1-f380af26fbb0 tempest-ServersNegativeTestJSON-891856253 tempest-ServersNegativeTestJSON-891856253-project-member] Releasing lock "refresh_cache-bdbaef31-7d37-40cb-ac3e-ad152b37e2d2" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1055.307948] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-62e5b121-52a2-4786-9b27-6a38716e35b3 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1055.314513] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-0535feeb-cc73-4034-b7b1-f380af26fbb0 tempest-ServersNegativeTestJSON-891856253 tempest-ServersNegativeTestJSON-891856253-project-member] [instance: bdbaef31-7d37-40cb-ac3e-ad152b37e2d2] Resuming the VM {{(pid=61806) resume /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1184}} [ 1055.314746] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-f5b18611-5245-4325-be92-b47169b36395 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1055.320660] env[61806]: DEBUG oslo_vmware.api [None req-0535feeb-cc73-4034-b7b1-f380af26fbb0 tempest-ServersNegativeTestJSON-891856253 tempest-ServersNegativeTestJSON-891856253-project-member] Waiting for the task: (returnval){ [ 1055.320660] env[61806]: value = "task-1295182" [ 1055.320660] env[61806]: _type = "Task" [ 1055.320660] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1055.329408] env[61806]: DEBUG oslo_vmware.api [None req-0535feeb-cc73-4034-b7b1-f380af26fbb0 tempest-ServersNegativeTestJSON-891856253 tempest-ServersNegativeTestJSON-891856253-project-member] Task: {'id': task-1295182, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1055.696549] env[61806]: DEBUG oslo_concurrency.lockutils [req-ae4c7206-ebe7-4922-809a-8d59dc6aa122 req-b34e6bb0-c724-4abe-94fe-5e24e2d8e487 service nova] Releasing lock "refresh_cache-ce2ad097-4c3f-4b55-b018-ac788aa5b662" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1055.759105] env[61806]: DEBUG nova.network.neutron [None req-5fe913ca-36c2-4202-9d2d-1c8428624cce tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] [instance: 7445d2c5-26bd-4f8d-8653-51c721ea801f] Updating instance_info_cache with network_info: [{"id": "773e4470-b54c-4449-b825-5bb680cf8d2f", "address": "fa:16:3e:09:73:90", "network": {"id": "b7e78559-d885-4b59-9c7f-9d966b424d04", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-1456721354-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7ab95fb72b5d46c3b6c7bebbccf897cd", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d891a662-7da0-4be5-8d0d-01b4ff70552a", "external-id": "nsx-vlan-transportzone-815", "segmentation_id": 815, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap773e4470-b5", "ovs_interfaceid": "773e4470-b54c-4449-b825-5bb680cf8d2f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61806) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1055.834577] env[61806]: DEBUG oslo_vmware.api [None req-0535feeb-cc73-4034-b7b1-f380af26fbb0 tempest-ServersNegativeTestJSON-891856253 tempest-ServersNegativeTestJSON-891856253-project-member] Task: {'id': task-1295182, 'name': PowerOnVM_Task} progress is 100%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1056.261617] env[61806]: DEBUG oslo_concurrency.lockutils [None req-5fe913ca-36c2-4202-9d2d-1c8428624cce tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Releasing lock "refresh_cache-7445d2c5-26bd-4f8d-8653-51c721ea801f" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1056.330963] env[61806]: DEBUG oslo_vmware.api [None req-0535feeb-cc73-4034-b7b1-f380af26fbb0 tempest-ServersNegativeTestJSON-891856253 tempest-ServersNegativeTestJSON-891856253-project-member] Task: {'id': task-1295182, 'name': PowerOnVM_Task, 'duration_secs': 0.539652} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1056.331335] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-0535feeb-cc73-4034-b7b1-f380af26fbb0 tempest-ServersNegativeTestJSON-891856253 tempest-ServersNegativeTestJSON-891856253-project-member] [instance: bdbaef31-7d37-40cb-ac3e-ad152b37e2d2] Resumed the VM {{(pid=61806) resume /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1189}} [ 1056.331580] env[61806]: DEBUG nova.compute.manager [None req-0535feeb-cc73-4034-b7b1-f380af26fbb0 tempest-ServersNegativeTestJSON-891856253 tempest-ServersNegativeTestJSON-891856253-project-member] [instance: bdbaef31-7d37-40cb-ac3e-ad152b37e2d2] Checking state {{(pid=61806) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1056.332484] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-18da20ea-8194-4c20-aef7-86ae50438de9 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1056.846619] env[61806]: DEBUG nova.compute.manager [None req-6e588c70-919c-4452-bb9f-4fb39e749d39 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] [instance: d0669232-6f33-4b30-97c2-2e4239af1ad0] Checking state {{(pid=61806) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1056.847666] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9d69f0a8-5887-477c-b8e7-329944d5ecca {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1057.358681] env[61806]: INFO nova.compute.manager [None req-6e588c70-919c-4452-bb9f-4fb39e749d39 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] [instance: d0669232-6f33-4b30-97c2-2e4239af1ad0] instance snapshotting [ 1057.358962] env[61806]: DEBUG nova.objects.instance [None req-6e588c70-919c-4452-bb9f-4fb39e749d39 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Lazy-loading 'flavor' on Instance uuid d0669232-6f33-4b30-97c2-2e4239af1ad0 {{(pid=61806) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1057.778625] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3a8aac7c-dd07-47bf-b0ae-8d270afe3c8e {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1057.796954] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-5fe913ca-36c2-4202-9d2d-1c8428624cce tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] [instance: 7445d2c5-26bd-4f8d-8653-51c721ea801f] Updating instance '7445d2c5-26bd-4f8d-8653-51c721ea801f' progress to 0 {{(pid=61806) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 1057.866563] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0e2810ec-c992-43b9-b058-b64d40fd8f98 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1057.886258] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-567d2390-18fe-45a8-8e0c-d8cf961695d4 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1058.302788] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [None req-5fe913ca-36c2-4202-9d2d-1c8428624cce tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] [instance: 7445d2c5-26bd-4f8d-8653-51c721ea801f] Powering off the VM {{(pid=61806) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1058.303149] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-cb476032-1c68-4a1d-9805-8c4f6687b02f {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1058.309962] env[61806]: DEBUG oslo_vmware.api [None req-5fe913ca-36c2-4202-9d2d-1c8428624cce tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Waiting for the task: (returnval){ [ 1058.309962] env[61806]: value = "task-1295183" [ 1058.309962] env[61806]: _type = "Task" [ 1058.309962] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1058.318109] env[61806]: DEBUG oslo_vmware.api [None req-5fe913ca-36c2-4202-9d2d-1c8428624cce tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Task: {'id': task-1295183, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1058.396490] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-6e588c70-919c-4452-bb9f-4fb39e749d39 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] [instance: d0669232-6f33-4b30-97c2-2e4239af1ad0] Creating Snapshot of the VM instance {{(pid=61806) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 1058.396838] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-a4404c1d-bc62-464f-b053-dd46f5c266df {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1058.404898] env[61806]: DEBUG oslo_vmware.api [None req-6e588c70-919c-4452-bb9f-4fb39e749d39 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Waiting for the task: (returnval){ [ 1058.404898] env[61806]: value = "task-1295184" [ 1058.404898] env[61806]: _type = "Task" [ 1058.404898] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1058.413875] env[61806]: DEBUG oslo_vmware.api [None req-6e588c70-919c-4452-bb9f-4fb39e749d39 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Task: {'id': task-1295184, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1058.820051] env[61806]: DEBUG oslo_vmware.api [None req-5fe913ca-36c2-4202-9d2d-1c8428624cce tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Task: {'id': task-1295183, 'name': PowerOffVM_Task, 'duration_secs': 0.185544} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1058.820051] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [None req-5fe913ca-36c2-4202-9d2d-1c8428624cce tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] [instance: 7445d2c5-26bd-4f8d-8653-51c721ea801f] Powered off the VM {{(pid=61806) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1058.820264] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-5fe913ca-36c2-4202-9d2d-1c8428624cce tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] [instance: 7445d2c5-26bd-4f8d-8653-51c721ea801f] Updating instance '7445d2c5-26bd-4f8d-8653-51c721ea801f' progress to 17 {{(pid=61806) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 1058.914576] env[61806]: DEBUG oslo_vmware.api [None req-6e588c70-919c-4452-bb9f-4fb39e749d39 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Task: {'id': task-1295184, 'name': CreateSnapshot_Task, 'duration_secs': 0.469566} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1058.914825] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-6e588c70-919c-4452-bb9f-4fb39e749d39 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] [instance: d0669232-6f33-4b30-97c2-2e4239af1ad0] Created Snapshot of the VM instance {{(pid=61806) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 1058.915615] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c5429621-bb3a-448e-9049-4b381248ed25 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1059.326888] env[61806]: DEBUG nova.virt.hardware [None req-5fe913ca-36c2-4202-9d2d-1c8428624cce tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-15T18:05:12Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='84',id=12,is_public=True,memory_mb=256,name='m1.micro',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=61806) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1059.327150] env[61806]: DEBUG nova.virt.hardware [None req-5fe913ca-36c2-4202-9d2d-1c8428624cce tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Flavor limits 0:0:0 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1059.327323] env[61806]: DEBUG nova.virt.hardware [None req-5fe913ca-36c2-4202-9d2d-1c8428624cce tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Image limits 0:0:0 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1059.327516] env[61806]: DEBUG nova.virt.hardware [None req-5fe913ca-36c2-4202-9d2d-1c8428624cce tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Flavor pref 0:0:0 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1059.327749] env[61806]: DEBUG nova.virt.hardware [None req-5fe913ca-36c2-4202-9d2d-1c8428624cce tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Image pref 0:0:0 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1059.327917] env[61806]: DEBUG nova.virt.hardware [None req-5fe913ca-36c2-4202-9d2d-1c8428624cce tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1059.328153] env[61806]: DEBUG nova.virt.hardware [None req-5fe913ca-36c2-4202-9d2d-1c8428624cce tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61806) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1059.328325] env[61806]: DEBUG nova.virt.hardware [None req-5fe913ca-36c2-4202-9d2d-1c8428624cce tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61806) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1059.328505] env[61806]: DEBUG nova.virt.hardware [None req-5fe913ca-36c2-4202-9d2d-1c8428624cce tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Got 1 possible topologies {{(pid=61806) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1059.328676] env[61806]: DEBUG nova.virt.hardware [None req-5fe913ca-36c2-4202-9d2d-1c8428624cce tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61806) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1059.328854] env[61806]: DEBUG nova.virt.hardware [None req-5fe913ca-36c2-4202-9d2d-1c8428624cce tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61806) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1059.333856] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-7ded4166-fb0a-437e-8b39-b4d484c1463e {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1059.350401] env[61806]: DEBUG oslo_vmware.api [None req-5fe913ca-36c2-4202-9d2d-1c8428624cce tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Waiting for the task: (returnval){ [ 1059.350401] env[61806]: value = "task-1295185" [ 1059.350401] env[61806]: _type = "Task" [ 1059.350401] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1059.361053] env[61806]: DEBUG oslo_vmware.api [None req-5fe913ca-36c2-4202-9d2d-1c8428624cce tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Task: {'id': task-1295185, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1059.431472] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-6e588c70-919c-4452-bb9f-4fb39e749d39 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] [instance: d0669232-6f33-4b30-97c2-2e4239af1ad0] Creating linked-clone VM from snapshot {{(pid=61806) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 1059.431784] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-354058cf-414b-4128-8ed6-6984fe1a25e2 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1059.440583] env[61806]: DEBUG oslo_vmware.api [None req-6e588c70-919c-4452-bb9f-4fb39e749d39 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Waiting for the task: (returnval){ [ 1059.440583] env[61806]: value = "task-1295186" [ 1059.440583] env[61806]: _type = "Task" [ 1059.440583] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1059.448515] env[61806]: DEBUG oslo_vmware.api [None req-6e588c70-919c-4452-bb9f-4fb39e749d39 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Task: {'id': task-1295186, 'name': CloneVM_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1059.860181] env[61806]: DEBUG oslo_vmware.api [None req-5fe913ca-36c2-4202-9d2d-1c8428624cce tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Task: {'id': task-1295185, 'name': ReconfigVM_Task, 'duration_secs': 0.401045} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1059.860525] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-5fe913ca-36c2-4202-9d2d-1c8428624cce tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] [instance: 7445d2c5-26bd-4f8d-8653-51c721ea801f] Updating instance '7445d2c5-26bd-4f8d-8653-51c721ea801f' progress to 33 {{(pid=61806) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 1059.953648] env[61806]: DEBUG oslo_vmware.api [None req-6e588c70-919c-4452-bb9f-4fb39e749d39 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Task: {'id': task-1295186, 'name': CloneVM_Task} progress is 94%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1060.367638] env[61806]: DEBUG nova.virt.hardware [None req-5fe913ca-36c2-4202-9d2d-1c8428624cce tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-15T18:05:10Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=61806) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1060.367917] env[61806]: DEBUG nova.virt.hardware [None req-5fe913ca-36c2-4202-9d2d-1c8428624cce tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Flavor limits 0:0:0 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1060.368176] env[61806]: DEBUG nova.virt.hardware [None req-5fe913ca-36c2-4202-9d2d-1c8428624cce tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Image limits 0:0:0 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1060.368500] env[61806]: DEBUG nova.virt.hardware [None req-5fe913ca-36c2-4202-9d2d-1c8428624cce tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Flavor pref 0:0:0 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1060.368776] env[61806]: DEBUG nova.virt.hardware [None req-5fe913ca-36c2-4202-9d2d-1c8428624cce tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Image pref 0:0:0 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1060.369085] env[61806]: DEBUG nova.virt.hardware [None req-5fe913ca-36c2-4202-9d2d-1c8428624cce tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1060.369443] env[61806]: DEBUG nova.virt.hardware [None req-5fe913ca-36c2-4202-9d2d-1c8428624cce tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61806) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1060.369673] env[61806]: DEBUG nova.virt.hardware [None req-5fe913ca-36c2-4202-9d2d-1c8428624cce tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61806) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1060.369908] env[61806]: DEBUG nova.virt.hardware [None req-5fe913ca-36c2-4202-9d2d-1c8428624cce tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Got 1 possible topologies {{(pid=61806) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1060.370108] env[61806]: DEBUG nova.virt.hardware [None req-5fe913ca-36c2-4202-9d2d-1c8428624cce tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61806) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1060.370319] env[61806]: DEBUG nova.virt.hardware [None req-5fe913ca-36c2-4202-9d2d-1c8428624cce tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61806) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1060.376151] env[61806]: DEBUG nova.virt.vmwareapi.volumeops [None req-5fe913ca-36c2-4202-9d2d-1c8428624cce tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] [instance: 7445d2c5-26bd-4f8d-8653-51c721ea801f] Reconfiguring VM instance instance-00000067 to detach disk 2000 {{(pid=61806) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1060.376500] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-5df5bae4-3c1c-4142-bdaf-4c81dcca695c {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1060.402859] env[61806]: DEBUG oslo_vmware.api [None req-5fe913ca-36c2-4202-9d2d-1c8428624cce tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Waiting for the task: (returnval){ [ 1060.402859] env[61806]: value = "task-1295187" [ 1060.402859] env[61806]: _type = "Task" [ 1060.402859] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1060.413044] env[61806]: DEBUG oslo_vmware.api [None req-5fe913ca-36c2-4202-9d2d-1c8428624cce tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Task: {'id': task-1295187, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1060.451274] env[61806]: DEBUG oslo_vmware.api [None req-6e588c70-919c-4452-bb9f-4fb39e749d39 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Task: {'id': task-1295186, 'name': CloneVM_Task, 'duration_secs': 0.993226} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1060.451663] env[61806]: INFO nova.virt.vmwareapi.vmops [None req-6e588c70-919c-4452-bb9f-4fb39e749d39 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] [instance: d0669232-6f33-4b30-97c2-2e4239af1ad0] Created linked-clone VM from snapshot [ 1060.452314] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-baaef98f-02e0-4fa4-9b72-aa7aee261d12 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1060.459599] env[61806]: DEBUG nova.virt.vmwareapi.images [None req-6e588c70-919c-4452-bb9f-4fb39e749d39 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] [instance: d0669232-6f33-4b30-97c2-2e4239af1ad0] Uploading image 301ac572-fe69-490a-b0d2-2012e697182e {{(pid=61806) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:442}} [ 1060.480845] env[61806]: DEBUG oslo_vmware.rw_handles [None req-6e588c70-919c-4452-bb9f-4fb39e749d39 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Creating HttpNfcLease lease for exporting VM: (result){ [ 1060.480845] env[61806]: value = "vm-277756" [ 1060.480845] env[61806]: _type = "VirtualMachine" [ 1060.480845] env[61806]: }. {{(pid=61806) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:478}} [ 1060.481182] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ExportVm with opID=oslo.vmware-8e48e21f-83e2-4962-8f45-a49de4750356 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1060.498080] env[61806]: DEBUG oslo_vmware.rw_handles [None req-6e588c70-919c-4452-bb9f-4fb39e749d39 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Lease: (returnval){ [ 1060.498080] env[61806]: value = "session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]52f07f03-0eaf-2a58-ac5d-e00ee83fbd0e" [ 1060.498080] env[61806]: _type = "HttpNfcLease" [ 1060.498080] env[61806]: } obtained for exporting VM: (result){ [ 1060.498080] env[61806]: value = "vm-277756" [ 1060.498080] env[61806]: _type = "VirtualMachine" [ 1060.498080] env[61806]: }. {{(pid=61806) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:481}} [ 1060.498416] env[61806]: DEBUG oslo_vmware.api [None req-6e588c70-919c-4452-bb9f-4fb39e749d39 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Waiting for the lease: (returnval){ [ 1060.498416] env[61806]: value = "session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]52f07f03-0eaf-2a58-ac5d-e00ee83fbd0e" [ 1060.498416] env[61806]: _type = "HttpNfcLease" [ 1060.498416] env[61806]: } to be ready. {{(pid=61806) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 1060.505537] env[61806]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1060.505537] env[61806]: value = "session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]52f07f03-0eaf-2a58-ac5d-e00ee83fbd0e" [ 1060.505537] env[61806]: _type = "HttpNfcLease" [ 1060.505537] env[61806]: } is initializing. {{(pid=61806) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 1060.912299] env[61806]: DEBUG oslo_vmware.api [None req-5fe913ca-36c2-4202-9d2d-1c8428624cce tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Task: {'id': task-1295187, 'name': ReconfigVM_Task, 'duration_secs': 0.161583} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1060.912598] env[61806]: DEBUG nova.virt.vmwareapi.volumeops [None req-5fe913ca-36c2-4202-9d2d-1c8428624cce tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] [instance: 7445d2c5-26bd-4f8d-8653-51c721ea801f] Reconfigured VM instance instance-00000067 to detach disk 2000 {{(pid=61806) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1060.913420] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b1e0b112-6eef-4ad4-b0b0-739f5d62e16f {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1060.935574] env[61806]: DEBUG nova.virt.vmwareapi.volumeops [None req-5fe913ca-36c2-4202-9d2d-1c8428624cce tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] [instance: 7445d2c5-26bd-4f8d-8653-51c721ea801f] Reconfiguring VM instance instance-00000067 to attach disk [datastore2] 7445d2c5-26bd-4f8d-8653-51c721ea801f/7445d2c5-26bd-4f8d-8653-51c721ea801f.vmdk or device None with type thin {{(pid=61806) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1060.935952] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-3b75e921-f0ae-4fe2-9891-f176389af7c4 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1060.954197] env[61806]: DEBUG oslo_vmware.api [None req-5fe913ca-36c2-4202-9d2d-1c8428624cce tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Waiting for the task: (returnval){ [ 1060.954197] env[61806]: value = "task-1295189" [ 1060.954197] env[61806]: _type = "Task" [ 1060.954197] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1060.961980] env[61806]: DEBUG oslo_vmware.api [None req-5fe913ca-36c2-4202-9d2d-1c8428624cce tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Task: {'id': task-1295189, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1061.005918] env[61806]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1061.005918] env[61806]: value = "session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]52f07f03-0eaf-2a58-ac5d-e00ee83fbd0e" [ 1061.005918] env[61806]: _type = "HttpNfcLease" [ 1061.005918] env[61806]: } is ready. {{(pid=61806) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 1061.006238] env[61806]: DEBUG oslo_vmware.rw_handles [None req-6e588c70-919c-4452-bb9f-4fb39e749d39 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 1061.006238] env[61806]: value = "session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]52f07f03-0eaf-2a58-ac5d-e00ee83fbd0e" [ 1061.006238] env[61806]: _type = "HttpNfcLease" [ 1061.006238] env[61806]: }. {{(pid=61806) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:486}} [ 1061.006964] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-369217fb-e606-4120-aa2c-4eb7bc4055fa {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1061.014311] env[61806]: DEBUG oslo_vmware.rw_handles [None req-6e588c70-919c-4452-bb9f-4fb39e749d39 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Found VMDK URL: https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/528fbe9b-12b5-468c-bc06-0a36a0b2be19/disk-0.vmdk from lease info. {{(pid=61806) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 1061.014493] env[61806]: DEBUG oslo_vmware.rw_handles [None req-6e588c70-919c-4452-bb9f-4fb39e749d39 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Opening URL: https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/528fbe9b-12b5-468c-bc06-0a36a0b2be19/disk-0.vmdk for reading. {{(pid=61806) _create_read_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:99}} [ 1061.112255] env[61806]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-a6a98f85-3701-4490-bf85-ee991d99b25d {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1061.376342] env[61806]: DEBUG oslo_concurrency.lockutils [None req-29044ecb-39c0-4f1e-a250-4d1988c06349 tempest-ServersNegativeTestJSON-891856253 tempest-ServersNegativeTestJSON-891856253-project-member] Acquiring lock "bdbaef31-7d37-40cb-ac3e-ad152b37e2d2" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1061.376666] env[61806]: DEBUG oslo_concurrency.lockutils [None req-29044ecb-39c0-4f1e-a250-4d1988c06349 tempest-ServersNegativeTestJSON-891856253 tempest-ServersNegativeTestJSON-891856253-project-member] Lock "bdbaef31-7d37-40cb-ac3e-ad152b37e2d2" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1061.376894] env[61806]: DEBUG oslo_concurrency.lockutils [None req-29044ecb-39c0-4f1e-a250-4d1988c06349 tempest-ServersNegativeTestJSON-891856253 tempest-ServersNegativeTestJSON-891856253-project-member] Acquiring lock "bdbaef31-7d37-40cb-ac3e-ad152b37e2d2-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1061.377147] env[61806]: DEBUG oslo_concurrency.lockutils [None req-29044ecb-39c0-4f1e-a250-4d1988c06349 tempest-ServersNegativeTestJSON-891856253 tempest-ServersNegativeTestJSON-891856253-project-member] Lock "bdbaef31-7d37-40cb-ac3e-ad152b37e2d2-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1061.377303] env[61806]: DEBUG oslo_concurrency.lockutils [None req-29044ecb-39c0-4f1e-a250-4d1988c06349 tempest-ServersNegativeTestJSON-891856253 tempest-ServersNegativeTestJSON-891856253-project-member] Lock "bdbaef31-7d37-40cb-ac3e-ad152b37e2d2-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1061.380205] env[61806]: INFO nova.compute.manager [None req-29044ecb-39c0-4f1e-a250-4d1988c06349 tempest-ServersNegativeTestJSON-891856253 tempest-ServersNegativeTestJSON-891856253-project-member] [instance: bdbaef31-7d37-40cb-ac3e-ad152b37e2d2] Terminating instance [ 1061.382366] env[61806]: DEBUG nova.compute.manager [None req-29044ecb-39c0-4f1e-a250-4d1988c06349 tempest-ServersNegativeTestJSON-891856253 tempest-ServersNegativeTestJSON-891856253-project-member] [instance: bdbaef31-7d37-40cb-ac3e-ad152b37e2d2] Start destroying the instance on the hypervisor. {{(pid=61806) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1061.382535] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-29044ecb-39c0-4f1e-a250-4d1988c06349 tempest-ServersNegativeTestJSON-891856253 tempest-ServersNegativeTestJSON-891856253-project-member] [instance: bdbaef31-7d37-40cb-ac3e-ad152b37e2d2] Destroying instance {{(pid=61806) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1061.383477] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b3a50e98-28b8-41af-8ef3-05f78f91bd7b {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1061.392160] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [None req-29044ecb-39c0-4f1e-a250-4d1988c06349 tempest-ServersNegativeTestJSON-891856253 tempest-ServersNegativeTestJSON-891856253-project-member] [instance: bdbaef31-7d37-40cb-ac3e-ad152b37e2d2] Powering off the VM {{(pid=61806) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1061.392479] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-28ffdea9-3fdd-42cd-8a1f-ee3cf25a3ab3 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1061.400358] env[61806]: DEBUG oslo_vmware.api [None req-29044ecb-39c0-4f1e-a250-4d1988c06349 tempest-ServersNegativeTestJSON-891856253 tempest-ServersNegativeTestJSON-891856253-project-member] Waiting for the task: (returnval){ [ 1061.400358] env[61806]: value = "task-1295190" [ 1061.400358] env[61806]: _type = "Task" [ 1061.400358] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1061.411245] env[61806]: DEBUG oslo_vmware.api [None req-29044ecb-39c0-4f1e-a250-4d1988c06349 tempest-ServersNegativeTestJSON-891856253 tempest-ServersNegativeTestJSON-891856253-project-member] Task: {'id': task-1295190, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1061.463521] env[61806]: DEBUG oslo_vmware.api [None req-5fe913ca-36c2-4202-9d2d-1c8428624cce tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Task: {'id': task-1295189, 'name': ReconfigVM_Task, 'duration_secs': 0.494904} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1061.463948] env[61806]: DEBUG nova.virt.vmwareapi.volumeops [None req-5fe913ca-36c2-4202-9d2d-1c8428624cce tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] [instance: 7445d2c5-26bd-4f8d-8653-51c721ea801f] Reconfigured VM instance instance-00000067 to attach disk [datastore2] 7445d2c5-26bd-4f8d-8653-51c721ea801f/7445d2c5-26bd-4f8d-8653-51c721ea801f.vmdk or device None with type thin {{(pid=61806) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1061.464262] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-5fe913ca-36c2-4202-9d2d-1c8428624cce tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] [instance: 7445d2c5-26bd-4f8d-8653-51c721ea801f] Updating instance '7445d2c5-26bd-4f8d-8653-51c721ea801f' progress to 50 {{(pid=61806) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 1061.910176] env[61806]: DEBUG oslo_vmware.api [None req-29044ecb-39c0-4f1e-a250-4d1988c06349 tempest-ServersNegativeTestJSON-891856253 tempest-ServersNegativeTestJSON-891856253-project-member] Task: {'id': task-1295190, 'name': PowerOffVM_Task, 'duration_secs': 0.306785} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1061.910561] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [None req-29044ecb-39c0-4f1e-a250-4d1988c06349 tempest-ServersNegativeTestJSON-891856253 tempest-ServersNegativeTestJSON-891856253-project-member] [instance: bdbaef31-7d37-40cb-ac3e-ad152b37e2d2] Powered off the VM {{(pid=61806) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1061.910807] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-29044ecb-39c0-4f1e-a250-4d1988c06349 tempest-ServersNegativeTestJSON-891856253 tempest-ServersNegativeTestJSON-891856253-project-member] [instance: bdbaef31-7d37-40cb-ac3e-ad152b37e2d2] Unregistering the VM {{(pid=61806) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1061.911156] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-bf606e27-75d1-4896-a3d0-dc62a1e77cf9 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1061.971523] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-08d0c2cc-3127-4bd4-8b8b-5c711bf93759 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1061.992686] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3f487686-2622-43d4-a139-acf1a5d12ad7 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1062.010794] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-5fe913ca-36c2-4202-9d2d-1c8428624cce tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] [instance: 7445d2c5-26bd-4f8d-8653-51c721ea801f] Updating instance '7445d2c5-26bd-4f8d-8653-51c721ea801f' progress to 67 {{(pid=61806) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 1062.263117] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-29044ecb-39c0-4f1e-a250-4d1988c06349 tempest-ServersNegativeTestJSON-891856253 tempest-ServersNegativeTestJSON-891856253-project-member] [instance: bdbaef31-7d37-40cb-ac3e-ad152b37e2d2] Unregistered the VM {{(pid=61806) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1062.263378] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-29044ecb-39c0-4f1e-a250-4d1988c06349 tempest-ServersNegativeTestJSON-891856253 tempest-ServersNegativeTestJSON-891856253-project-member] [instance: bdbaef31-7d37-40cb-ac3e-ad152b37e2d2] Deleting contents of the VM from datastore datastore1 {{(pid=61806) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1062.263653] env[61806]: DEBUG nova.virt.vmwareapi.ds_util [None req-29044ecb-39c0-4f1e-a250-4d1988c06349 tempest-ServersNegativeTestJSON-891856253 tempest-ServersNegativeTestJSON-891856253-project-member] Deleting the datastore file [datastore1] bdbaef31-7d37-40cb-ac3e-ad152b37e2d2 {{(pid=61806) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1062.263988] env[61806]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-1eeb3d1b-2779-4711-b582-c3f35c021bba {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1062.270920] env[61806]: DEBUG oslo_vmware.api [None req-29044ecb-39c0-4f1e-a250-4d1988c06349 tempest-ServersNegativeTestJSON-891856253 tempest-ServersNegativeTestJSON-891856253-project-member] Waiting for the task: (returnval){ [ 1062.270920] env[61806]: value = "task-1295192" [ 1062.270920] env[61806]: _type = "Task" [ 1062.270920] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1062.278533] env[61806]: DEBUG oslo_vmware.api [None req-29044ecb-39c0-4f1e-a250-4d1988c06349 tempest-ServersNegativeTestJSON-891856253 tempest-ServersNegativeTestJSON-891856253-project-member] Task: {'id': task-1295192, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1062.552852] env[61806]: DEBUG nova.network.neutron [None req-5fe913ca-36c2-4202-9d2d-1c8428624cce tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] [instance: 7445d2c5-26bd-4f8d-8653-51c721ea801f] Port 773e4470-b54c-4449-b825-5bb680cf8d2f binding to destination host cpu-1 is already ACTIVE {{(pid=61806) migrate_instance_start /opt/stack/nova/nova/network/neutron.py:3171}} [ 1062.780348] env[61806]: DEBUG oslo_vmware.api [None req-29044ecb-39c0-4f1e-a250-4d1988c06349 tempest-ServersNegativeTestJSON-891856253 tempest-ServersNegativeTestJSON-891856253-project-member] Task: {'id': task-1295192, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.177995} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1062.780652] env[61806]: DEBUG nova.virt.vmwareapi.ds_util [None req-29044ecb-39c0-4f1e-a250-4d1988c06349 tempest-ServersNegativeTestJSON-891856253 tempest-ServersNegativeTestJSON-891856253-project-member] Deleted the datastore file {{(pid=61806) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1062.780838] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-29044ecb-39c0-4f1e-a250-4d1988c06349 tempest-ServersNegativeTestJSON-891856253 tempest-ServersNegativeTestJSON-891856253-project-member] [instance: bdbaef31-7d37-40cb-ac3e-ad152b37e2d2] Deleted contents of the VM from datastore datastore1 {{(pid=61806) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1062.781178] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-29044ecb-39c0-4f1e-a250-4d1988c06349 tempest-ServersNegativeTestJSON-891856253 tempest-ServersNegativeTestJSON-891856253-project-member] [instance: bdbaef31-7d37-40cb-ac3e-ad152b37e2d2] Instance destroyed {{(pid=61806) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1062.781367] env[61806]: INFO nova.compute.manager [None req-29044ecb-39c0-4f1e-a250-4d1988c06349 tempest-ServersNegativeTestJSON-891856253 tempest-ServersNegativeTestJSON-891856253-project-member] [instance: bdbaef31-7d37-40cb-ac3e-ad152b37e2d2] Took 1.40 seconds to destroy the instance on the hypervisor. [ 1062.781799] env[61806]: DEBUG oslo.service.loopingcall [None req-29044ecb-39c0-4f1e-a250-4d1988c06349 tempest-ServersNegativeTestJSON-891856253 tempest-ServersNegativeTestJSON-891856253-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61806) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1062.782070] env[61806]: DEBUG nova.compute.manager [-] [instance: bdbaef31-7d37-40cb-ac3e-ad152b37e2d2] Deallocating network for instance {{(pid=61806) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1062.782224] env[61806]: DEBUG nova.network.neutron [-] [instance: bdbaef31-7d37-40cb-ac3e-ad152b37e2d2] deallocate_for_instance() {{(pid=61806) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1063.095153] env[61806]: DEBUG nova.compute.manager [req-7b13967b-199a-4dad-9482-f7122407cc68 req-be1e61e9-77e7-4ecb-a4f2-fb6bc88bae17 service nova] [instance: bdbaef31-7d37-40cb-ac3e-ad152b37e2d2] Received event network-vif-deleted-b8b714a4-b17c-4b3d-889e-407bec6c30ae {{(pid=61806) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1063.095448] env[61806]: INFO nova.compute.manager [req-7b13967b-199a-4dad-9482-f7122407cc68 req-be1e61e9-77e7-4ecb-a4f2-fb6bc88bae17 service nova] [instance: bdbaef31-7d37-40cb-ac3e-ad152b37e2d2] Neutron deleted interface b8b714a4-b17c-4b3d-889e-407bec6c30ae; detaching it from the instance and deleting it from the info cache [ 1063.095645] env[61806]: DEBUG nova.network.neutron [req-7b13967b-199a-4dad-9482-f7122407cc68 req-be1e61e9-77e7-4ecb-a4f2-fb6bc88bae17 service nova] [instance: bdbaef31-7d37-40cb-ac3e-ad152b37e2d2] Updating instance_info_cache with network_info: [] {{(pid=61806) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1063.571229] env[61806]: DEBUG nova.network.neutron [-] [instance: bdbaef31-7d37-40cb-ac3e-ad152b37e2d2] Updating instance_info_cache with network_info: [] {{(pid=61806) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1063.580104] env[61806]: DEBUG oslo_concurrency.lockutils [None req-5fe913ca-36c2-4202-9d2d-1c8428624cce tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Acquiring lock "7445d2c5-26bd-4f8d-8653-51c721ea801f-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1063.580341] env[61806]: DEBUG oslo_concurrency.lockutils [None req-5fe913ca-36c2-4202-9d2d-1c8428624cce tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Lock "7445d2c5-26bd-4f8d-8653-51c721ea801f-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1063.580506] env[61806]: DEBUG oslo_concurrency.lockutils [None req-5fe913ca-36c2-4202-9d2d-1c8428624cce tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Lock "7445d2c5-26bd-4f8d-8653-51c721ea801f-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1063.598766] env[61806]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-ac1661a8-886e-4e98-bbe6-3925326b8df2 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1063.608426] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-56feb237-ae65-41b4-9cc5-24d9e0141009 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1063.634300] env[61806]: DEBUG nova.compute.manager [req-7b13967b-199a-4dad-9482-f7122407cc68 req-be1e61e9-77e7-4ecb-a4f2-fb6bc88bae17 service nova] [instance: bdbaef31-7d37-40cb-ac3e-ad152b37e2d2] Detach interface failed, port_id=b8b714a4-b17c-4b3d-889e-407bec6c30ae, reason: Instance bdbaef31-7d37-40cb-ac3e-ad152b37e2d2 could not be found. {{(pid=61806) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10963}} [ 1064.081375] env[61806]: INFO nova.compute.manager [-] [instance: bdbaef31-7d37-40cb-ac3e-ad152b37e2d2] Took 1.30 seconds to deallocate network for instance. [ 1064.593048] env[61806]: DEBUG oslo_concurrency.lockutils [None req-29044ecb-39c0-4f1e-a250-4d1988c06349 tempest-ServersNegativeTestJSON-891856253 tempest-ServersNegativeTestJSON-891856253-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1064.593048] env[61806]: DEBUG oslo_concurrency.lockutils [None req-29044ecb-39c0-4f1e-a250-4d1988c06349 tempest-ServersNegativeTestJSON-891856253 tempest-ServersNegativeTestJSON-891856253-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.000s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1064.593048] env[61806]: DEBUG nova.objects.instance [None req-29044ecb-39c0-4f1e-a250-4d1988c06349 tempest-ServersNegativeTestJSON-891856253 tempest-ServersNegativeTestJSON-891856253-project-member] Lazy-loading 'resources' on Instance uuid bdbaef31-7d37-40cb-ac3e-ad152b37e2d2 {{(pid=61806) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1064.633847] env[61806]: DEBUG oslo_concurrency.lockutils [None req-5fe913ca-36c2-4202-9d2d-1c8428624cce tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Acquiring lock "refresh_cache-7445d2c5-26bd-4f8d-8653-51c721ea801f" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1064.634068] env[61806]: DEBUG oslo_concurrency.lockutils [None req-5fe913ca-36c2-4202-9d2d-1c8428624cce tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Acquired lock "refresh_cache-7445d2c5-26bd-4f8d-8653-51c721ea801f" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1064.634258] env[61806]: DEBUG nova.network.neutron [None req-5fe913ca-36c2-4202-9d2d-1c8428624cce tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] [instance: 7445d2c5-26bd-4f8d-8653-51c721ea801f] Building network info cache for instance {{(pid=61806) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1065.194612] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2fed0e5c-2de5-453e-8ad8-67aa270d5ab1 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1065.204284] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0e17c0e0-0c1f-45f4-9d81-9bc469124bc6 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1065.238676] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e992cd54-00c0-445f-ba5e-a5b54add8d2e {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1065.246552] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5744efab-a91f-4c7c-911d-7bc97aa3a9a4 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1065.260605] env[61806]: DEBUG nova.compute.provider_tree [None req-29044ecb-39c0-4f1e-a250-4d1988c06349 tempest-ServersNegativeTestJSON-891856253 tempest-ServersNegativeTestJSON-891856253-project-member] Inventory has not changed in ProviderTree for provider: a2858be1-fd22-4e08-979e-87ad25293407 {{(pid=61806) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1065.370498] env[61806]: DEBUG nova.network.neutron [None req-5fe913ca-36c2-4202-9d2d-1c8428624cce tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] [instance: 7445d2c5-26bd-4f8d-8653-51c721ea801f] Updating instance_info_cache with network_info: [{"id": "773e4470-b54c-4449-b825-5bb680cf8d2f", "address": "fa:16:3e:09:73:90", "network": {"id": "b7e78559-d885-4b59-9c7f-9d966b424d04", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-1456721354-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7ab95fb72b5d46c3b6c7bebbccf897cd", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d891a662-7da0-4be5-8d0d-01b4ff70552a", "external-id": "nsx-vlan-transportzone-815", "segmentation_id": 815, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap773e4470-b5", "ovs_interfaceid": "773e4470-b54c-4449-b825-5bb680cf8d2f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61806) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1065.765563] env[61806]: DEBUG nova.scheduler.client.report [None req-29044ecb-39c0-4f1e-a250-4d1988c06349 tempest-ServersNegativeTestJSON-891856253 tempest-ServersNegativeTestJSON-891856253-project-member] Inventory has not changed for provider a2858be1-fd22-4e08-979e-87ad25293407 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 138, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61806) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1065.873832] env[61806]: DEBUG oslo_concurrency.lockutils [None req-5fe913ca-36c2-4202-9d2d-1c8428624cce tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Releasing lock "refresh_cache-7445d2c5-26bd-4f8d-8653-51c721ea801f" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1066.270850] env[61806]: DEBUG oslo_concurrency.lockutils [None req-29044ecb-39c0-4f1e-a250-4d1988c06349 tempest-ServersNegativeTestJSON-891856253 tempest-ServersNegativeTestJSON-891856253-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.679s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1066.294408] env[61806]: INFO nova.scheduler.client.report [None req-29044ecb-39c0-4f1e-a250-4d1988c06349 tempest-ServersNegativeTestJSON-891856253 tempest-ServersNegativeTestJSON-891856253-project-member] Deleted allocations for instance bdbaef31-7d37-40cb-ac3e-ad152b37e2d2 [ 1066.395852] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7b7a52d5-710b-4eba-8ca2-92cdf2497ed1 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1066.417096] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-985b9b80-c089-4c3a-85c8-75736d0c5d24 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1066.424682] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-5fe913ca-36c2-4202-9d2d-1c8428624cce tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] [instance: 7445d2c5-26bd-4f8d-8653-51c721ea801f] Updating instance '7445d2c5-26bd-4f8d-8653-51c721ea801f' progress to 83 {{(pid=61806) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 1066.803053] env[61806]: DEBUG oslo_concurrency.lockutils [None req-29044ecb-39c0-4f1e-a250-4d1988c06349 tempest-ServersNegativeTestJSON-891856253 tempest-ServersNegativeTestJSON-891856253-project-member] Lock "bdbaef31-7d37-40cb-ac3e-ad152b37e2d2" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.426s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1066.932025] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [None req-5fe913ca-36c2-4202-9d2d-1c8428624cce tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] [instance: 7445d2c5-26bd-4f8d-8653-51c721ea801f] Powering on the VM {{(pid=61806) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1066.932025] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-008b26f8-a14d-4db2-a9d5-44bde412ed36 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1066.939538] env[61806]: DEBUG oslo_vmware.api [None req-5fe913ca-36c2-4202-9d2d-1c8428624cce tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Waiting for the task: (returnval){ [ 1066.939538] env[61806]: value = "task-1295193" [ 1066.939538] env[61806]: _type = "Task" [ 1066.939538] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1066.948277] env[61806]: DEBUG oslo_vmware.api [None req-5fe913ca-36c2-4202-9d2d-1c8428624cce tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Task: {'id': task-1295193, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1067.022916] env[61806]: DEBUG oslo_service.periodic_task [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Running periodic task ComputeManager._cleanup_expired_console_auth_tokens {{(pid=61806) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1067.450464] env[61806]: DEBUG oslo_vmware.api [None req-5fe913ca-36c2-4202-9d2d-1c8428624cce tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Task: {'id': task-1295193, 'name': PowerOnVM_Task, 'duration_secs': 0.466492} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1067.451993] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [None req-5fe913ca-36c2-4202-9d2d-1c8428624cce tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] [instance: 7445d2c5-26bd-4f8d-8653-51c721ea801f] Powered on the VM {{(pid=61806) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1067.451993] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-5fe913ca-36c2-4202-9d2d-1c8428624cce tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] [instance: 7445d2c5-26bd-4f8d-8653-51c721ea801f] Updating instance '7445d2c5-26bd-4f8d-8653-51c721ea801f' progress to 100 {{(pid=61806) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 1069.273246] env[61806]: DEBUG oslo_vmware.rw_handles [None req-6e588c70-919c-4452-bb9f-4fb39e749d39 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Getting lease state for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/528fbe9b-12b5-468c-bc06-0a36a0b2be19/disk-0.vmdk. {{(pid=61806) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 1069.274136] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4928c952-2913-4b56-ac2f-81d6574097a0 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1069.280447] env[61806]: DEBUG oslo_vmware.rw_handles [None req-6e588c70-919c-4452-bb9f-4fb39e749d39 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/528fbe9b-12b5-468c-bc06-0a36a0b2be19/disk-0.vmdk is in state: ready. {{(pid=61806) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 1069.280620] env[61806]: ERROR oslo_vmware.rw_handles [None req-6e588c70-919c-4452-bb9f-4fb39e749d39 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Aborting lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/528fbe9b-12b5-468c-bc06-0a36a0b2be19/disk-0.vmdk due to incomplete transfer. [ 1069.280844] env[61806]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseAbort with opID=oslo.vmware-46fb37b6-3128-43bc-acfb-b59279b201a3 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1069.287613] env[61806]: DEBUG oslo_vmware.rw_handles [None req-6e588c70-919c-4452-bb9f-4fb39e749d39 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Closed VMDK read handle for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/528fbe9b-12b5-468c-bc06-0a36a0b2be19/disk-0.vmdk. {{(pid=61806) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:735}} [ 1069.287822] env[61806]: DEBUG nova.virt.vmwareapi.images [None req-6e588c70-919c-4452-bb9f-4fb39e749d39 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] [instance: d0669232-6f33-4b30-97c2-2e4239af1ad0] Uploaded image 301ac572-fe69-490a-b0d2-2012e697182e to the Glance image server {{(pid=61806) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:473}} [ 1069.289957] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [None req-6e588c70-919c-4452-bb9f-4fb39e749d39 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] [instance: d0669232-6f33-4b30-97c2-2e4239af1ad0] Destroying the VM {{(pid=61806) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1358}} [ 1069.290210] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-a72fe92d-45d7-4473-bab7-7fb25a8e99e7 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1069.295761] env[61806]: DEBUG oslo_vmware.api [None req-6e588c70-919c-4452-bb9f-4fb39e749d39 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Waiting for the task: (returnval){ [ 1069.295761] env[61806]: value = "task-1295194" [ 1069.295761] env[61806]: _type = "Task" [ 1069.295761] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1069.303048] env[61806]: DEBUG oslo_vmware.api [None req-6e588c70-919c-4452-bb9f-4fb39e749d39 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Task: {'id': task-1295194, 'name': Destroy_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1069.808060] env[61806]: DEBUG oslo_vmware.api [None req-6e588c70-919c-4452-bb9f-4fb39e749d39 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Task: {'id': task-1295194, 'name': Destroy_Task} progress is 33%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1070.307420] env[61806]: DEBUG oslo_vmware.api [None req-6e588c70-919c-4452-bb9f-4fb39e749d39 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Task: {'id': task-1295194, 'name': Destroy_Task} progress is 33%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1070.392090] env[61806]: DEBUG oslo_concurrency.lockutils [None req-1f1225dc-2b77-47cd-9aaa-80fb8ebe5df6 tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Acquiring lock "7445d2c5-26bd-4f8d-8653-51c721ea801f" by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1070.392406] env[61806]: DEBUG oslo_concurrency.lockutils [None req-1f1225dc-2b77-47cd-9aaa-80fb8ebe5df6 tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Lock "7445d2c5-26bd-4f8d-8653-51c721ea801f" acquired by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" :: waited 0.000s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1070.392610] env[61806]: DEBUG nova.compute.manager [None req-1f1225dc-2b77-47cd-9aaa-80fb8ebe5df6 tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] [instance: 7445d2c5-26bd-4f8d-8653-51c721ea801f] Going to confirm migration 4 {{(pid=61806) do_confirm_resize /opt/stack/nova/nova/compute/manager.py:4783}} [ 1070.520847] env[61806]: DEBUG oslo_service.periodic_task [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=61806) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1070.521116] env[61806]: DEBUG oslo_service.periodic_task [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=61806) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1070.521270] env[61806]: DEBUG nova.compute.manager [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Starting heal instance info cache {{(pid=61806) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9926}} [ 1070.521392] env[61806]: DEBUG nova.compute.manager [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Rebuilding the list of instances to heal {{(pid=61806) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9930}} [ 1070.807851] env[61806]: DEBUG oslo_vmware.api [None req-6e588c70-919c-4452-bb9f-4fb39e749d39 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Task: {'id': task-1295194, 'name': Destroy_Task, 'duration_secs': 1.280517} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1070.810961] env[61806]: INFO nova.virt.vmwareapi.vm_util [None req-6e588c70-919c-4452-bb9f-4fb39e749d39 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] [instance: d0669232-6f33-4b30-97c2-2e4239af1ad0] Destroyed the VM [ 1070.810961] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-6e588c70-919c-4452-bb9f-4fb39e749d39 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] [instance: d0669232-6f33-4b30-97c2-2e4239af1ad0] Deleting Snapshot of the VM instance {{(pid=61806) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 1070.810961] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-3debe975-1e05-4d28-a3a3-e8d63e94b3b4 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1070.818021] env[61806]: DEBUG oslo_vmware.api [None req-6e588c70-919c-4452-bb9f-4fb39e749d39 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Waiting for the task: (returnval){ [ 1070.818021] env[61806]: value = "task-1295195" [ 1070.818021] env[61806]: _type = "Task" [ 1070.818021] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1070.830031] env[61806]: DEBUG oslo_vmware.api [None req-6e588c70-919c-4452-bb9f-4fb39e749d39 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Task: {'id': task-1295195, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1070.961939] env[61806]: DEBUG oslo_concurrency.lockutils [None req-1f1225dc-2b77-47cd-9aaa-80fb8ebe5df6 tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Acquiring lock "refresh_cache-7445d2c5-26bd-4f8d-8653-51c721ea801f" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1070.961939] env[61806]: DEBUG oslo_concurrency.lockutils [None req-1f1225dc-2b77-47cd-9aaa-80fb8ebe5df6 tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Acquired lock "refresh_cache-7445d2c5-26bd-4f8d-8653-51c721ea801f" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1070.961939] env[61806]: DEBUG nova.network.neutron [None req-1f1225dc-2b77-47cd-9aaa-80fb8ebe5df6 tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] [instance: 7445d2c5-26bd-4f8d-8653-51c721ea801f] Building network info cache for instance {{(pid=61806) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1070.961939] env[61806]: DEBUG nova.objects.instance [None req-1f1225dc-2b77-47cd-9aaa-80fb8ebe5df6 tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Lazy-loading 'info_cache' on Instance uuid 7445d2c5-26bd-4f8d-8653-51c721ea801f {{(pid=61806) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1071.026591] env[61806]: DEBUG nova.compute.manager [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] [instance: 7445d2c5-26bd-4f8d-8653-51c721ea801f] Skipping network cache update for instance because it is being deleted. {{(pid=61806) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9943}} [ 1071.080157] env[61806]: DEBUG oslo_concurrency.lockutils [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Acquiring lock "refresh_cache-fd9e520a-2e7b-43d4-b84d-8933c32cc6a6" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1071.080157] env[61806]: DEBUG oslo_concurrency.lockutils [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Acquired lock "refresh_cache-fd9e520a-2e7b-43d4-b84d-8933c32cc6a6" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1071.080157] env[61806]: DEBUG nova.network.neutron [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] [instance: fd9e520a-2e7b-43d4-b84d-8933c32cc6a6] Forcefully refreshing network info cache for instance {{(pid=61806) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2004}} [ 1071.080157] env[61806]: DEBUG nova.objects.instance [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Lazy-loading 'info_cache' on Instance uuid fd9e520a-2e7b-43d4-b84d-8933c32cc6a6 {{(pid=61806) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1071.326103] env[61806]: DEBUG oslo_vmware.api [None req-6e588c70-919c-4452-bb9f-4fb39e749d39 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Task: {'id': task-1295195, 'name': RemoveSnapshot_Task} progress is 100%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1071.743562] env[61806]: DEBUG oslo_concurrency.lockutils [None req-cf2c392a-60ea-4b4e-a759-3c8cd900848d tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] Acquiring lock "fd9e520a-2e7b-43d4-b84d-8933c32cc6a6" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1071.744329] env[61806]: DEBUG oslo_concurrency.lockutils [None req-cf2c392a-60ea-4b4e-a759-3c8cd900848d tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] Lock "fd9e520a-2e7b-43d4-b84d-8933c32cc6a6" acquired by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: waited 0.001s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1071.827286] env[61806]: DEBUG oslo_vmware.api [None req-6e588c70-919c-4452-bb9f-4fb39e749d39 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Task: {'id': task-1295195, 'name': RemoveSnapshot_Task, 'duration_secs': 0.512512} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1071.828477] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-6e588c70-919c-4452-bb9f-4fb39e749d39 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] [instance: d0669232-6f33-4b30-97c2-2e4239af1ad0] Deleted Snapshot of the VM instance {{(pid=61806) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 1071.828477] env[61806]: INFO nova.compute.manager [None req-6e588c70-919c-4452-bb9f-4fb39e749d39 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] [instance: d0669232-6f33-4b30-97c2-2e4239af1ad0] Took 13.96 seconds to snapshot the instance on the hypervisor. [ 1072.247810] env[61806]: INFO nova.compute.manager [None req-cf2c392a-60ea-4b4e-a759-3c8cd900848d tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] [instance: fd9e520a-2e7b-43d4-b84d-8933c32cc6a6] Detaching volume dd35bef0-37b9-4821-9a9d-7f5c7789746c [ 1072.296719] env[61806]: INFO nova.virt.block_device [None req-cf2c392a-60ea-4b4e-a759-3c8cd900848d tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] [instance: fd9e520a-2e7b-43d4-b84d-8933c32cc6a6] Attempting to driver detach volume dd35bef0-37b9-4821-9a9d-7f5c7789746c from mountpoint /dev/sdb [ 1072.296987] env[61806]: DEBUG nova.virt.vmwareapi.volumeops [None req-cf2c392a-60ea-4b4e-a759-3c8cd900848d tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] [instance: fd9e520a-2e7b-43d4-b84d-8933c32cc6a6] Volume detach. Driver type: vmdk {{(pid=61806) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 1072.297199] env[61806]: DEBUG nova.virt.vmwareapi.volumeops [None req-cf2c392a-60ea-4b4e-a759-3c8cd900848d tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] [instance: fd9e520a-2e7b-43d4-b84d-8933c32cc6a6] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-277740', 'volume_id': 'dd35bef0-37b9-4821-9a9d-7f5c7789746c', 'name': 'volume-dd35bef0-37b9-4821-9a9d-7f5c7789746c', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'attached', 'instance': 'fd9e520a-2e7b-43d4-b84d-8933c32cc6a6', 'attached_at': '', 'detached_at': '', 'volume_id': 'dd35bef0-37b9-4821-9a9d-7f5c7789746c', 'serial': 'dd35bef0-37b9-4821-9a9d-7f5c7789746c'} {{(pid=61806) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 1072.298495] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a2290d8a-b213-4973-9863-70aa16c881f4 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1072.326303] env[61806]: DEBUG nova.network.neutron [None req-1f1225dc-2b77-47cd-9aaa-80fb8ebe5df6 tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] [instance: 7445d2c5-26bd-4f8d-8653-51c721ea801f] Updating instance_info_cache with network_info: [{"id": "773e4470-b54c-4449-b825-5bb680cf8d2f", "address": "fa:16:3e:09:73:90", "network": {"id": "b7e78559-d885-4b59-9c7f-9d966b424d04", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-1456721354-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7ab95fb72b5d46c3b6c7bebbccf897cd", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d891a662-7da0-4be5-8d0d-01b4ff70552a", "external-id": "nsx-vlan-transportzone-815", "segmentation_id": 815, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap773e4470-b5", "ovs_interfaceid": "773e4470-b54c-4449-b825-5bb680cf8d2f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61806) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1072.328031] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-745dafcc-ebb4-4e02-86be-a540c82739d7 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1072.338746] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e19588a0-c621-48a5-ba9d-2b2b854d81b6 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1072.364414] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-196e6de6-8fe2-4fff-a762-11f186e803ac {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1072.383234] env[61806]: DEBUG nova.virt.vmwareapi.volumeops [None req-cf2c392a-60ea-4b4e-a759-3c8cd900848d tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] The volume has not been displaced from its original location: [datastore2] volume-dd35bef0-37b9-4821-9a9d-7f5c7789746c/volume-dd35bef0-37b9-4821-9a9d-7f5c7789746c.vmdk. No consolidation needed. {{(pid=61806) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 1072.389922] env[61806]: DEBUG nova.virt.vmwareapi.volumeops [None req-cf2c392a-60ea-4b4e-a759-3c8cd900848d tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] [instance: fd9e520a-2e7b-43d4-b84d-8933c32cc6a6] Reconfiguring VM instance instance-00000057 to detach disk 2001 {{(pid=61806) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1072.392481] env[61806]: DEBUG nova.compute.manager [None req-6e588c70-919c-4452-bb9f-4fb39e749d39 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] [instance: d0669232-6f33-4b30-97c2-2e4239af1ad0] Found 1 images (rotation: 2) {{(pid=61806) _rotate_backups /opt/stack/nova/nova/compute/manager.py:4554}} [ 1072.393498] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-6f2574c4-154d-4152-ba2d-d0f0e9e911c5 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1072.413596] env[61806]: DEBUG oslo_vmware.api [None req-cf2c392a-60ea-4b4e-a759-3c8cd900848d tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] Waiting for the task: (returnval){ [ 1072.413596] env[61806]: value = "task-1295196" [ 1072.413596] env[61806]: _type = "Task" [ 1072.413596] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1072.424151] env[61806]: DEBUG oslo_vmware.api [None req-cf2c392a-60ea-4b4e-a759-3c8cd900848d tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] Task: {'id': task-1295196, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1072.832593] env[61806]: DEBUG oslo_concurrency.lockutils [None req-1f1225dc-2b77-47cd-9aaa-80fb8ebe5df6 tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Releasing lock "refresh_cache-7445d2c5-26bd-4f8d-8653-51c721ea801f" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1072.833464] env[61806]: DEBUG nova.objects.instance [None req-1f1225dc-2b77-47cd-9aaa-80fb8ebe5df6 tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Lazy-loading 'migration_context' on Instance uuid 7445d2c5-26bd-4f8d-8653-51c721ea801f {{(pid=61806) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1072.925547] env[61806]: DEBUG oslo_vmware.api [None req-cf2c392a-60ea-4b4e-a759-3c8cd900848d tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] Task: {'id': task-1295196, 'name': ReconfigVM_Task, 'duration_secs': 0.251216} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1072.925843] env[61806]: DEBUG nova.virt.vmwareapi.volumeops [None req-cf2c392a-60ea-4b4e-a759-3c8cd900848d tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] [instance: fd9e520a-2e7b-43d4-b84d-8933c32cc6a6] Reconfigured VM instance instance-00000057 to detach disk 2001 {{(pid=61806) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1072.931232] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-2c68ee18-53da-45ce-a263-ee7a32fcb760 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1072.950236] env[61806]: DEBUG oslo_vmware.api [None req-cf2c392a-60ea-4b4e-a759-3c8cd900848d tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] Waiting for the task: (returnval){ [ 1072.950236] env[61806]: value = "task-1295197" [ 1072.950236] env[61806]: _type = "Task" [ 1072.950236] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1072.954060] env[61806]: DEBUG nova.network.neutron [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] [instance: fd9e520a-2e7b-43d4-b84d-8933c32cc6a6] Updating instance_info_cache with network_info: [{"id": "670f75e3-58f9-493b-9148-416e2e9cea5d", "address": "fa:16:3e:76:be:e1", "network": {"id": "a9d993bc-0817-4ee6-ab6e-1ef9432a110f", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-939246736-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.226", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "bbbb09a49203421d807f733616c854df", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ad4c015b-4a8b-46ca-9556-74bad8db9fb3", "external-id": "nsx-vlan-transportzone-246", "segmentation_id": 246, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap670f75e3-58", "ovs_interfaceid": "670f75e3-58f9-493b-9148-416e2e9cea5d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61806) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1072.961603] env[61806]: DEBUG oslo_vmware.api [None req-cf2c392a-60ea-4b4e-a759-3c8cd900848d tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] Task: {'id': task-1295197, 'name': ReconfigVM_Task} progress is 10%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1073.085379] env[61806]: DEBUG nova.compute.manager [None req-5fb160b1-8322-4bcb-adab-57abc5e44515 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] [instance: d0669232-6f33-4b30-97c2-2e4239af1ad0] Checking state {{(pid=61806) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1073.085981] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-00c5aef2-3426-4528-b0ed-4c4b143c7efe {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1073.336516] env[61806]: DEBUG nova.objects.base [None req-1f1225dc-2b77-47cd-9aaa-80fb8ebe5df6 tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Object Instance<7445d2c5-26bd-4f8d-8653-51c721ea801f> lazy-loaded attributes: info_cache,migration_context {{(pid=61806) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 1073.337473] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a6d7362e-6990-4a47-b415-e3fd0e1c28a9 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1073.356608] env[61806]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e67e24e3-a0b9-49c0-b662-131326f7450d {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1073.361817] env[61806]: DEBUG oslo_vmware.api [None req-1f1225dc-2b77-47cd-9aaa-80fb8ebe5df6 tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Waiting for the task: (returnval){ [ 1073.361817] env[61806]: value = "session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]52609b2b-104a-6317-beb3-635786eca8b9" [ 1073.361817] env[61806]: _type = "Task" [ 1073.361817] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1073.369286] env[61806]: DEBUG oslo_vmware.api [None req-1f1225dc-2b77-47cd-9aaa-80fb8ebe5df6 tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Task: {'id': session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]52609b2b-104a-6317-beb3-635786eca8b9, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1073.458051] env[61806]: DEBUG oslo_concurrency.lockutils [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Releasing lock "refresh_cache-fd9e520a-2e7b-43d4-b84d-8933c32cc6a6" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1073.458321] env[61806]: DEBUG nova.compute.manager [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] [instance: fd9e520a-2e7b-43d4-b84d-8933c32cc6a6] Updated the network info_cache for instance {{(pid=61806) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9997}} [ 1073.458519] env[61806]: DEBUG oslo_vmware.api [None req-cf2c392a-60ea-4b4e-a759-3c8cd900848d tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] Task: {'id': task-1295197, 'name': ReconfigVM_Task, 'duration_secs': 0.142058} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1073.458721] env[61806]: DEBUG oslo_service.periodic_task [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=61806) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1073.458976] env[61806]: DEBUG nova.virt.vmwareapi.volumeops [None req-cf2c392a-60ea-4b4e-a759-3c8cd900848d tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] [instance: fd9e520a-2e7b-43d4-b84d-8933c32cc6a6] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-277740', 'volume_id': 'dd35bef0-37b9-4821-9a9d-7f5c7789746c', 'name': 'volume-dd35bef0-37b9-4821-9a9d-7f5c7789746c', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'attached', 'instance': 'fd9e520a-2e7b-43d4-b84d-8933c32cc6a6', 'attached_at': '', 'detached_at': '', 'volume_id': 'dd35bef0-37b9-4821-9a9d-7f5c7789746c', 'serial': 'dd35bef0-37b9-4821-9a9d-7f5c7789746c'} {{(pid=61806) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 1073.461216] env[61806]: DEBUG oslo_service.periodic_task [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=61806) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1073.461391] env[61806]: DEBUG oslo_service.periodic_task [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=61806) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1073.461550] env[61806]: DEBUG oslo_service.periodic_task [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Running periodic task ComputeManager.update_available_resource {{(pid=61806) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1073.597859] env[61806]: INFO nova.compute.manager [None req-5fb160b1-8322-4bcb-adab-57abc5e44515 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] [instance: d0669232-6f33-4b30-97c2-2e4239af1ad0] instance snapshotting [ 1073.598566] env[61806]: DEBUG nova.objects.instance [None req-5fb160b1-8322-4bcb-adab-57abc5e44515 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Lazy-loading 'flavor' on Instance uuid d0669232-6f33-4b30-97c2-2e4239af1ad0 {{(pid=61806) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1073.872510] env[61806]: DEBUG oslo_vmware.api [None req-1f1225dc-2b77-47cd-9aaa-80fb8ebe5df6 tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Task: {'id': session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]52609b2b-104a-6317-beb3-635786eca8b9, 'name': SearchDatastore_Task, 'duration_secs': 0.008196} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1073.873531] env[61806]: DEBUG oslo_concurrency.lockutils [None req-1f1225dc-2b77-47cd-9aaa-80fb8ebe5df6 tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1073.873531] env[61806]: DEBUG oslo_concurrency.lockutils [None req-1f1225dc-2b77-47cd-9aaa-80fb8ebe5df6 tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" :: waited 0.000s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1073.965233] env[61806]: DEBUG oslo_concurrency.lockutils [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1074.013427] env[61806]: DEBUG nova.objects.instance [None req-cf2c392a-60ea-4b4e-a759-3c8cd900848d tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] Lazy-loading 'flavor' on Instance uuid fd9e520a-2e7b-43d4-b84d-8933c32cc6a6 {{(pid=61806) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1074.110925] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5f1d2e55-52fc-4708-8c09-9aa3a1fad2d6 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1074.131472] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5590efb1-2c06-45bb-ba05-77b66b352739 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1074.461618] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f065bdfc-4a22-4704-8bbc-9aaad72539fd {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1074.470852] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b6e4cc96-d6ff-4265-b334-70904746403d {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1074.501833] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f0801292-b19c-4499-b576-fe655f962977 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1074.509499] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5861d113-d0cc-4c34-8c43-ca33fa87086a {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1074.524716] env[61806]: DEBUG nova.compute.provider_tree [None req-1f1225dc-2b77-47cd-9aaa-80fb8ebe5df6 tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Inventory has not changed in ProviderTree for provider: a2858be1-fd22-4e08-979e-87ad25293407 {{(pid=61806) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1074.594081] env[61806]: DEBUG oslo_concurrency.lockutils [None req-f066ed90-7998-4480-842c-9e7fb5165059 tempest-ServersNegativeTestMultiTenantJSON-775824527 tempest-ServersNegativeTestMultiTenantJSON-775824527-project-member] Acquiring lock "cb9e9bac-d06c-4479-88e3-616b0f26ad7e" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1074.594381] env[61806]: DEBUG oslo_concurrency.lockutils [None req-f066ed90-7998-4480-842c-9e7fb5165059 tempest-ServersNegativeTestMultiTenantJSON-775824527 tempest-ServersNegativeTestMultiTenantJSON-775824527-project-member] Lock "cb9e9bac-d06c-4479-88e3-616b0f26ad7e" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1074.642757] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-5fb160b1-8322-4bcb-adab-57abc5e44515 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] [instance: d0669232-6f33-4b30-97c2-2e4239af1ad0] Creating Snapshot of the VM instance {{(pid=61806) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 1074.643141] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-84bbb371-2f94-4870-8d77-69c37fd380b6 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1074.652317] env[61806]: DEBUG oslo_vmware.api [None req-5fb160b1-8322-4bcb-adab-57abc5e44515 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Waiting for the task: (returnval){ [ 1074.652317] env[61806]: value = "task-1295198" [ 1074.652317] env[61806]: _type = "Task" [ 1074.652317] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1074.661063] env[61806]: DEBUG oslo_vmware.api [None req-5fb160b1-8322-4bcb-adab-57abc5e44515 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Task: {'id': task-1295198, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1075.028304] env[61806]: DEBUG nova.scheduler.client.report [None req-1f1225dc-2b77-47cd-9aaa-80fb8ebe5df6 tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Inventory has not changed for provider a2858be1-fd22-4e08-979e-87ad25293407 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 138, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61806) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1075.032598] env[61806]: DEBUG oslo_concurrency.lockutils [None req-cf2c392a-60ea-4b4e-a759-3c8cd900848d tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] Lock "fd9e520a-2e7b-43d4-b84d-8933c32cc6a6" "released" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: held 3.288s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1075.096886] env[61806]: DEBUG nova.compute.manager [None req-f066ed90-7998-4480-842c-9e7fb5165059 tempest-ServersNegativeTestMultiTenantJSON-775824527 tempest-ServersNegativeTestMultiTenantJSON-775824527-project-member] [instance: cb9e9bac-d06c-4479-88e3-616b0f26ad7e] Starting instance... {{(pid=61806) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1075.162237] env[61806]: DEBUG oslo_vmware.api [None req-5fb160b1-8322-4bcb-adab-57abc5e44515 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Task: {'id': task-1295198, 'name': CreateSnapshot_Task, 'duration_secs': 0.399144} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1075.162525] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-5fb160b1-8322-4bcb-adab-57abc5e44515 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] [instance: d0669232-6f33-4b30-97c2-2e4239af1ad0] Created Snapshot of the VM instance {{(pid=61806) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 1075.164217] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d5b1e6e8-5ae8-4e99-a2f9-050d9eccb339 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1075.620049] env[61806]: DEBUG oslo_concurrency.lockutils [None req-f066ed90-7998-4480-842c-9e7fb5165059 tempest-ServersNegativeTestMultiTenantJSON-775824527 tempest-ServersNegativeTestMultiTenantJSON-775824527-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1075.670296] env[61806]: DEBUG oslo_concurrency.lockutils [None req-cfb899a7-3b9e-475d-ac76-e219c5dfcd7e tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] Acquiring lock "fd9e520a-2e7b-43d4-b84d-8933c32cc6a6" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1075.670577] env[61806]: DEBUG oslo_concurrency.lockutils [None req-cfb899a7-3b9e-475d-ac76-e219c5dfcd7e tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] Lock "fd9e520a-2e7b-43d4-b84d-8933c32cc6a6" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1075.670825] env[61806]: DEBUG oslo_concurrency.lockutils [None req-cfb899a7-3b9e-475d-ac76-e219c5dfcd7e tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] Acquiring lock "fd9e520a-2e7b-43d4-b84d-8933c32cc6a6-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1075.671040] env[61806]: DEBUG oslo_concurrency.lockutils [None req-cfb899a7-3b9e-475d-ac76-e219c5dfcd7e tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] Lock "fd9e520a-2e7b-43d4-b84d-8933c32cc6a6-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1075.671226] env[61806]: DEBUG oslo_concurrency.lockutils [None req-cfb899a7-3b9e-475d-ac76-e219c5dfcd7e tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] Lock "fd9e520a-2e7b-43d4-b84d-8933c32cc6a6-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1075.673465] env[61806]: INFO nova.compute.manager [None req-cfb899a7-3b9e-475d-ac76-e219c5dfcd7e tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] [instance: fd9e520a-2e7b-43d4-b84d-8933c32cc6a6] Terminating instance [ 1075.675335] env[61806]: DEBUG nova.compute.manager [None req-cfb899a7-3b9e-475d-ac76-e219c5dfcd7e tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] [instance: fd9e520a-2e7b-43d4-b84d-8933c32cc6a6] Start destroying the instance on the hypervisor. {{(pid=61806) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1075.675547] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-cfb899a7-3b9e-475d-ac76-e219c5dfcd7e tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] [instance: fd9e520a-2e7b-43d4-b84d-8933c32cc6a6] Destroying instance {{(pid=61806) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1075.676398] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-afba76cc-b147-40dd-b163-cf781564740b {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1075.686217] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-5fb160b1-8322-4bcb-adab-57abc5e44515 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] [instance: d0669232-6f33-4b30-97c2-2e4239af1ad0] Creating linked-clone VM from snapshot {{(pid=61806) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 1075.686758] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-d01cf82c-6808-4ce4-93a7-34434f957b26 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1075.694977] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [None req-cfb899a7-3b9e-475d-ac76-e219c5dfcd7e tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] [instance: fd9e520a-2e7b-43d4-b84d-8933c32cc6a6] Powering off the VM {{(pid=61806) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1075.696142] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-c680c325-25c0-4b6c-890b-c81bea312cbc {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1075.697678] env[61806]: DEBUG oslo_vmware.api [None req-5fb160b1-8322-4bcb-adab-57abc5e44515 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Waiting for the task: (returnval){ [ 1075.697678] env[61806]: value = "task-1295199" [ 1075.697678] env[61806]: _type = "Task" [ 1075.697678] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1075.701751] env[61806]: DEBUG oslo_vmware.api [None req-cfb899a7-3b9e-475d-ac76-e219c5dfcd7e tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] Waiting for the task: (returnval){ [ 1075.701751] env[61806]: value = "task-1295200" [ 1075.701751] env[61806]: _type = "Task" [ 1075.701751] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1075.707927] env[61806]: DEBUG oslo_vmware.api [None req-5fb160b1-8322-4bcb-adab-57abc5e44515 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Task: {'id': task-1295199, 'name': CloneVM_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1075.712817] env[61806]: DEBUG oslo_vmware.api [None req-cfb899a7-3b9e-475d-ac76-e219c5dfcd7e tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] Task: {'id': task-1295200, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1076.040679] env[61806]: DEBUG oslo_concurrency.lockutils [None req-1f1225dc-2b77-47cd-9aaa-80fb8ebe5df6 tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" :: held 2.167s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1076.043951] env[61806]: DEBUG oslo_concurrency.lockutils [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 2.079s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1076.044159] env[61806]: DEBUG oslo_concurrency.lockutils [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1076.044329] env[61806]: DEBUG nova.compute.resource_tracker [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=61806) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 1076.044681] env[61806]: DEBUG oslo_concurrency.lockutils [None req-f066ed90-7998-4480-842c-9e7fb5165059 tempest-ServersNegativeTestMultiTenantJSON-775824527 tempest-ServersNegativeTestMultiTenantJSON-775824527-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.425s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1076.046613] env[61806]: INFO nova.compute.claims [None req-f066ed90-7998-4480-842c-9e7fb5165059 tempest-ServersNegativeTestMultiTenantJSON-775824527 tempest-ServersNegativeTestMultiTenantJSON-775824527-project-member] [instance: cb9e9bac-d06c-4479-88e3-616b0f26ad7e] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1076.050189] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e1bdf4ea-8ae2-4103-803f-31fb626ffcd5 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1076.065695] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2fe1e552-3b64-4dbc-b440-3452e09d6993 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1076.085489] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-43431c09-646a-4c9d-989d-ec0a52df94aa {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1076.093542] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d0256a09-7f7e-4d3a-99ec-16de93ec125f {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1076.131895] env[61806]: DEBUG nova.compute.resource_tracker [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=180598MB free_disk=138GB free_vcpus=48 pci_devices=None {{(pid=61806) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 1076.132165] env[61806]: DEBUG oslo_concurrency.lockutils [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1076.209102] env[61806]: DEBUG oslo_vmware.api [None req-5fb160b1-8322-4bcb-adab-57abc5e44515 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Task: {'id': task-1295199, 'name': CloneVM_Task} progress is 94%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1076.214323] env[61806]: DEBUG oslo_vmware.api [None req-cfb899a7-3b9e-475d-ac76-e219c5dfcd7e tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] Task: {'id': task-1295200, 'name': PowerOffVM_Task, 'duration_secs': 0.495295} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1076.214672] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [None req-cfb899a7-3b9e-475d-ac76-e219c5dfcd7e tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] [instance: fd9e520a-2e7b-43d4-b84d-8933c32cc6a6] Powered off the VM {{(pid=61806) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1076.214903] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-cfb899a7-3b9e-475d-ac76-e219c5dfcd7e tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] [instance: fd9e520a-2e7b-43d4-b84d-8933c32cc6a6] Unregistering the VM {{(pid=61806) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1076.215428] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-8eedc961-3338-484f-ac74-c1d65fddb4d3 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1076.370353] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-cfb899a7-3b9e-475d-ac76-e219c5dfcd7e tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] [instance: fd9e520a-2e7b-43d4-b84d-8933c32cc6a6] Unregistered the VM {{(pid=61806) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1076.370649] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-cfb899a7-3b9e-475d-ac76-e219c5dfcd7e tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] [instance: fd9e520a-2e7b-43d4-b84d-8933c32cc6a6] Deleting contents of the VM from datastore datastore2 {{(pid=61806) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1076.370862] env[61806]: DEBUG nova.virt.vmwareapi.ds_util [None req-cfb899a7-3b9e-475d-ac76-e219c5dfcd7e tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] Deleting the datastore file [datastore2] fd9e520a-2e7b-43d4-b84d-8933c32cc6a6 {{(pid=61806) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1076.371181] env[61806]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-b4f98fb3-8660-4b37-a89f-79607ba426da {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1076.378163] env[61806]: DEBUG oslo_vmware.api [None req-cfb899a7-3b9e-475d-ac76-e219c5dfcd7e tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] Waiting for the task: (returnval){ [ 1076.378163] env[61806]: value = "task-1295202" [ 1076.378163] env[61806]: _type = "Task" [ 1076.378163] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1076.387409] env[61806]: DEBUG oslo_vmware.api [None req-cfb899a7-3b9e-475d-ac76-e219c5dfcd7e tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] Task: {'id': task-1295202, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1076.607368] env[61806]: INFO nova.scheduler.client.report [None req-1f1225dc-2b77-47cd-9aaa-80fb8ebe5df6 tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Deleted allocation for migration 523713f3-d834-4268-aee1-f7185bcad0bb [ 1076.709163] env[61806]: DEBUG oslo_vmware.api [None req-5fb160b1-8322-4bcb-adab-57abc5e44515 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Task: {'id': task-1295199, 'name': CloneVM_Task} progress is 100%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1076.887962] env[61806]: DEBUG oslo_vmware.api [None req-cfb899a7-3b9e-475d-ac76-e219c5dfcd7e tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] Task: {'id': task-1295202, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.143677} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1076.888260] env[61806]: DEBUG nova.virt.vmwareapi.ds_util [None req-cfb899a7-3b9e-475d-ac76-e219c5dfcd7e tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] Deleted the datastore file {{(pid=61806) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1076.888461] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-cfb899a7-3b9e-475d-ac76-e219c5dfcd7e tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] [instance: fd9e520a-2e7b-43d4-b84d-8933c32cc6a6] Deleted contents of the VM from datastore datastore2 {{(pid=61806) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1076.888650] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-cfb899a7-3b9e-475d-ac76-e219c5dfcd7e tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] [instance: fd9e520a-2e7b-43d4-b84d-8933c32cc6a6] Instance destroyed {{(pid=61806) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1076.888821] env[61806]: INFO nova.compute.manager [None req-cfb899a7-3b9e-475d-ac76-e219c5dfcd7e tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] [instance: fd9e520a-2e7b-43d4-b84d-8933c32cc6a6] Took 1.21 seconds to destroy the instance on the hypervisor. [ 1076.889092] env[61806]: DEBUG oslo.service.loopingcall [None req-cfb899a7-3b9e-475d-ac76-e219c5dfcd7e tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61806) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1076.889301] env[61806]: DEBUG nova.compute.manager [-] [instance: fd9e520a-2e7b-43d4-b84d-8933c32cc6a6] Deallocating network for instance {{(pid=61806) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1076.889398] env[61806]: DEBUG nova.network.neutron [-] [instance: fd9e520a-2e7b-43d4-b84d-8933c32cc6a6] deallocate_for_instance() {{(pid=61806) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1077.112643] env[61806]: DEBUG oslo_concurrency.lockutils [None req-1f1225dc-2b77-47cd-9aaa-80fb8ebe5df6 tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Lock "7445d2c5-26bd-4f8d-8653-51c721ea801f" "released" by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" :: held 6.720s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1077.155305] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0f4c6a4c-8cc5-465b-b0cc-543163478d36 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1077.163407] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9a706069-3376-4b12-8a74-d94032a7aba5 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1077.195248] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3b839181-e67c-40d7-bbc8-dc38294ea81e {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1077.205974] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e986b32f-fdc3-4ee1-97e4-19c10e4b8c4b {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1077.212595] env[61806]: DEBUG oslo_vmware.api [None req-5fb160b1-8322-4bcb-adab-57abc5e44515 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Task: {'id': task-1295199, 'name': CloneVM_Task, 'duration_secs': 1.076272} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1077.213242] env[61806]: INFO nova.virt.vmwareapi.vmops [None req-5fb160b1-8322-4bcb-adab-57abc5e44515 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] [instance: d0669232-6f33-4b30-97c2-2e4239af1ad0] Created linked-clone VM from snapshot [ 1077.213967] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ecfada43-b3cf-4cd8-98ef-f923c0320236 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1077.223607] env[61806]: DEBUG nova.compute.provider_tree [None req-f066ed90-7998-4480-842c-9e7fb5165059 tempest-ServersNegativeTestMultiTenantJSON-775824527 tempest-ServersNegativeTestMultiTenantJSON-775824527-project-member] Inventory has not changed in ProviderTree for provider: a2858be1-fd22-4e08-979e-87ad25293407 {{(pid=61806) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1077.229837] env[61806]: DEBUG nova.virt.vmwareapi.images [None req-5fb160b1-8322-4bcb-adab-57abc5e44515 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] [instance: d0669232-6f33-4b30-97c2-2e4239af1ad0] Uploading image 87519589-60c5-40c3-8c84-791ed1a73a25 {{(pid=61806) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:442}} [ 1077.251657] env[61806]: DEBUG oslo_vmware.rw_handles [None req-5fb160b1-8322-4bcb-adab-57abc5e44515 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Creating HttpNfcLease lease for exporting VM: (result){ [ 1077.251657] env[61806]: value = "vm-277758" [ 1077.251657] env[61806]: _type = "VirtualMachine" [ 1077.251657] env[61806]: }. {{(pid=61806) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:478}} [ 1077.252294] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ExportVm with opID=oslo.vmware-ea578508-4f24-4311-a78a-28986d684d9e {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1077.260326] env[61806]: DEBUG oslo_vmware.rw_handles [None req-5fb160b1-8322-4bcb-adab-57abc5e44515 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Lease: (returnval){ [ 1077.260326] env[61806]: value = "session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]5208eee8-6411-43b3-20e0-cea71fef0223" [ 1077.260326] env[61806]: _type = "HttpNfcLease" [ 1077.260326] env[61806]: } obtained for exporting VM: (result){ [ 1077.260326] env[61806]: value = "vm-277758" [ 1077.260326] env[61806]: _type = "VirtualMachine" [ 1077.260326] env[61806]: }. {{(pid=61806) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:481}} [ 1077.260588] env[61806]: DEBUG oslo_vmware.api [None req-5fb160b1-8322-4bcb-adab-57abc5e44515 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Waiting for the lease: (returnval){ [ 1077.260588] env[61806]: value = "session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]5208eee8-6411-43b3-20e0-cea71fef0223" [ 1077.260588] env[61806]: _type = "HttpNfcLease" [ 1077.260588] env[61806]: } to be ready. {{(pid=61806) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 1077.266963] env[61806]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1077.266963] env[61806]: value = "session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]5208eee8-6411-43b3-20e0-cea71fef0223" [ 1077.266963] env[61806]: _type = "HttpNfcLease" [ 1077.266963] env[61806]: } is initializing. {{(pid=61806) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 1077.333809] env[61806]: DEBUG nova.compute.manager [req-7ae6d753-0893-47f2-9f7e-9e86c6e77d9a req-2f942325-e731-4161-8e67-8a6bccb63f6d service nova] [instance: fd9e520a-2e7b-43d4-b84d-8933c32cc6a6] Received event network-vif-deleted-670f75e3-58f9-493b-9148-416e2e9cea5d {{(pid=61806) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1077.334045] env[61806]: INFO nova.compute.manager [req-7ae6d753-0893-47f2-9f7e-9e86c6e77d9a req-2f942325-e731-4161-8e67-8a6bccb63f6d service nova] [instance: fd9e520a-2e7b-43d4-b84d-8933c32cc6a6] Neutron deleted interface 670f75e3-58f9-493b-9148-416e2e9cea5d; detaching it from the instance and deleting it from the info cache [ 1077.334231] env[61806]: DEBUG nova.network.neutron [req-7ae6d753-0893-47f2-9f7e-9e86c6e77d9a req-2f942325-e731-4161-8e67-8a6bccb63f6d service nova] [instance: fd9e520a-2e7b-43d4-b84d-8933c32cc6a6] Updating instance_info_cache with network_info: [] {{(pid=61806) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1077.726469] env[61806]: DEBUG nova.scheduler.client.report [None req-f066ed90-7998-4480-842c-9e7fb5165059 tempest-ServersNegativeTestMultiTenantJSON-775824527 tempest-ServersNegativeTestMultiTenantJSON-775824527-project-member] Inventory has not changed for provider a2858be1-fd22-4e08-979e-87ad25293407 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 138, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61806) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1077.768431] env[61806]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1077.768431] env[61806]: value = "session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]5208eee8-6411-43b3-20e0-cea71fef0223" [ 1077.768431] env[61806]: _type = "HttpNfcLease" [ 1077.768431] env[61806]: } is ready. {{(pid=61806) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 1077.768739] env[61806]: DEBUG oslo_vmware.rw_handles [None req-5fb160b1-8322-4bcb-adab-57abc5e44515 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 1077.768739] env[61806]: value = "session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]5208eee8-6411-43b3-20e0-cea71fef0223" [ 1077.768739] env[61806]: _type = "HttpNfcLease" [ 1077.768739] env[61806]: }. {{(pid=61806) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:486}} [ 1077.769474] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-336dc734-45fc-48e6-ae0f-3897890463b6 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1077.776745] env[61806]: DEBUG oslo_vmware.rw_handles [None req-5fb160b1-8322-4bcb-adab-57abc5e44515 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Found VMDK URL: https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52ca7cdb-8ede-5cf9-2c14-5ecf127f4778/disk-0.vmdk from lease info. {{(pid=61806) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 1077.776968] env[61806]: DEBUG oslo_vmware.rw_handles [None req-5fb160b1-8322-4bcb-adab-57abc5e44515 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Opening URL: https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52ca7cdb-8ede-5cf9-2c14-5ecf127f4778/disk-0.vmdk for reading. {{(pid=61806) _create_read_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:99}} [ 1077.833503] env[61806]: DEBUG nova.network.neutron [-] [instance: fd9e520a-2e7b-43d4-b84d-8933c32cc6a6] Updating instance_info_cache with network_info: [] {{(pid=61806) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1077.837752] env[61806]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-9f990687-8f6e-46c6-a926-b54267ea56d2 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1077.847566] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c55c70d5-40ca-4657-a8bc-88644ed57979 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1077.867031] env[61806]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-ed480e72-8d54-4a8e-bf1c-b3b47aef7c96 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1077.877704] env[61806]: DEBUG nova.compute.manager [req-7ae6d753-0893-47f2-9f7e-9e86c6e77d9a req-2f942325-e731-4161-8e67-8a6bccb63f6d service nova] [instance: fd9e520a-2e7b-43d4-b84d-8933c32cc6a6] Detach interface failed, port_id=670f75e3-58f9-493b-9148-416e2e9cea5d, reason: Instance fd9e520a-2e7b-43d4-b84d-8933c32cc6a6 could not be found. {{(pid=61806) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10963}} [ 1078.231456] env[61806]: DEBUG oslo_concurrency.lockutils [None req-f066ed90-7998-4480-842c-9e7fb5165059 tempest-ServersNegativeTestMultiTenantJSON-775824527 tempest-ServersNegativeTestMultiTenantJSON-775824527-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.187s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1078.232084] env[61806]: DEBUG nova.compute.manager [None req-f066ed90-7998-4480-842c-9e7fb5165059 tempest-ServersNegativeTestMultiTenantJSON-775824527 tempest-ServersNegativeTestMultiTenantJSON-775824527-project-member] [instance: cb9e9bac-d06c-4479-88e3-616b0f26ad7e] Start building networks asynchronously for instance. {{(pid=61806) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1078.235091] env[61806]: DEBUG oslo_concurrency.lockutils [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 2.103s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1078.337021] env[61806]: INFO nova.compute.manager [-] [instance: fd9e520a-2e7b-43d4-b84d-8933c32cc6a6] Took 1.45 seconds to deallocate network for instance. [ 1078.421363] env[61806]: DEBUG oslo_concurrency.lockutils [None req-1f1225dc-2b77-47cd-9aaa-80fb8ebe5df6 tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Acquiring lock "7445d2c5-26bd-4f8d-8653-51c721ea801f" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1078.421495] env[61806]: DEBUG oslo_concurrency.lockutils [None req-1f1225dc-2b77-47cd-9aaa-80fb8ebe5df6 tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Lock "7445d2c5-26bd-4f8d-8653-51c721ea801f" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1078.422307] env[61806]: DEBUG oslo_concurrency.lockutils [None req-1f1225dc-2b77-47cd-9aaa-80fb8ebe5df6 tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Acquiring lock "7445d2c5-26bd-4f8d-8653-51c721ea801f-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1078.422307] env[61806]: DEBUG oslo_concurrency.lockutils [None req-1f1225dc-2b77-47cd-9aaa-80fb8ebe5df6 tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Lock "7445d2c5-26bd-4f8d-8653-51c721ea801f-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1078.422474] env[61806]: DEBUG oslo_concurrency.lockutils [None req-1f1225dc-2b77-47cd-9aaa-80fb8ebe5df6 tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Lock "7445d2c5-26bd-4f8d-8653-51c721ea801f-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1078.425143] env[61806]: INFO nova.compute.manager [None req-1f1225dc-2b77-47cd-9aaa-80fb8ebe5df6 tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] [instance: 7445d2c5-26bd-4f8d-8653-51c721ea801f] Terminating instance [ 1078.427525] env[61806]: DEBUG nova.compute.manager [None req-1f1225dc-2b77-47cd-9aaa-80fb8ebe5df6 tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] [instance: 7445d2c5-26bd-4f8d-8653-51c721ea801f] Start destroying the instance on the hypervisor. {{(pid=61806) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1078.427828] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-1f1225dc-2b77-47cd-9aaa-80fb8ebe5df6 tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] [instance: 7445d2c5-26bd-4f8d-8653-51c721ea801f] Destroying instance {{(pid=61806) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1078.428997] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9d44a84e-a4b3-4a38-bf5c-c8266e106405 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1078.437925] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [None req-1f1225dc-2b77-47cd-9aaa-80fb8ebe5df6 tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] [instance: 7445d2c5-26bd-4f8d-8653-51c721ea801f] Powering off the VM {{(pid=61806) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1078.438586] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-896a0a96-3958-4f59-a718-d95524720da0 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1078.444778] env[61806]: DEBUG oslo_vmware.api [None req-1f1225dc-2b77-47cd-9aaa-80fb8ebe5df6 tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Waiting for the task: (returnval){ [ 1078.444778] env[61806]: value = "task-1295204" [ 1078.444778] env[61806]: _type = "Task" [ 1078.444778] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1078.455137] env[61806]: DEBUG oslo_vmware.api [None req-1f1225dc-2b77-47cd-9aaa-80fb8ebe5df6 tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Task: {'id': task-1295204, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1078.740555] env[61806]: DEBUG nova.compute.utils [None req-f066ed90-7998-4480-842c-9e7fb5165059 tempest-ServersNegativeTestMultiTenantJSON-775824527 tempest-ServersNegativeTestMultiTenantJSON-775824527-project-member] Using /dev/sd instead of None {{(pid=61806) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1078.745204] env[61806]: DEBUG nova.compute.manager [None req-f066ed90-7998-4480-842c-9e7fb5165059 tempest-ServersNegativeTestMultiTenantJSON-775824527 tempest-ServersNegativeTestMultiTenantJSON-775824527-project-member] [instance: cb9e9bac-d06c-4479-88e3-616b0f26ad7e] Allocating IP information in the background. {{(pid=61806) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1078.745458] env[61806]: DEBUG nova.network.neutron [None req-f066ed90-7998-4480-842c-9e7fb5165059 tempest-ServersNegativeTestMultiTenantJSON-775824527 tempest-ServersNegativeTestMultiTenantJSON-775824527-project-member] [instance: cb9e9bac-d06c-4479-88e3-616b0f26ad7e] allocate_for_instance() {{(pid=61806) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 1078.796505] env[61806]: DEBUG nova.policy [None req-f066ed90-7998-4480-842c-9e7fb5165059 tempest-ServersNegativeTestMultiTenantJSON-775824527 tempest-ServersNegativeTestMultiTenantJSON-775824527-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '06e1c864350e4ddca05e7754532d5c96', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '008da83c5c894402a1158ead2cc19896', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61806) authorize /opt/stack/nova/nova/policy.py:201}} [ 1078.845989] env[61806]: DEBUG oslo_concurrency.lockutils [None req-cfb899a7-3b9e-475d-ac76-e219c5dfcd7e tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1078.954540] env[61806]: DEBUG oslo_vmware.api [None req-1f1225dc-2b77-47cd-9aaa-80fb8ebe5df6 tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Task: {'id': task-1295204, 'name': PowerOffVM_Task, 'duration_secs': 0.217301} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1078.954882] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [None req-1f1225dc-2b77-47cd-9aaa-80fb8ebe5df6 tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] [instance: 7445d2c5-26bd-4f8d-8653-51c721ea801f] Powered off the VM {{(pid=61806) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1078.955177] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-1f1225dc-2b77-47cd-9aaa-80fb8ebe5df6 tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] [instance: 7445d2c5-26bd-4f8d-8653-51c721ea801f] Unregistering the VM {{(pid=61806) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1078.955587] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-1c688872-c178-4428-8758-7570d898c3fc {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1079.020698] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-1f1225dc-2b77-47cd-9aaa-80fb8ebe5df6 tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] [instance: 7445d2c5-26bd-4f8d-8653-51c721ea801f] Unregistered the VM {{(pid=61806) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1079.020942] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-1f1225dc-2b77-47cd-9aaa-80fb8ebe5df6 tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] [instance: 7445d2c5-26bd-4f8d-8653-51c721ea801f] Deleting contents of the VM from datastore datastore2 {{(pid=61806) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1079.021145] env[61806]: DEBUG nova.virt.vmwareapi.ds_util [None req-1f1225dc-2b77-47cd-9aaa-80fb8ebe5df6 tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Deleting the datastore file [datastore2] 7445d2c5-26bd-4f8d-8653-51c721ea801f {{(pid=61806) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1079.021433] env[61806]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-ba34a8dc-4dfe-4863-be01-e3a84b6c6e5c {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1079.030602] env[61806]: DEBUG oslo_vmware.api [None req-1f1225dc-2b77-47cd-9aaa-80fb8ebe5df6 tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Waiting for the task: (returnval){ [ 1079.030602] env[61806]: value = "task-1295206" [ 1079.030602] env[61806]: _type = "Task" [ 1079.030602] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1079.040226] env[61806]: DEBUG oslo_vmware.api [None req-1f1225dc-2b77-47cd-9aaa-80fb8ebe5df6 tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Task: {'id': task-1295206, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1079.052881] env[61806]: DEBUG nova.network.neutron [None req-f066ed90-7998-4480-842c-9e7fb5165059 tempest-ServersNegativeTestMultiTenantJSON-775824527 tempest-ServersNegativeTestMultiTenantJSON-775824527-project-member] [instance: cb9e9bac-d06c-4479-88e3-616b0f26ad7e] Successfully created port: ed5dd9bf-96f6-4cad-8f5d-0ea62887f3b9 {{(pid=61806) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1079.249023] env[61806]: DEBUG nova.compute.manager [None req-f066ed90-7998-4480-842c-9e7fb5165059 tempest-ServersNegativeTestMultiTenantJSON-775824527 tempest-ServersNegativeTestMultiTenantJSON-775824527-project-member] [instance: cb9e9bac-d06c-4479-88e3-616b0f26ad7e] Start building block device mappings for instance. {{(pid=61806) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1079.359176] env[61806]: DEBUG nova.compute.resource_tracker [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Instance fd9e520a-2e7b-43d4-b84d-8933c32cc6a6 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61806) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1079.359347] env[61806]: DEBUG nova.compute.resource_tracker [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Instance d0669232-6f33-4b30-97c2-2e4239af1ad0 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61806) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1079.360333] env[61806]: DEBUG nova.compute.resource_tracker [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Instance 06658134-4c6d-4911-9a55-b6805f555c78 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61806) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1079.360509] env[61806]: DEBUG nova.compute.resource_tracker [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Instance ce2ad097-4c3f-4b55-b018-ac788aa5b662 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61806) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1079.360641] env[61806]: DEBUG nova.compute.resource_tracker [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Instance 7445d2c5-26bd-4f8d-8653-51c721ea801f actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 256, 'VCPU': 1}}. {{(pid=61806) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1079.360811] env[61806]: DEBUG nova.compute.resource_tracker [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Instance cb9e9bac-d06c-4479-88e3-616b0f26ad7e actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61806) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1079.360983] env[61806]: DEBUG nova.compute.resource_tracker [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Total usable vcpus: 48, total allocated vcpus: 6 {{(pid=61806) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 1079.361157] env[61806]: DEBUG nova.compute.resource_tracker [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=1728MB phys_disk=200GB used_disk=6GB total_vcpus=48 used_vcpus=6 pci_stats=[] {{(pid=61806) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 1079.460176] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b5d8b7fb-9a89-463a-ba19-71f1e0f98c73 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1079.468639] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fbb33656-e1e4-4047-8fb1-546ac3c4afa4 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1079.499665] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-16efc85a-d8aa-464b-8b8d-74049280265a {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1079.507751] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6cc29cc5-6fd8-4e48-b022-d5233fa80633 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1079.524767] env[61806]: DEBUG nova.compute.provider_tree [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Inventory has not changed in ProviderTree for provider: a2858be1-fd22-4e08-979e-87ad25293407 {{(pid=61806) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1079.541587] env[61806]: DEBUG oslo_vmware.api [None req-1f1225dc-2b77-47cd-9aaa-80fb8ebe5df6 tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Task: {'id': task-1295206, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.177417} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1079.542204] env[61806]: DEBUG nova.virt.vmwareapi.ds_util [None req-1f1225dc-2b77-47cd-9aaa-80fb8ebe5df6 tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Deleted the datastore file {{(pid=61806) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1079.542558] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-1f1225dc-2b77-47cd-9aaa-80fb8ebe5df6 tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] [instance: 7445d2c5-26bd-4f8d-8653-51c721ea801f] Deleted contents of the VM from datastore datastore2 {{(pid=61806) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1079.542883] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-1f1225dc-2b77-47cd-9aaa-80fb8ebe5df6 tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] [instance: 7445d2c5-26bd-4f8d-8653-51c721ea801f] Instance destroyed {{(pid=61806) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1079.543703] env[61806]: INFO nova.compute.manager [None req-1f1225dc-2b77-47cd-9aaa-80fb8ebe5df6 tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] [instance: 7445d2c5-26bd-4f8d-8653-51c721ea801f] Took 1.12 seconds to destroy the instance on the hypervisor. [ 1079.543703] env[61806]: DEBUG oslo.service.loopingcall [None req-1f1225dc-2b77-47cd-9aaa-80fb8ebe5df6 tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61806) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1079.543703] env[61806]: DEBUG nova.compute.manager [-] [instance: 7445d2c5-26bd-4f8d-8653-51c721ea801f] Deallocating network for instance {{(pid=61806) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1079.543888] env[61806]: DEBUG nova.network.neutron [-] [instance: 7445d2c5-26bd-4f8d-8653-51c721ea801f] deallocate_for_instance() {{(pid=61806) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1079.815943] env[61806]: DEBUG nova.compute.manager [req-99120017-7368-4164-bce3-d86595ef0290 req-6afa9218-9c54-49f1-b216-8e7f8f698bae service nova] [instance: 7445d2c5-26bd-4f8d-8653-51c721ea801f] Received event network-vif-deleted-773e4470-b54c-4449-b825-5bb680cf8d2f {{(pid=61806) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1079.816207] env[61806]: INFO nova.compute.manager [req-99120017-7368-4164-bce3-d86595ef0290 req-6afa9218-9c54-49f1-b216-8e7f8f698bae service nova] [instance: 7445d2c5-26bd-4f8d-8653-51c721ea801f] Neutron deleted interface 773e4470-b54c-4449-b825-5bb680cf8d2f; detaching it from the instance and deleting it from the info cache [ 1079.816530] env[61806]: DEBUG nova.network.neutron [req-99120017-7368-4164-bce3-d86595ef0290 req-6afa9218-9c54-49f1-b216-8e7f8f698bae service nova] [instance: 7445d2c5-26bd-4f8d-8653-51c721ea801f] Updating instance_info_cache with network_info: [] {{(pid=61806) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1080.029041] env[61806]: DEBUG nova.scheduler.client.report [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Inventory has not changed for provider a2858be1-fd22-4e08-979e-87ad25293407 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 138, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61806) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1080.260616] env[61806]: DEBUG nova.compute.manager [None req-f066ed90-7998-4480-842c-9e7fb5165059 tempest-ServersNegativeTestMultiTenantJSON-775824527 tempest-ServersNegativeTestMultiTenantJSON-775824527-project-member] [instance: cb9e9bac-d06c-4479-88e3-616b0f26ad7e] Start spawning the instance on the hypervisor. {{(pid=61806) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1080.282252] env[61806]: DEBUG nova.virt.hardware [None req-f066ed90-7998-4480-842c-9e7fb5165059 tempest-ServersNegativeTestMultiTenantJSON-775824527 tempest-ServersNegativeTestMultiTenantJSON-775824527-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-15T18:05:10Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-15T18:04:52Z,direct_url=,disk_format='vmdk',id=ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='84ba83fce2c349a6988173c5d6fc3b07',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-15T18:04:53Z,virtual_size=,visibility=), allow threads: False {{(pid=61806) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1080.282574] env[61806]: DEBUG nova.virt.hardware [None req-f066ed90-7998-4480-842c-9e7fb5165059 tempest-ServersNegativeTestMultiTenantJSON-775824527 tempest-ServersNegativeTestMultiTenantJSON-775824527-project-member] Flavor limits 0:0:0 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1080.282739] env[61806]: DEBUG nova.virt.hardware [None req-f066ed90-7998-4480-842c-9e7fb5165059 tempest-ServersNegativeTestMultiTenantJSON-775824527 tempest-ServersNegativeTestMultiTenantJSON-775824527-project-member] Image limits 0:0:0 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1080.282927] env[61806]: DEBUG nova.virt.hardware [None req-f066ed90-7998-4480-842c-9e7fb5165059 tempest-ServersNegativeTestMultiTenantJSON-775824527 tempest-ServersNegativeTestMultiTenantJSON-775824527-project-member] Flavor pref 0:0:0 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1080.283100] env[61806]: DEBUG nova.virt.hardware [None req-f066ed90-7998-4480-842c-9e7fb5165059 tempest-ServersNegativeTestMultiTenantJSON-775824527 tempest-ServersNegativeTestMultiTenantJSON-775824527-project-member] Image pref 0:0:0 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1080.283256] env[61806]: DEBUG nova.virt.hardware [None req-f066ed90-7998-4480-842c-9e7fb5165059 tempest-ServersNegativeTestMultiTenantJSON-775824527 tempest-ServersNegativeTestMultiTenantJSON-775824527-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1080.283466] env[61806]: DEBUG nova.virt.hardware [None req-f066ed90-7998-4480-842c-9e7fb5165059 tempest-ServersNegativeTestMultiTenantJSON-775824527 tempest-ServersNegativeTestMultiTenantJSON-775824527-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61806) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1080.283631] env[61806]: DEBUG nova.virt.hardware [None req-f066ed90-7998-4480-842c-9e7fb5165059 tempest-ServersNegativeTestMultiTenantJSON-775824527 tempest-ServersNegativeTestMultiTenantJSON-775824527-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61806) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1080.283805] env[61806]: DEBUG nova.virt.hardware [None req-f066ed90-7998-4480-842c-9e7fb5165059 tempest-ServersNegativeTestMultiTenantJSON-775824527 tempest-ServersNegativeTestMultiTenantJSON-775824527-project-member] Got 1 possible topologies {{(pid=61806) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1080.283973] env[61806]: DEBUG nova.virt.hardware [None req-f066ed90-7998-4480-842c-9e7fb5165059 tempest-ServersNegativeTestMultiTenantJSON-775824527 tempest-ServersNegativeTestMultiTenantJSON-775824527-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61806) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1080.284178] env[61806]: DEBUG nova.virt.hardware [None req-f066ed90-7998-4480-842c-9e7fb5165059 tempest-ServersNegativeTestMultiTenantJSON-775824527 tempest-ServersNegativeTestMultiTenantJSON-775824527-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61806) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1080.285088] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-24f03e49-0ec3-4ba1-8578-3a384a098cd0 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1080.292718] env[61806]: DEBUG nova.network.neutron [-] [instance: 7445d2c5-26bd-4f8d-8653-51c721ea801f] Updating instance_info_cache with network_info: [] {{(pid=61806) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1080.294895] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-48ce2619-5549-4d94-8d03-c939be9d4c68 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1080.319397] env[61806]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-516f0d42-0871-4ff9-b92f-e2316d8e4039 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1080.329088] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-84153213-ab47-4880-a47a-4ddded392eba {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1080.354240] env[61806]: DEBUG nova.compute.manager [req-99120017-7368-4164-bce3-d86595ef0290 req-6afa9218-9c54-49f1-b216-8e7f8f698bae service nova] [instance: 7445d2c5-26bd-4f8d-8653-51c721ea801f] Detach interface failed, port_id=773e4470-b54c-4449-b825-5bb680cf8d2f, reason: Instance 7445d2c5-26bd-4f8d-8653-51c721ea801f could not be found. {{(pid=61806) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10963}} [ 1080.469430] env[61806]: DEBUG nova.compute.manager [req-5ff6d484-fdb8-4154-8a78-7b0bdead4560 req-9fb060a2-19e1-4a93-a396-6b67d54ded65 service nova] [instance: cb9e9bac-d06c-4479-88e3-616b0f26ad7e] Received event network-vif-plugged-ed5dd9bf-96f6-4cad-8f5d-0ea62887f3b9 {{(pid=61806) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1080.469713] env[61806]: DEBUG oslo_concurrency.lockutils [req-5ff6d484-fdb8-4154-8a78-7b0bdead4560 req-9fb060a2-19e1-4a93-a396-6b67d54ded65 service nova] Acquiring lock "cb9e9bac-d06c-4479-88e3-616b0f26ad7e-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1080.469896] env[61806]: DEBUG oslo_concurrency.lockutils [req-5ff6d484-fdb8-4154-8a78-7b0bdead4560 req-9fb060a2-19e1-4a93-a396-6b67d54ded65 service nova] Lock "cb9e9bac-d06c-4479-88e3-616b0f26ad7e-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1080.470082] env[61806]: DEBUG oslo_concurrency.lockutils [req-5ff6d484-fdb8-4154-8a78-7b0bdead4560 req-9fb060a2-19e1-4a93-a396-6b67d54ded65 service nova] Lock "cb9e9bac-d06c-4479-88e3-616b0f26ad7e-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1080.470254] env[61806]: DEBUG nova.compute.manager [req-5ff6d484-fdb8-4154-8a78-7b0bdead4560 req-9fb060a2-19e1-4a93-a396-6b67d54ded65 service nova] [instance: cb9e9bac-d06c-4479-88e3-616b0f26ad7e] No waiting events found dispatching network-vif-plugged-ed5dd9bf-96f6-4cad-8f5d-0ea62887f3b9 {{(pid=61806) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1080.470422] env[61806]: WARNING nova.compute.manager [req-5ff6d484-fdb8-4154-8a78-7b0bdead4560 req-9fb060a2-19e1-4a93-a396-6b67d54ded65 service nova] [instance: cb9e9bac-d06c-4479-88e3-616b0f26ad7e] Received unexpected event network-vif-plugged-ed5dd9bf-96f6-4cad-8f5d-0ea62887f3b9 for instance with vm_state building and task_state spawning. [ 1080.534383] env[61806]: DEBUG nova.compute.resource_tracker [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=61806) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 1080.534658] env[61806]: DEBUG oslo_concurrency.lockutils [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 2.300s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1080.535386] env[61806]: DEBUG oslo_concurrency.lockutils [None req-cfb899a7-3b9e-475d-ac76-e219c5dfcd7e tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 1.690s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1080.535705] env[61806]: DEBUG nova.objects.instance [None req-cfb899a7-3b9e-475d-ac76-e219c5dfcd7e tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] Lazy-loading 'resources' on Instance uuid fd9e520a-2e7b-43d4-b84d-8933c32cc6a6 {{(pid=61806) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1080.536735] env[61806]: DEBUG oslo_service.periodic_task [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Running periodic task ComputeManager._run_pending_deletes {{(pid=61806) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1080.536897] env[61806]: DEBUG nova.compute.manager [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Cleaning up deleted instances {{(pid=61806) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11226}} [ 1080.562583] env[61806]: DEBUG nova.network.neutron [None req-f066ed90-7998-4480-842c-9e7fb5165059 tempest-ServersNegativeTestMultiTenantJSON-775824527 tempest-ServersNegativeTestMultiTenantJSON-775824527-project-member] [instance: cb9e9bac-d06c-4479-88e3-616b0f26ad7e] Successfully updated port: ed5dd9bf-96f6-4cad-8f5d-0ea62887f3b9 {{(pid=61806) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1080.800705] env[61806]: INFO nova.compute.manager [-] [instance: 7445d2c5-26bd-4f8d-8653-51c721ea801f] Took 1.26 seconds to deallocate network for instance. [ 1081.054392] env[61806]: DEBUG nova.compute.manager [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] There are 46 instances to clean {{(pid=61806) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11235}} [ 1081.054655] env[61806]: DEBUG nova.compute.manager [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] [instance: a61abfd9-3cc1-4a2b-b68b-a446094daf87] Instance has had 0 of 5 cleanup attempts {{(pid=61806) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1081.066038] env[61806]: DEBUG oslo_concurrency.lockutils [None req-f066ed90-7998-4480-842c-9e7fb5165059 tempest-ServersNegativeTestMultiTenantJSON-775824527 tempest-ServersNegativeTestMultiTenantJSON-775824527-project-member] Acquiring lock "refresh_cache-cb9e9bac-d06c-4479-88e3-616b0f26ad7e" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1081.066196] env[61806]: DEBUG oslo_concurrency.lockutils [None req-f066ed90-7998-4480-842c-9e7fb5165059 tempest-ServersNegativeTestMultiTenantJSON-775824527 tempest-ServersNegativeTestMultiTenantJSON-775824527-project-member] Acquired lock "refresh_cache-cb9e9bac-d06c-4479-88e3-616b0f26ad7e" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1081.066347] env[61806]: DEBUG nova.network.neutron [None req-f066ed90-7998-4480-842c-9e7fb5165059 tempest-ServersNegativeTestMultiTenantJSON-775824527 tempest-ServersNegativeTestMultiTenantJSON-775824527-project-member] [instance: cb9e9bac-d06c-4479-88e3-616b0f26ad7e] Building network info cache for instance {{(pid=61806) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1081.186269] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f1a6b754-6e87-42e3-aa39-fa5808bd1c80 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1081.194834] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-607cb355-8dd1-48c1-9f5f-6b07c73ba033 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1081.225342] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b9017d0b-9c41-4f3e-a81c-7ec33e4b7ade {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1081.233086] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9d81edfd-4ff2-496d-a44f-3d33911f253f {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1081.247218] env[61806]: DEBUG nova.compute.provider_tree [None req-cfb899a7-3b9e-475d-ac76-e219c5dfcd7e tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] Inventory has not changed in ProviderTree for provider: a2858be1-fd22-4e08-979e-87ad25293407 {{(pid=61806) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1081.309614] env[61806]: DEBUG oslo_concurrency.lockutils [None req-1f1225dc-2b77-47cd-9aaa-80fb8ebe5df6 tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1081.580775] env[61806]: DEBUG nova.compute.manager [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] [instance: 713d4e6e-6f0e-4844-9b55-cc010030eda6] Instance has had 0 of 5 cleanup attempts {{(pid=61806) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1081.611903] env[61806]: DEBUG nova.network.neutron [None req-f066ed90-7998-4480-842c-9e7fb5165059 tempest-ServersNegativeTestMultiTenantJSON-775824527 tempest-ServersNegativeTestMultiTenantJSON-775824527-project-member] [instance: cb9e9bac-d06c-4479-88e3-616b0f26ad7e] Instance cache missing network info. {{(pid=61806) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 1081.752027] env[61806]: DEBUG nova.scheduler.client.report [None req-cfb899a7-3b9e-475d-ac76-e219c5dfcd7e tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] Inventory has not changed for provider a2858be1-fd22-4e08-979e-87ad25293407 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 138, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61806) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1081.793287] env[61806]: DEBUG nova.network.neutron [None req-f066ed90-7998-4480-842c-9e7fb5165059 tempest-ServersNegativeTestMultiTenantJSON-775824527 tempest-ServersNegativeTestMultiTenantJSON-775824527-project-member] [instance: cb9e9bac-d06c-4479-88e3-616b0f26ad7e] Updating instance_info_cache with network_info: [{"id": "ed5dd9bf-96f6-4cad-8f5d-0ea62887f3b9", "address": "fa:16:3e:2e:c4:2e", "network": {"id": "40634ddd-4d4e-456e-8215-9e3dd0f31c5e", "bridge": "br-int", "label": "tempest-ServersNegativeTestMultiTenantJSON-596273102-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "008da83c5c894402a1158ead2cc19896", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "b5c34919-7d52-4a52-bab1-81af4c8182ef", "external-id": "nsx-vlan-transportzone-458", "segmentation_id": 458, "bound_drivers": {"0": "nsxv3"}}, "devname": "taped5dd9bf-96", "ovs_interfaceid": "ed5dd9bf-96f6-4cad-8f5d-0ea62887f3b9", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61806) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1082.084880] env[61806]: DEBUG nova.compute.manager [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] [instance: 71999b04-bf6e-4e4c-87c9-d5ec1932a80f] Instance has had 0 of 5 cleanup attempts {{(pid=61806) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1082.256281] env[61806]: DEBUG oslo_concurrency.lockutils [None req-cfb899a7-3b9e-475d-ac76-e219c5dfcd7e tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.721s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1082.258924] env[61806]: DEBUG oslo_concurrency.lockutils [None req-1f1225dc-2b77-47cd-9aaa-80fb8ebe5df6 tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.949s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1082.259237] env[61806]: DEBUG nova.objects.instance [None req-1f1225dc-2b77-47cd-9aaa-80fb8ebe5df6 tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Lazy-loading 'resources' on Instance uuid 7445d2c5-26bd-4f8d-8653-51c721ea801f {{(pid=61806) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1082.275255] env[61806]: INFO nova.scheduler.client.report [None req-cfb899a7-3b9e-475d-ac76-e219c5dfcd7e tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] Deleted allocations for instance fd9e520a-2e7b-43d4-b84d-8933c32cc6a6 [ 1082.295869] env[61806]: DEBUG oslo_concurrency.lockutils [None req-f066ed90-7998-4480-842c-9e7fb5165059 tempest-ServersNegativeTestMultiTenantJSON-775824527 tempest-ServersNegativeTestMultiTenantJSON-775824527-project-member] Releasing lock "refresh_cache-cb9e9bac-d06c-4479-88e3-616b0f26ad7e" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1082.296214] env[61806]: DEBUG nova.compute.manager [None req-f066ed90-7998-4480-842c-9e7fb5165059 tempest-ServersNegativeTestMultiTenantJSON-775824527 tempest-ServersNegativeTestMultiTenantJSON-775824527-project-member] [instance: cb9e9bac-d06c-4479-88e3-616b0f26ad7e] Instance network_info: |[{"id": "ed5dd9bf-96f6-4cad-8f5d-0ea62887f3b9", "address": "fa:16:3e:2e:c4:2e", "network": {"id": "40634ddd-4d4e-456e-8215-9e3dd0f31c5e", "bridge": "br-int", "label": "tempest-ServersNegativeTestMultiTenantJSON-596273102-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "008da83c5c894402a1158ead2cc19896", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "b5c34919-7d52-4a52-bab1-81af4c8182ef", "external-id": "nsx-vlan-transportzone-458", "segmentation_id": 458, "bound_drivers": {"0": "nsxv3"}}, "devname": "taped5dd9bf-96", "ovs_interfaceid": "ed5dd9bf-96f6-4cad-8f5d-0ea62887f3b9", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61806) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1082.296665] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-f066ed90-7998-4480-842c-9e7fb5165059 tempest-ServersNegativeTestMultiTenantJSON-775824527 tempest-ServersNegativeTestMultiTenantJSON-775824527-project-member] [instance: cb9e9bac-d06c-4479-88e3-616b0f26ad7e] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:2e:c4:2e', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'b5c34919-7d52-4a52-bab1-81af4c8182ef', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'ed5dd9bf-96f6-4cad-8f5d-0ea62887f3b9', 'vif_model': 'vmxnet3'}] {{(pid=61806) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1082.304110] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [None req-f066ed90-7998-4480-842c-9e7fb5165059 tempest-ServersNegativeTestMultiTenantJSON-775824527 tempest-ServersNegativeTestMultiTenantJSON-775824527-project-member] Creating folder: Project (008da83c5c894402a1158ead2cc19896). Parent ref: group-v277609. {{(pid=61806) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 1082.304722] env[61806]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-dc6d4bf9-a0c2-4fbf-95bc-1afcbf204151 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1082.315525] env[61806]: INFO nova.virt.vmwareapi.vm_util [None req-f066ed90-7998-4480-842c-9e7fb5165059 tempest-ServersNegativeTestMultiTenantJSON-775824527 tempest-ServersNegativeTestMultiTenantJSON-775824527-project-member] Created folder: Project (008da83c5c894402a1158ead2cc19896) in parent group-v277609. [ 1082.315743] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [None req-f066ed90-7998-4480-842c-9e7fb5165059 tempest-ServersNegativeTestMultiTenantJSON-775824527 tempest-ServersNegativeTestMultiTenantJSON-775824527-project-member] Creating folder: Instances. Parent ref: group-v277759. {{(pid=61806) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 1082.316593] env[61806]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-324fd331-ed91-4d82-9911-6861bdb6816a {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1082.326238] env[61806]: INFO nova.virt.vmwareapi.vm_util [None req-f066ed90-7998-4480-842c-9e7fb5165059 tempest-ServersNegativeTestMultiTenantJSON-775824527 tempest-ServersNegativeTestMultiTenantJSON-775824527-project-member] Created folder: Instances in parent group-v277759. [ 1082.326485] env[61806]: DEBUG oslo.service.loopingcall [None req-f066ed90-7998-4480-842c-9e7fb5165059 tempest-ServersNegativeTestMultiTenantJSON-775824527 tempest-ServersNegativeTestMultiTenantJSON-775824527-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61806) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1082.326689] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: cb9e9bac-d06c-4479-88e3-616b0f26ad7e] Creating VM on the ESX host {{(pid=61806) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 1082.326919] env[61806]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-6888b160-2a11-4ead-81a4-fd42432bae88 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1082.346198] env[61806]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1082.346198] env[61806]: value = "task-1295209" [ 1082.346198] env[61806]: _type = "Task" [ 1082.346198] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1082.354010] env[61806]: DEBUG oslo_vmware.api [-] Task: {'id': task-1295209, 'name': CreateVM_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1082.497220] env[61806]: DEBUG nova.compute.manager [req-eacb548a-57bd-472b-997b-e4c9da5f9d4c req-0795e3af-3a6c-4c6d-bbd8-326297f5c94d service nova] [instance: cb9e9bac-d06c-4479-88e3-616b0f26ad7e] Received event network-changed-ed5dd9bf-96f6-4cad-8f5d-0ea62887f3b9 {{(pid=61806) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1082.497432] env[61806]: DEBUG nova.compute.manager [req-eacb548a-57bd-472b-997b-e4c9da5f9d4c req-0795e3af-3a6c-4c6d-bbd8-326297f5c94d service nova] [instance: cb9e9bac-d06c-4479-88e3-616b0f26ad7e] Refreshing instance network info cache due to event network-changed-ed5dd9bf-96f6-4cad-8f5d-0ea62887f3b9. {{(pid=61806) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 1082.497659] env[61806]: DEBUG oslo_concurrency.lockutils [req-eacb548a-57bd-472b-997b-e4c9da5f9d4c req-0795e3af-3a6c-4c6d-bbd8-326297f5c94d service nova] Acquiring lock "refresh_cache-cb9e9bac-d06c-4479-88e3-616b0f26ad7e" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1082.497811] env[61806]: DEBUG oslo_concurrency.lockutils [req-eacb548a-57bd-472b-997b-e4c9da5f9d4c req-0795e3af-3a6c-4c6d-bbd8-326297f5c94d service nova] Acquired lock "refresh_cache-cb9e9bac-d06c-4479-88e3-616b0f26ad7e" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1082.497979] env[61806]: DEBUG nova.network.neutron [req-eacb548a-57bd-472b-997b-e4c9da5f9d4c req-0795e3af-3a6c-4c6d-bbd8-326297f5c94d service nova] [instance: cb9e9bac-d06c-4479-88e3-616b0f26ad7e] Refreshing network info cache for port ed5dd9bf-96f6-4cad-8f5d-0ea62887f3b9 {{(pid=61806) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1082.588579] env[61806]: DEBUG nova.compute.manager [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] [instance: fe1d4ace-7de3-423b-b423-9d7397aa2413] Instance has had 0 of 5 cleanup attempts {{(pid=61806) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1082.784275] env[61806]: DEBUG oslo_concurrency.lockutils [None req-cfb899a7-3b9e-475d-ac76-e219c5dfcd7e tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] Lock "fd9e520a-2e7b-43d4-b84d-8933c32cc6a6" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 7.113s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1082.842455] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-57852125-d6d5-4f8b-95c1-b588fd2b8d89 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1082.853326] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eaed309d-c61f-491d-b018-2c11d68c3e9c {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1082.860727] env[61806]: DEBUG oslo_vmware.api [-] Task: {'id': task-1295209, 'name': CreateVM_Task, 'duration_secs': 0.445889} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1082.885253] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: cb9e9bac-d06c-4479-88e3-616b0f26ad7e] Created VM on the ESX host {{(pid=61806) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 1082.886541] env[61806]: DEBUG oslo_concurrency.lockutils [None req-f066ed90-7998-4480-842c-9e7fb5165059 tempest-ServersNegativeTestMultiTenantJSON-775824527 tempest-ServersNegativeTestMultiTenantJSON-775824527-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1082.886748] env[61806]: DEBUG oslo_concurrency.lockutils [None req-f066ed90-7998-4480-842c-9e7fb5165059 tempest-ServersNegativeTestMultiTenantJSON-775824527 tempest-ServersNegativeTestMultiTenantJSON-775824527-project-member] Acquired lock "[datastore1] devstack-image-cache_base/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1082.887099] env[61806]: DEBUG oslo_concurrency.lockutils [None req-f066ed90-7998-4480-842c-9e7fb5165059 tempest-ServersNegativeTestMultiTenantJSON-775824527 tempest-ServersNegativeTestMultiTenantJSON-775824527-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1082.887839] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a7477a79-c132-4184-853d-104f7524e972 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1082.890357] env[61806]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e81598df-f789-470c-a9d9-6e0a9fae469f {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1082.898458] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3e6a1ec8-6c22-4bc4-b893-68378748dd19 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1082.902412] env[61806]: DEBUG oslo_vmware.api [None req-f066ed90-7998-4480-842c-9e7fb5165059 tempest-ServersNegativeTestMultiTenantJSON-775824527 tempest-ServersNegativeTestMultiTenantJSON-775824527-project-member] Waiting for the task: (returnval){ [ 1082.902412] env[61806]: value = "session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]520434e0-7662-d3e8-20d9-6b5e3204861c" [ 1082.902412] env[61806]: _type = "Task" [ 1082.902412] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1082.913717] env[61806]: DEBUG nova.compute.provider_tree [None req-1f1225dc-2b77-47cd-9aaa-80fb8ebe5df6 tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Inventory has not changed in ProviderTree for provider: a2858be1-fd22-4e08-979e-87ad25293407 {{(pid=61806) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1082.920075] env[61806]: DEBUG oslo_vmware.api [None req-f066ed90-7998-4480-842c-9e7fb5165059 tempest-ServersNegativeTestMultiTenantJSON-775824527 tempest-ServersNegativeTestMultiTenantJSON-775824527-project-member] Task: {'id': session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]520434e0-7662-d3e8-20d9-6b5e3204861c, 'name': SearchDatastore_Task, 'duration_secs': 0.010508} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1082.920369] env[61806]: DEBUG oslo_concurrency.lockutils [None req-f066ed90-7998-4480-842c-9e7fb5165059 tempest-ServersNegativeTestMultiTenantJSON-775824527 tempest-ServersNegativeTestMultiTenantJSON-775824527-project-member] Releasing lock "[datastore1] devstack-image-cache_base/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1082.920597] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-f066ed90-7998-4480-842c-9e7fb5165059 tempest-ServersNegativeTestMultiTenantJSON-775824527 tempest-ServersNegativeTestMultiTenantJSON-775824527-project-member] [instance: cb9e9bac-d06c-4479-88e3-616b0f26ad7e] Processing image ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa {{(pid=61806) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1082.920836] env[61806]: DEBUG oslo_concurrency.lockutils [None req-f066ed90-7998-4480-842c-9e7fb5165059 tempest-ServersNegativeTestMultiTenantJSON-775824527 tempest-ServersNegativeTestMultiTenantJSON-775824527-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa.vmdk" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1082.920991] env[61806]: DEBUG oslo_concurrency.lockutils [None req-f066ed90-7998-4480-842c-9e7fb5165059 tempest-ServersNegativeTestMultiTenantJSON-775824527 tempest-ServersNegativeTestMultiTenantJSON-775824527-project-member] Acquired lock "[datastore1] devstack-image-cache_base/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa.vmdk" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1082.921200] env[61806]: DEBUG nova.virt.vmwareapi.ds_util [None req-f066ed90-7998-4480-842c-9e7fb5165059 tempest-ServersNegativeTestMultiTenantJSON-775824527 tempest-ServersNegativeTestMultiTenantJSON-775824527-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61806) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1082.921978] env[61806]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-c32fec05-4f3c-456b-a252-b5d85ea2e4d6 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1082.929755] env[61806]: DEBUG nova.virt.vmwareapi.ds_util [None req-f066ed90-7998-4480-842c-9e7fb5165059 tempest-ServersNegativeTestMultiTenantJSON-775824527 tempest-ServersNegativeTestMultiTenantJSON-775824527-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61806) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1082.930539] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-f066ed90-7998-4480-842c-9e7fb5165059 tempest-ServersNegativeTestMultiTenantJSON-775824527 tempest-ServersNegativeTestMultiTenantJSON-775824527-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61806) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 1082.930961] env[61806]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-a8e9f420-2583-4c46-985c-73552084d85d {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1082.936650] env[61806]: DEBUG oslo_vmware.api [None req-f066ed90-7998-4480-842c-9e7fb5165059 tempest-ServersNegativeTestMultiTenantJSON-775824527 tempest-ServersNegativeTestMultiTenantJSON-775824527-project-member] Waiting for the task: (returnval){ [ 1082.936650] env[61806]: value = "session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]527f1b59-6689-7a2a-d555-53ecc9bf91a9" [ 1082.936650] env[61806]: _type = "Task" [ 1082.936650] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1082.945141] env[61806]: DEBUG oslo_vmware.api [None req-f066ed90-7998-4480-842c-9e7fb5165059 tempest-ServersNegativeTestMultiTenantJSON-775824527 tempest-ServersNegativeTestMultiTenantJSON-775824527-project-member] Task: {'id': session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]527f1b59-6689-7a2a-d555-53ecc9bf91a9, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1083.092723] env[61806]: DEBUG nova.compute.manager [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] [instance: 9357e61f-1628-43bd-ab46-de13c1529f51] Instance has had 0 of 5 cleanup attempts {{(pid=61806) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1083.204971] env[61806]: DEBUG nova.network.neutron [req-eacb548a-57bd-472b-997b-e4c9da5f9d4c req-0795e3af-3a6c-4c6d-bbd8-326297f5c94d service nova] [instance: cb9e9bac-d06c-4479-88e3-616b0f26ad7e] Updated VIF entry in instance network info cache for port ed5dd9bf-96f6-4cad-8f5d-0ea62887f3b9. {{(pid=61806) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1083.204971] env[61806]: DEBUG nova.network.neutron [req-eacb548a-57bd-472b-997b-e4c9da5f9d4c req-0795e3af-3a6c-4c6d-bbd8-326297f5c94d service nova] [instance: cb9e9bac-d06c-4479-88e3-616b0f26ad7e] Updating instance_info_cache with network_info: [{"id": "ed5dd9bf-96f6-4cad-8f5d-0ea62887f3b9", "address": "fa:16:3e:2e:c4:2e", "network": {"id": "40634ddd-4d4e-456e-8215-9e3dd0f31c5e", "bridge": "br-int", "label": "tempest-ServersNegativeTestMultiTenantJSON-596273102-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "008da83c5c894402a1158ead2cc19896", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "b5c34919-7d52-4a52-bab1-81af4c8182ef", "external-id": "nsx-vlan-transportzone-458", "segmentation_id": 458, "bound_drivers": {"0": "nsxv3"}}, "devname": "taped5dd9bf-96", "ovs_interfaceid": "ed5dd9bf-96f6-4cad-8f5d-0ea62887f3b9", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61806) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1083.417126] env[61806]: DEBUG nova.scheduler.client.report [None req-1f1225dc-2b77-47cd-9aaa-80fb8ebe5df6 tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Inventory has not changed for provider a2858be1-fd22-4e08-979e-87ad25293407 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 138, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61806) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1083.448176] env[61806]: DEBUG oslo_vmware.api [None req-f066ed90-7998-4480-842c-9e7fb5165059 tempest-ServersNegativeTestMultiTenantJSON-775824527 tempest-ServersNegativeTestMultiTenantJSON-775824527-project-member] Task: {'id': session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]527f1b59-6689-7a2a-d555-53ecc9bf91a9, 'name': SearchDatastore_Task, 'duration_secs': 0.009318} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1083.449051] env[61806]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-8fad893b-c4c0-4fde-8985-7c4e3b848d14 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1083.454590] env[61806]: DEBUG oslo_vmware.api [None req-f066ed90-7998-4480-842c-9e7fb5165059 tempest-ServersNegativeTestMultiTenantJSON-775824527 tempest-ServersNegativeTestMultiTenantJSON-775824527-project-member] Waiting for the task: (returnval){ [ 1083.454590] env[61806]: value = "session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]52d041d5-fd60-3e71-f58f-149c3fe87a56" [ 1083.454590] env[61806]: _type = "Task" [ 1083.454590] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1083.462517] env[61806]: DEBUG oslo_vmware.api [None req-f066ed90-7998-4480-842c-9e7fb5165059 tempest-ServersNegativeTestMultiTenantJSON-775824527 tempest-ServersNegativeTestMultiTenantJSON-775824527-project-member] Task: {'id': session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]52d041d5-fd60-3e71-f58f-149c3fe87a56, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1083.596460] env[61806]: DEBUG nova.compute.manager [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] [instance: 962d3d88-03ee-4ff2-9d79-15091671b330] Instance has had 0 of 5 cleanup attempts {{(pid=61806) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1083.710890] env[61806]: DEBUG oslo_concurrency.lockutils [req-eacb548a-57bd-472b-997b-e4c9da5f9d4c req-0795e3af-3a6c-4c6d-bbd8-326297f5c94d service nova] Releasing lock "refresh_cache-cb9e9bac-d06c-4479-88e3-616b0f26ad7e" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1083.922902] env[61806]: DEBUG oslo_concurrency.lockutils [None req-1f1225dc-2b77-47cd-9aaa-80fb8ebe5df6 tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.664s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1083.951490] env[61806]: INFO nova.scheduler.client.report [None req-1f1225dc-2b77-47cd-9aaa-80fb8ebe5df6 tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Deleted allocations for instance 7445d2c5-26bd-4f8d-8653-51c721ea801f [ 1083.965664] env[61806]: DEBUG oslo_vmware.api [None req-f066ed90-7998-4480-842c-9e7fb5165059 tempest-ServersNegativeTestMultiTenantJSON-775824527 tempest-ServersNegativeTestMultiTenantJSON-775824527-project-member] Task: {'id': session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]52d041d5-fd60-3e71-f58f-149c3fe87a56, 'name': SearchDatastore_Task, 'duration_secs': 0.01528} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1083.966015] env[61806]: DEBUG oslo_concurrency.lockutils [None req-f066ed90-7998-4480-842c-9e7fb5165059 tempest-ServersNegativeTestMultiTenantJSON-775824527 tempest-ServersNegativeTestMultiTenantJSON-775824527-project-member] Releasing lock "[datastore1] devstack-image-cache_base/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa.vmdk" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1083.966302] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [None req-f066ed90-7998-4480-842c-9e7fb5165059 tempest-ServersNegativeTestMultiTenantJSON-775824527 tempest-ServersNegativeTestMultiTenantJSON-775824527-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa.vmdk to [datastore1] cb9e9bac-d06c-4479-88e3-616b0f26ad7e/cb9e9bac-d06c-4479-88e3-616b0f26ad7e.vmdk {{(pid=61806) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 1083.966611] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-d35e4d97-53c5-40ae-ab0e-2ecfb0f57437 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1083.973252] env[61806]: DEBUG oslo_vmware.api [None req-f066ed90-7998-4480-842c-9e7fb5165059 tempest-ServersNegativeTestMultiTenantJSON-775824527 tempest-ServersNegativeTestMultiTenantJSON-775824527-project-member] Waiting for the task: (returnval){ [ 1083.973252] env[61806]: value = "task-1295210" [ 1083.973252] env[61806]: _type = "Task" [ 1083.973252] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1083.981766] env[61806]: DEBUG oslo_vmware.api [None req-f066ed90-7998-4480-842c-9e7fb5165059 tempest-ServersNegativeTestMultiTenantJSON-775824527 tempest-ServersNegativeTestMultiTenantJSON-775824527-project-member] Task: {'id': task-1295210, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1084.099615] env[61806]: DEBUG nova.compute.manager [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] [instance: e146f7e5-bbf8-4adf-a2e9-b7e753438976] Instance has had 0 of 5 cleanup attempts {{(pid=61806) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1084.463519] env[61806]: DEBUG oslo_concurrency.lockutils [None req-1f1225dc-2b77-47cd-9aaa-80fb8ebe5df6 tempest-DeleteServersTestJSON-1007282177 tempest-DeleteServersTestJSON-1007282177-project-member] Lock "7445d2c5-26bd-4f8d-8653-51c721ea801f" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 6.042s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1084.483980] env[61806]: DEBUG oslo_vmware.api [None req-f066ed90-7998-4480-842c-9e7fb5165059 tempest-ServersNegativeTestMultiTenantJSON-775824527 tempest-ServersNegativeTestMultiTenantJSON-775824527-project-member] Task: {'id': task-1295210, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1084.603596] env[61806]: DEBUG nova.compute.manager [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] [instance: e4638faa-7be1-4909-a595-c437837fa314] Instance has had 0 of 5 cleanup attempts {{(pid=61806) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1084.729088] env[61806]: DEBUG oslo_concurrency.lockutils [None req-b9e61586-635a-432e-ae51-c0a6d111aaf8 tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] Acquiring lock "29210bcd-0fe1-4ed9-a459-abc8016a4255" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1084.729343] env[61806]: DEBUG oslo_concurrency.lockutils [None req-b9e61586-635a-432e-ae51-c0a6d111aaf8 tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] Lock "29210bcd-0fe1-4ed9-a459-abc8016a4255" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1084.949860] env[61806]: DEBUG oslo_vmware.rw_handles [None req-5fb160b1-8322-4bcb-adab-57abc5e44515 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Getting lease state for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52ca7cdb-8ede-5cf9-2c14-5ecf127f4778/disk-0.vmdk. {{(pid=61806) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 1084.950893] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e93e65aa-7ebd-4b6d-b2ca-adf6486c459d {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1084.957084] env[61806]: DEBUG oslo_vmware.rw_handles [None req-5fb160b1-8322-4bcb-adab-57abc5e44515 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52ca7cdb-8ede-5cf9-2c14-5ecf127f4778/disk-0.vmdk is in state: ready. {{(pid=61806) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 1084.957263] env[61806]: ERROR oslo_vmware.rw_handles [None req-5fb160b1-8322-4bcb-adab-57abc5e44515 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Aborting lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52ca7cdb-8ede-5cf9-2c14-5ecf127f4778/disk-0.vmdk due to incomplete transfer. [ 1084.957489] env[61806]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseAbort with opID=oslo.vmware-857289ea-027a-456d-b151-0fb619506aa8 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1084.963949] env[61806]: DEBUG oslo_vmware.rw_handles [None req-5fb160b1-8322-4bcb-adab-57abc5e44515 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Closed VMDK read handle for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52ca7cdb-8ede-5cf9-2c14-5ecf127f4778/disk-0.vmdk. {{(pid=61806) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:735}} [ 1084.964207] env[61806]: DEBUG nova.virt.vmwareapi.images [None req-5fb160b1-8322-4bcb-adab-57abc5e44515 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] [instance: d0669232-6f33-4b30-97c2-2e4239af1ad0] Uploaded image 87519589-60c5-40c3-8c84-791ed1a73a25 to the Glance image server {{(pid=61806) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:473}} [ 1084.966566] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [None req-5fb160b1-8322-4bcb-adab-57abc5e44515 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] [instance: d0669232-6f33-4b30-97c2-2e4239af1ad0] Destroying the VM {{(pid=61806) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1358}} [ 1084.967009] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-24749d61-9bee-4dbc-9d52-20ca369e3921 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1084.973434] env[61806]: DEBUG oslo_vmware.api [None req-5fb160b1-8322-4bcb-adab-57abc5e44515 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Waiting for the task: (returnval){ [ 1084.973434] env[61806]: value = "task-1295212" [ 1084.973434] env[61806]: _type = "Task" [ 1084.973434] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1084.983946] env[61806]: DEBUG oslo_vmware.api [None req-5fb160b1-8322-4bcb-adab-57abc5e44515 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Task: {'id': task-1295212, 'name': Destroy_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1084.987133] env[61806]: DEBUG oslo_vmware.api [None req-f066ed90-7998-4480-842c-9e7fb5165059 tempest-ServersNegativeTestMultiTenantJSON-775824527 tempest-ServersNegativeTestMultiTenantJSON-775824527-project-member] Task: {'id': task-1295210, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.616242} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1084.987409] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [None req-f066ed90-7998-4480-842c-9e7fb5165059 tempest-ServersNegativeTestMultiTenantJSON-775824527 tempest-ServersNegativeTestMultiTenantJSON-775824527-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa.vmdk to [datastore1] cb9e9bac-d06c-4479-88e3-616b0f26ad7e/cb9e9bac-d06c-4479-88e3-616b0f26ad7e.vmdk {{(pid=61806) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 1084.987630] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-f066ed90-7998-4480-842c-9e7fb5165059 tempest-ServersNegativeTestMultiTenantJSON-775824527 tempest-ServersNegativeTestMultiTenantJSON-775824527-project-member] [instance: cb9e9bac-d06c-4479-88e3-616b0f26ad7e] Extending root virtual disk to 1048576 {{(pid=61806) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1084.987894] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-1342b962-086e-4ea8-8a73-378d8b2e2485 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1084.993762] env[61806]: DEBUG oslo_vmware.api [None req-f066ed90-7998-4480-842c-9e7fb5165059 tempest-ServersNegativeTestMultiTenantJSON-775824527 tempest-ServersNegativeTestMultiTenantJSON-775824527-project-member] Waiting for the task: (returnval){ [ 1084.993762] env[61806]: value = "task-1295213" [ 1084.993762] env[61806]: _type = "Task" [ 1084.993762] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1085.001301] env[61806]: DEBUG oslo_vmware.api [None req-f066ed90-7998-4480-842c-9e7fb5165059 tempest-ServersNegativeTestMultiTenantJSON-775824527 tempest-ServersNegativeTestMultiTenantJSON-775824527-project-member] Task: {'id': task-1295213, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1085.107085] env[61806]: DEBUG nova.compute.manager [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] [instance: 0bac3048-759a-47d9-a6ef-a0386a49a974] Instance has had 0 of 5 cleanup attempts {{(pid=61806) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1085.232650] env[61806]: DEBUG nova.compute.manager [None req-b9e61586-635a-432e-ae51-c0a6d111aaf8 tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] [instance: 29210bcd-0fe1-4ed9-a459-abc8016a4255] Starting instance... {{(pid=61806) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1085.483509] env[61806]: DEBUG oslo_vmware.api [None req-5fb160b1-8322-4bcb-adab-57abc5e44515 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Task: {'id': task-1295212, 'name': Destroy_Task, 'duration_secs': 0.374601} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1085.483733] env[61806]: INFO nova.virt.vmwareapi.vm_util [None req-5fb160b1-8322-4bcb-adab-57abc5e44515 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] [instance: d0669232-6f33-4b30-97c2-2e4239af1ad0] Destroyed the VM [ 1085.483974] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-5fb160b1-8322-4bcb-adab-57abc5e44515 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] [instance: d0669232-6f33-4b30-97c2-2e4239af1ad0] Deleting Snapshot of the VM instance {{(pid=61806) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 1085.484244] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-97e010dc-a871-4995-886e-8fcf79e835cc {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1085.490972] env[61806]: DEBUG oslo_vmware.api [None req-5fb160b1-8322-4bcb-adab-57abc5e44515 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Waiting for the task: (returnval){ [ 1085.490972] env[61806]: value = "task-1295214" [ 1085.490972] env[61806]: _type = "Task" [ 1085.490972] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1085.500132] env[61806]: DEBUG oslo_vmware.api [None req-5fb160b1-8322-4bcb-adab-57abc5e44515 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Task: {'id': task-1295214, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1085.504976] env[61806]: DEBUG oslo_vmware.api [None req-f066ed90-7998-4480-842c-9e7fb5165059 tempest-ServersNegativeTestMultiTenantJSON-775824527 tempest-ServersNegativeTestMultiTenantJSON-775824527-project-member] Task: {'id': task-1295213, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.064629} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1085.505328] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-f066ed90-7998-4480-842c-9e7fb5165059 tempest-ServersNegativeTestMultiTenantJSON-775824527 tempest-ServersNegativeTestMultiTenantJSON-775824527-project-member] [instance: cb9e9bac-d06c-4479-88e3-616b0f26ad7e] Extended root virtual disk {{(pid=61806) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1085.506271] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ab240163-895a-4e57-81bd-ca0ada4b307d {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1085.528630] env[61806]: DEBUG nova.virt.vmwareapi.volumeops [None req-f066ed90-7998-4480-842c-9e7fb5165059 tempest-ServersNegativeTestMultiTenantJSON-775824527 tempest-ServersNegativeTestMultiTenantJSON-775824527-project-member] [instance: cb9e9bac-d06c-4479-88e3-616b0f26ad7e] Reconfiguring VM instance instance-00000069 to attach disk [datastore1] cb9e9bac-d06c-4479-88e3-616b0f26ad7e/cb9e9bac-d06c-4479-88e3-616b0f26ad7e.vmdk or device None with type sparse {{(pid=61806) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1085.528630] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-915b12d4-84ee-4c86-890b-794377cbb9d8 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1085.550247] env[61806]: DEBUG oslo_vmware.api [None req-f066ed90-7998-4480-842c-9e7fb5165059 tempest-ServersNegativeTestMultiTenantJSON-775824527 tempest-ServersNegativeTestMultiTenantJSON-775824527-project-member] Waiting for the task: (returnval){ [ 1085.550247] env[61806]: value = "task-1295215" [ 1085.550247] env[61806]: _type = "Task" [ 1085.550247] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1085.558048] env[61806]: DEBUG oslo_vmware.api [None req-f066ed90-7998-4480-842c-9e7fb5165059 tempest-ServersNegativeTestMultiTenantJSON-775824527 tempest-ServersNegativeTestMultiTenantJSON-775824527-project-member] Task: {'id': task-1295215, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1085.610712] env[61806]: DEBUG nova.compute.manager [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] [instance: 461f98f2-80ba-4088-9e7f-1ba2a9b0c1d9] Instance has had 0 of 5 cleanup attempts {{(pid=61806) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1085.751711] env[61806]: DEBUG oslo_concurrency.lockutils [None req-b9e61586-635a-432e-ae51-c0a6d111aaf8 tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1085.752064] env[61806]: DEBUG oslo_concurrency.lockutils [None req-b9e61586-635a-432e-ae51-c0a6d111aaf8 tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.000s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1085.754258] env[61806]: INFO nova.compute.claims [None req-b9e61586-635a-432e-ae51-c0a6d111aaf8 tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] [instance: 29210bcd-0fe1-4ed9-a459-abc8016a4255] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1086.000790] env[61806]: DEBUG oslo_vmware.api [None req-5fb160b1-8322-4bcb-adab-57abc5e44515 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Task: {'id': task-1295214, 'name': RemoveSnapshot_Task} progress is 17%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1086.061574] env[61806]: DEBUG oslo_vmware.api [None req-f066ed90-7998-4480-842c-9e7fb5165059 tempest-ServersNegativeTestMultiTenantJSON-775824527 tempest-ServersNegativeTestMultiTenantJSON-775824527-project-member] Task: {'id': task-1295215, 'name': ReconfigVM_Task, 'duration_secs': 0.297217} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1086.063054] env[61806]: DEBUG nova.virt.vmwareapi.volumeops [None req-f066ed90-7998-4480-842c-9e7fb5165059 tempest-ServersNegativeTestMultiTenantJSON-775824527 tempest-ServersNegativeTestMultiTenantJSON-775824527-project-member] [instance: cb9e9bac-d06c-4479-88e3-616b0f26ad7e] Reconfigured VM instance instance-00000069 to attach disk [datastore1] cb9e9bac-d06c-4479-88e3-616b0f26ad7e/cb9e9bac-d06c-4479-88e3-616b0f26ad7e.vmdk or device None with type sparse {{(pid=61806) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1086.063054] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-d14f6f07-4522-41bc-b119-f31af63f948d {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1086.068636] env[61806]: DEBUG oslo_vmware.api [None req-f066ed90-7998-4480-842c-9e7fb5165059 tempest-ServersNegativeTestMultiTenantJSON-775824527 tempest-ServersNegativeTestMultiTenantJSON-775824527-project-member] Waiting for the task: (returnval){ [ 1086.068636] env[61806]: value = "task-1295216" [ 1086.068636] env[61806]: _type = "Task" [ 1086.068636] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1086.076334] env[61806]: DEBUG oslo_vmware.api [None req-f066ed90-7998-4480-842c-9e7fb5165059 tempest-ServersNegativeTestMultiTenantJSON-775824527 tempest-ServersNegativeTestMultiTenantJSON-775824527-project-member] Task: {'id': task-1295216, 'name': Rename_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1086.114058] env[61806]: DEBUG nova.compute.manager [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] [instance: bdbaef31-7d37-40cb-ac3e-ad152b37e2d2] Instance has had 0 of 5 cleanup attempts {{(pid=61806) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1086.501552] env[61806]: DEBUG oslo_vmware.api [None req-5fb160b1-8322-4bcb-adab-57abc5e44515 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Task: {'id': task-1295214, 'name': RemoveSnapshot_Task, 'duration_secs': 0.551722} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1086.502311] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-5fb160b1-8322-4bcb-adab-57abc5e44515 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] [instance: d0669232-6f33-4b30-97c2-2e4239af1ad0] Deleted Snapshot of the VM instance {{(pid=61806) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 1086.502402] env[61806]: INFO nova.compute.manager [None req-5fb160b1-8322-4bcb-adab-57abc5e44515 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] [instance: d0669232-6f33-4b30-97c2-2e4239af1ad0] Took 12.39 seconds to snapshot the instance on the hypervisor. [ 1086.578942] env[61806]: DEBUG oslo_vmware.api [None req-f066ed90-7998-4480-842c-9e7fb5165059 tempest-ServersNegativeTestMultiTenantJSON-775824527 tempest-ServersNegativeTestMultiTenantJSON-775824527-project-member] Task: {'id': task-1295216, 'name': Rename_Task, 'duration_secs': 0.150366} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1086.579198] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [None req-f066ed90-7998-4480-842c-9e7fb5165059 tempest-ServersNegativeTestMultiTenantJSON-775824527 tempest-ServersNegativeTestMultiTenantJSON-775824527-project-member] [instance: cb9e9bac-d06c-4479-88e3-616b0f26ad7e] Powering on the VM {{(pid=61806) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1086.579447] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-b466eb27-2cea-44b7-af51-10445dd9e0b6 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1086.585110] env[61806]: DEBUG oslo_vmware.api [None req-f066ed90-7998-4480-842c-9e7fb5165059 tempest-ServersNegativeTestMultiTenantJSON-775824527 tempest-ServersNegativeTestMultiTenantJSON-775824527-project-member] Waiting for the task: (returnval){ [ 1086.585110] env[61806]: value = "task-1295217" [ 1086.585110] env[61806]: _type = "Task" [ 1086.585110] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1086.592713] env[61806]: DEBUG oslo_vmware.api [None req-f066ed90-7998-4480-842c-9e7fb5165059 tempest-ServersNegativeTestMultiTenantJSON-775824527 tempest-ServersNegativeTestMultiTenantJSON-775824527-project-member] Task: {'id': task-1295217, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1086.616837] env[61806]: DEBUG nova.compute.manager [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] [instance: 38e96e22-4200-4175-9085-31977357da2d] Instance has had 0 of 5 cleanup attempts {{(pid=61806) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1086.700839] env[61806]: DEBUG nova.compute.manager [None req-7f52d370-4c9e-4ec0-a23f-5733d9a2d432 tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] [instance: 06658134-4c6d-4911-9a55-b6805f555c78] Stashing vm_state: active {{(pid=61806) _prep_resize /opt/stack/nova/nova/compute/manager.py:5624}} [ 1086.848304] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cc5ce2fb-aab2-49e8-b342-2fd4da3a34ed {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1086.855773] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e016354c-d089-4dfe-b774-f43dd9f20ec1 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1086.885642] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6ac3efd9-e52b-4240-b9fe-5925c1e177d5 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1086.893670] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1c1e4b3d-ac02-4466-9ad8-054a07212019 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1086.907641] env[61806]: DEBUG nova.compute.provider_tree [None req-b9e61586-635a-432e-ae51-c0a6d111aaf8 tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] Inventory has not changed in ProviderTree for provider: a2858be1-fd22-4e08-979e-87ad25293407 {{(pid=61806) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1087.059632] env[61806]: DEBUG nova.compute.manager [None req-5fb160b1-8322-4bcb-adab-57abc5e44515 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] [instance: d0669232-6f33-4b30-97c2-2e4239af1ad0] Found 2 images (rotation: 2) {{(pid=61806) _rotate_backups /opt/stack/nova/nova/compute/manager.py:4554}} [ 1087.095085] env[61806]: DEBUG oslo_vmware.api [None req-f066ed90-7998-4480-842c-9e7fb5165059 tempest-ServersNegativeTestMultiTenantJSON-775824527 tempest-ServersNegativeTestMultiTenantJSON-775824527-project-member] Task: {'id': task-1295217, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1087.119551] env[61806]: DEBUG nova.compute.manager [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] [instance: 676f18b3-3da1-4a53-b96d-fe64ee8f9101] Instance has had 0 of 5 cleanup attempts {{(pid=61806) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1087.217623] env[61806]: DEBUG oslo_concurrency.lockutils [None req-7f52d370-4c9e-4ec0-a23f-5733d9a2d432 tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1087.410889] env[61806]: DEBUG nova.scheduler.client.report [None req-b9e61586-635a-432e-ae51-c0a6d111aaf8 tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] Inventory has not changed for provider a2858be1-fd22-4e08-979e-87ad25293407 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 138, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61806) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1087.595403] env[61806]: DEBUG oslo_vmware.api [None req-f066ed90-7998-4480-842c-9e7fb5165059 tempest-ServersNegativeTestMultiTenantJSON-775824527 tempest-ServersNegativeTestMultiTenantJSON-775824527-project-member] Task: {'id': task-1295217, 'name': PowerOnVM_Task, 'duration_secs': 0.81234} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1087.597827] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [None req-f066ed90-7998-4480-842c-9e7fb5165059 tempest-ServersNegativeTestMultiTenantJSON-775824527 tempest-ServersNegativeTestMultiTenantJSON-775824527-project-member] [instance: cb9e9bac-d06c-4479-88e3-616b0f26ad7e] Powered on the VM {{(pid=61806) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1087.597827] env[61806]: INFO nova.compute.manager [None req-f066ed90-7998-4480-842c-9e7fb5165059 tempest-ServersNegativeTestMultiTenantJSON-775824527 tempest-ServersNegativeTestMultiTenantJSON-775824527-project-member] [instance: cb9e9bac-d06c-4479-88e3-616b0f26ad7e] Took 7.34 seconds to spawn the instance on the hypervisor. [ 1087.597827] env[61806]: DEBUG nova.compute.manager [None req-f066ed90-7998-4480-842c-9e7fb5165059 tempest-ServersNegativeTestMultiTenantJSON-775824527 tempest-ServersNegativeTestMultiTenantJSON-775824527-project-member] [instance: cb9e9bac-d06c-4479-88e3-616b0f26ad7e] Checking state {{(pid=61806) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1087.597827] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-72244bf2-93c5-4ec9-a703-66088dcee372 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1087.623381] env[61806]: DEBUG nova.compute.manager [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] [instance: 3f4fa1ba-a6ad-4e0f-aebc-973816ee09d5] Instance has had 0 of 5 cleanup attempts {{(pid=61806) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1087.915252] env[61806]: DEBUG oslo_concurrency.lockutils [None req-b9e61586-635a-432e-ae51-c0a6d111aaf8 tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.163s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1087.915834] env[61806]: DEBUG nova.compute.manager [None req-b9e61586-635a-432e-ae51-c0a6d111aaf8 tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] [instance: 29210bcd-0fe1-4ed9-a459-abc8016a4255] Start building networks asynchronously for instance. {{(pid=61806) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1087.918418] env[61806]: DEBUG oslo_concurrency.lockutils [None req-7f52d370-4c9e-4ec0-a23f-5733d9a2d432 tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: waited 0.701s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1087.986208] env[61806]: DEBUG nova.compute.manager [None req-4284f75d-699f-4bae-83bf-967ca811ad7c tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] [instance: d0669232-6f33-4b30-97c2-2e4239af1ad0] Checking state {{(pid=61806) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1087.986873] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3106f292-b6a2-4321-852b-5789143f4112 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1088.116293] env[61806]: INFO nova.compute.manager [None req-f066ed90-7998-4480-842c-9e7fb5165059 tempest-ServersNegativeTestMultiTenantJSON-775824527 tempest-ServersNegativeTestMultiTenantJSON-775824527-project-member] [instance: cb9e9bac-d06c-4479-88e3-616b0f26ad7e] Took 12.51 seconds to build instance. [ 1088.125859] env[61806]: DEBUG nova.compute.manager [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] [instance: c486c00d-61fc-4795-9299-e6548b686ee8] Instance has had 0 of 5 cleanup attempts {{(pid=61806) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1088.421688] env[61806]: DEBUG nova.compute.utils [None req-b9e61586-635a-432e-ae51-c0a6d111aaf8 tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] Using /dev/sd instead of None {{(pid=61806) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1088.423141] env[61806]: DEBUG nova.compute.manager [None req-b9e61586-635a-432e-ae51-c0a6d111aaf8 tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] [instance: 29210bcd-0fe1-4ed9-a459-abc8016a4255] Allocating IP information in the background. {{(pid=61806) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1088.423324] env[61806]: DEBUG nova.network.neutron [None req-b9e61586-635a-432e-ae51-c0a6d111aaf8 tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] [instance: 29210bcd-0fe1-4ed9-a459-abc8016a4255] allocate_for_instance() {{(pid=61806) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 1088.428087] env[61806]: INFO nova.compute.claims [None req-7f52d370-4c9e-4ec0-a23f-5733d9a2d432 tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] [instance: 06658134-4c6d-4911-9a55-b6805f555c78] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1088.494845] env[61806]: DEBUG nova.policy [None req-b9e61586-635a-432e-ae51-c0a6d111aaf8 tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '50a051926b2b4ea9b22828762d548ce6', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'bbbb09a49203421d807f733616c854df', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61806) authorize /opt/stack/nova/nova/policy.py:201}} [ 1088.496981] env[61806]: INFO nova.compute.manager [None req-4284f75d-699f-4bae-83bf-967ca811ad7c tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] [instance: d0669232-6f33-4b30-97c2-2e4239af1ad0] instance snapshotting [ 1088.497511] env[61806]: DEBUG nova.objects.instance [None req-4284f75d-699f-4bae-83bf-967ca811ad7c tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Lazy-loading 'flavor' on Instance uuid d0669232-6f33-4b30-97c2-2e4239af1ad0 {{(pid=61806) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1088.619338] env[61806]: DEBUG oslo_concurrency.lockutils [None req-f066ed90-7998-4480-842c-9e7fb5165059 tempest-ServersNegativeTestMultiTenantJSON-775824527 tempest-ServersNegativeTestMultiTenantJSON-775824527-project-member] Lock "cb9e9bac-d06c-4479-88e3-616b0f26ad7e" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 14.025s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1088.629094] env[61806]: DEBUG nova.compute.manager [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] [instance: d9caee4a-a00c-4e50-b01c-99a8f796a09f] Instance has had 0 of 5 cleanup attempts {{(pid=61806) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1088.927535] env[61806]: DEBUG nova.compute.manager [None req-b9e61586-635a-432e-ae51-c0a6d111aaf8 tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] [instance: 29210bcd-0fe1-4ed9-a459-abc8016a4255] Start building block device mappings for instance. {{(pid=61806) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1088.934142] env[61806]: INFO nova.compute.resource_tracker [None req-7f52d370-4c9e-4ec0-a23f-5733d9a2d432 tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] [instance: 06658134-4c6d-4911-9a55-b6805f555c78] Updating resource usage from migration 9dbaf4c6-2a66-44c7-ac3c-b6fb57d4d009 [ 1089.006072] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ad1c545d-3cab-40bb-8a6b-5cd3657ed584 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1089.033073] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b51e3f67-f560-4f40-9eb7-b0d15b168cb0 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1089.056637] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-acfff485-c154-4d91-b842-029e57e4b7ef {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1089.063755] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f6cc8202-5dc4-411d-82e5-8039467fa4d4 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1089.095475] env[61806]: DEBUG nova.network.neutron [None req-b9e61586-635a-432e-ae51-c0a6d111aaf8 tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] [instance: 29210bcd-0fe1-4ed9-a459-abc8016a4255] Successfully created port: 8aad34b7-4309-44b6-b88d-ae5961da22c2 {{(pid=61806) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1089.098615] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0c4918a5-a4dd-417a-97d2-652fe59826ae {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1089.106636] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-40dc786f-4ee3-410c-928c-4984fef2a2c5 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1089.121191] env[61806]: DEBUG nova.compute.provider_tree [None req-7f52d370-4c9e-4ec0-a23f-5733d9a2d432 tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Inventory has not changed in ProviderTree for provider: a2858be1-fd22-4e08-979e-87ad25293407 {{(pid=61806) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1089.135518] env[61806]: DEBUG nova.compute.manager [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] [instance: 75f1aded-d1da-4ddc-be29-f1fb93799364] Instance has had 0 of 5 cleanup attempts {{(pid=61806) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1089.403081] env[61806]: DEBUG oslo_concurrency.lockutils [None req-af8a470a-2cd4-4dfe-86a8-20bd182afa23 tempest-ServersNegativeTestMultiTenantJSON-775824527 tempest-ServersNegativeTestMultiTenantJSON-775824527-project-member] Acquiring lock "cb9e9bac-d06c-4479-88e3-616b0f26ad7e" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1089.403081] env[61806]: DEBUG oslo_concurrency.lockutils [None req-af8a470a-2cd4-4dfe-86a8-20bd182afa23 tempest-ServersNegativeTestMultiTenantJSON-775824527 tempest-ServersNegativeTestMultiTenantJSON-775824527-project-member] Lock "cb9e9bac-d06c-4479-88e3-616b0f26ad7e" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1089.403314] env[61806]: DEBUG oslo_concurrency.lockutils [None req-af8a470a-2cd4-4dfe-86a8-20bd182afa23 tempest-ServersNegativeTestMultiTenantJSON-775824527 tempest-ServersNegativeTestMultiTenantJSON-775824527-project-member] Acquiring lock "cb9e9bac-d06c-4479-88e3-616b0f26ad7e-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1089.403503] env[61806]: DEBUG oslo_concurrency.lockutils [None req-af8a470a-2cd4-4dfe-86a8-20bd182afa23 tempest-ServersNegativeTestMultiTenantJSON-775824527 tempest-ServersNegativeTestMultiTenantJSON-775824527-project-member] Lock "cb9e9bac-d06c-4479-88e3-616b0f26ad7e-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1089.403766] env[61806]: DEBUG oslo_concurrency.lockutils [None req-af8a470a-2cd4-4dfe-86a8-20bd182afa23 tempest-ServersNegativeTestMultiTenantJSON-775824527 tempest-ServersNegativeTestMultiTenantJSON-775824527-project-member] Lock "cb9e9bac-d06c-4479-88e3-616b0f26ad7e-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1089.406412] env[61806]: INFO nova.compute.manager [None req-af8a470a-2cd4-4dfe-86a8-20bd182afa23 tempest-ServersNegativeTestMultiTenantJSON-775824527 tempest-ServersNegativeTestMultiTenantJSON-775824527-project-member] [instance: cb9e9bac-d06c-4479-88e3-616b0f26ad7e] Terminating instance [ 1089.408653] env[61806]: DEBUG nova.compute.manager [None req-af8a470a-2cd4-4dfe-86a8-20bd182afa23 tempest-ServersNegativeTestMultiTenantJSON-775824527 tempest-ServersNegativeTestMultiTenantJSON-775824527-project-member] [instance: cb9e9bac-d06c-4479-88e3-616b0f26ad7e] Start destroying the instance on the hypervisor. {{(pid=61806) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1089.408872] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-af8a470a-2cd4-4dfe-86a8-20bd182afa23 tempest-ServersNegativeTestMultiTenantJSON-775824527 tempest-ServersNegativeTestMultiTenantJSON-775824527-project-member] [instance: cb9e9bac-d06c-4479-88e3-616b0f26ad7e] Destroying instance {{(pid=61806) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1089.409810] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eb21cbbf-b6b8-4b9d-979d-24a63e04073f {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1089.418284] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [None req-af8a470a-2cd4-4dfe-86a8-20bd182afa23 tempest-ServersNegativeTestMultiTenantJSON-775824527 tempest-ServersNegativeTestMultiTenantJSON-775824527-project-member] [instance: cb9e9bac-d06c-4479-88e3-616b0f26ad7e] Powering off the VM {{(pid=61806) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1089.418547] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-2a750f8e-8092-47b5-9ce9-9126a9fb124f {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1089.425173] env[61806]: DEBUG oslo_vmware.api [None req-af8a470a-2cd4-4dfe-86a8-20bd182afa23 tempest-ServersNegativeTestMultiTenantJSON-775824527 tempest-ServersNegativeTestMultiTenantJSON-775824527-project-member] Waiting for the task: (returnval){ [ 1089.425173] env[61806]: value = "task-1295218" [ 1089.425173] env[61806]: _type = "Task" [ 1089.425173] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1089.438421] env[61806]: DEBUG oslo_vmware.api [None req-af8a470a-2cd4-4dfe-86a8-20bd182afa23 tempest-ServersNegativeTestMultiTenantJSON-775824527 tempest-ServersNegativeTestMultiTenantJSON-775824527-project-member] Task: {'id': task-1295218, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1089.547463] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-4284f75d-699f-4bae-83bf-967ca811ad7c tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] [instance: d0669232-6f33-4b30-97c2-2e4239af1ad0] Creating Snapshot of the VM instance {{(pid=61806) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 1089.547463] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-1586d25f-27dc-4d68-ab93-7ed356e224e8 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1089.555033] env[61806]: DEBUG oslo_vmware.api [None req-4284f75d-699f-4bae-83bf-967ca811ad7c tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Waiting for the task: (returnval){ [ 1089.555033] env[61806]: value = "task-1295219" [ 1089.555033] env[61806]: _type = "Task" [ 1089.555033] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1089.562074] env[61806]: DEBUG oslo_vmware.api [None req-4284f75d-699f-4bae-83bf-967ca811ad7c tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Task: {'id': task-1295219, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1089.624754] env[61806]: DEBUG nova.scheduler.client.report [None req-7f52d370-4c9e-4ec0-a23f-5733d9a2d432 tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Inventory has not changed for provider a2858be1-fd22-4e08-979e-87ad25293407 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 138, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61806) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1089.639171] env[61806]: DEBUG nova.compute.manager [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] [instance: ee002e7c-8965-494f-a1af-0b6f3b74eceb] Instance has had 0 of 5 cleanup attempts {{(pid=61806) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1089.820921] env[61806]: DEBUG oslo_concurrency.lockutils [None req-8a03425c-8b40-4033-9e76-3245e7519202 tempest-AttachVolumeNegativeTest-1683033704 tempest-AttachVolumeNegativeTest-1683033704-project-member] Acquiring lock "ce2ad097-4c3f-4b55-b018-ac788aa5b662" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1089.821335] env[61806]: DEBUG oslo_concurrency.lockutils [None req-8a03425c-8b40-4033-9e76-3245e7519202 tempest-AttachVolumeNegativeTest-1683033704 tempest-AttachVolumeNegativeTest-1683033704-project-member] Lock "ce2ad097-4c3f-4b55-b018-ac788aa5b662" acquired by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: waited 0.000s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1089.938396] env[61806]: DEBUG oslo_vmware.api [None req-af8a470a-2cd4-4dfe-86a8-20bd182afa23 tempest-ServersNegativeTestMultiTenantJSON-775824527 tempest-ServersNegativeTestMultiTenantJSON-775824527-project-member] Task: {'id': task-1295218, 'name': PowerOffVM_Task, 'duration_secs': 0.184707} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1089.938855] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [None req-af8a470a-2cd4-4dfe-86a8-20bd182afa23 tempest-ServersNegativeTestMultiTenantJSON-775824527 tempest-ServersNegativeTestMultiTenantJSON-775824527-project-member] [instance: cb9e9bac-d06c-4479-88e3-616b0f26ad7e] Powered off the VM {{(pid=61806) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1089.939424] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-af8a470a-2cd4-4dfe-86a8-20bd182afa23 tempest-ServersNegativeTestMultiTenantJSON-775824527 tempest-ServersNegativeTestMultiTenantJSON-775824527-project-member] [instance: cb9e9bac-d06c-4479-88e3-616b0f26ad7e] Unregistering the VM {{(pid=61806) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1089.940803] env[61806]: DEBUG nova.compute.manager [None req-b9e61586-635a-432e-ae51-c0a6d111aaf8 tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] [instance: 29210bcd-0fe1-4ed9-a459-abc8016a4255] Start spawning the instance on the hypervisor. {{(pid=61806) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1089.942989] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-a17315e2-6708-4fd7-8f51-bdaa944d9d7b {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1089.980546] env[61806]: DEBUG nova.virt.hardware [None req-b9e61586-635a-432e-ae51-c0a6d111aaf8 tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-15T18:05:10Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-15T18:04:52Z,direct_url=,disk_format='vmdk',id=ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='84ba83fce2c349a6988173c5d6fc3b07',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-15T18:04:53Z,virtual_size=,visibility=), allow threads: False {{(pid=61806) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1089.980874] env[61806]: DEBUG nova.virt.hardware [None req-b9e61586-635a-432e-ae51-c0a6d111aaf8 tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] Flavor limits 0:0:0 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1089.981070] env[61806]: DEBUG nova.virt.hardware [None req-b9e61586-635a-432e-ae51-c0a6d111aaf8 tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] Image limits 0:0:0 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1089.981273] env[61806]: DEBUG nova.virt.hardware [None req-b9e61586-635a-432e-ae51-c0a6d111aaf8 tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] Flavor pref 0:0:0 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1089.981427] env[61806]: DEBUG nova.virt.hardware [None req-b9e61586-635a-432e-ae51-c0a6d111aaf8 tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] Image pref 0:0:0 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1089.981601] env[61806]: DEBUG nova.virt.hardware [None req-b9e61586-635a-432e-ae51-c0a6d111aaf8 tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1089.981849] env[61806]: DEBUG nova.virt.hardware [None req-b9e61586-635a-432e-ae51-c0a6d111aaf8 tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61806) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1089.982050] env[61806]: DEBUG nova.virt.hardware [None req-b9e61586-635a-432e-ae51-c0a6d111aaf8 tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61806) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1089.982607] env[61806]: DEBUG nova.virt.hardware [None req-b9e61586-635a-432e-ae51-c0a6d111aaf8 tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] Got 1 possible topologies {{(pid=61806) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1089.982607] env[61806]: DEBUG nova.virt.hardware [None req-b9e61586-635a-432e-ae51-c0a6d111aaf8 tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61806) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1089.982607] env[61806]: DEBUG nova.virt.hardware [None req-b9e61586-635a-432e-ae51-c0a6d111aaf8 tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61806) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1089.983592] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-558e9ba2-d449-4ff6-87be-661d957938e0 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1089.992456] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d6ea8bd2-252e-42c5-a83c-4002eacdcebf {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1090.022368] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-af8a470a-2cd4-4dfe-86a8-20bd182afa23 tempest-ServersNegativeTestMultiTenantJSON-775824527 tempest-ServersNegativeTestMultiTenantJSON-775824527-project-member] [instance: cb9e9bac-d06c-4479-88e3-616b0f26ad7e] Unregistered the VM {{(pid=61806) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1090.022639] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-af8a470a-2cd4-4dfe-86a8-20bd182afa23 tempest-ServersNegativeTestMultiTenantJSON-775824527 tempest-ServersNegativeTestMultiTenantJSON-775824527-project-member] [instance: cb9e9bac-d06c-4479-88e3-616b0f26ad7e] Deleting contents of the VM from datastore datastore1 {{(pid=61806) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1090.022777] env[61806]: DEBUG nova.virt.vmwareapi.ds_util [None req-af8a470a-2cd4-4dfe-86a8-20bd182afa23 tempest-ServersNegativeTestMultiTenantJSON-775824527 tempest-ServersNegativeTestMultiTenantJSON-775824527-project-member] Deleting the datastore file [datastore1] cb9e9bac-d06c-4479-88e3-616b0f26ad7e {{(pid=61806) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1090.023093] env[61806]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-29530a37-7630-4f73-8b30-4eeef87fa015 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1090.030886] env[61806]: DEBUG oslo_vmware.api [None req-af8a470a-2cd4-4dfe-86a8-20bd182afa23 tempest-ServersNegativeTestMultiTenantJSON-775824527 tempest-ServersNegativeTestMultiTenantJSON-775824527-project-member] Waiting for the task: (returnval){ [ 1090.030886] env[61806]: value = "task-1295221" [ 1090.030886] env[61806]: _type = "Task" [ 1090.030886] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1090.042168] env[61806]: DEBUG oslo_vmware.api [None req-af8a470a-2cd4-4dfe-86a8-20bd182afa23 tempest-ServersNegativeTestMultiTenantJSON-775824527 tempest-ServersNegativeTestMultiTenantJSON-775824527-project-member] Task: {'id': task-1295221, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1090.064031] env[61806]: DEBUG oslo_vmware.api [None req-4284f75d-699f-4bae-83bf-967ca811ad7c tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Task: {'id': task-1295219, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1090.129914] env[61806]: DEBUG oslo_concurrency.lockutils [None req-7f52d370-4c9e-4ec0-a23f-5733d9a2d432 tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: held 2.211s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1090.130217] env[61806]: INFO nova.compute.manager [None req-7f52d370-4c9e-4ec0-a23f-5733d9a2d432 tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] [instance: 06658134-4c6d-4911-9a55-b6805f555c78] Migrating [ 1090.146159] env[61806]: DEBUG nova.compute.manager [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] [instance: 6f6ba57a-e2d6-4749-a53a-e263861cb1c0] Instance has had 0 of 5 cleanup attempts {{(pid=61806) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1090.324832] env[61806]: DEBUG nova.compute.utils [None req-8a03425c-8b40-4033-9e76-3245e7519202 tempest-AttachVolumeNegativeTest-1683033704 tempest-AttachVolumeNegativeTest-1683033704-project-member] Using /dev/sd instead of None {{(pid=61806) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1090.539918] env[61806]: DEBUG nova.compute.manager [req-1602c283-146a-47b0-a4f5-5d3e88914a69 req-f0c72ec7-6c82-4709-9d6e-950967a9f81e service nova] [instance: 29210bcd-0fe1-4ed9-a459-abc8016a4255] Received event network-vif-plugged-8aad34b7-4309-44b6-b88d-ae5961da22c2 {{(pid=61806) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1090.540183] env[61806]: DEBUG oslo_concurrency.lockutils [req-1602c283-146a-47b0-a4f5-5d3e88914a69 req-f0c72ec7-6c82-4709-9d6e-950967a9f81e service nova] Acquiring lock "29210bcd-0fe1-4ed9-a459-abc8016a4255-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1090.540404] env[61806]: DEBUG oslo_concurrency.lockutils [req-1602c283-146a-47b0-a4f5-5d3e88914a69 req-f0c72ec7-6c82-4709-9d6e-950967a9f81e service nova] Lock "29210bcd-0fe1-4ed9-a459-abc8016a4255-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1090.540578] env[61806]: DEBUG oslo_concurrency.lockutils [req-1602c283-146a-47b0-a4f5-5d3e88914a69 req-f0c72ec7-6c82-4709-9d6e-950967a9f81e service nova] Lock "29210bcd-0fe1-4ed9-a459-abc8016a4255-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1090.540895] env[61806]: DEBUG nova.compute.manager [req-1602c283-146a-47b0-a4f5-5d3e88914a69 req-f0c72ec7-6c82-4709-9d6e-950967a9f81e service nova] [instance: 29210bcd-0fe1-4ed9-a459-abc8016a4255] No waiting events found dispatching network-vif-plugged-8aad34b7-4309-44b6-b88d-ae5961da22c2 {{(pid=61806) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1090.541220] env[61806]: WARNING nova.compute.manager [req-1602c283-146a-47b0-a4f5-5d3e88914a69 req-f0c72ec7-6c82-4709-9d6e-950967a9f81e service nova] [instance: 29210bcd-0fe1-4ed9-a459-abc8016a4255] Received unexpected event network-vif-plugged-8aad34b7-4309-44b6-b88d-ae5961da22c2 for instance with vm_state building and task_state spawning. [ 1090.545826] env[61806]: DEBUG oslo_vmware.api [None req-af8a470a-2cd4-4dfe-86a8-20bd182afa23 tempest-ServersNegativeTestMultiTenantJSON-775824527 tempest-ServersNegativeTestMultiTenantJSON-775824527-project-member] Task: {'id': task-1295221, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.200698} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1090.546113] env[61806]: DEBUG nova.virt.vmwareapi.ds_util [None req-af8a470a-2cd4-4dfe-86a8-20bd182afa23 tempest-ServersNegativeTestMultiTenantJSON-775824527 tempest-ServersNegativeTestMultiTenantJSON-775824527-project-member] Deleted the datastore file {{(pid=61806) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1090.546306] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-af8a470a-2cd4-4dfe-86a8-20bd182afa23 tempest-ServersNegativeTestMultiTenantJSON-775824527 tempest-ServersNegativeTestMultiTenantJSON-775824527-project-member] [instance: cb9e9bac-d06c-4479-88e3-616b0f26ad7e] Deleted contents of the VM from datastore datastore1 {{(pid=61806) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1090.546489] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-af8a470a-2cd4-4dfe-86a8-20bd182afa23 tempest-ServersNegativeTestMultiTenantJSON-775824527 tempest-ServersNegativeTestMultiTenantJSON-775824527-project-member] [instance: cb9e9bac-d06c-4479-88e3-616b0f26ad7e] Instance destroyed {{(pid=61806) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1090.546828] env[61806]: INFO nova.compute.manager [None req-af8a470a-2cd4-4dfe-86a8-20bd182afa23 tempest-ServersNegativeTestMultiTenantJSON-775824527 tempest-ServersNegativeTestMultiTenantJSON-775824527-project-member] [instance: cb9e9bac-d06c-4479-88e3-616b0f26ad7e] Took 1.14 seconds to destroy the instance on the hypervisor. [ 1090.547315] env[61806]: DEBUG oslo.service.loopingcall [None req-af8a470a-2cd4-4dfe-86a8-20bd182afa23 tempest-ServersNegativeTestMultiTenantJSON-775824527 tempest-ServersNegativeTestMultiTenantJSON-775824527-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61806) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1090.547570] env[61806]: DEBUG nova.compute.manager [-] [instance: cb9e9bac-d06c-4479-88e3-616b0f26ad7e] Deallocating network for instance {{(pid=61806) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1090.547670] env[61806]: DEBUG nova.network.neutron [-] [instance: cb9e9bac-d06c-4479-88e3-616b0f26ad7e] deallocate_for_instance() {{(pid=61806) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1090.566414] env[61806]: DEBUG oslo_vmware.api [None req-4284f75d-699f-4bae-83bf-967ca811ad7c tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Task: {'id': task-1295219, 'name': CreateSnapshot_Task, 'duration_secs': 0.781079} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1090.566773] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-4284f75d-699f-4bae-83bf-967ca811ad7c tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] [instance: d0669232-6f33-4b30-97c2-2e4239af1ad0] Created Snapshot of the VM instance {{(pid=61806) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 1090.567802] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a506128c-c3c7-4012-9436-531eb8089c1b {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1090.648489] env[61806]: DEBUG oslo_concurrency.lockutils [None req-7f52d370-4c9e-4ec0-a23f-5733d9a2d432 tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Acquiring lock "refresh_cache-06658134-4c6d-4911-9a55-b6805f555c78" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1090.648701] env[61806]: DEBUG oslo_concurrency.lockutils [None req-7f52d370-4c9e-4ec0-a23f-5733d9a2d432 tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Acquired lock "refresh_cache-06658134-4c6d-4911-9a55-b6805f555c78" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1090.648884] env[61806]: DEBUG nova.network.neutron [None req-7f52d370-4c9e-4ec0-a23f-5733d9a2d432 tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] [instance: 06658134-4c6d-4911-9a55-b6805f555c78] Building network info cache for instance {{(pid=61806) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1090.650352] env[61806]: DEBUG nova.compute.manager [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] [instance: daff849e-eb4f-411e-92f8-9362e3bd91e3] Instance has had 0 of 5 cleanup attempts {{(pid=61806) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1090.656425] env[61806]: DEBUG nova.network.neutron [None req-b9e61586-635a-432e-ae51-c0a6d111aaf8 tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] [instance: 29210bcd-0fe1-4ed9-a459-abc8016a4255] Successfully updated port: 8aad34b7-4309-44b6-b88d-ae5961da22c2 {{(pid=61806) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1090.829511] env[61806]: DEBUG oslo_concurrency.lockutils [None req-8a03425c-8b40-4033-9e76-3245e7519202 tempest-AttachVolumeNegativeTest-1683033704 tempest-AttachVolumeNegativeTest-1683033704-project-member] Lock "ce2ad097-4c3f-4b55-b018-ac788aa5b662" "released" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: held 1.007s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1090.856383] env[61806]: DEBUG nova.compute.manager [req-617545b4-8a6b-40e2-b9ab-787839bad2f3 req-4c5af4b8-7446-4ee1-8325-9ae179b45ebe service nova] [instance: cb9e9bac-d06c-4479-88e3-616b0f26ad7e] Received event network-vif-deleted-ed5dd9bf-96f6-4cad-8f5d-0ea62887f3b9 {{(pid=61806) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1090.856640] env[61806]: INFO nova.compute.manager [req-617545b4-8a6b-40e2-b9ab-787839bad2f3 req-4c5af4b8-7446-4ee1-8325-9ae179b45ebe service nova] [instance: cb9e9bac-d06c-4479-88e3-616b0f26ad7e] Neutron deleted interface ed5dd9bf-96f6-4cad-8f5d-0ea62887f3b9; detaching it from the instance and deleting it from the info cache [ 1090.856899] env[61806]: DEBUG nova.network.neutron [req-617545b4-8a6b-40e2-b9ab-787839bad2f3 req-4c5af4b8-7446-4ee1-8325-9ae179b45ebe service nova] [instance: cb9e9bac-d06c-4479-88e3-616b0f26ad7e] Updating instance_info_cache with network_info: [] {{(pid=61806) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1091.087045] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-4284f75d-699f-4bae-83bf-967ca811ad7c tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] [instance: d0669232-6f33-4b30-97c2-2e4239af1ad0] Creating linked-clone VM from snapshot {{(pid=61806) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 1091.087355] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-2dbc0b3e-45b4-43fa-9072-43df74717ad4 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1091.096740] env[61806]: DEBUG oslo_vmware.api [None req-4284f75d-699f-4bae-83bf-967ca811ad7c tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Waiting for the task: (returnval){ [ 1091.096740] env[61806]: value = "task-1295222" [ 1091.096740] env[61806]: _type = "Task" [ 1091.096740] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1091.106912] env[61806]: DEBUG oslo_vmware.api [None req-4284f75d-699f-4bae-83bf-967ca811ad7c tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Task: {'id': task-1295222, 'name': CloneVM_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1091.162697] env[61806]: DEBUG oslo_concurrency.lockutils [None req-b9e61586-635a-432e-ae51-c0a6d111aaf8 tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] Acquiring lock "refresh_cache-29210bcd-0fe1-4ed9-a459-abc8016a4255" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1091.163180] env[61806]: DEBUG oslo_concurrency.lockutils [None req-b9e61586-635a-432e-ae51-c0a6d111aaf8 tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] Acquired lock "refresh_cache-29210bcd-0fe1-4ed9-a459-abc8016a4255" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1091.163180] env[61806]: DEBUG nova.network.neutron [None req-b9e61586-635a-432e-ae51-c0a6d111aaf8 tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] [instance: 29210bcd-0fe1-4ed9-a459-abc8016a4255] Building network info cache for instance {{(pid=61806) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1091.164356] env[61806]: DEBUG nova.compute.manager [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] [instance: 1a73d994-5f21-4f80-8f33-5830d860a7ee] Instance has had 0 of 5 cleanup attempts {{(pid=61806) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1091.333179] env[61806]: DEBUG nova.network.neutron [-] [instance: cb9e9bac-d06c-4479-88e3-616b0f26ad7e] Updating instance_info_cache with network_info: [] {{(pid=61806) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1091.359068] env[61806]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-55bfbf16-7e49-4f37-bf9a-b7ced99be946 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1091.372471] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-41641d03-8c16-4596-8527-4f2928b55696 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1091.399308] env[61806]: DEBUG nova.compute.manager [req-617545b4-8a6b-40e2-b9ab-787839bad2f3 req-4c5af4b8-7446-4ee1-8325-9ae179b45ebe service nova] [instance: cb9e9bac-d06c-4479-88e3-616b0f26ad7e] Detach interface failed, port_id=ed5dd9bf-96f6-4cad-8f5d-0ea62887f3b9, reason: Instance cb9e9bac-d06c-4479-88e3-616b0f26ad7e could not be found. {{(pid=61806) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10963}} [ 1091.456062] env[61806]: DEBUG nova.network.neutron [None req-7f52d370-4c9e-4ec0-a23f-5733d9a2d432 tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] [instance: 06658134-4c6d-4911-9a55-b6805f555c78] Updating instance_info_cache with network_info: [{"id": "783dae1f-d1c0-4958-b039-ed57b202b64a", "address": "fa:16:3e:89:24:35", "network": {"id": "746a03b6-c1a6-4d27-98b9-0bee6376634f", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-1971382587-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.245", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c1f2d161c50f4b70987b90e6364c3779", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6a31b224-77d7-48a4-af87-312758611b8e", "external-id": "nsx-vlan-transportzone-761", "segmentation_id": 761, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap783dae1f-d1", "ovs_interfaceid": "783dae1f-d1c0-4958-b039-ed57b202b64a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61806) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1091.606282] env[61806]: DEBUG oslo_vmware.api [None req-4284f75d-699f-4bae-83bf-967ca811ad7c tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Task: {'id': task-1295222, 'name': CloneVM_Task} progress is 94%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1091.669498] env[61806]: DEBUG nova.compute.manager [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] [instance: 5a526856-0dcd-4b3f-8359-13a5c3b9bfc4] Instance has had 0 of 5 cleanup attempts {{(pid=61806) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1091.707461] env[61806]: DEBUG nova.network.neutron [None req-b9e61586-635a-432e-ae51-c0a6d111aaf8 tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] [instance: 29210bcd-0fe1-4ed9-a459-abc8016a4255] Instance cache missing network info. {{(pid=61806) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 1091.839399] env[61806]: INFO nova.compute.manager [-] [instance: cb9e9bac-d06c-4479-88e3-616b0f26ad7e] Took 1.29 seconds to deallocate network for instance. [ 1091.864036] env[61806]: DEBUG nova.network.neutron [None req-b9e61586-635a-432e-ae51-c0a6d111aaf8 tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] [instance: 29210bcd-0fe1-4ed9-a459-abc8016a4255] Updating instance_info_cache with network_info: [{"id": "8aad34b7-4309-44b6-b88d-ae5961da22c2", "address": "fa:16:3e:c0:a7:c5", "network": {"id": "a9d993bc-0817-4ee6-ab6e-1ef9432a110f", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-939246736-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "bbbb09a49203421d807f733616c854df", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ad4c015b-4a8b-46ca-9556-74bad8db9fb3", "external-id": "nsx-vlan-transportzone-246", "segmentation_id": 246, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap8aad34b7-43", "ovs_interfaceid": "8aad34b7-4309-44b6-b88d-ae5961da22c2", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61806) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1091.904645] env[61806]: DEBUG oslo_concurrency.lockutils [None req-8a03425c-8b40-4033-9e76-3245e7519202 tempest-AttachVolumeNegativeTest-1683033704 tempest-AttachVolumeNegativeTest-1683033704-project-member] Acquiring lock "ce2ad097-4c3f-4b55-b018-ac788aa5b662" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1091.904966] env[61806]: DEBUG oslo_concurrency.lockutils [None req-8a03425c-8b40-4033-9e76-3245e7519202 tempest-AttachVolumeNegativeTest-1683033704 tempest-AttachVolumeNegativeTest-1683033704-project-member] Lock "ce2ad097-4c3f-4b55-b018-ac788aa5b662" acquired by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: waited 0.000s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1091.905258] env[61806]: INFO nova.compute.manager [None req-8a03425c-8b40-4033-9e76-3245e7519202 tempest-AttachVolumeNegativeTest-1683033704 tempest-AttachVolumeNegativeTest-1683033704-project-member] [instance: ce2ad097-4c3f-4b55-b018-ac788aa5b662] Attaching volume cb0179b0-05d9-4062-b3db-f8484769394d to /dev/sdb [ 1091.939647] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-021c308d-c1c5-49b5-b82c-306beb99c953 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1091.947725] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-95cc4d3c-e64d-400c-9845-e7eb318ab171 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1091.961568] env[61806]: DEBUG oslo_concurrency.lockutils [None req-7f52d370-4c9e-4ec0-a23f-5733d9a2d432 tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Releasing lock "refresh_cache-06658134-4c6d-4911-9a55-b6805f555c78" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1091.963027] env[61806]: DEBUG nova.virt.block_device [None req-8a03425c-8b40-4033-9e76-3245e7519202 tempest-AttachVolumeNegativeTest-1683033704 tempest-AttachVolumeNegativeTest-1683033704-project-member] [instance: ce2ad097-4c3f-4b55-b018-ac788aa5b662] Updating existing volume attachment record: 8a89a781-77ab-4466-a2c1-06a7c69f96db {{(pid=61806) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 1092.107838] env[61806]: DEBUG oslo_vmware.api [None req-4284f75d-699f-4bae-83bf-967ca811ad7c tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Task: {'id': task-1295222, 'name': CloneVM_Task} progress is 94%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1092.173786] env[61806]: DEBUG nova.compute.manager [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] [instance: 9c20d337-b8f5-4965-b707-0f87c4ab6fc4] Instance has had 0 of 5 cleanup attempts {{(pid=61806) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1092.346170] env[61806]: DEBUG oslo_concurrency.lockutils [None req-af8a470a-2cd4-4dfe-86a8-20bd182afa23 tempest-ServersNegativeTestMultiTenantJSON-775824527 tempest-ServersNegativeTestMultiTenantJSON-775824527-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1092.346420] env[61806]: DEBUG oslo_concurrency.lockutils [None req-af8a470a-2cd4-4dfe-86a8-20bd182afa23 tempest-ServersNegativeTestMultiTenantJSON-775824527 tempest-ServersNegativeTestMultiTenantJSON-775824527-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.001s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1092.346763] env[61806]: DEBUG nova.objects.instance [None req-af8a470a-2cd4-4dfe-86a8-20bd182afa23 tempest-ServersNegativeTestMultiTenantJSON-775824527 tempest-ServersNegativeTestMultiTenantJSON-775824527-project-member] Lazy-loading 'resources' on Instance uuid cb9e9bac-d06c-4479-88e3-616b0f26ad7e {{(pid=61806) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1092.366774] env[61806]: DEBUG oslo_concurrency.lockutils [None req-b9e61586-635a-432e-ae51-c0a6d111aaf8 tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] Releasing lock "refresh_cache-29210bcd-0fe1-4ed9-a459-abc8016a4255" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1092.367129] env[61806]: DEBUG nova.compute.manager [None req-b9e61586-635a-432e-ae51-c0a6d111aaf8 tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] [instance: 29210bcd-0fe1-4ed9-a459-abc8016a4255] Instance network_info: |[{"id": "8aad34b7-4309-44b6-b88d-ae5961da22c2", "address": "fa:16:3e:c0:a7:c5", "network": {"id": "a9d993bc-0817-4ee6-ab6e-1ef9432a110f", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-939246736-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "bbbb09a49203421d807f733616c854df", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ad4c015b-4a8b-46ca-9556-74bad8db9fb3", "external-id": "nsx-vlan-transportzone-246", "segmentation_id": 246, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap8aad34b7-43", "ovs_interfaceid": "8aad34b7-4309-44b6-b88d-ae5961da22c2", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61806) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1092.367595] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-b9e61586-635a-432e-ae51-c0a6d111aaf8 tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] [instance: 29210bcd-0fe1-4ed9-a459-abc8016a4255] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:c0:a7:c5', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'ad4c015b-4a8b-46ca-9556-74bad8db9fb3', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '8aad34b7-4309-44b6-b88d-ae5961da22c2', 'vif_model': 'vmxnet3'}] {{(pid=61806) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1092.375543] env[61806]: DEBUG oslo.service.loopingcall [None req-b9e61586-635a-432e-ae51-c0a6d111aaf8 tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61806) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1092.376173] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 29210bcd-0fe1-4ed9-a459-abc8016a4255] Creating VM on the ESX host {{(pid=61806) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 1092.376472] env[61806]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-99a0e82e-288c-4f39-8efc-07310bfffaad {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1092.397025] env[61806]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1092.397025] env[61806]: value = "task-1295224" [ 1092.397025] env[61806]: _type = "Task" [ 1092.397025] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1092.405580] env[61806]: DEBUG oslo_vmware.api [-] Task: {'id': task-1295224, 'name': CreateVM_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1092.567503] env[61806]: DEBUG nova.compute.manager [req-368efe55-1c5d-42bf-bed9-d55bc7c4e004 req-a9b123fd-a69b-4e6a-b33e-0175c2bf7758 service nova] [instance: 29210bcd-0fe1-4ed9-a459-abc8016a4255] Received event network-changed-8aad34b7-4309-44b6-b88d-ae5961da22c2 {{(pid=61806) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1092.567737] env[61806]: DEBUG nova.compute.manager [req-368efe55-1c5d-42bf-bed9-d55bc7c4e004 req-a9b123fd-a69b-4e6a-b33e-0175c2bf7758 service nova] [instance: 29210bcd-0fe1-4ed9-a459-abc8016a4255] Refreshing instance network info cache due to event network-changed-8aad34b7-4309-44b6-b88d-ae5961da22c2. {{(pid=61806) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 1092.567963] env[61806]: DEBUG oslo_concurrency.lockutils [req-368efe55-1c5d-42bf-bed9-d55bc7c4e004 req-a9b123fd-a69b-4e6a-b33e-0175c2bf7758 service nova] Acquiring lock "refresh_cache-29210bcd-0fe1-4ed9-a459-abc8016a4255" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1092.568163] env[61806]: DEBUG oslo_concurrency.lockutils [req-368efe55-1c5d-42bf-bed9-d55bc7c4e004 req-a9b123fd-a69b-4e6a-b33e-0175c2bf7758 service nova] Acquired lock "refresh_cache-29210bcd-0fe1-4ed9-a459-abc8016a4255" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1092.568372] env[61806]: DEBUG nova.network.neutron [req-368efe55-1c5d-42bf-bed9-d55bc7c4e004 req-a9b123fd-a69b-4e6a-b33e-0175c2bf7758 service nova] [instance: 29210bcd-0fe1-4ed9-a459-abc8016a4255] Refreshing network info cache for port 8aad34b7-4309-44b6-b88d-ae5961da22c2 {{(pid=61806) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1092.607328] env[61806]: DEBUG oslo_vmware.api [None req-4284f75d-699f-4bae-83bf-967ca811ad7c tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Task: {'id': task-1295222, 'name': CloneVM_Task, 'duration_secs': 1.394222} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1092.607618] env[61806]: INFO nova.virt.vmwareapi.vmops [None req-4284f75d-699f-4bae-83bf-967ca811ad7c tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] [instance: d0669232-6f33-4b30-97c2-2e4239af1ad0] Created linked-clone VM from snapshot [ 1092.608435] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-828a27c9-cac5-4211-a01a-e0ec710c3f31 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1092.615955] env[61806]: DEBUG nova.virt.vmwareapi.images [None req-4284f75d-699f-4bae-83bf-967ca811ad7c tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] [instance: d0669232-6f33-4b30-97c2-2e4239af1ad0] Uploading image 7e155d45-9a75-4ca6-97c4-ac1fbb21e996 {{(pid=61806) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:442}} [ 1092.644200] env[61806]: DEBUG oslo_vmware.rw_handles [None req-4284f75d-699f-4bae-83bf-967ca811ad7c tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Creating HttpNfcLease lease for exporting VM: (result){ [ 1092.644200] env[61806]: value = "vm-277763" [ 1092.644200] env[61806]: _type = "VirtualMachine" [ 1092.644200] env[61806]: }. {{(pid=61806) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:478}} [ 1092.644896] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ExportVm with opID=oslo.vmware-61a00b74-19f0-4344-b41b-a3bcddbf64f8 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1092.652066] env[61806]: DEBUG oslo_vmware.rw_handles [None req-4284f75d-699f-4bae-83bf-967ca811ad7c tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Lease: (returnval){ [ 1092.652066] env[61806]: value = "session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]52bcd10e-8b58-8ca2-6235-b7dc6705bd4f" [ 1092.652066] env[61806]: _type = "HttpNfcLease" [ 1092.652066] env[61806]: } obtained for exporting VM: (result){ [ 1092.652066] env[61806]: value = "vm-277763" [ 1092.652066] env[61806]: _type = "VirtualMachine" [ 1092.652066] env[61806]: }. {{(pid=61806) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:481}} [ 1092.652318] env[61806]: DEBUG oslo_vmware.api [None req-4284f75d-699f-4bae-83bf-967ca811ad7c tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Waiting for the lease: (returnval){ [ 1092.652318] env[61806]: value = "session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]52bcd10e-8b58-8ca2-6235-b7dc6705bd4f" [ 1092.652318] env[61806]: _type = "HttpNfcLease" [ 1092.652318] env[61806]: } to be ready. {{(pid=61806) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 1092.658459] env[61806]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1092.658459] env[61806]: value = "session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]52bcd10e-8b58-8ca2-6235-b7dc6705bd4f" [ 1092.658459] env[61806]: _type = "HttpNfcLease" [ 1092.658459] env[61806]: } is initializing. {{(pid=61806) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 1092.677052] env[61806]: DEBUG nova.compute.manager [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] [instance: 59c22c16-3d1c-4e64-bfac-7f8886a1927e] Instance has had 0 of 5 cleanup attempts {{(pid=61806) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1092.908323] env[61806]: DEBUG oslo_vmware.api [-] Task: {'id': task-1295224, 'name': CreateVM_Task, 'duration_secs': 0.328617} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1092.908489] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 29210bcd-0fe1-4ed9-a459-abc8016a4255] Created VM on the ESX host {{(pid=61806) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 1092.909174] env[61806]: DEBUG oslo_concurrency.lockutils [None req-b9e61586-635a-432e-ae51-c0a6d111aaf8 tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1092.909353] env[61806]: DEBUG oslo_concurrency.lockutils [None req-b9e61586-635a-432e-ae51-c0a6d111aaf8 tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] Acquired lock "[datastore1] devstack-image-cache_base/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1092.909692] env[61806]: DEBUG oslo_concurrency.lockutils [None req-b9e61586-635a-432e-ae51-c0a6d111aaf8 tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1092.909950] env[61806]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-9f58c874-5847-4315-a2c2-c1f2ee662fa8 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1092.917625] env[61806]: DEBUG oslo_vmware.api [None req-b9e61586-635a-432e-ae51-c0a6d111aaf8 tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] Waiting for the task: (returnval){ [ 1092.917625] env[61806]: value = "session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]52280166-3d3f-dea9-82ba-a63976b30efa" [ 1092.917625] env[61806]: _type = "Task" [ 1092.917625] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1092.926468] env[61806]: DEBUG oslo_vmware.api [None req-b9e61586-635a-432e-ae51-c0a6d111aaf8 tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] Task: {'id': session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]52280166-3d3f-dea9-82ba-a63976b30efa, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1092.937442] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d2d269a3-a3e2-4e63-8572-adf343c22b1f {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1092.944623] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c0ea655d-7a8f-426b-b5c1-bc2c8a63e361 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1092.977971] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ee25f487-5334-41cd-a794-560250b87464 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1092.985906] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6dd39582-26ce-4547-b17b-40065094b00f {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1092.999119] env[61806]: DEBUG nova.compute.provider_tree [None req-af8a470a-2cd4-4dfe-86a8-20bd182afa23 tempest-ServersNegativeTestMultiTenantJSON-775824527 tempest-ServersNegativeTestMultiTenantJSON-775824527-project-member] Inventory has not changed in ProviderTree for provider: a2858be1-fd22-4e08-979e-87ad25293407 {{(pid=61806) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1093.159567] env[61806]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1093.159567] env[61806]: value = "session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]52bcd10e-8b58-8ca2-6235-b7dc6705bd4f" [ 1093.159567] env[61806]: _type = "HttpNfcLease" [ 1093.159567] env[61806]: } is ready. {{(pid=61806) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 1093.159931] env[61806]: DEBUG oslo_vmware.rw_handles [None req-4284f75d-699f-4bae-83bf-967ca811ad7c tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 1093.159931] env[61806]: value = "session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]52bcd10e-8b58-8ca2-6235-b7dc6705bd4f" [ 1093.159931] env[61806]: _type = "HttpNfcLease" [ 1093.159931] env[61806]: }. {{(pid=61806) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:486}} [ 1093.160667] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-59588403-3983-4579-8151-4dee1ab6fded {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1093.167615] env[61806]: DEBUG oslo_vmware.rw_handles [None req-4284f75d-699f-4bae-83bf-967ca811ad7c tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Found VMDK URL: https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/5255ffd5-142b-628e-f764-280f309ea5c7/disk-0.vmdk from lease info. {{(pid=61806) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 1093.167800] env[61806]: DEBUG oslo_vmware.rw_handles [None req-4284f75d-699f-4bae-83bf-967ca811ad7c tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Opening URL: https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/5255ffd5-142b-628e-f764-280f309ea5c7/disk-0.vmdk for reading. {{(pid=61806) _create_read_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:99}} [ 1093.223918] env[61806]: DEBUG nova.compute.manager [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] [instance: d1802da4-a670-427d-84d7-0fcb4717e18f] Instance has had 0 of 5 cleanup attempts {{(pid=61806) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1093.266290] env[61806]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-9a0e83b4-e2eb-4a99-b540-16279d525e73 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1093.329735] env[61806]: DEBUG nova.network.neutron [req-368efe55-1c5d-42bf-bed9-d55bc7c4e004 req-a9b123fd-a69b-4e6a-b33e-0175c2bf7758 service nova] [instance: 29210bcd-0fe1-4ed9-a459-abc8016a4255] Updated VIF entry in instance network info cache for port 8aad34b7-4309-44b6-b88d-ae5961da22c2. {{(pid=61806) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1093.330249] env[61806]: DEBUG nova.network.neutron [req-368efe55-1c5d-42bf-bed9-d55bc7c4e004 req-a9b123fd-a69b-4e6a-b33e-0175c2bf7758 service nova] [instance: 29210bcd-0fe1-4ed9-a459-abc8016a4255] Updating instance_info_cache with network_info: [{"id": "8aad34b7-4309-44b6-b88d-ae5961da22c2", "address": "fa:16:3e:c0:a7:c5", "network": {"id": "a9d993bc-0817-4ee6-ab6e-1ef9432a110f", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-939246736-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "bbbb09a49203421d807f733616c854df", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ad4c015b-4a8b-46ca-9556-74bad8db9fb3", "external-id": "nsx-vlan-transportzone-246", "segmentation_id": 246, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap8aad34b7-43", "ovs_interfaceid": "8aad34b7-4309-44b6-b88d-ae5961da22c2", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61806) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1093.428896] env[61806]: DEBUG oslo_vmware.api [None req-b9e61586-635a-432e-ae51-c0a6d111aaf8 tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] Task: {'id': session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]52280166-3d3f-dea9-82ba-a63976b30efa, 'name': SearchDatastore_Task, 'duration_secs': 0.014999} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1093.429325] env[61806]: DEBUG oslo_concurrency.lockutils [None req-b9e61586-635a-432e-ae51-c0a6d111aaf8 tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] Releasing lock "[datastore1] devstack-image-cache_base/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1093.429601] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-b9e61586-635a-432e-ae51-c0a6d111aaf8 tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] [instance: 29210bcd-0fe1-4ed9-a459-abc8016a4255] Processing image ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa {{(pid=61806) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1093.429942] env[61806]: DEBUG oslo_concurrency.lockutils [None req-b9e61586-635a-432e-ae51-c0a6d111aaf8 tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa.vmdk" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1093.430145] env[61806]: DEBUG oslo_concurrency.lockutils [None req-b9e61586-635a-432e-ae51-c0a6d111aaf8 tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] Acquired lock "[datastore1] devstack-image-cache_base/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa.vmdk" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1093.430341] env[61806]: DEBUG nova.virt.vmwareapi.ds_util [None req-b9e61586-635a-432e-ae51-c0a6d111aaf8 tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61806) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1093.430608] env[61806]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-97a6aa36-e709-4978-9af7-5933dd95b53f {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1093.439303] env[61806]: DEBUG nova.virt.vmwareapi.ds_util [None req-b9e61586-635a-432e-ae51-c0a6d111aaf8 tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61806) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1093.439566] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-b9e61586-635a-432e-ae51-c0a6d111aaf8 tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61806) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 1093.440391] env[61806]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-4c1ea53a-6bd2-4297-a098-5fbb8c3e22f0 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1093.445774] env[61806]: DEBUG oslo_vmware.api [None req-b9e61586-635a-432e-ae51-c0a6d111aaf8 tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] Waiting for the task: (returnval){ [ 1093.445774] env[61806]: value = "session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]52fae65c-b75d-8279-a4c0-c22e0c04a258" [ 1093.445774] env[61806]: _type = "Task" [ 1093.445774] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1093.453259] env[61806]: DEBUG oslo_vmware.api [None req-b9e61586-635a-432e-ae51-c0a6d111aaf8 tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] Task: {'id': session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]52fae65c-b75d-8279-a4c0-c22e0c04a258, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1093.485101] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6f6cbb65-97c2-4f51-a024-4d00929c41c5 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1093.506040] env[61806]: DEBUG nova.scheduler.client.report [None req-af8a470a-2cd4-4dfe-86a8-20bd182afa23 tempest-ServersNegativeTestMultiTenantJSON-775824527 tempest-ServersNegativeTestMultiTenantJSON-775824527-project-member] Inventory has not changed for provider a2858be1-fd22-4e08-979e-87ad25293407 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 138, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61806) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1093.509563] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-7f52d370-4c9e-4ec0-a23f-5733d9a2d432 tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] [instance: 06658134-4c6d-4911-9a55-b6805f555c78] Updating instance '06658134-4c6d-4911-9a55-b6805f555c78' progress to 0 {{(pid=61806) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 1093.731142] env[61806]: DEBUG nova.compute.manager [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] [instance: efa19aac-3ad7-42d5-a6d8-859050de6e63] Instance has had 0 of 5 cleanup attempts {{(pid=61806) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1093.833211] env[61806]: DEBUG oslo_concurrency.lockutils [req-368efe55-1c5d-42bf-bed9-d55bc7c4e004 req-a9b123fd-a69b-4e6a-b33e-0175c2bf7758 service nova] Releasing lock "refresh_cache-29210bcd-0fe1-4ed9-a459-abc8016a4255" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1093.956278] env[61806]: DEBUG oslo_vmware.api [None req-b9e61586-635a-432e-ae51-c0a6d111aaf8 tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] Task: {'id': session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]52fae65c-b75d-8279-a4c0-c22e0c04a258, 'name': SearchDatastore_Task, 'duration_secs': 0.00916} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1093.957307] env[61806]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b8edbb78-6131-4e37-b34e-e87315ae2030 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1093.962974] env[61806]: DEBUG oslo_vmware.api [None req-b9e61586-635a-432e-ae51-c0a6d111aaf8 tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] Waiting for the task: (returnval){ [ 1093.962974] env[61806]: value = "session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]528797fd-aa34-9367-0de2-c335b29dba6a" [ 1093.962974] env[61806]: _type = "Task" [ 1093.962974] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1093.970993] env[61806]: DEBUG oslo_vmware.api [None req-b9e61586-635a-432e-ae51-c0a6d111aaf8 tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] Task: {'id': session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]528797fd-aa34-9367-0de2-c335b29dba6a, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1094.014089] env[61806]: DEBUG oslo_concurrency.lockutils [None req-af8a470a-2cd4-4dfe-86a8-20bd182afa23 tempest-ServersNegativeTestMultiTenantJSON-775824527 tempest-ServersNegativeTestMultiTenantJSON-775824527-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.668s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1094.017872] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [None req-7f52d370-4c9e-4ec0-a23f-5733d9a2d432 tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] [instance: 06658134-4c6d-4911-9a55-b6805f555c78] Powering off the VM {{(pid=61806) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1094.018757] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-a44d26eb-54d6-4345-ab5f-113eab44b677 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1094.026450] env[61806]: DEBUG oslo_vmware.api [None req-7f52d370-4c9e-4ec0-a23f-5733d9a2d432 tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Waiting for the task: (returnval){ [ 1094.026450] env[61806]: value = "task-1295226" [ 1094.026450] env[61806]: _type = "Task" [ 1094.026450] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1094.035476] env[61806]: DEBUG oslo_vmware.api [None req-7f52d370-4c9e-4ec0-a23f-5733d9a2d432 tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Task: {'id': task-1295226, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1094.036598] env[61806]: INFO nova.scheduler.client.report [None req-af8a470a-2cd4-4dfe-86a8-20bd182afa23 tempest-ServersNegativeTestMultiTenantJSON-775824527 tempest-ServersNegativeTestMultiTenantJSON-775824527-project-member] Deleted allocations for instance cb9e9bac-d06c-4479-88e3-616b0f26ad7e [ 1094.235240] env[61806]: DEBUG nova.compute.manager [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] [instance: 28e0baab-8516-42e3-acc2-9b8eb5192f57] Instance has had 0 of 5 cleanup attempts {{(pid=61806) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1094.475088] env[61806]: DEBUG oslo_vmware.api [None req-b9e61586-635a-432e-ae51-c0a6d111aaf8 tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] Task: {'id': session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]528797fd-aa34-9367-0de2-c335b29dba6a, 'name': SearchDatastore_Task, 'duration_secs': 0.010802} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1094.475496] env[61806]: DEBUG oslo_concurrency.lockutils [None req-b9e61586-635a-432e-ae51-c0a6d111aaf8 tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] Releasing lock "[datastore1] devstack-image-cache_base/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa.vmdk" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1094.475930] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [None req-b9e61586-635a-432e-ae51-c0a6d111aaf8 tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa.vmdk to [datastore1] 29210bcd-0fe1-4ed9-a459-abc8016a4255/29210bcd-0fe1-4ed9-a459-abc8016a4255.vmdk {{(pid=61806) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 1094.477015] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-b2bd50e7-36b1-4bee-8c94-a0895339f238 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1094.482586] env[61806]: DEBUG oslo_vmware.api [None req-b9e61586-635a-432e-ae51-c0a6d111aaf8 tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] Waiting for the task: (returnval){ [ 1094.482586] env[61806]: value = "task-1295228" [ 1094.482586] env[61806]: _type = "Task" [ 1094.482586] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1094.490690] env[61806]: DEBUG oslo_vmware.api [None req-b9e61586-635a-432e-ae51-c0a6d111aaf8 tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] Task: {'id': task-1295228, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1094.536278] env[61806]: DEBUG oslo_vmware.api [None req-7f52d370-4c9e-4ec0-a23f-5733d9a2d432 tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Task: {'id': task-1295226, 'name': PowerOffVM_Task, 'duration_secs': 0.182229} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1094.536864] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [None req-7f52d370-4c9e-4ec0-a23f-5733d9a2d432 tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] [instance: 06658134-4c6d-4911-9a55-b6805f555c78] Powered off the VM {{(pid=61806) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1094.539120] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-7f52d370-4c9e-4ec0-a23f-5733d9a2d432 tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] [instance: 06658134-4c6d-4911-9a55-b6805f555c78] Updating instance '06658134-4c6d-4911-9a55-b6805f555c78' progress to 17 {{(pid=61806) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 1094.544902] env[61806]: DEBUG oslo_concurrency.lockutils [None req-af8a470a-2cd4-4dfe-86a8-20bd182afa23 tempest-ServersNegativeTestMultiTenantJSON-775824527 tempest-ServersNegativeTestMultiTenantJSON-775824527-project-member] Lock "cb9e9bac-d06c-4479-88e3-616b0f26ad7e" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.142s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1094.738901] env[61806]: DEBUG nova.compute.manager [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] [instance: be19fc44-ecbe-489c-9b6e-be8957bb1dd0] Instance has had 0 of 5 cleanup attempts {{(pid=61806) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1094.995105] env[61806]: DEBUG oslo_vmware.api [None req-b9e61586-635a-432e-ae51-c0a6d111aaf8 tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] Task: {'id': task-1295228, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1095.043431] env[61806]: DEBUG nova.virt.hardware [None req-7f52d370-4c9e-4ec0-a23f-5733d9a2d432 tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-15T18:05:12Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='84',id=12,is_public=True,memory_mb=256,name='m1.micro',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=61806) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1095.043718] env[61806]: DEBUG nova.virt.hardware [None req-7f52d370-4c9e-4ec0-a23f-5733d9a2d432 tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Flavor limits 0:0:0 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1095.043917] env[61806]: DEBUG nova.virt.hardware [None req-7f52d370-4c9e-4ec0-a23f-5733d9a2d432 tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Image limits 0:0:0 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1095.044210] env[61806]: DEBUG nova.virt.hardware [None req-7f52d370-4c9e-4ec0-a23f-5733d9a2d432 tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Flavor pref 0:0:0 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1095.044391] env[61806]: DEBUG nova.virt.hardware [None req-7f52d370-4c9e-4ec0-a23f-5733d9a2d432 tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Image pref 0:0:0 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1095.044562] env[61806]: DEBUG nova.virt.hardware [None req-7f52d370-4c9e-4ec0-a23f-5733d9a2d432 tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1095.044809] env[61806]: DEBUG nova.virt.hardware [None req-7f52d370-4c9e-4ec0-a23f-5733d9a2d432 tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61806) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1095.044993] env[61806]: DEBUG nova.virt.hardware [None req-7f52d370-4c9e-4ec0-a23f-5733d9a2d432 tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61806) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1095.045824] env[61806]: DEBUG nova.virt.hardware [None req-7f52d370-4c9e-4ec0-a23f-5733d9a2d432 tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Got 1 possible topologies {{(pid=61806) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1095.046066] env[61806]: DEBUG nova.virt.hardware [None req-7f52d370-4c9e-4ec0-a23f-5733d9a2d432 tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61806) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1095.046289] env[61806]: DEBUG nova.virt.hardware [None req-7f52d370-4c9e-4ec0-a23f-5733d9a2d432 tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61806) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1095.053542] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-736e7074-2cc3-423c-a8fa-3dd940f585ba {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1095.070869] env[61806]: DEBUG oslo_vmware.api [None req-7f52d370-4c9e-4ec0-a23f-5733d9a2d432 tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Waiting for the task: (returnval){ [ 1095.070869] env[61806]: value = "task-1295229" [ 1095.070869] env[61806]: _type = "Task" [ 1095.070869] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1095.080666] env[61806]: DEBUG oslo_vmware.api [None req-7f52d370-4c9e-4ec0-a23f-5733d9a2d432 tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Task: {'id': task-1295229, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1095.243049] env[61806]: DEBUG nova.compute.manager [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] [instance: 7f4907d0-178d-452a-8149-030becde8779] Instance has had 0 of 5 cleanup attempts {{(pid=61806) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1095.496625] env[61806]: DEBUG oslo_vmware.api [None req-b9e61586-635a-432e-ae51-c0a6d111aaf8 tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] Task: {'id': task-1295228, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.608366} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1095.497036] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [None req-b9e61586-635a-432e-ae51-c0a6d111aaf8 tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa.vmdk to [datastore1] 29210bcd-0fe1-4ed9-a459-abc8016a4255/29210bcd-0fe1-4ed9-a459-abc8016a4255.vmdk {{(pid=61806) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 1095.497149] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-b9e61586-635a-432e-ae51-c0a6d111aaf8 tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] [instance: 29210bcd-0fe1-4ed9-a459-abc8016a4255] Extending root virtual disk to 1048576 {{(pid=61806) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1095.497698] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-83d0bcec-69d0-459f-b14c-e4628c8a7ef4 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1095.505646] env[61806]: DEBUG oslo_vmware.api [None req-b9e61586-635a-432e-ae51-c0a6d111aaf8 tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] Waiting for the task: (returnval){ [ 1095.505646] env[61806]: value = "task-1295230" [ 1095.505646] env[61806]: _type = "Task" [ 1095.505646] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1095.514787] env[61806]: DEBUG oslo_vmware.api [None req-b9e61586-635a-432e-ae51-c0a6d111aaf8 tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] Task: {'id': task-1295230, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1095.582037] env[61806]: DEBUG oslo_vmware.api [None req-7f52d370-4c9e-4ec0-a23f-5733d9a2d432 tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Task: {'id': task-1295229, 'name': ReconfigVM_Task, 'duration_secs': 0.217553} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1095.582371] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-7f52d370-4c9e-4ec0-a23f-5733d9a2d432 tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] [instance: 06658134-4c6d-4911-9a55-b6805f555c78] Updating instance '06658134-4c6d-4911-9a55-b6805f555c78' progress to 33 {{(pid=61806) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 1095.746902] env[61806]: DEBUG nova.compute.manager [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] [instance: c8033551-6591-4e37-ae78-4efe7145b10b] Instance has had 0 of 5 cleanup attempts {{(pid=61806) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1096.015460] env[61806]: DEBUG oslo_vmware.api [None req-b9e61586-635a-432e-ae51-c0a6d111aaf8 tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] Task: {'id': task-1295230, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.072891} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1096.015821] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-b9e61586-635a-432e-ae51-c0a6d111aaf8 tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] [instance: 29210bcd-0fe1-4ed9-a459-abc8016a4255] Extended root virtual disk {{(pid=61806) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1096.016764] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-81b28484-430a-492e-ac18-62b2d46bdc09 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1096.040820] env[61806]: DEBUG nova.virt.vmwareapi.volumeops [None req-b9e61586-635a-432e-ae51-c0a6d111aaf8 tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] [instance: 29210bcd-0fe1-4ed9-a459-abc8016a4255] Reconfiguring VM instance instance-0000006a to attach disk [datastore1] 29210bcd-0fe1-4ed9-a459-abc8016a4255/29210bcd-0fe1-4ed9-a459-abc8016a4255.vmdk or device None with type sparse {{(pid=61806) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1096.042014] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-7c246e36-87c1-4a94-9cdd-549b66db2025 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1096.062767] env[61806]: DEBUG oslo_vmware.api [None req-b9e61586-635a-432e-ae51-c0a6d111aaf8 tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] Waiting for the task: (returnval){ [ 1096.062767] env[61806]: value = "task-1295231" [ 1096.062767] env[61806]: _type = "Task" [ 1096.062767] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1096.071722] env[61806]: DEBUG oslo_vmware.api [None req-b9e61586-635a-432e-ae51-c0a6d111aaf8 tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] Task: {'id': task-1295231, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1096.093378] env[61806]: DEBUG nova.virt.hardware [None req-7f52d370-4c9e-4ec0-a23f-5733d9a2d432 tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-15T18:05:10Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=61806) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1096.093718] env[61806]: DEBUG nova.virt.hardware [None req-7f52d370-4c9e-4ec0-a23f-5733d9a2d432 tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Flavor limits 0:0:0 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1096.093938] env[61806]: DEBUG nova.virt.hardware [None req-7f52d370-4c9e-4ec0-a23f-5733d9a2d432 tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Image limits 0:0:0 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1096.094155] env[61806]: DEBUG nova.virt.hardware [None req-7f52d370-4c9e-4ec0-a23f-5733d9a2d432 tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Flavor pref 0:0:0 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1096.095163] env[61806]: DEBUG nova.virt.hardware [None req-7f52d370-4c9e-4ec0-a23f-5733d9a2d432 tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Image pref 0:0:0 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1096.095163] env[61806]: DEBUG nova.virt.hardware [None req-7f52d370-4c9e-4ec0-a23f-5733d9a2d432 tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1096.095163] env[61806]: DEBUG nova.virt.hardware [None req-7f52d370-4c9e-4ec0-a23f-5733d9a2d432 tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61806) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1096.095163] env[61806]: DEBUG nova.virt.hardware [None req-7f52d370-4c9e-4ec0-a23f-5733d9a2d432 tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61806) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1096.095163] env[61806]: DEBUG nova.virt.hardware [None req-7f52d370-4c9e-4ec0-a23f-5733d9a2d432 tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Got 1 possible topologies {{(pid=61806) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1096.095359] env[61806]: DEBUG nova.virt.hardware [None req-7f52d370-4c9e-4ec0-a23f-5733d9a2d432 tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61806) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1096.096019] env[61806]: DEBUG nova.virt.hardware [None req-7f52d370-4c9e-4ec0-a23f-5733d9a2d432 tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61806) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1096.101090] env[61806]: DEBUG nova.virt.vmwareapi.volumeops [None req-7f52d370-4c9e-4ec0-a23f-5733d9a2d432 tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] [instance: 06658134-4c6d-4911-9a55-b6805f555c78] Reconfiguring VM instance instance-00000066 to detach disk 2000 {{(pid=61806) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1096.101406] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-b83e6e3f-7932-4f01-b05f-1a19c2d084d9 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1096.121798] env[61806]: DEBUG oslo_vmware.api [None req-7f52d370-4c9e-4ec0-a23f-5733d9a2d432 tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Waiting for the task: (returnval){ [ 1096.121798] env[61806]: value = "task-1295232" [ 1096.121798] env[61806]: _type = "Task" [ 1096.121798] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1096.130567] env[61806]: DEBUG oslo_vmware.api [None req-7f52d370-4c9e-4ec0-a23f-5733d9a2d432 tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Task: {'id': task-1295232, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1096.250557] env[61806]: DEBUG nova.compute.manager [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] [instance: 117139cd-4ed4-4bea-8ad3-f4ac32c692e4] Instance has had 0 of 5 cleanup attempts {{(pid=61806) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1096.511354] env[61806]: DEBUG nova.virt.vmwareapi.volumeops [None req-8a03425c-8b40-4033-9e76-3245e7519202 tempest-AttachVolumeNegativeTest-1683033704 tempest-AttachVolumeNegativeTest-1683033704-project-member] [instance: ce2ad097-4c3f-4b55-b018-ac788aa5b662] Volume attach. Driver type: vmdk {{(pid=61806) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 1096.511849] env[61806]: DEBUG nova.virt.vmwareapi.volumeops [None req-8a03425c-8b40-4033-9e76-3245e7519202 tempest-AttachVolumeNegativeTest-1683033704 tempest-AttachVolumeNegativeTest-1683033704-project-member] [instance: ce2ad097-4c3f-4b55-b018-ac788aa5b662] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-277764', 'volume_id': 'cb0179b0-05d9-4062-b3db-f8484769394d', 'name': 'volume-cb0179b0-05d9-4062-b3db-f8484769394d', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'ce2ad097-4c3f-4b55-b018-ac788aa5b662', 'attached_at': '', 'detached_at': '', 'volume_id': 'cb0179b0-05d9-4062-b3db-f8484769394d', 'serial': 'cb0179b0-05d9-4062-b3db-f8484769394d'} {{(pid=61806) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 1096.513174] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-79095ea3-01ac-49d2-a0a5-72200b94feec {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1096.532484] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9dd162ba-9ff6-4120-815d-ff0407d60e5a {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1096.561317] env[61806]: DEBUG nova.virt.vmwareapi.volumeops [None req-8a03425c-8b40-4033-9e76-3245e7519202 tempest-AttachVolumeNegativeTest-1683033704 tempest-AttachVolumeNegativeTest-1683033704-project-member] [instance: ce2ad097-4c3f-4b55-b018-ac788aa5b662] Reconfiguring VM instance instance-00000068 to attach disk [datastore1] volume-cb0179b0-05d9-4062-b3db-f8484769394d/volume-cb0179b0-05d9-4062-b3db-f8484769394d.vmdk or device None with type thin {{(pid=61806) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1096.561844] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-053e02c5-4663-481c-8f5c-061d61f8554b {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1096.591677] env[61806]: DEBUG oslo_vmware.api [None req-b9e61586-635a-432e-ae51-c0a6d111aaf8 tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] Task: {'id': task-1295231, 'name': ReconfigVM_Task, 'duration_secs': 0.349915} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1096.593170] env[61806]: DEBUG nova.virt.vmwareapi.volumeops [None req-b9e61586-635a-432e-ae51-c0a6d111aaf8 tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] [instance: 29210bcd-0fe1-4ed9-a459-abc8016a4255] Reconfigured VM instance instance-0000006a to attach disk [datastore1] 29210bcd-0fe1-4ed9-a459-abc8016a4255/29210bcd-0fe1-4ed9-a459-abc8016a4255.vmdk or device None with type sparse {{(pid=61806) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1096.593907] env[61806]: DEBUG oslo_vmware.api [None req-8a03425c-8b40-4033-9e76-3245e7519202 tempest-AttachVolumeNegativeTest-1683033704 tempest-AttachVolumeNegativeTest-1683033704-project-member] Waiting for the task: (returnval){ [ 1096.593907] env[61806]: value = "task-1295233" [ 1096.593907] env[61806]: _type = "Task" [ 1096.593907] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1096.594125] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-c63c233c-3c39-4072-b127-8ec2ac874233 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1096.605123] env[61806]: DEBUG oslo_vmware.api [None req-8a03425c-8b40-4033-9e76-3245e7519202 tempest-AttachVolumeNegativeTest-1683033704 tempest-AttachVolumeNegativeTest-1683033704-project-member] Task: {'id': task-1295233, 'name': ReconfigVM_Task} progress is 10%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1096.606623] env[61806]: DEBUG oslo_vmware.api [None req-b9e61586-635a-432e-ae51-c0a6d111aaf8 tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] Waiting for the task: (returnval){ [ 1096.606623] env[61806]: value = "task-1295234" [ 1096.606623] env[61806]: _type = "Task" [ 1096.606623] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1096.615316] env[61806]: DEBUG oslo_vmware.api [None req-b9e61586-635a-432e-ae51-c0a6d111aaf8 tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] Task: {'id': task-1295234, 'name': Rename_Task} progress is 5%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1096.631923] env[61806]: DEBUG oslo_vmware.api [None req-7f52d370-4c9e-4ec0-a23f-5733d9a2d432 tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Task: {'id': task-1295232, 'name': ReconfigVM_Task, 'duration_secs': 0.190536} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1096.632333] env[61806]: DEBUG nova.virt.vmwareapi.volumeops [None req-7f52d370-4c9e-4ec0-a23f-5733d9a2d432 tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] [instance: 06658134-4c6d-4911-9a55-b6805f555c78] Reconfigured VM instance instance-00000066 to detach disk 2000 {{(pid=61806) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1096.633266] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eb54cd3d-0f39-466b-9a8a-3273ffbd8a21 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1096.656877] env[61806]: DEBUG nova.virt.vmwareapi.volumeops [None req-7f52d370-4c9e-4ec0-a23f-5733d9a2d432 tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] [instance: 06658134-4c6d-4911-9a55-b6805f555c78] Reconfiguring VM instance instance-00000066 to attach disk [datastore2] 06658134-4c6d-4911-9a55-b6805f555c78/06658134-4c6d-4911-9a55-b6805f555c78.vmdk or device None with type thin {{(pid=61806) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1096.657283] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-787441a2-fe69-4cae-93fd-940134fc8445 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1096.677970] env[61806]: DEBUG oslo_vmware.api [None req-7f52d370-4c9e-4ec0-a23f-5733d9a2d432 tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Waiting for the task: (returnval){ [ 1096.677970] env[61806]: value = "task-1295235" [ 1096.677970] env[61806]: _type = "Task" [ 1096.677970] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1096.686404] env[61806]: DEBUG oslo_vmware.api [None req-7f52d370-4c9e-4ec0-a23f-5733d9a2d432 tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Task: {'id': task-1295235, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1096.754259] env[61806]: DEBUG nova.compute.manager [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] [instance: ef502a7d-7dc0-4319-8ce3-6a8701288237] Instance has had 0 of 5 cleanup attempts {{(pid=61806) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1097.107141] env[61806]: DEBUG oslo_vmware.api [None req-8a03425c-8b40-4033-9e76-3245e7519202 tempest-AttachVolumeNegativeTest-1683033704 tempest-AttachVolumeNegativeTest-1683033704-project-member] Task: {'id': task-1295233, 'name': ReconfigVM_Task, 'duration_secs': 0.435541} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1097.107708] env[61806]: DEBUG nova.virt.vmwareapi.volumeops [None req-8a03425c-8b40-4033-9e76-3245e7519202 tempest-AttachVolumeNegativeTest-1683033704 tempest-AttachVolumeNegativeTest-1683033704-project-member] [instance: ce2ad097-4c3f-4b55-b018-ac788aa5b662] Reconfigured VM instance instance-00000068 to attach disk [datastore1] volume-cb0179b0-05d9-4062-b3db-f8484769394d/volume-cb0179b0-05d9-4062-b3db-f8484769394d.vmdk or device None with type thin {{(pid=61806) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1097.115495] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-0f7c77ec-612b-44da-bcc5-ea91a0a8e190 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1097.133108] env[61806]: DEBUG oslo_vmware.api [None req-b9e61586-635a-432e-ae51-c0a6d111aaf8 tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] Task: {'id': task-1295234, 'name': Rename_Task, 'duration_secs': 0.167967} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1097.133108] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [None req-b9e61586-635a-432e-ae51-c0a6d111aaf8 tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] [instance: 29210bcd-0fe1-4ed9-a459-abc8016a4255] Powering on the VM {{(pid=61806) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1097.133408] env[61806]: DEBUG oslo_vmware.api [None req-8a03425c-8b40-4033-9e76-3245e7519202 tempest-AttachVolumeNegativeTest-1683033704 tempest-AttachVolumeNegativeTest-1683033704-project-member] Waiting for the task: (returnval){ [ 1097.133408] env[61806]: value = "task-1295236" [ 1097.133408] env[61806]: _type = "Task" [ 1097.133408] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1097.133627] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-1b43e8b2-7bec-4254-ad4c-e8617da119d1 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1097.144116] env[61806]: DEBUG oslo_vmware.api [None req-8a03425c-8b40-4033-9e76-3245e7519202 tempest-AttachVolumeNegativeTest-1683033704 tempest-AttachVolumeNegativeTest-1683033704-project-member] Task: {'id': task-1295236, 'name': ReconfigVM_Task} progress is 10%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1097.145520] env[61806]: DEBUG oslo_vmware.api [None req-b9e61586-635a-432e-ae51-c0a6d111aaf8 tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] Waiting for the task: (returnval){ [ 1097.145520] env[61806]: value = "task-1295237" [ 1097.145520] env[61806]: _type = "Task" [ 1097.145520] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1097.155925] env[61806]: DEBUG oslo_vmware.api [None req-b9e61586-635a-432e-ae51-c0a6d111aaf8 tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] Task: {'id': task-1295237, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1097.186796] env[61806]: DEBUG oslo_vmware.api [None req-7f52d370-4c9e-4ec0-a23f-5733d9a2d432 tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Task: {'id': task-1295235, 'name': ReconfigVM_Task, 'duration_secs': 0.351451} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1097.187240] env[61806]: DEBUG nova.virt.vmwareapi.volumeops [None req-7f52d370-4c9e-4ec0-a23f-5733d9a2d432 tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] [instance: 06658134-4c6d-4911-9a55-b6805f555c78] Reconfigured VM instance instance-00000066 to attach disk [datastore2] 06658134-4c6d-4911-9a55-b6805f555c78/06658134-4c6d-4911-9a55-b6805f555c78.vmdk or device None with type thin {{(pid=61806) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1097.187564] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-7f52d370-4c9e-4ec0-a23f-5733d9a2d432 tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] [instance: 06658134-4c6d-4911-9a55-b6805f555c78] Updating instance '06658134-4c6d-4911-9a55-b6805f555c78' progress to 50 {{(pid=61806) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 1097.258198] env[61806]: DEBUG nova.compute.manager [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] [instance: 25218e83-c1ab-4b97-ade1-7c32b6f99305] Instance has had 0 of 5 cleanup attempts {{(pid=61806) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1097.646120] env[61806]: DEBUG oslo_vmware.api [None req-8a03425c-8b40-4033-9e76-3245e7519202 tempest-AttachVolumeNegativeTest-1683033704 tempest-AttachVolumeNegativeTest-1683033704-project-member] Task: {'id': task-1295236, 'name': ReconfigVM_Task, 'duration_secs': 0.154532} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1097.646120] env[61806]: DEBUG nova.virt.vmwareapi.volumeops [None req-8a03425c-8b40-4033-9e76-3245e7519202 tempest-AttachVolumeNegativeTest-1683033704 tempest-AttachVolumeNegativeTest-1683033704-project-member] [instance: ce2ad097-4c3f-4b55-b018-ac788aa5b662] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-277764', 'volume_id': 'cb0179b0-05d9-4062-b3db-f8484769394d', 'name': 'volume-cb0179b0-05d9-4062-b3db-f8484769394d', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'ce2ad097-4c3f-4b55-b018-ac788aa5b662', 'attached_at': '', 'detached_at': '', 'volume_id': 'cb0179b0-05d9-4062-b3db-f8484769394d', 'serial': 'cb0179b0-05d9-4062-b3db-f8484769394d'} {{(pid=61806) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 1097.656311] env[61806]: DEBUG oslo_vmware.api [None req-b9e61586-635a-432e-ae51-c0a6d111aaf8 tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] Task: {'id': task-1295237, 'name': PowerOnVM_Task, 'duration_secs': 0.494194} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1097.656995] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [None req-b9e61586-635a-432e-ae51-c0a6d111aaf8 tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] [instance: 29210bcd-0fe1-4ed9-a459-abc8016a4255] Powered on the VM {{(pid=61806) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1097.656995] env[61806]: INFO nova.compute.manager [None req-b9e61586-635a-432e-ae51-c0a6d111aaf8 tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] [instance: 29210bcd-0fe1-4ed9-a459-abc8016a4255] Took 7.72 seconds to spawn the instance on the hypervisor. [ 1097.657218] env[61806]: DEBUG nova.compute.manager [None req-b9e61586-635a-432e-ae51-c0a6d111aaf8 tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] [instance: 29210bcd-0fe1-4ed9-a459-abc8016a4255] Checking state {{(pid=61806) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1097.658034] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7c3b4be8-7773-472c-8e75-f4902a7de00d {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1097.694867] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e095512e-b68d-40f7-b9b7-f35b81d96ed1 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1097.715671] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-edbbbd14-6023-4d37-9793-7dd3c9979c75 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1097.734972] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-7f52d370-4c9e-4ec0-a23f-5733d9a2d432 tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] [instance: 06658134-4c6d-4911-9a55-b6805f555c78] Updating instance '06658134-4c6d-4911-9a55-b6805f555c78' progress to 67 {{(pid=61806) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 1097.763075] env[61806]: DEBUG nova.compute.manager [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] [instance: 67d1bdd8-cf8f-4fb5-bd3c-42008e4d60e3] Instance has had 0 of 5 cleanup attempts {{(pid=61806) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1098.184081] env[61806]: INFO nova.compute.manager [None req-b9e61586-635a-432e-ae51-c0a6d111aaf8 tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] [instance: 29210bcd-0fe1-4ed9-a459-abc8016a4255] Took 12.45 seconds to build instance. [ 1098.265599] env[61806]: DEBUG nova.compute.manager [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] [instance: b41531f2-f28c-4d82-9682-0b557bbaa491] Instance has had 0 of 5 cleanup attempts {{(pid=61806) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1098.286576] env[61806]: DEBUG nova.network.neutron [None req-7f52d370-4c9e-4ec0-a23f-5733d9a2d432 tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] [instance: 06658134-4c6d-4911-9a55-b6805f555c78] Port 783dae1f-d1c0-4958-b039-ed57b202b64a binding to destination host cpu-1 is already ACTIVE {{(pid=61806) migrate_instance_start /opt/stack/nova/nova/network/neutron.py:3171}} [ 1098.686854] env[61806]: DEBUG oslo_concurrency.lockutils [None req-b9e61586-635a-432e-ae51-c0a6d111aaf8 tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] Lock "29210bcd-0fe1-4ed9-a459-abc8016a4255" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 13.957s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1098.690009] env[61806]: DEBUG nova.objects.instance [None req-8a03425c-8b40-4033-9e76-3245e7519202 tempest-AttachVolumeNegativeTest-1683033704 tempest-AttachVolumeNegativeTest-1683033704-project-member] Lazy-loading 'flavor' on Instance uuid ce2ad097-4c3f-4b55-b018-ac788aa5b662 {{(pid=61806) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1098.769521] env[61806]: DEBUG nova.compute.manager [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] [instance: 43dcce0d-4fdb-4ee1-8309-92e92d2331a9] Instance has had 0 of 5 cleanup attempts {{(pid=61806) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1099.195609] env[61806]: DEBUG oslo_concurrency.lockutils [None req-8a03425c-8b40-4033-9e76-3245e7519202 tempest-AttachVolumeNegativeTest-1683033704 tempest-AttachVolumeNegativeTest-1683033704-project-member] Lock "ce2ad097-4c3f-4b55-b018-ac788aa5b662" "released" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: held 7.290s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1099.272736] env[61806]: DEBUG nova.compute.manager [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] [instance: a842df44-d8a9-4376-b9fc-5ca19a68a4b7] Instance has had 0 of 5 cleanup attempts {{(pid=61806) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1099.310866] env[61806]: DEBUG oslo_concurrency.lockutils [None req-7f52d370-4c9e-4ec0-a23f-5733d9a2d432 tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Acquiring lock "06658134-4c6d-4911-9a55-b6805f555c78-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1099.310866] env[61806]: DEBUG oslo_concurrency.lockutils [None req-7f52d370-4c9e-4ec0-a23f-5733d9a2d432 tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Lock "06658134-4c6d-4911-9a55-b6805f555c78-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1099.310866] env[61806]: DEBUG oslo_concurrency.lockutils [None req-7f52d370-4c9e-4ec0-a23f-5733d9a2d432 tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Lock "06658134-4c6d-4911-9a55-b6805f555c78-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1099.397987] env[61806]: DEBUG oslo_concurrency.lockutils [None req-f685fa1b-ce5d-413e-89f4-9a33f5f03e13 tempest-AttachVolumeNegativeTest-1683033704 tempest-AttachVolumeNegativeTest-1683033704-project-member] Acquiring lock "ce2ad097-4c3f-4b55-b018-ac788aa5b662" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1099.398321] env[61806]: DEBUG oslo_concurrency.lockutils [None req-f685fa1b-ce5d-413e-89f4-9a33f5f03e13 tempest-AttachVolumeNegativeTest-1683033704 tempest-AttachVolumeNegativeTest-1683033704-project-member] Lock "ce2ad097-4c3f-4b55-b018-ac788aa5b662" acquired by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: waited 0.001s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1099.524337] env[61806]: DEBUG nova.compute.manager [req-704b46e2-e103-439d-b6c0-f287ca546de9 req-dbd7de39-ac8c-4c5f-8488-05497ee3296e service nova] [instance: 29210bcd-0fe1-4ed9-a459-abc8016a4255] Received event network-changed-8aad34b7-4309-44b6-b88d-ae5961da22c2 {{(pid=61806) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1099.524544] env[61806]: DEBUG nova.compute.manager [req-704b46e2-e103-439d-b6c0-f287ca546de9 req-dbd7de39-ac8c-4c5f-8488-05497ee3296e service nova] [instance: 29210bcd-0fe1-4ed9-a459-abc8016a4255] Refreshing instance network info cache due to event network-changed-8aad34b7-4309-44b6-b88d-ae5961da22c2. {{(pid=61806) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 1099.524807] env[61806]: DEBUG oslo_concurrency.lockutils [req-704b46e2-e103-439d-b6c0-f287ca546de9 req-dbd7de39-ac8c-4c5f-8488-05497ee3296e service nova] Acquiring lock "refresh_cache-29210bcd-0fe1-4ed9-a459-abc8016a4255" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1099.525021] env[61806]: DEBUG oslo_concurrency.lockutils [req-704b46e2-e103-439d-b6c0-f287ca546de9 req-dbd7de39-ac8c-4c5f-8488-05497ee3296e service nova] Acquired lock "refresh_cache-29210bcd-0fe1-4ed9-a459-abc8016a4255" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1099.525229] env[61806]: DEBUG nova.network.neutron [req-704b46e2-e103-439d-b6c0-f287ca546de9 req-dbd7de39-ac8c-4c5f-8488-05497ee3296e service nova] [instance: 29210bcd-0fe1-4ed9-a459-abc8016a4255] Refreshing network info cache for port 8aad34b7-4309-44b6-b88d-ae5961da22c2 {{(pid=61806) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1099.777352] env[61806]: DEBUG nova.compute.manager [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] [instance: 4373b735-31cf-4b53-b655-38555cf212a5] Instance has had 0 of 5 cleanup attempts {{(pid=61806) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1099.901391] env[61806]: INFO nova.compute.manager [None req-f685fa1b-ce5d-413e-89f4-9a33f5f03e13 tempest-AttachVolumeNegativeTest-1683033704 tempest-AttachVolumeNegativeTest-1683033704-project-member] [instance: ce2ad097-4c3f-4b55-b018-ac788aa5b662] Detaching volume cb0179b0-05d9-4062-b3db-f8484769394d [ 1099.977273] env[61806]: INFO nova.virt.block_device [None req-f685fa1b-ce5d-413e-89f4-9a33f5f03e13 tempest-AttachVolumeNegativeTest-1683033704 tempest-AttachVolumeNegativeTest-1683033704-project-member] [instance: ce2ad097-4c3f-4b55-b018-ac788aa5b662] Attempting to driver detach volume cb0179b0-05d9-4062-b3db-f8484769394d from mountpoint /dev/sdb [ 1099.977535] env[61806]: DEBUG nova.virt.vmwareapi.volumeops [None req-f685fa1b-ce5d-413e-89f4-9a33f5f03e13 tempest-AttachVolumeNegativeTest-1683033704 tempest-AttachVolumeNegativeTest-1683033704-project-member] [instance: ce2ad097-4c3f-4b55-b018-ac788aa5b662] Volume detach. Driver type: vmdk {{(pid=61806) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 1099.977731] env[61806]: DEBUG nova.virt.vmwareapi.volumeops [None req-f685fa1b-ce5d-413e-89f4-9a33f5f03e13 tempest-AttachVolumeNegativeTest-1683033704 tempest-AttachVolumeNegativeTest-1683033704-project-member] [instance: ce2ad097-4c3f-4b55-b018-ac788aa5b662] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-277764', 'volume_id': 'cb0179b0-05d9-4062-b3db-f8484769394d', 'name': 'volume-cb0179b0-05d9-4062-b3db-f8484769394d', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'ce2ad097-4c3f-4b55-b018-ac788aa5b662', 'attached_at': '', 'detached_at': '', 'volume_id': 'cb0179b0-05d9-4062-b3db-f8484769394d', 'serial': 'cb0179b0-05d9-4062-b3db-f8484769394d'} {{(pid=61806) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 1099.978677] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e612741a-cae9-43e2-b2cc-badd21a9fd4d {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1100.001507] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dd1ad84d-3458-43f9-9dc0-7c6083107971 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1100.012046] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2f1be2ff-44df-4085-b68a-deb534ecded8 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1100.035485] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-70fbba54-f218-40f9-bc93-fec45f16f337 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1100.053062] env[61806]: DEBUG nova.virt.vmwareapi.volumeops [None req-f685fa1b-ce5d-413e-89f4-9a33f5f03e13 tempest-AttachVolumeNegativeTest-1683033704 tempest-AttachVolumeNegativeTest-1683033704-project-member] The volume has not been displaced from its original location: [datastore1] volume-cb0179b0-05d9-4062-b3db-f8484769394d/volume-cb0179b0-05d9-4062-b3db-f8484769394d.vmdk. No consolidation needed. {{(pid=61806) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 1100.058589] env[61806]: DEBUG nova.virt.vmwareapi.volumeops [None req-f685fa1b-ce5d-413e-89f4-9a33f5f03e13 tempest-AttachVolumeNegativeTest-1683033704 tempest-AttachVolumeNegativeTest-1683033704-project-member] [instance: ce2ad097-4c3f-4b55-b018-ac788aa5b662] Reconfiguring VM instance instance-00000068 to detach disk 2001 {{(pid=61806) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1100.059380] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-9fb9d788-16b3-4b2e-8f28-80b1dc47299a {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1100.081333] env[61806]: DEBUG oslo_vmware.api [None req-f685fa1b-ce5d-413e-89f4-9a33f5f03e13 tempest-AttachVolumeNegativeTest-1683033704 tempest-AttachVolumeNegativeTest-1683033704-project-member] Waiting for the task: (returnval){ [ 1100.081333] env[61806]: value = "task-1295238" [ 1100.081333] env[61806]: _type = "Task" [ 1100.081333] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1100.091076] env[61806]: DEBUG oslo_vmware.api [None req-f685fa1b-ce5d-413e-89f4-9a33f5f03e13 tempest-AttachVolumeNegativeTest-1683033704 tempest-AttachVolumeNegativeTest-1683033704-project-member] Task: {'id': task-1295238, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1100.274385] env[61806]: DEBUG nova.network.neutron [req-704b46e2-e103-439d-b6c0-f287ca546de9 req-dbd7de39-ac8c-4c5f-8488-05497ee3296e service nova] [instance: 29210bcd-0fe1-4ed9-a459-abc8016a4255] Updated VIF entry in instance network info cache for port 8aad34b7-4309-44b6-b88d-ae5961da22c2. {{(pid=61806) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1100.274731] env[61806]: DEBUG nova.network.neutron [req-704b46e2-e103-439d-b6c0-f287ca546de9 req-dbd7de39-ac8c-4c5f-8488-05497ee3296e service nova] [instance: 29210bcd-0fe1-4ed9-a459-abc8016a4255] Updating instance_info_cache with network_info: [{"id": "8aad34b7-4309-44b6-b88d-ae5961da22c2", "address": "fa:16:3e:c0:a7:c5", "network": {"id": "a9d993bc-0817-4ee6-ab6e-1ef9432a110f", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-939246736-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.196", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "bbbb09a49203421d807f733616c854df", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ad4c015b-4a8b-46ca-9556-74bad8db9fb3", "external-id": "nsx-vlan-transportzone-246", "segmentation_id": 246, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap8aad34b7-43", "ovs_interfaceid": "8aad34b7-4309-44b6-b88d-ae5961da22c2", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61806) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1100.280755] env[61806]: DEBUG nova.compute.manager [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] [instance: 0758901a-7093-41d3-b0e2-5c519333abdd] Instance has had 0 of 5 cleanup attempts {{(pid=61806) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1100.332572] env[61806]: DEBUG oslo_vmware.rw_handles [None req-4284f75d-699f-4bae-83bf-967ca811ad7c tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Getting lease state for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/5255ffd5-142b-628e-f764-280f309ea5c7/disk-0.vmdk. {{(pid=61806) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 1100.333525] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e5387e80-bbd0-40e7-a541-81f26a77475d {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1100.339527] env[61806]: DEBUG oslo_vmware.rw_handles [None req-4284f75d-699f-4bae-83bf-967ca811ad7c tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/5255ffd5-142b-628e-f764-280f309ea5c7/disk-0.vmdk is in state: ready. {{(pid=61806) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 1100.339722] env[61806]: ERROR oslo_vmware.rw_handles [None req-4284f75d-699f-4bae-83bf-967ca811ad7c tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Aborting lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/5255ffd5-142b-628e-f764-280f309ea5c7/disk-0.vmdk due to incomplete transfer. [ 1100.339954] env[61806]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseAbort with opID=oslo.vmware-3c2216be-36b2-43cd-9877-adaf4c8245cf {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1100.342874] env[61806]: DEBUG oslo_concurrency.lockutils [None req-7f52d370-4c9e-4ec0-a23f-5733d9a2d432 tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Acquiring lock "refresh_cache-06658134-4c6d-4911-9a55-b6805f555c78" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1100.343069] env[61806]: DEBUG oslo_concurrency.lockutils [None req-7f52d370-4c9e-4ec0-a23f-5733d9a2d432 tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Acquired lock "refresh_cache-06658134-4c6d-4911-9a55-b6805f555c78" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1100.343344] env[61806]: DEBUG nova.network.neutron [None req-7f52d370-4c9e-4ec0-a23f-5733d9a2d432 tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] [instance: 06658134-4c6d-4911-9a55-b6805f555c78] Building network info cache for instance {{(pid=61806) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1100.347281] env[61806]: DEBUG oslo_vmware.rw_handles [None req-4284f75d-699f-4bae-83bf-967ca811ad7c tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Closed VMDK read handle for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/5255ffd5-142b-628e-f764-280f309ea5c7/disk-0.vmdk. {{(pid=61806) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:735}} [ 1100.347470] env[61806]: DEBUG nova.virt.vmwareapi.images [None req-4284f75d-699f-4bae-83bf-967ca811ad7c tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] [instance: d0669232-6f33-4b30-97c2-2e4239af1ad0] Uploaded image 7e155d45-9a75-4ca6-97c4-ac1fbb21e996 to the Glance image server {{(pid=61806) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:473}} [ 1100.349940] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [None req-4284f75d-699f-4bae-83bf-967ca811ad7c tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] [instance: d0669232-6f33-4b30-97c2-2e4239af1ad0] Destroying the VM {{(pid=61806) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1358}} [ 1100.350217] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-6a4b2156-7c8e-4456-979e-4494522171ce {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1100.356388] env[61806]: DEBUG oslo_vmware.api [None req-4284f75d-699f-4bae-83bf-967ca811ad7c tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Waiting for the task: (returnval){ [ 1100.356388] env[61806]: value = "task-1295239" [ 1100.356388] env[61806]: _type = "Task" [ 1100.356388] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1100.364372] env[61806]: DEBUG oslo_vmware.api [None req-4284f75d-699f-4bae-83bf-967ca811ad7c tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Task: {'id': task-1295239, 'name': Destroy_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1100.592284] env[61806]: DEBUG oslo_vmware.api [None req-f685fa1b-ce5d-413e-89f4-9a33f5f03e13 tempest-AttachVolumeNegativeTest-1683033704 tempest-AttachVolumeNegativeTest-1683033704-project-member] Task: {'id': task-1295238, 'name': ReconfigVM_Task, 'duration_secs': 0.253483} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1100.592545] env[61806]: DEBUG nova.virt.vmwareapi.volumeops [None req-f685fa1b-ce5d-413e-89f4-9a33f5f03e13 tempest-AttachVolumeNegativeTest-1683033704 tempest-AttachVolumeNegativeTest-1683033704-project-member] [instance: ce2ad097-4c3f-4b55-b018-ac788aa5b662] Reconfigured VM instance instance-00000068 to detach disk 2001 {{(pid=61806) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1100.597504] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-21568b07-3e7a-4fe7-9575-8f824c6eb423 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1100.613655] env[61806]: DEBUG oslo_vmware.api [None req-f685fa1b-ce5d-413e-89f4-9a33f5f03e13 tempest-AttachVolumeNegativeTest-1683033704 tempest-AttachVolumeNegativeTest-1683033704-project-member] Waiting for the task: (returnval){ [ 1100.613655] env[61806]: value = "task-1295240" [ 1100.613655] env[61806]: _type = "Task" [ 1100.613655] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1100.622428] env[61806]: DEBUG oslo_vmware.api [None req-f685fa1b-ce5d-413e-89f4-9a33f5f03e13 tempest-AttachVolumeNegativeTest-1683033704 tempest-AttachVolumeNegativeTest-1683033704-project-member] Task: {'id': task-1295240, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1100.777598] env[61806]: DEBUG oslo_concurrency.lockutils [req-704b46e2-e103-439d-b6c0-f287ca546de9 req-dbd7de39-ac8c-4c5f-8488-05497ee3296e service nova] Releasing lock "refresh_cache-29210bcd-0fe1-4ed9-a459-abc8016a4255" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1100.783934] env[61806]: DEBUG nova.compute.manager [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] [instance: 616ec206-9804-469e-ab5c-41aea7f048aa] Instance has had 0 of 5 cleanup attempts {{(pid=61806) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1100.865858] env[61806]: DEBUG oslo_vmware.api [None req-4284f75d-699f-4bae-83bf-967ca811ad7c tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Task: {'id': task-1295239, 'name': Destroy_Task, 'duration_secs': 0.333569} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1100.866354] env[61806]: INFO nova.virt.vmwareapi.vm_util [None req-4284f75d-699f-4bae-83bf-967ca811ad7c tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] [instance: d0669232-6f33-4b30-97c2-2e4239af1ad0] Destroyed the VM [ 1100.866452] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-4284f75d-699f-4bae-83bf-967ca811ad7c tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] [instance: d0669232-6f33-4b30-97c2-2e4239af1ad0] Deleting Snapshot of the VM instance {{(pid=61806) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 1100.866628] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-599feac9-60e7-4e2a-a8e0-c28134ab4949 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1100.872359] env[61806]: DEBUG oslo_vmware.api [None req-4284f75d-699f-4bae-83bf-967ca811ad7c tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Waiting for the task: (returnval){ [ 1100.872359] env[61806]: value = "task-1295241" [ 1100.872359] env[61806]: _type = "Task" [ 1100.872359] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1100.880266] env[61806]: DEBUG oslo_vmware.api [None req-4284f75d-699f-4bae-83bf-967ca811ad7c tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Task: {'id': task-1295241, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1101.072532] env[61806]: DEBUG nova.network.neutron [None req-7f52d370-4c9e-4ec0-a23f-5733d9a2d432 tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] [instance: 06658134-4c6d-4911-9a55-b6805f555c78] Updating instance_info_cache with network_info: [{"id": "783dae1f-d1c0-4958-b039-ed57b202b64a", "address": "fa:16:3e:89:24:35", "network": {"id": "746a03b6-c1a6-4d27-98b9-0bee6376634f", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-1971382587-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.245", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c1f2d161c50f4b70987b90e6364c3779", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6a31b224-77d7-48a4-af87-312758611b8e", "external-id": "nsx-vlan-transportzone-761", "segmentation_id": 761, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap783dae1f-d1", "ovs_interfaceid": "783dae1f-d1c0-4958-b039-ed57b202b64a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61806) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1101.123997] env[61806]: DEBUG oslo_vmware.api [None req-f685fa1b-ce5d-413e-89f4-9a33f5f03e13 tempest-AttachVolumeNegativeTest-1683033704 tempest-AttachVolumeNegativeTest-1683033704-project-member] Task: {'id': task-1295240, 'name': ReconfigVM_Task} progress is 99%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1101.286878] env[61806]: DEBUG nova.compute.manager [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] [instance: a934d02d-26aa-4900-b473-a58489e5629e] Instance has had 0 of 5 cleanup attempts {{(pid=61806) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1101.382606] env[61806]: DEBUG oslo_vmware.api [None req-4284f75d-699f-4bae-83bf-967ca811ad7c tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Task: {'id': task-1295241, 'name': RemoveSnapshot_Task, 'duration_secs': 0.482496} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1101.382606] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-4284f75d-699f-4bae-83bf-967ca811ad7c tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] [instance: d0669232-6f33-4b30-97c2-2e4239af1ad0] Deleted Snapshot of the VM instance {{(pid=61806) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 1101.382801] env[61806]: INFO nova.compute.manager [None req-4284f75d-699f-4bae-83bf-967ca811ad7c tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] [instance: d0669232-6f33-4b30-97c2-2e4239af1ad0] Took 12.38 seconds to snapshot the instance on the hypervisor. [ 1101.575803] env[61806]: DEBUG oslo_concurrency.lockutils [None req-7f52d370-4c9e-4ec0-a23f-5733d9a2d432 tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Releasing lock "refresh_cache-06658134-4c6d-4911-9a55-b6805f555c78" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1101.625066] env[61806]: DEBUG oslo_vmware.api [None req-f685fa1b-ce5d-413e-89f4-9a33f5f03e13 tempest-AttachVolumeNegativeTest-1683033704 tempest-AttachVolumeNegativeTest-1683033704-project-member] Task: {'id': task-1295240, 'name': ReconfigVM_Task} progress is 99%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1101.790491] env[61806]: DEBUG nova.compute.manager [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] [instance: 4febd093-9f2b-494e-b175-e4693b0e3e0d] Instance has had 0 of 5 cleanup attempts {{(pid=61806) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1101.929154] env[61806]: DEBUG nova.compute.manager [None req-4284f75d-699f-4bae-83bf-967ca811ad7c tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] [instance: d0669232-6f33-4b30-97c2-2e4239af1ad0] Found 3 images (rotation: 2) {{(pid=61806) _rotate_backups /opt/stack/nova/nova/compute/manager.py:4554}} [ 1101.929377] env[61806]: DEBUG nova.compute.manager [None req-4284f75d-699f-4bae-83bf-967ca811ad7c tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] [instance: d0669232-6f33-4b30-97c2-2e4239af1ad0] Rotating out 1 backups {{(pid=61806) _rotate_backups /opt/stack/nova/nova/compute/manager.py:4562}} [ 1101.929548] env[61806]: DEBUG nova.compute.manager [None req-4284f75d-699f-4bae-83bf-967ca811ad7c tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] [instance: d0669232-6f33-4b30-97c2-2e4239af1ad0] Deleting image 301ac572-fe69-490a-b0d2-2012e697182e {{(pid=61806) _rotate_backups /opt/stack/nova/nova/compute/manager.py:4567}} [ 1102.103841] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e939d5aa-1a93-40d4-8b73-f7c945801a0a {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1102.125267] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5dc7c0eb-bfb0-4d7f-a1e5-aa850d1a1215 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1102.132606] env[61806]: DEBUG oslo_vmware.api [None req-f685fa1b-ce5d-413e-89f4-9a33f5f03e13 tempest-AttachVolumeNegativeTest-1683033704 tempest-AttachVolumeNegativeTest-1683033704-project-member] Task: {'id': task-1295240, 'name': ReconfigVM_Task, 'duration_secs': 1.239182} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1102.135106] env[61806]: DEBUG nova.virt.vmwareapi.volumeops [None req-f685fa1b-ce5d-413e-89f4-9a33f5f03e13 tempest-AttachVolumeNegativeTest-1683033704 tempest-AttachVolumeNegativeTest-1683033704-project-member] [instance: ce2ad097-4c3f-4b55-b018-ac788aa5b662] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-277764', 'volume_id': 'cb0179b0-05d9-4062-b3db-f8484769394d', 'name': 'volume-cb0179b0-05d9-4062-b3db-f8484769394d', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'ce2ad097-4c3f-4b55-b018-ac788aa5b662', 'attached_at': '', 'detached_at': '', 'volume_id': 'cb0179b0-05d9-4062-b3db-f8484769394d', 'serial': 'cb0179b0-05d9-4062-b3db-f8484769394d'} {{(pid=61806) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 1102.137180] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-7f52d370-4c9e-4ec0-a23f-5733d9a2d432 tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] [instance: 06658134-4c6d-4911-9a55-b6805f555c78] Updating instance '06658134-4c6d-4911-9a55-b6805f555c78' progress to 83 {{(pid=61806) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 1102.293853] env[61806]: DEBUG nova.compute.manager [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] [instance: 7c3fd2ff-ebd5-454f-a743-7fbae1088941] Instance has had 0 of 5 cleanup attempts {{(pid=61806) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1102.643701] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [None req-7f52d370-4c9e-4ec0-a23f-5733d9a2d432 tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] [instance: 06658134-4c6d-4911-9a55-b6805f555c78] Powering on the VM {{(pid=61806) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1102.644105] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-b415a96f-e2df-4c0b-aee7-0a4fbda1af88 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1102.652362] env[61806]: DEBUG oslo_vmware.api [None req-7f52d370-4c9e-4ec0-a23f-5733d9a2d432 tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Waiting for the task: (returnval){ [ 1102.652362] env[61806]: value = "task-1295242" [ 1102.652362] env[61806]: _type = "Task" [ 1102.652362] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1102.660051] env[61806]: DEBUG oslo_vmware.api [None req-7f52d370-4c9e-4ec0-a23f-5733d9a2d432 tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Task: {'id': task-1295242, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1102.684644] env[61806]: DEBUG nova.objects.instance [None req-f685fa1b-ce5d-413e-89f4-9a33f5f03e13 tempest-AttachVolumeNegativeTest-1683033704 tempest-AttachVolumeNegativeTest-1683033704-project-member] Lazy-loading 'flavor' on Instance uuid ce2ad097-4c3f-4b55-b018-ac788aa5b662 {{(pid=61806) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1102.798085] env[61806]: DEBUG nova.compute.manager [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] [instance: 4184e263-6f56-4bc1-99b2-a2c460531516] Instance has had 0 of 5 cleanup attempts {{(pid=61806) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1103.162837] env[61806]: DEBUG oslo_vmware.api [None req-7f52d370-4c9e-4ec0-a23f-5733d9a2d432 tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Task: {'id': task-1295242, 'name': PowerOnVM_Task, 'duration_secs': 0.403985} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1103.163176] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [None req-7f52d370-4c9e-4ec0-a23f-5733d9a2d432 tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] [instance: 06658134-4c6d-4911-9a55-b6805f555c78] Powered on the VM {{(pid=61806) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1103.163400] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-7f52d370-4c9e-4ec0-a23f-5733d9a2d432 tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] [instance: 06658134-4c6d-4911-9a55-b6805f555c78] Updating instance '06658134-4c6d-4911-9a55-b6805f555c78' progress to 100 {{(pid=61806) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 1103.301169] env[61806]: DEBUG nova.compute.manager [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] [instance: 2a3f81e3-b84f-4370-a598-2f6a607c3d47] Instance has had 0 of 5 cleanup attempts {{(pid=61806) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1103.694495] env[61806]: DEBUG oslo_concurrency.lockutils [None req-f685fa1b-ce5d-413e-89f4-9a33f5f03e13 tempest-AttachVolumeNegativeTest-1683033704 tempest-AttachVolumeNegativeTest-1683033704-project-member] Lock "ce2ad097-4c3f-4b55-b018-ac788aa5b662" "released" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: held 4.296s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1103.804117] env[61806]: DEBUG nova.compute.manager [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] [instance: e0ef0a35-82a5-495b-9d5c-5805e8306390] Instance has had 0 of 5 cleanup attempts {{(pid=61806) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1103.829725] env[61806]: DEBUG oslo_concurrency.lockutils [None req-0267dd86-554f-40ed-a8f9-73fb6bfb3cd2 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Acquiring lock "d0669232-6f33-4b30-97c2-2e4239af1ad0" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1103.829987] env[61806]: DEBUG oslo_concurrency.lockutils [None req-0267dd86-554f-40ed-a8f9-73fb6bfb3cd2 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Lock "d0669232-6f33-4b30-97c2-2e4239af1ad0" acquired by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: waited 0.000s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1103.830189] env[61806]: DEBUG nova.compute.manager [None req-0267dd86-554f-40ed-a8f9-73fb6bfb3cd2 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] [instance: d0669232-6f33-4b30-97c2-2e4239af1ad0] Checking state {{(pid=61806) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1103.831326] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-749e9ca7-f743-4ac3-9f30-acc48eb4794b {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1103.838471] env[61806]: DEBUG nova.compute.manager [None req-0267dd86-554f-40ed-a8f9-73fb6bfb3cd2 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] [instance: d0669232-6f33-4b30-97c2-2e4239af1ad0] Stopping instance; current vm_state: active, current task_state: powering-off, current DB power_state: 1, current VM power_state: 1 {{(pid=61806) do_stop_instance /opt/stack/nova/nova/compute/manager.py:3368}} [ 1103.839040] env[61806]: DEBUG nova.objects.instance [None req-0267dd86-554f-40ed-a8f9-73fb6bfb3cd2 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Lazy-loading 'flavor' on Instance uuid d0669232-6f33-4b30-97c2-2e4239af1ad0 {{(pid=61806) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1104.306820] env[61806]: DEBUG oslo_service.periodic_task [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Running periodic task ComputeManager._cleanup_incomplete_migrations {{(pid=61806) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1104.307114] env[61806]: DEBUG nova.compute.manager [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Cleaning up deleted instances with incomplete migration {{(pid=61806) _cleanup_incomplete_migrations /opt/stack/nova/nova/compute/manager.py:11264}} [ 1104.343286] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [None req-0267dd86-554f-40ed-a8f9-73fb6bfb3cd2 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] [instance: d0669232-6f33-4b30-97c2-2e4239af1ad0] Powering off the VM {{(pid=61806) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1104.343522] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-9b567cf3-aec6-45cc-bd09-040a931d75b6 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1104.350595] env[61806]: DEBUG oslo_vmware.api [None req-0267dd86-554f-40ed-a8f9-73fb6bfb3cd2 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Waiting for the task: (returnval){ [ 1104.350595] env[61806]: value = "task-1295243" [ 1104.350595] env[61806]: _type = "Task" [ 1104.350595] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1104.358135] env[61806]: DEBUG oslo_vmware.api [None req-0267dd86-554f-40ed-a8f9-73fb6bfb3cd2 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Task: {'id': task-1295243, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1104.685022] env[61806]: DEBUG oslo_concurrency.lockutils [None req-18d3b39e-bf74-4159-88ef-012a6899869a tempest-AttachVolumeNegativeTest-1683033704 tempest-AttachVolumeNegativeTest-1683033704-project-member] Acquiring lock "ce2ad097-4c3f-4b55-b018-ac788aa5b662" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1104.685022] env[61806]: DEBUG oslo_concurrency.lockutils [None req-18d3b39e-bf74-4159-88ef-012a6899869a tempest-AttachVolumeNegativeTest-1683033704 tempest-AttachVolumeNegativeTest-1683033704-project-member] Lock "ce2ad097-4c3f-4b55-b018-ac788aa5b662" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1104.685022] env[61806]: DEBUG oslo_concurrency.lockutils [None req-18d3b39e-bf74-4159-88ef-012a6899869a tempest-AttachVolumeNegativeTest-1683033704 tempest-AttachVolumeNegativeTest-1683033704-project-member] Acquiring lock "ce2ad097-4c3f-4b55-b018-ac788aa5b662-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1104.685022] env[61806]: DEBUG oslo_concurrency.lockutils [None req-18d3b39e-bf74-4159-88ef-012a6899869a tempest-AttachVolumeNegativeTest-1683033704 tempest-AttachVolumeNegativeTest-1683033704-project-member] Lock "ce2ad097-4c3f-4b55-b018-ac788aa5b662-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1104.685022] env[61806]: DEBUG oslo_concurrency.lockutils [None req-18d3b39e-bf74-4159-88ef-012a6899869a tempest-AttachVolumeNegativeTest-1683033704 tempest-AttachVolumeNegativeTest-1683033704-project-member] Lock "ce2ad097-4c3f-4b55-b018-ac788aa5b662-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1104.686949] env[61806]: INFO nova.compute.manager [None req-18d3b39e-bf74-4159-88ef-012a6899869a tempest-AttachVolumeNegativeTest-1683033704 tempest-AttachVolumeNegativeTest-1683033704-project-member] [instance: ce2ad097-4c3f-4b55-b018-ac788aa5b662] Terminating instance [ 1104.689473] env[61806]: DEBUG nova.compute.manager [None req-18d3b39e-bf74-4159-88ef-012a6899869a tempest-AttachVolumeNegativeTest-1683033704 tempest-AttachVolumeNegativeTest-1683033704-project-member] [instance: ce2ad097-4c3f-4b55-b018-ac788aa5b662] Start destroying the instance on the hypervisor. {{(pid=61806) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1104.689788] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-18d3b39e-bf74-4159-88ef-012a6899869a tempest-AttachVolumeNegativeTest-1683033704 tempest-AttachVolumeNegativeTest-1683033704-project-member] [instance: ce2ad097-4c3f-4b55-b018-ac788aa5b662] Destroying instance {{(pid=61806) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1104.690726] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6e3ecc0d-cb27-4126-9c60-63fc5a605e96 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1104.698795] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [None req-18d3b39e-bf74-4159-88ef-012a6899869a tempest-AttachVolumeNegativeTest-1683033704 tempest-AttachVolumeNegativeTest-1683033704-project-member] [instance: ce2ad097-4c3f-4b55-b018-ac788aa5b662] Powering off the VM {{(pid=61806) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1104.699612] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-048d76af-1035-45d2-b458-c4d6b3acdbb4 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1104.706037] env[61806]: DEBUG oslo_vmware.api [None req-18d3b39e-bf74-4159-88ef-012a6899869a tempest-AttachVolumeNegativeTest-1683033704 tempest-AttachVolumeNegativeTest-1683033704-project-member] Waiting for the task: (returnval){ [ 1104.706037] env[61806]: value = "task-1295244" [ 1104.706037] env[61806]: _type = "Task" [ 1104.706037] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1104.714301] env[61806]: DEBUG oslo_vmware.api [None req-18d3b39e-bf74-4159-88ef-012a6899869a tempest-AttachVolumeNegativeTest-1683033704 tempest-AttachVolumeNegativeTest-1683033704-project-member] Task: {'id': task-1295244, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1104.865709] env[61806]: DEBUG oslo_vmware.api [None req-0267dd86-554f-40ed-a8f9-73fb6bfb3cd2 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Task: {'id': task-1295243, 'name': PowerOffVM_Task, 'duration_secs': 0.192061} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1104.866112] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [None req-0267dd86-554f-40ed-a8f9-73fb6bfb3cd2 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] [instance: d0669232-6f33-4b30-97c2-2e4239af1ad0] Powered off the VM {{(pid=61806) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1104.866390] env[61806]: DEBUG nova.compute.manager [None req-0267dd86-554f-40ed-a8f9-73fb6bfb3cd2 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] [instance: d0669232-6f33-4b30-97c2-2e4239af1ad0] Checking state {{(pid=61806) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1104.867878] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-89516cd9-34e9-4990-852f-1fb82ed823c6 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1105.218164] env[61806]: DEBUG oslo_vmware.api [None req-18d3b39e-bf74-4159-88ef-012a6899869a tempest-AttachVolumeNegativeTest-1683033704 tempest-AttachVolumeNegativeTest-1683033704-project-member] Task: {'id': task-1295244, 'name': PowerOffVM_Task, 'duration_secs': 0.225252} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1105.218496] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [None req-18d3b39e-bf74-4159-88ef-012a6899869a tempest-AttachVolumeNegativeTest-1683033704 tempest-AttachVolumeNegativeTest-1683033704-project-member] [instance: ce2ad097-4c3f-4b55-b018-ac788aa5b662] Powered off the VM {{(pid=61806) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1105.218708] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-18d3b39e-bf74-4159-88ef-012a6899869a tempest-AttachVolumeNegativeTest-1683033704 tempest-AttachVolumeNegativeTest-1683033704-project-member] [instance: ce2ad097-4c3f-4b55-b018-ac788aa5b662] Unregistering the VM {{(pid=61806) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1105.218978] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-cf3d9698-cbc3-4aa7-8f36-7c9da486dbae {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1105.284847] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-18d3b39e-bf74-4159-88ef-012a6899869a tempest-AttachVolumeNegativeTest-1683033704 tempest-AttachVolumeNegativeTest-1683033704-project-member] [instance: ce2ad097-4c3f-4b55-b018-ac788aa5b662] Unregistered the VM {{(pid=61806) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1105.285102] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-18d3b39e-bf74-4159-88ef-012a6899869a tempest-AttachVolumeNegativeTest-1683033704 tempest-AttachVolumeNegativeTest-1683033704-project-member] [instance: ce2ad097-4c3f-4b55-b018-ac788aa5b662] Deleting contents of the VM from datastore datastore1 {{(pid=61806) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1105.285303] env[61806]: DEBUG nova.virt.vmwareapi.ds_util [None req-18d3b39e-bf74-4159-88ef-012a6899869a tempest-AttachVolumeNegativeTest-1683033704 tempest-AttachVolumeNegativeTest-1683033704-project-member] Deleting the datastore file [datastore1] ce2ad097-4c3f-4b55-b018-ac788aa5b662 {{(pid=61806) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1105.285590] env[61806]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-60114cb3-f471-492a-bfe9-43ca48779dc0 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1105.292809] env[61806]: DEBUG oslo_vmware.api [None req-18d3b39e-bf74-4159-88ef-012a6899869a tempest-AttachVolumeNegativeTest-1683033704 tempest-AttachVolumeNegativeTest-1683033704-project-member] Waiting for the task: (returnval){ [ 1105.292809] env[61806]: value = "task-1295246" [ 1105.292809] env[61806]: _type = "Task" [ 1105.292809] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1105.296714] env[61806]: DEBUG nova.network.neutron [None req-705f3985-4e14-42b8-a840-12f7c3a060f8 tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] [instance: 06658134-4c6d-4911-9a55-b6805f555c78] Port 783dae1f-d1c0-4958-b039-ed57b202b64a binding to destination host cpu-1 is already ACTIVE {{(pid=61806) migrate_instance_start /opt/stack/nova/nova/network/neutron.py:3171}} [ 1105.296966] env[61806]: DEBUG oslo_concurrency.lockutils [None req-705f3985-4e14-42b8-a840-12f7c3a060f8 tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Acquiring lock "refresh_cache-06658134-4c6d-4911-9a55-b6805f555c78" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1105.297137] env[61806]: DEBUG oslo_concurrency.lockutils [None req-705f3985-4e14-42b8-a840-12f7c3a060f8 tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Acquired lock "refresh_cache-06658134-4c6d-4911-9a55-b6805f555c78" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1105.297306] env[61806]: DEBUG nova.network.neutron [None req-705f3985-4e14-42b8-a840-12f7c3a060f8 tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] [instance: 06658134-4c6d-4911-9a55-b6805f555c78] Building network info cache for instance {{(pid=61806) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1105.302020] env[61806]: DEBUG oslo_vmware.api [None req-18d3b39e-bf74-4159-88ef-012a6899869a tempest-AttachVolumeNegativeTest-1683033704 tempest-AttachVolumeNegativeTest-1683033704-project-member] Task: {'id': task-1295246, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1105.382290] env[61806]: DEBUG oslo_concurrency.lockutils [None req-0267dd86-554f-40ed-a8f9-73fb6bfb3cd2 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Lock "d0669232-6f33-4b30-97c2-2e4239af1ad0" "released" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: held 1.552s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1105.804692] env[61806]: DEBUG oslo_vmware.api [None req-18d3b39e-bf74-4159-88ef-012a6899869a tempest-AttachVolumeNegativeTest-1683033704 tempest-AttachVolumeNegativeTest-1683033704-project-member] Task: {'id': task-1295246, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.219244} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1105.805012] env[61806]: DEBUG nova.virt.vmwareapi.ds_util [None req-18d3b39e-bf74-4159-88ef-012a6899869a tempest-AttachVolumeNegativeTest-1683033704 tempest-AttachVolumeNegativeTest-1683033704-project-member] Deleted the datastore file {{(pid=61806) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1105.805205] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-18d3b39e-bf74-4159-88ef-012a6899869a tempest-AttachVolumeNegativeTest-1683033704 tempest-AttachVolumeNegativeTest-1683033704-project-member] [instance: ce2ad097-4c3f-4b55-b018-ac788aa5b662] Deleted contents of the VM from datastore datastore1 {{(pid=61806) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1105.805390] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-18d3b39e-bf74-4159-88ef-012a6899869a tempest-AttachVolumeNegativeTest-1683033704 tempest-AttachVolumeNegativeTest-1683033704-project-member] [instance: ce2ad097-4c3f-4b55-b018-ac788aa5b662] Instance destroyed {{(pid=61806) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1105.805679] env[61806]: INFO nova.compute.manager [None req-18d3b39e-bf74-4159-88ef-012a6899869a tempest-AttachVolumeNegativeTest-1683033704 tempest-AttachVolumeNegativeTest-1683033704-project-member] [instance: ce2ad097-4c3f-4b55-b018-ac788aa5b662] Took 1.12 seconds to destroy the instance on the hypervisor. [ 1105.805982] env[61806]: DEBUG oslo.service.loopingcall [None req-18d3b39e-bf74-4159-88ef-012a6899869a tempest-AttachVolumeNegativeTest-1683033704 tempest-AttachVolumeNegativeTest-1683033704-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61806) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1105.807015] env[61806]: DEBUG nova.compute.manager [-] [instance: ce2ad097-4c3f-4b55-b018-ac788aa5b662] Deallocating network for instance {{(pid=61806) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1105.807015] env[61806]: DEBUG nova.network.neutron [-] [instance: ce2ad097-4c3f-4b55-b018-ac788aa5b662] deallocate_for_instance() {{(pid=61806) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1106.058533] env[61806]: DEBUG nova.network.neutron [None req-705f3985-4e14-42b8-a840-12f7c3a060f8 tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] [instance: 06658134-4c6d-4911-9a55-b6805f555c78] Updating instance_info_cache with network_info: [{"id": "783dae1f-d1c0-4958-b039-ed57b202b64a", "address": "fa:16:3e:89:24:35", "network": {"id": "746a03b6-c1a6-4d27-98b9-0bee6376634f", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-1971382587-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.245", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c1f2d161c50f4b70987b90e6364c3779", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6a31b224-77d7-48a4-af87-312758611b8e", "external-id": "nsx-vlan-transportzone-761", "segmentation_id": 761, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap783dae1f-d1", "ovs_interfaceid": "783dae1f-d1c0-4958-b039-ed57b202b64a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61806) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1106.371392] env[61806]: DEBUG oslo_service.periodic_task [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Running periodic task ComputeManager._sync_scheduler_instance_info {{(pid=61806) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1106.459417] env[61806]: DEBUG nova.compute.manager [req-1f0038d0-1a00-4fa5-b581-e5ffe86b6df4 req-b85169ef-74dc-4488-810c-ace5d12cf0aa service nova] [instance: ce2ad097-4c3f-4b55-b018-ac788aa5b662] Received event network-vif-deleted-6792e7fd-4056-4d2f-86a5-45ea41172f3f {{(pid=61806) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1106.459691] env[61806]: INFO nova.compute.manager [req-1f0038d0-1a00-4fa5-b581-e5ffe86b6df4 req-b85169ef-74dc-4488-810c-ace5d12cf0aa service nova] [instance: ce2ad097-4c3f-4b55-b018-ac788aa5b662] Neutron deleted interface 6792e7fd-4056-4d2f-86a5-45ea41172f3f; detaching it from the instance and deleting it from the info cache [ 1106.459864] env[61806]: DEBUG nova.network.neutron [req-1f0038d0-1a00-4fa5-b581-e5ffe86b6df4 req-b85169ef-74dc-4488-810c-ace5d12cf0aa service nova] [instance: ce2ad097-4c3f-4b55-b018-ac788aa5b662] Updating instance_info_cache with network_info: [] {{(pid=61806) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1106.561123] env[61806]: DEBUG oslo_concurrency.lockutils [None req-705f3985-4e14-42b8-a840-12f7c3a060f8 tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Releasing lock "refresh_cache-06658134-4c6d-4911-9a55-b6805f555c78" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1106.875552] env[61806]: DEBUG oslo_service.periodic_task [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=61806) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1106.875744] env[61806]: DEBUG oslo_service.periodic_task [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=61806) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1106.875932] env[61806]: DEBUG oslo_service.periodic_task [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=61806) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1106.876092] env[61806]: DEBUG nova.compute.manager [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=61806) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10545}} [ 1106.894330] env[61806]: DEBUG nova.compute.manager [None req-77073341-4390-4aa8-b9e9-ac30959d4690 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] [instance: d0669232-6f33-4b30-97c2-2e4239af1ad0] Stashing vm_state: stopped {{(pid=61806) _prep_resize /opt/stack/nova/nova/compute/manager.py:5624}} [ 1106.933563] env[61806]: DEBUG nova.network.neutron [-] [instance: ce2ad097-4c3f-4b55-b018-ac788aa5b662] Updating instance_info_cache with network_info: [] {{(pid=61806) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1106.965710] env[61806]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-cba462c9-ac26-400b-b90e-db98cb394cf9 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1106.975199] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8f6b38c8-cddb-4450-9f6e-5aa813d6f672 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1106.999965] env[61806]: DEBUG nova.compute.manager [req-1f0038d0-1a00-4fa5-b581-e5ffe86b6df4 req-b85169ef-74dc-4488-810c-ace5d12cf0aa service nova] [instance: ce2ad097-4c3f-4b55-b018-ac788aa5b662] Detach interface failed, port_id=6792e7fd-4056-4d2f-86a5-45ea41172f3f, reason: Instance ce2ad097-4c3f-4b55-b018-ac788aa5b662 could not be found. {{(pid=61806) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10963}} [ 1107.064998] env[61806]: DEBUG nova.compute.manager [None req-705f3985-4e14-42b8-a840-12f7c3a060f8 tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] [instance: 06658134-4c6d-4911-9a55-b6805f555c78] Hypervisor driver does not support instance shared storage check, assuming it's not on shared storage {{(pid=61806) _is_instance_storage_shared /opt/stack/nova/nova/compute/manager.py:897}} [ 1107.065161] env[61806]: DEBUG oslo_concurrency.lockutils [None req-705f3985-4e14-42b8-a840-12f7c3a060f8 tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_dest" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1107.065403] env[61806]: DEBUG oslo_concurrency.lockutils [None req-705f3985-4e14-42b8-a840-12f7c3a060f8 tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_dest" :: waited 0.000s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1107.413555] env[61806]: DEBUG oslo_concurrency.lockutils [None req-77073341-4390-4aa8-b9e9-ac30959d4690 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1107.436297] env[61806]: INFO nova.compute.manager [-] [instance: ce2ad097-4c3f-4b55-b018-ac788aa5b662] Took 1.63 seconds to deallocate network for instance. [ 1107.568805] env[61806]: DEBUG nova.objects.instance [None req-705f3985-4e14-42b8-a840-12f7c3a060f8 tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Lazy-loading 'migration_context' on Instance uuid 06658134-4c6d-4911-9a55-b6805f555c78 {{(pid=61806) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1107.943407] env[61806]: DEBUG oslo_concurrency.lockutils [None req-18d3b39e-bf74-4159-88ef-012a6899869a tempest-AttachVolumeNegativeTest-1683033704 tempest-AttachVolumeNegativeTest-1683033704-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1108.090281] env[61806]: DEBUG nova.scheduler.client.report [None req-705f3985-4e14-42b8-a840-12f7c3a060f8 tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Refreshing inventories for resource provider a2858be1-fd22-4e08-979e-87ad25293407 {{(pid=61806) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:818}} [ 1108.104152] env[61806]: DEBUG nova.scheduler.client.report [None req-705f3985-4e14-42b8-a840-12f7c3a060f8 tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Updating ProviderTree inventory for provider a2858be1-fd22-4e08-979e-87ad25293407 from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 138, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61806) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:782}} [ 1108.104377] env[61806]: DEBUG nova.compute.provider_tree [None req-705f3985-4e14-42b8-a840-12f7c3a060f8 tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Updating inventory in ProviderTree for provider a2858be1-fd22-4e08-979e-87ad25293407 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 138, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61806) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1108.116046] env[61806]: DEBUG nova.scheduler.client.report [None req-705f3985-4e14-42b8-a840-12f7c3a060f8 tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Refreshing aggregate associations for resource provider a2858be1-fd22-4e08-979e-87ad25293407, aggregates: None {{(pid=61806) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:827}} [ 1108.133613] env[61806]: DEBUG nova.scheduler.client.report [None req-705f3985-4e14-42b8-a840-12f7c3a060f8 tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Refreshing trait associations for resource provider a2858be1-fd22-4e08-979e-87ad25293407, traits: COMPUTE_IMAGE_TYPE_ISO,COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_SAME_HOST_COLD_MIGRATE,COMPUTE_NET_ATTACH_INTERFACE,COMPUTE_NODE {{(pid=61806) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:839}} [ 1108.207853] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c8c058fa-d2cc-4f4c-a08d-4026cd67e7d0 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1108.214877] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0bd27c8c-be68-4047-8667-ae4abe644a26 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1108.244065] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-391478b7-c60c-4bb4-af96-6cbcf7b7bc32 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1108.252395] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a7ef7c21-9626-4fb8-afdb-c5401b75f873 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1108.267360] env[61806]: DEBUG nova.compute.provider_tree [None req-705f3985-4e14-42b8-a840-12f7c3a060f8 tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Inventory has not changed in ProviderTree for provider: a2858be1-fd22-4e08-979e-87ad25293407 {{(pid=61806) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1108.770907] env[61806]: DEBUG nova.scheduler.client.report [None req-705f3985-4e14-42b8-a840-12f7c3a060f8 tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Inventory has not changed for provider a2858be1-fd22-4e08-979e-87ad25293407 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 138, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61806) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1109.782678] env[61806]: DEBUG oslo_concurrency.lockutils [None req-705f3985-4e14-42b8-a840-12f7c3a060f8 tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_dest" :: held 2.717s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1109.788293] env[61806]: DEBUG oslo_concurrency.lockutils [None req-77073341-4390-4aa8-b9e9-ac30959d4690 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: waited 2.375s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1110.297428] env[61806]: INFO nova.compute.claims [None req-77073341-4390-4aa8-b9e9-ac30959d4690 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] [instance: d0669232-6f33-4b30-97c2-2e4239af1ad0] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1110.804571] env[61806]: INFO nova.compute.resource_tracker [None req-77073341-4390-4aa8-b9e9-ac30959d4690 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] [instance: d0669232-6f33-4b30-97c2-2e4239af1ad0] Updating resource usage from migration 46ba7ac1-5345-483f-b3ae-3d5aed8a7aec [ 1110.881258] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-87c8d09c-7ff6-425f-8e56-c96c747c7ced {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1110.888922] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-48679c1a-0edc-4fde-bed7-3dc7a9dfa2a6 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1110.919486] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a25777d3-24db-4c6b-bbd2-65f4267f0248 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1110.926273] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3d9e6d2b-340b-481b-97e1-95dc6fda51fa {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1110.938891] env[61806]: DEBUG nova.compute.provider_tree [None req-77073341-4390-4aa8-b9e9-ac30959d4690 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Inventory has not changed in ProviderTree for provider: a2858be1-fd22-4e08-979e-87ad25293407 {{(pid=61806) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1111.327407] env[61806]: INFO nova.compute.manager [None req-705f3985-4e14-42b8-a840-12f7c3a060f8 tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] [instance: 06658134-4c6d-4911-9a55-b6805f555c78] Swapping old allocation on dict_keys(['a2858be1-fd22-4e08-979e-87ad25293407']) held by migration 9dbaf4c6-2a66-44c7-ac3c-b6fb57d4d009 for instance [ 1111.347877] env[61806]: DEBUG nova.scheduler.client.report [None req-705f3985-4e14-42b8-a840-12f7c3a060f8 tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Overwriting current allocation {'allocations': {'a2858be1-fd22-4e08-979e-87ad25293407': {'resources': {'DISK_GB': 1, 'MEMORY_MB': 256, 'VCPU': 1}, 'generation': 140}}, 'project_id': 'c1f2d161c50f4b70987b90e6364c3779', 'user_id': '921aeb4e90e64f89888a9c65f873d66b', 'consumer_generation': 1} on consumer 06658134-4c6d-4911-9a55-b6805f555c78 {{(pid=61806) move_allocations /opt/stack/nova/nova/scheduler/client/report.py:2032}} [ 1111.420837] env[61806]: DEBUG oslo_concurrency.lockutils [None req-705f3985-4e14-42b8-a840-12f7c3a060f8 tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Acquiring lock "refresh_cache-06658134-4c6d-4911-9a55-b6805f555c78" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1111.421052] env[61806]: DEBUG oslo_concurrency.lockutils [None req-705f3985-4e14-42b8-a840-12f7c3a060f8 tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Acquired lock "refresh_cache-06658134-4c6d-4911-9a55-b6805f555c78" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1111.421241] env[61806]: DEBUG nova.network.neutron [None req-705f3985-4e14-42b8-a840-12f7c3a060f8 tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] [instance: 06658134-4c6d-4911-9a55-b6805f555c78] Building network info cache for instance {{(pid=61806) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1111.441509] env[61806]: DEBUG nova.scheduler.client.report [None req-77073341-4390-4aa8-b9e9-ac30959d4690 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Inventory has not changed for provider a2858be1-fd22-4e08-979e-87ad25293407 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 138, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61806) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1111.946526] env[61806]: DEBUG oslo_concurrency.lockutils [None req-77073341-4390-4aa8-b9e9-ac30959d4690 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: held 2.158s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1111.947020] env[61806]: INFO nova.compute.manager [None req-77073341-4390-4aa8-b9e9-ac30959d4690 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] [instance: d0669232-6f33-4b30-97c2-2e4239af1ad0] Migrating [ 1111.958900] env[61806]: DEBUG oslo_concurrency.lockutils [None req-18d3b39e-bf74-4159-88ef-012a6899869a tempest-AttachVolumeNegativeTest-1683033704 tempest-AttachVolumeNegativeTest-1683033704-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 4.015s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1111.959222] env[61806]: DEBUG nova.objects.instance [None req-18d3b39e-bf74-4159-88ef-012a6899869a tempest-AttachVolumeNegativeTest-1683033704 tempest-AttachVolumeNegativeTest-1683033704-project-member] Lazy-loading 'resources' on Instance uuid ce2ad097-4c3f-4b55-b018-ac788aa5b662 {{(pid=61806) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1112.147434] env[61806]: DEBUG nova.network.neutron [None req-705f3985-4e14-42b8-a840-12f7c3a060f8 tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] [instance: 06658134-4c6d-4911-9a55-b6805f555c78] Updating instance_info_cache with network_info: [{"id": "783dae1f-d1c0-4958-b039-ed57b202b64a", "address": "fa:16:3e:89:24:35", "network": {"id": "746a03b6-c1a6-4d27-98b9-0bee6376634f", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-1971382587-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.245", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c1f2d161c50f4b70987b90e6364c3779", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6a31b224-77d7-48a4-af87-312758611b8e", "external-id": "nsx-vlan-transportzone-761", "segmentation_id": 761, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap783dae1f-d1", "ovs_interfaceid": "783dae1f-d1c0-4958-b039-ed57b202b64a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61806) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1112.465871] env[61806]: DEBUG oslo_concurrency.lockutils [None req-77073341-4390-4aa8-b9e9-ac30959d4690 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Acquiring lock "refresh_cache-d0669232-6f33-4b30-97c2-2e4239af1ad0" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1112.466085] env[61806]: DEBUG oslo_concurrency.lockutils [None req-77073341-4390-4aa8-b9e9-ac30959d4690 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Acquired lock "refresh_cache-d0669232-6f33-4b30-97c2-2e4239af1ad0" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1112.466272] env[61806]: DEBUG nova.network.neutron [None req-77073341-4390-4aa8-b9e9-ac30959d4690 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] [instance: d0669232-6f33-4b30-97c2-2e4239af1ad0] Building network info cache for instance {{(pid=61806) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1112.526543] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3fe451af-58a6-4218-8681-c769c5b188ef {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1112.534462] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1c6abf77-6850-4860-a6b8-c9ef985bed4f {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1112.565641] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1a85a9d7-574e-4aae-86ae-ee552c926576 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1112.573428] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-87d7479e-18f3-4efa-9be9-cbe5316819f3 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1112.586645] env[61806]: DEBUG nova.compute.provider_tree [None req-18d3b39e-bf74-4159-88ef-012a6899869a tempest-AttachVolumeNegativeTest-1683033704 tempest-AttachVolumeNegativeTest-1683033704-project-member] Inventory has not changed in ProviderTree for provider: a2858be1-fd22-4e08-979e-87ad25293407 {{(pid=61806) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1112.650157] env[61806]: DEBUG oslo_concurrency.lockutils [None req-705f3985-4e14-42b8-a840-12f7c3a060f8 tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Releasing lock "refresh_cache-06658134-4c6d-4911-9a55-b6805f555c78" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1112.650647] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [None req-705f3985-4e14-42b8-a840-12f7c3a060f8 tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] [instance: 06658134-4c6d-4911-9a55-b6805f555c78] Powering off the VM {{(pid=61806) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1112.650933] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-088d2c6b-ded8-473a-b4fa-cca471cb0230 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1112.659397] env[61806]: DEBUG oslo_vmware.api [None req-705f3985-4e14-42b8-a840-12f7c3a060f8 tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Waiting for the task: (returnval){ [ 1112.659397] env[61806]: value = "task-1295247" [ 1112.659397] env[61806]: _type = "Task" [ 1112.659397] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1112.668868] env[61806]: DEBUG oslo_vmware.api [None req-705f3985-4e14-42b8-a840-12f7c3a060f8 tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Task: {'id': task-1295247, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1113.089919] env[61806]: DEBUG nova.scheduler.client.report [None req-18d3b39e-bf74-4159-88ef-012a6899869a tempest-AttachVolumeNegativeTest-1683033704 tempest-AttachVolumeNegativeTest-1683033704-project-member] Inventory has not changed for provider a2858be1-fd22-4e08-979e-87ad25293407 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 138, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61806) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1113.159867] env[61806]: DEBUG nova.network.neutron [None req-77073341-4390-4aa8-b9e9-ac30959d4690 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] [instance: d0669232-6f33-4b30-97c2-2e4239af1ad0] Updating instance_info_cache with network_info: [{"id": "f958e9a3-f8d6-4665-a59f-876aab66edfb", "address": "fa:16:3e:fc:93:23", "network": {"id": "33bf5b21-82b0-432e-ae3d-9c2ca75161d0", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-1294159336-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.187", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "1c55cc0455584fbb8cda88d33b8f8b7e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "5b0fa7a2-ebd9-4788-8904-7bf250ce466c", "external-id": "nsx-vlan-transportzone-669", "segmentation_id": 669, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf958e9a3-f8", "ovs_interfaceid": "f958e9a3-f8d6-4665-a59f-876aab66edfb", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61806) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1113.170333] env[61806]: DEBUG oslo_vmware.api [None req-705f3985-4e14-42b8-a840-12f7c3a060f8 tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Task: {'id': task-1295247, 'name': PowerOffVM_Task, 'duration_secs': 0.231752} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1113.170587] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [None req-705f3985-4e14-42b8-a840-12f7c3a060f8 tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] [instance: 06658134-4c6d-4911-9a55-b6805f555c78] Powered off the VM {{(pid=61806) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1113.171267] env[61806]: DEBUG nova.virt.hardware [None req-705f3985-4e14-42b8-a840-12f7c3a060f8 tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-15T18:05:10Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=61806) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1113.171496] env[61806]: DEBUG nova.virt.hardware [None req-705f3985-4e14-42b8-a840-12f7c3a060f8 tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Flavor limits 0:0:0 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1113.171659] env[61806]: DEBUG nova.virt.hardware [None req-705f3985-4e14-42b8-a840-12f7c3a060f8 tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Image limits 0:0:0 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1113.171850] env[61806]: DEBUG nova.virt.hardware [None req-705f3985-4e14-42b8-a840-12f7c3a060f8 tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Flavor pref 0:0:0 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1113.172009] env[61806]: DEBUG nova.virt.hardware [None req-705f3985-4e14-42b8-a840-12f7c3a060f8 tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Image pref 0:0:0 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1113.172174] env[61806]: DEBUG nova.virt.hardware [None req-705f3985-4e14-42b8-a840-12f7c3a060f8 tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1113.172381] env[61806]: DEBUG nova.virt.hardware [None req-705f3985-4e14-42b8-a840-12f7c3a060f8 tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61806) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1113.172541] env[61806]: DEBUG nova.virt.hardware [None req-705f3985-4e14-42b8-a840-12f7c3a060f8 tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61806) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1113.172707] env[61806]: DEBUG nova.virt.hardware [None req-705f3985-4e14-42b8-a840-12f7c3a060f8 tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Got 1 possible topologies {{(pid=61806) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1113.172872] env[61806]: DEBUG nova.virt.hardware [None req-705f3985-4e14-42b8-a840-12f7c3a060f8 tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61806) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1113.173061] env[61806]: DEBUG nova.virt.hardware [None req-705f3985-4e14-42b8-a840-12f7c3a060f8 tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61806) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1113.178147] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-ba584788-8421-441a-8262-927892cbaeb3 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1113.193438] env[61806]: DEBUG oslo_vmware.api [None req-705f3985-4e14-42b8-a840-12f7c3a060f8 tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Waiting for the task: (returnval){ [ 1113.193438] env[61806]: value = "task-1295248" [ 1113.193438] env[61806]: _type = "Task" [ 1113.193438] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1113.201138] env[61806]: DEBUG oslo_vmware.api [None req-705f3985-4e14-42b8-a840-12f7c3a060f8 tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Task: {'id': task-1295248, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1113.596626] env[61806]: DEBUG oslo_concurrency.lockutils [None req-18d3b39e-bf74-4159-88ef-012a6899869a tempest-AttachVolumeNegativeTest-1683033704 tempest-AttachVolumeNegativeTest-1683033704-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.638s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1113.612932] env[61806]: INFO nova.scheduler.client.report [None req-18d3b39e-bf74-4159-88ef-012a6899869a tempest-AttachVolumeNegativeTest-1683033704 tempest-AttachVolumeNegativeTest-1683033704-project-member] Deleted allocations for instance ce2ad097-4c3f-4b55-b018-ac788aa5b662 [ 1113.664682] env[61806]: DEBUG oslo_concurrency.lockutils [None req-77073341-4390-4aa8-b9e9-ac30959d4690 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Releasing lock "refresh_cache-d0669232-6f33-4b30-97c2-2e4239af1ad0" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1113.704630] env[61806]: DEBUG oslo_vmware.api [None req-705f3985-4e14-42b8-a840-12f7c3a060f8 tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Task: {'id': task-1295248, 'name': ReconfigVM_Task, 'duration_secs': 0.14107} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1113.705453] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c63af92a-a95a-4f6b-aa16-6dec7d09d207 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1113.723109] env[61806]: DEBUG nova.virt.hardware [None req-705f3985-4e14-42b8-a840-12f7c3a060f8 tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-15T18:05:10Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=61806) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1113.723394] env[61806]: DEBUG nova.virt.hardware [None req-705f3985-4e14-42b8-a840-12f7c3a060f8 tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Flavor limits 0:0:0 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1113.723565] env[61806]: DEBUG nova.virt.hardware [None req-705f3985-4e14-42b8-a840-12f7c3a060f8 tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Image limits 0:0:0 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1113.723752] env[61806]: DEBUG nova.virt.hardware [None req-705f3985-4e14-42b8-a840-12f7c3a060f8 tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Flavor pref 0:0:0 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1113.723906] env[61806]: DEBUG nova.virt.hardware [None req-705f3985-4e14-42b8-a840-12f7c3a060f8 tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Image pref 0:0:0 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1113.724073] env[61806]: DEBUG nova.virt.hardware [None req-705f3985-4e14-42b8-a840-12f7c3a060f8 tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1113.724288] env[61806]: DEBUG nova.virt.hardware [None req-705f3985-4e14-42b8-a840-12f7c3a060f8 tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61806) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1113.724484] env[61806]: DEBUG nova.virt.hardware [None req-705f3985-4e14-42b8-a840-12f7c3a060f8 tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61806) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1113.724665] env[61806]: DEBUG nova.virt.hardware [None req-705f3985-4e14-42b8-a840-12f7c3a060f8 tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Got 1 possible topologies {{(pid=61806) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1113.724833] env[61806]: DEBUG nova.virt.hardware [None req-705f3985-4e14-42b8-a840-12f7c3a060f8 tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61806) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1113.725028] env[61806]: DEBUG nova.virt.hardware [None req-705f3985-4e14-42b8-a840-12f7c3a060f8 tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61806) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1113.725748] env[61806]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-25a2392b-de86-481b-8130-e3a3a91bf19a {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1113.730871] env[61806]: DEBUG oslo_vmware.api [None req-705f3985-4e14-42b8-a840-12f7c3a060f8 tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Waiting for the task: (returnval){ [ 1113.730871] env[61806]: value = "session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]52db3ae8-b221-f086-51cc-e81bc38f1f22" [ 1113.730871] env[61806]: _type = "Task" [ 1113.730871] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1113.738137] env[61806]: DEBUG oslo_vmware.api [None req-705f3985-4e14-42b8-a840-12f7c3a060f8 tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Task: {'id': session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]52db3ae8-b221-f086-51cc-e81bc38f1f22, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1114.121559] env[61806]: DEBUG oslo_concurrency.lockutils [None req-18d3b39e-bf74-4159-88ef-012a6899869a tempest-AttachVolumeNegativeTest-1683033704 tempest-AttachVolumeNegativeTest-1683033704-project-member] Lock "ce2ad097-4c3f-4b55-b018-ac788aa5b662" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 9.439s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1114.241250] env[61806]: DEBUG oslo_vmware.api [None req-705f3985-4e14-42b8-a840-12f7c3a060f8 tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Task: {'id': session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]52db3ae8-b221-f086-51cc-e81bc38f1f22, 'name': SearchDatastore_Task, 'duration_secs': 0.006225} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1114.246544] env[61806]: DEBUG nova.virt.vmwareapi.volumeops [None req-705f3985-4e14-42b8-a840-12f7c3a060f8 tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] [instance: 06658134-4c6d-4911-9a55-b6805f555c78] Reconfiguring VM instance instance-00000066 to detach disk 2000 {{(pid=61806) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1114.247138] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-d743f3aa-8bbb-4d0f-b778-5cebe629452b {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1114.265823] env[61806]: DEBUG oslo_vmware.api [None req-705f3985-4e14-42b8-a840-12f7c3a060f8 tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Waiting for the task: (returnval){ [ 1114.265823] env[61806]: value = "task-1295249" [ 1114.265823] env[61806]: _type = "Task" [ 1114.265823] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1114.273999] env[61806]: DEBUG oslo_vmware.api [None req-705f3985-4e14-42b8-a840-12f7c3a060f8 tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Task: {'id': task-1295249, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1114.775991] env[61806]: DEBUG oslo_vmware.api [None req-705f3985-4e14-42b8-a840-12f7c3a060f8 tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Task: {'id': task-1295249, 'name': ReconfigVM_Task, 'duration_secs': 0.174511} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1114.776363] env[61806]: DEBUG nova.virt.vmwareapi.volumeops [None req-705f3985-4e14-42b8-a840-12f7c3a060f8 tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] [instance: 06658134-4c6d-4911-9a55-b6805f555c78] Reconfigured VM instance instance-00000066 to detach disk 2000 {{(pid=61806) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1114.777147] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7721b508-4de9-4923-8b6c-814569e71cd7 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1114.798548] env[61806]: DEBUG nova.virt.vmwareapi.volumeops [None req-705f3985-4e14-42b8-a840-12f7c3a060f8 tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] [instance: 06658134-4c6d-4911-9a55-b6805f555c78] Reconfiguring VM instance instance-00000066 to attach disk [datastore2] 06658134-4c6d-4911-9a55-b6805f555c78/06658134-4c6d-4911-9a55-b6805f555c78.vmdk or device None with type thin {{(pid=61806) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1114.798798] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-7f59ca72-77dd-4868-b8e4-a5ed4341d381 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1114.816254] env[61806]: DEBUG oslo_vmware.api [None req-705f3985-4e14-42b8-a840-12f7c3a060f8 tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Waiting for the task: (returnval){ [ 1114.816254] env[61806]: value = "task-1295251" [ 1114.816254] env[61806]: _type = "Task" [ 1114.816254] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1114.823545] env[61806]: DEBUG oslo_vmware.api [None req-705f3985-4e14-42b8-a840-12f7c3a060f8 tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Task: {'id': task-1295251, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1115.180057] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5c772916-0639-4165-8aac-4b7a8e78b07a {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1115.199739] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-77073341-4390-4aa8-b9e9-ac30959d4690 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] [instance: d0669232-6f33-4b30-97c2-2e4239af1ad0] Updating instance 'd0669232-6f33-4b30-97c2-2e4239af1ad0' progress to 0 {{(pid=61806) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 1115.325875] env[61806]: DEBUG oslo_vmware.api [None req-705f3985-4e14-42b8-a840-12f7c3a060f8 tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Task: {'id': task-1295251, 'name': ReconfigVM_Task, 'duration_secs': 0.257908} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1115.326211] env[61806]: DEBUG nova.virt.vmwareapi.volumeops [None req-705f3985-4e14-42b8-a840-12f7c3a060f8 tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] [instance: 06658134-4c6d-4911-9a55-b6805f555c78] Reconfigured VM instance instance-00000066 to attach disk [datastore2] 06658134-4c6d-4911-9a55-b6805f555c78/06658134-4c6d-4911-9a55-b6805f555c78.vmdk or device None with type thin {{(pid=61806) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1115.327036] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-514fbc76-5ed5-4f64-a5b1-3ba9cbed70ab {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1115.344206] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6a4c8c33-2c8d-4ee8-b65b-8ef1a175aa3c {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1115.361399] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e8ea7c3e-67b6-40af-a338-aea48899118b {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1115.378707] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-289c81e8-bc20-49ca-a7f9-cd8a4a050a51 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1115.384742] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [None req-705f3985-4e14-42b8-a840-12f7c3a060f8 tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] [instance: 06658134-4c6d-4911-9a55-b6805f555c78] Powering on the VM {{(pid=61806) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1115.384964] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-69070061-118f-494d-8802-5ee216c35506 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1115.390172] env[61806]: DEBUG oslo_vmware.api [None req-705f3985-4e14-42b8-a840-12f7c3a060f8 tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Waiting for the task: (returnval){ [ 1115.390172] env[61806]: value = "task-1295252" [ 1115.390172] env[61806]: _type = "Task" [ 1115.390172] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1115.396934] env[61806]: DEBUG oslo_vmware.api [None req-705f3985-4e14-42b8-a840-12f7c3a060f8 tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Task: {'id': task-1295252, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1115.706115] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [None req-77073341-4390-4aa8-b9e9-ac30959d4690 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] [instance: d0669232-6f33-4b30-97c2-2e4239af1ad0] Powering off the VM {{(pid=61806) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1115.706520] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-3225b9de-d3bf-435c-87a5-a790591e7552 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1115.714837] env[61806]: DEBUG oslo_vmware.api [None req-77073341-4390-4aa8-b9e9-ac30959d4690 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Waiting for the task: (returnval){ [ 1115.714837] env[61806]: value = "task-1295253" [ 1115.714837] env[61806]: _type = "Task" [ 1115.714837] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1115.725348] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [None req-77073341-4390-4aa8-b9e9-ac30959d4690 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] [instance: d0669232-6f33-4b30-97c2-2e4239af1ad0] VM already powered off {{(pid=61806) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1509}} [ 1115.725657] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-77073341-4390-4aa8-b9e9-ac30959d4690 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] [instance: d0669232-6f33-4b30-97c2-2e4239af1ad0] Updating instance 'd0669232-6f33-4b30-97c2-2e4239af1ad0' progress to 17 {{(pid=61806) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 1115.899808] env[61806]: DEBUG oslo_vmware.api [None req-705f3985-4e14-42b8-a840-12f7c3a060f8 tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Task: {'id': task-1295252, 'name': PowerOnVM_Task, 'duration_secs': 0.362087} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1115.900090] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [None req-705f3985-4e14-42b8-a840-12f7c3a060f8 tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] [instance: 06658134-4c6d-4911-9a55-b6805f555c78] Powered on the VM {{(pid=61806) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1116.232216] env[61806]: DEBUG nova.virt.hardware [None req-77073341-4390-4aa8-b9e9-ac30959d4690 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-15T18:05:12Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='84',id=12,is_public=True,memory_mb=256,name='m1.micro',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=61806) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1116.232578] env[61806]: DEBUG nova.virt.hardware [None req-77073341-4390-4aa8-b9e9-ac30959d4690 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Flavor limits 0:0:0 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1116.232578] env[61806]: DEBUG nova.virt.hardware [None req-77073341-4390-4aa8-b9e9-ac30959d4690 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Image limits 0:0:0 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1116.232737] env[61806]: DEBUG nova.virt.hardware [None req-77073341-4390-4aa8-b9e9-ac30959d4690 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Flavor pref 0:0:0 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1116.232862] env[61806]: DEBUG nova.virt.hardware [None req-77073341-4390-4aa8-b9e9-ac30959d4690 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Image pref 0:0:0 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1116.233018] env[61806]: DEBUG nova.virt.hardware [None req-77073341-4390-4aa8-b9e9-ac30959d4690 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1116.233234] env[61806]: DEBUG nova.virt.hardware [None req-77073341-4390-4aa8-b9e9-ac30959d4690 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61806) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1116.233399] env[61806]: DEBUG nova.virt.hardware [None req-77073341-4390-4aa8-b9e9-ac30959d4690 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61806) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1116.233569] env[61806]: DEBUG nova.virt.hardware [None req-77073341-4390-4aa8-b9e9-ac30959d4690 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Got 1 possible topologies {{(pid=61806) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1116.233737] env[61806]: DEBUG nova.virt.hardware [None req-77073341-4390-4aa8-b9e9-ac30959d4690 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61806) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1116.233944] env[61806]: DEBUG nova.virt.hardware [None req-77073341-4390-4aa8-b9e9-ac30959d4690 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61806) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1116.239168] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-1f1a7ce9-bcef-478e-95e4-48238de86977 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1116.254985] env[61806]: DEBUG oslo_vmware.api [None req-77073341-4390-4aa8-b9e9-ac30959d4690 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Waiting for the task: (returnval){ [ 1116.254985] env[61806]: value = "task-1295254" [ 1116.254985] env[61806]: _type = "Task" [ 1116.254985] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1116.264712] env[61806]: DEBUG oslo_vmware.api [None req-77073341-4390-4aa8-b9e9-ac30959d4690 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Task: {'id': task-1295254, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1116.764750] env[61806]: DEBUG oslo_vmware.api [None req-77073341-4390-4aa8-b9e9-ac30959d4690 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Task: {'id': task-1295254, 'name': ReconfigVM_Task, 'duration_secs': 0.144325} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1116.765081] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-77073341-4390-4aa8-b9e9-ac30959d4690 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] [instance: d0669232-6f33-4b30-97c2-2e4239af1ad0] Updating instance 'd0669232-6f33-4b30-97c2-2e4239af1ad0' progress to 33 {{(pid=61806) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 1116.910693] env[61806]: INFO nova.compute.manager [None req-705f3985-4e14-42b8-a840-12f7c3a060f8 tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] [instance: 06658134-4c6d-4911-9a55-b6805f555c78] Updating instance to original state: 'active' [ 1117.271126] env[61806]: DEBUG nova.virt.hardware [None req-77073341-4390-4aa8-b9e9-ac30959d4690 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-15T18:05:10Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=61806) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1117.271429] env[61806]: DEBUG nova.virt.hardware [None req-77073341-4390-4aa8-b9e9-ac30959d4690 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Flavor limits 0:0:0 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1117.271787] env[61806]: DEBUG nova.virt.hardware [None req-77073341-4390-4aa8-b9e9-ac30959d4690 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Image limits 0:0:0 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1117.271787] env[61806]: DEBUG nova.virt.hardware [None req-77073341-4390-4aa8-b9e9-ac30959d4690 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Flavor pref 0:0:0 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1117.271988] env[61806]: DEBUG nova.virt.hardware [None req-77073341-4390-4aa8-b9e9-ac30959d4690 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Image pref 0:0:0 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1117.272182] env[61806]: DEBUG nova.virt.hardware [None req-77073341-4390-4aa8-b9e9-ac30959d4690 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1117.272395] env[61806]: DEBUG nova.virt.hardware [None req-77073341-4390-4aa8-b9e9-ac30959d4690 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61806) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1117.272561] env[61806]: DEBUG nova.virt.hardware [None req-77073341-4390-4aa8-b9e9-ac30959d4690 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61806) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1117.272735] env[61806]: DEBUG nova.virt.hardware [None req-77073341-4390-4aa8-b9e9-ac30959d4690 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Got 1 possible topologies {{(pid=61806) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1117.272904] env[61806]: DEBUG nova.virt.hardware [None req-77073341-4390-4aa8-b9e9-ac30959d4690 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61806) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1117.273100] env[61806]: DEBUG nova.virt.hardware [None req-77073341-4390-4aa8-b9e9-ac30959d4690 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61806) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1117.278585] env[61806]: DEBUG nova.virt.vmwareapi.volumeops [None req-77073341-4390-4aa8-b9e9-ac30959d4690 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] [instance: d0669232-6f33-4b30-97c2-2e4239af1ad0] Reconfiguring VM instance instance-00000063 to detach disk 2000 {{(pid=61806) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1117.278883] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-1a00e09c-2659-46a7-970c-160e45d55efb {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1117.298675] env[61806]: DEBUG oslo_vmware.api [None req-77073341-4390-4aa8-b9e9-ac30959d4690 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Waiting for the task: (returnval){ [ 1117.298675] env[61806]: value = "task-1295255" [ 1117.298675] env[61806]: _type = "Task" [ 1117.298675] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1117.306261] env[61806]: DEBUG oslo_vmware.api [None req-77073341-4390-4aa8-b9e9-ac30959d4690 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Task: {'id': task-1295255, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1117.808579] env[61806]: DEBUG oslo_vmware.api [None req-77073341-4390-4aa8-b9e9-ac30959d4690 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Task: {'id': task-1295255, 'name': ReconfigVM_Task, 'duration_secs': 0.15303} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1117.808856] env[61806]: DEBUG nova.virt.vmwareapi.volumeops [None req-77073341-4390-4aa8-b9e9-ac30959d4690 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] [instance: d0669232-6f33-4b30-97c2-2e4239af1ad0] Reconfigured VM instance instance-00000063 to detach disk 2000 {{(pid=61806) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1117.809623] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d90d9e75-67c0-407a-8bea-3707b85086cf {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1117.830548] env[61806]: DEBUG nova.virt.vmwareapi.volumeops [None req-77073341-4390-4aa8-b9e9-ac30959d4690 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] [instance: d0669232-6f33-4b30-97c2-2e4239af1ad0] Reconfiguring VM instance instance-00000063 to attach disk [datastore2] d0669232-6f33-4b30-97c2-2e4239af1ad0/d0669232-6f33-4b30-97c2-2e4239af1ad0.vmdk or device None with type thin {{(pid=61806) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1117.830790] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-ba1e5439-8612-443b-8807-d2337927de3f {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1117.847702] env[61806]: DEBUG oslo_vmware.api [None req-77073341-4390-4aa8-b9e9-ac30959d4690 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Waiting for the task: (returnval){ [ 1117.847702] env[61806]: value = "task-1295257" [ 1117.847702] env[61806]: _type = "Task" [ 1117.847702] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1117.854979] env[61806]: DEBUG oslo_vmware.api [None req-77073341-4390-4aa8-b9e9-ac30959d4690 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Task: {'id': task-1295257, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1117.958402] env[61806]: DEBUG oslo_concurrency.lockutils [None req-75c8929d-8206-4645-ac5b-f11d5b220500 tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Acquiring lock "06658134-4c6d-4911-9a55-b6805f555c78" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1117.959070] env[61806]: DEBUG oslo_concurrency.lockutils [None req-75c8929d-8206-4645-ac5b-f11d5b220500 tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Lock "06658134-4c6d-4911-9a55-b6805f555c78" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1117.959070] env[61806]: DEBUG oslo_concurrency.lockutils [None req-75c8929d-8206-4645-ac5b-f11d5b220500 tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Acquiring lock "06658134-4c6d-4911-9a55-b6805f555c78-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1117.959242] env[61806]: DEBUG oslo_concurrency.lockutils [None req-75c8929d-8206-4645-ac5b-f11d5b220500 tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Lock "06658134-4c6d-4911-9a55-b6805f555c78-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1117.959468] env[61806]: DEBUG oslo_concurrency.lockutils [None req-75c8929d-8206-4645-ac5b-f11d5b220500 tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Lock "06658134-4c6d-4911-9a55-b6805f555c78-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1117.962956] env[61806]: INFO nova.compute.manager [None req-75c8929d-8206-4645-ac5b-f11d5b220500 tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] [instance: 06658134-4c6d-4911-9a55-b6805f555c78] Terminating instance [ 1117.965121] env[61806]: DEBUG nova.compute.manager [None req-75c8929d-8206-4645-ac5b-f11d5b220500 tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] [instance: 06658134-4c6d-4911-9a55-b6805f555c78] Start destroying the instance on the hypervisor. {{(pid=61806) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1117.965397] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-75c8929d-8206-4645-ac5b-f11d5b220500 tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] [instance: 06658134-4c6d-4911-9a55-b6805f555c78] Destroying instance {{(pid=61806) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1117.966417] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-81a3c9ad-8c82-4f39-8a10-b41f35015fe3 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1117.974360] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [None req-75c8929d-8206-4645-ac5b-f11d5b220500 tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] [instance: 06658134-4c6d-4911-9a55-b6805f555c78] Powering off the VM {{(pid=61806) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1117.974590] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-a7ec8d5d-65e5-493a-9325-18baed6d20c7 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1117.980722] env[61806]: DEBUG oslo_vmware.api [None req-75c8929d-8206-4645-ac5b-f11d5b220500 tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Waiting for the task: (returnval){ [ 1117.980722] env[61806]: value = "task-1295258" [ 1117.980722] env[61806]: _type = "Task" [ 1117.980722] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1117.989693] env[61806]: DEBUG oslo_vmware.api [None req-75c8929d-8206-4645-ac5b-f11d5b220500 tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Task: {'id': task-1295258, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1118.357784] env[61806]: DEBUG oslo_vmware.api [None req-77073341-4390-4aa8-b9e9-ac30959d4690 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Task: {'id': task-1295257, 'name': ReconfigVM_Task, 'duration_secs': 0.270799} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1118.358124] env[61806]: DEBUG nova.virt.vmwareapi.volumeops [None req-77073341-4390-4aa8-b9e9-ac30959d4690 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] [instance: d0669232-6f33-4b30-97c2-2e4239af1ad0] Reconfigured VM instance instance-00000063 to attach disk [datastore2] d0669232-6f33-4b30-97c2-2e4239af1ad0/d0669232-6f33-4b30-97c2-2e4239af1ad0.vmdk or device None with type thin {{(pid=61806) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1118.358353] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-77073341-4390-4aa8-b9e9-ac30959d4690 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] [instance: d0669232-6f33-4b30-97c2-2e4239af1ad0] Updating instance 'd0669232-6f33-4b30-97c2-2e4239af1ad0' progress to 50 {{(pid=61806) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 1118.490723] env[61806]: DEBUG oslo_vmware.api [None req-75c8929d-8206-4645-ac5b-f11d5b220500 tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Task: {'id': task-1295258, 'name': PowerOffVM_Task, 'duration_secs': 0.185115} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1118.490995] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [None req-75c8929d-8206-4645-ac5b-f11d5b220500 tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] [instance: 06658134-4c6d-4911-9a55-b6805f555c78] Powered off the VM {{(pid=61806) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1118.491188] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-75c8929d-8206-4645-ac5b-f11d5b220500 tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] [instance: 06658134-4c6d-4911-9a55-b6805f555c78] Unregistering the VM {{(pid=61806) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1118.491442] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-8112002f-aa73-4327-8db3-1d46d5d583e0 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1118.554539] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-75c8929d-8206-4645-ac5b-f11d5b220500 tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] [instance: 06658134-4c6d-4911-9a55-b6805f555c78] Unregistered the VM {{(pid=61806) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1118.554713] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-75c8929d-8206-4645-ac5b-f11d5b220500 tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] [instance: 06658134-4c6d-4911-9a55-b6805f555c78] Deleting contents of the VM from datastore datastore2 {{(pid=61806) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1118.554899] env[61806]: DEBUG nova.virt.vmwareapi.ds_util [None req-75c8929d-8206-4645-ac5b-f11d5b220500 tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Deleting the datastore file [datastore2] 06658134-4c6d-4911-9a55-b6805f555c78 {{(pid=61806) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1118.555182] env[61806]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-279bfbf9-5ae4-47b7-85f9-b1a7cfbd9ad0 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1118.561890] env[61806]: DEBUG oslo_vmware.api [None req-75c8929d-8206-4645-ac5b-f11d5b220500 tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Waiting for the task: (returnval){ [ 1118.561890] env[61806]: value = "task-1295260" [ 1118.561890] env[61806]: _type = "Task" [ 1118.561890] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1118.569674] env[61806]: DEBUG oslo_vmware.api [None req-75c8929d-8206-4645-ac5b-f11d5b220500 tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Task: {'id': task-1295260, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1118.865109] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ccb1ec64-0e30-4e32-96c3-65961eebb9dd {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1118.883685] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dbdfa45a-bb4b-4dd5-8114-26e3806f2624 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1118.900383] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-77073341-4390-4aa8-b9e9-ac30959d4690 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] [instance: d0669232-6f33-4b30-97c2-2e4239af1ad0] Updating instance 'd0669232-6f33-4b30-97c2-2e4239af1ad0' progress to 67 {{(pid=61806) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 1119.072348] env[61806]: DEBUG oslo_vmware.api [None req-75c8929d-8206-4645-ac5b-f11d5b220500 tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Task: {'id': task-1295260, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.146667} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1119.072676] env[61806]: DEBUG nova.virt.vmwareapi.ds_util [None req-75c8929d-8206-4645-ac5b-f11d5b220500 tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Deleted the datastore file {{(pid=61806) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1119.072871] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-75c8929d-8206-4645-ac5b-f11d5b220500 tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] [instance: 06658134-4c6d-4911-9a55-b6805f555c78] Deleted contents of the VM from datastore datastore2 {{(pid=61806) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1119.073068] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-75c8929d-8206-4645-ac5b-f11d5b220500 tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] [instance: 06658134-4c6d-4911-9a55-b6805f555c78] Instance destroyed {{(pid=61806) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1119.073251] env[61806]: INFO nova.compute.manager [None req-75c8929d-8206-4645-ac5b-f11d5b220500 tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] [instance: 06658134-4c6d-4911-9a55-b6805f555c78] Took 1.11 seconds to destroy the instance on the hypervisor. [ 1119.073493] env[61806]: DEBUG oslo.service.loopingcall [None req-75c8929d-8206-4645-ac5b-f11d5b220500 tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61806) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1119.073686] env[61806]: DEBUG nova.compute.manager [-] [instance: 06658134-4c6d-4911-9a55-b6805f555c78] Deallocating network for instance {{(pid=61806) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1119.073782] env[61806]: DEBUG nova.network.neutron [-] [instance: 06658134-4c6d-4911-9a55-b6805f555c78] deallocate_for_instance() {{(pid=61806) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1119.463396] env[61806]: DEBUG nova.network.neutron [None req-77073341-4390-4aa8-b9e9-ac30959d4690 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] [instance: d0669232-6f33-4b30-97c2-2e4239af1ad0] Port f958e9a3-f8d6-4665-a59f-876aab66edfb binding to destination host cpu-1 is already ACTIVE {{(pid=61806) migrate_instance_start /opt/stack/nova/nova/network/neutron.py:3171}} [ 1119.564220] env[61806]: DEBUG nova.compute.manager [req-5338c47f-8566-403d-992e-96e5d6596c13 req-7dd35e1e-1dbe-4b56-94cc-6dd77ac16fd4 service nova] [instance: 06658134-4c6d-4911-9a55-b6805f555c78] Received event network-vif-deleted-783dae1f-d1c0-4958-b039-ed57b202b64a {{(pid=61806) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1119.564482] env[61806]: INFO nova.compute.manager [req-5338c47f-8566-403d-992e-96e5d6596c13 req-7dd35e1e-1dbe-4b56-94cc-6dd77ac16fd4 service nova] [instance: 06658134-4c6d-4911-9a55-b6805f555c78] Neutron deleted interface 783dae1f-d1c0-4958-b039-ed57b202b64a; detaching it from the instance and deleting it from the info cache [ 1119.564701] env[61806]: DEBUG nova.network.neutron [req-5338c47f-8566-403d-992e-96e5d6596c13 req-7dd35e1e-1dbe-4b56-94cc-6dd77ac16fd4 service nova] [instance: 06658134-4c6d-4911-9a55-b6805f555c78] Updating instance_info_cache with network_info: [] {{(pid=61806) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1120.041222] env[61806]: DEBUG nova.network.neutron [-] [instance: 06658134-4c6d-4911-9a55-b6805f555c78] Updating instance_info_cache with network_info: [] {{(pid=61806) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1120.067210] env[61806]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-786195bb-fd6f-4de1-8d09-b58a6cc07420 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1120.076410] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a52e43eb-19f2-4881-95f0-f6a2262fa0c1 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1120.100375] env[61806]: DEBUG nova.compute.manager [req-5338c47f-8566-403d-992e-96e5d6596c13 req-7dd35e1e-1dbe-4b56-94cc-6dd77ac16fd4 service nova] [instance: 06658134-4c6d-4911-9a55-b6805f555c78] Detach interface failed, port_id=783dae1f-d1c0-4958-b039-ed57b202b64a, reason: Instance 06658134-4c6d-4911-9a55-b6805f555c78 could not be found. {{(pid=61806) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10963}} [ 1120.487174] env[61806]: DEBUG oslo_concurrency.lockutils [None req-77073341-4390-4aa8-b9e9-ac30959d4690 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Acquiring lock "d0669232-6f33-4b30-97c2-2e4239af1ad0-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1120.487434] env[61806]: DEBUG oslo_concurrency.lockutils [None req-77073341-4390-4aa8-b9e9-ac30959d4690 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Lock "d0669232-6f33-4b30-97c2-2e4239af1ad0-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1120.487634] env[61806]: DEBUG oslo_concurrency.lockutils [None req-77073341-4390-4aa8-b9e9-ac30959d4690 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Lock "d0669232-6f33-4b30-97c2-2e4239af1ad0-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1120.544048] env[61806]: INFO nova.compute.manager [-] [instance: 06658134-4c6d-4911-9a55-b6805f555c78] Took 1.47 seconds to deallocate network for instance. [ 1120.580523] env[61806]: DEBUG oslo_service.periodic_task [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Running periodic task ComputeManager._sync_power_states {{(pid=61806) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1121.050763] env[61806]: DEBUG oslo_concurrency.lockutils [None req-75c8929d-8206-4645-ac5b-f11d5b220500 tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1121.050982] env[61806]: DEBUG oslo_concurrency.lockutils [None req-75c8929d-8206-4645-ac5b-f11d5b220500 tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.000s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1121.051195] env[61806]: DEBUG oslo_concurrency.lockutils [None req-75c8929d-8206-4645-ac5b-f11d5b220500 tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1121.077286] env[61806]: INFO nova.scheduler.client.report [None req-75c8929d-8206-4645-ac5b-f11d5b220500 tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Deleted allocations for instance 06658134-4c6d-4911-9a55-b6805f555c78 [ 1121.084078] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Getting list of instances from cluster (obj){ [ 1121.084078] env[61806]: value = "domain-c8" [ 1121.084078] env[61806]: _type = "ClusterComputeResource" [ 1121.084078] env[61806]: } {{(pid=61806) list_instances /opt/stack/nova/nova/virt/vmwareapi/vmops.py:2122}} [ 1121.085096] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8337605a-fde0-411e-b599-3a86317f27d8 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1121.095325] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Got total of 2 instances {{(pid=61806) list_instances /opt/stack/nova/nova/virt/vmwareapi/vmops.py:2131}} [ 1121.095472] env[61806]: WARNING nova.compute.manager [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] While synchronizing instance power states, found 3 instances in the database and 2 instances on the hypervisor. [ 1121.095612] env[61806]: DEBUG nova.compute.manager [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Triggering sync for uuid d0669232-6f33-4b30-97c2-2e4239af1ad0 {{(pid=61806) _sync_power_states /opt/stack/nova/nova/compute/manager.py:10336}} [ 1121.095794] env[61806]: DEBUG nova.compute.manager [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Triggering sync for uuid 06658134-4c6d-4911-9a55-b6805f555c78 {{(pid=61806) _sync_power_states /opt/stack/nova/nova/compute/manager.py:10336}} [ 1121.095947] env[61806]: DEBUG nova.compute.manager [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Triggering sync for uuid 29210bcd-0fe1-4ed9-a459-abc8016a4255 {{(pid=61806) _sync_power_states /opt/stack/nova/nova/compute/manager.py:10336}} [ 1121.096285] env[61806]: DEBUG oslo_concurrency.lockutils [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Acquiring lock "d0669232-6f33-4b30-97c2-2e4239af1ad0" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1121.096503] env[61806]: DEBUG oslo_concurrency.lockutils [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Lock "d0669232-6f33-4b30-97c2-2e4239af1ad0" acquired by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: waited 0.000s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1121.096675] env[61806]: INFO nova.compute.manager [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] [instance: d0669232-6f33-4b30-97c2-2e4239af1ad0] During sync_power_state the instance has a pending task (resize_migrated). Skip. [ 1121.096843] env[61806]: DEBUG oslo_concurrency.lockutils [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Lock "d0669232-6f33-4b30-97c2-2e4239af1ad0" "released" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: held 0.000s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1121.097058] env[61806]: DEBUG oslo_concurrency.lockutils [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Acquiring lock "06658134-4c6d-4911-9a55-b6805f555c78" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1121.097313] env[61806]: DEBUG oslo_concurrency.lockutils [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Acquiring lock "29210bcd-0fe1-4ed9-a459-abc8016a4255" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1121.097507] env[61806]: DEBUG oslo_concurrency.lockutils [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Lock "29210bcd-0fe1-4ed9-a459-abc8016a4255" acquired by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: waited 0.000s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1121.098356] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aabf24f1-4e86-4e5c-a1c1-b7b06e7999b3 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1121.520474] env[61806]: DEBUG oslo_concurrency.lockutils [None req-77073341-4390-4aa8-b9e9-ac30959d4690 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Acquiring lock "refresh_cache-d0669232-6f33-4b30-97c2-2e4239af1ad0" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1121.520740] env[61806]: DEBUG oslo_concurrency.lockutils [None req-77073341-4390-4aa8-b9e9-ac30959d4690 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Acquired lock "refresh_cache-d0669232-6f33-4b30-97c2-2e4239af1ad0" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1121.520851] env[61806]: DEBUG nova.network.neutron [None req-77073341-4390-4aa8-b9e9-ac30959d4690 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] [instance: d0669232-6f33-4b30-97c2-2e4239af1ad0] Building network info cache for instance {{(pid=61806) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1121.584375] env[61806]: DEBUG oslo_concurrency.lockutils [None req-75c8929d-8206-4645-ac5b-f11d5b220500 tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Lock "06658134-4c6d-4911-9a55-b6805f555c78" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 3.626s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1121.585286] env[61806]: DEBUG oslo_concurrency.lockutils [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Lock "06658134-4c6d-4911-9a55-b6805f555c78" acquired by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: waited 0.488s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1121.585719] env[61806]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-b10df8b6-178c-4a47-a2c9-a0e020de0793 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1121.597547] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1990a2e7-6bac-4b1b-9dfa-9f982f58a888 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1121.608373] env[61806]: DEBUG oslo_concurrency.lockutils [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Lock "29210bcd-0fe1-4ed9-a459-abc8016a4255" "released" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: held 0.511s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1122.123668] env[61806]: DEBUG oslo_concurrency.lockutils [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Lock "06658134-4c6d-4911-9a55-b6805f555c78" "released" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: held 0.538s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1122.278251] env[61806]: DEBUG nova.network.neutron [None req-77073341-4390-4aa8-b9e9-ac30959d4690 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] [instance: d0669232-6f33-4b30-97c2-2e4239af1ad0] Updating instance_info_cache with network_info: [{"id": "f958e9a3-f8d6-4665-a59f-876aab66edfb", "address": "fa:16:3e:fc:93:23", "network": {"id": "33bf5b21-82b0-432e-ae3d-9c2ca75161d0", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-1294159336-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.187", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "1c55cc0455584fbb8cda88d33b8f8b7e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "5b0fa7a2-ebd9-4788-8904-7bf250ce466c", "external-id": "nsx-vlan-transportzone-669", "segmentation_id": 669, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf958e9a3-f8", "ovs_interfaceid": "f958e9a3-f8d6-4665-a59f-876aab66edfb", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61806) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1122.781104] env[61806]: DEBUG oslo_concurrency.lockutils [None req-77073341-4390-4aa8-b9e9-ac30959d4690 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Releasing lock "refresh_cache-d0669232-6f33-4b30-97c2-2e4239af1ad0" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1122.905440] env[61806]: DEBUG oslo_concurrency.lockutils [None req-bf8f4ec5-63a0-4ddc-bc14-7d6705c06bed tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Acquiring lock "c66fe2e7-7af0-48fa-8509-fcc24c1a437b" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1122.905699] env[61806]: DEBUG oslo_concurrency.lockutils [None req-bf8f4ec5-63a0-4ddc-bc14-7d6705c06bed tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Lock "c66fe2e7-7af0-48fa-8509-fcc24c1a437b" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1123.305245] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1de03cc4-baf0-4813-9897-71e069417468 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1123.324291] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-69c3e963-8ad1-4204-900b-ad54c5a80457 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1123.331320] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-77073341-4390-4aa8-b9e9-ac30959d4690 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] [instance: d0669232-6f33-4b30-97c2-2e4239af1ad0] Updating instance 'd0669232-6f33-4b30-97c2-2e4239af1ad0' progress to 83 {{(pid=61806) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 1123.408284] env[61806]: DEBUG nova.compute.manager [None req-bf8f4ec5-63a0-4ddc-bc14-7d6705c06bed tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] [instance: c66fe2e7-7af0-48fa-8509-fcc24c1a437b] Starting instance... {{(pid=61806) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1123.839220] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-77073341-4390-4aa8-b9e9-ac30959d4690 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] [instance: d0669232-6f33-4b30-97c2-2e4239af1ad0] Updating instance 'd0669232-6f33-4b30-97c2-2e4239af1ad0' progress to 100 {{(pid=61806) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 1123.929065] env[61806]: DEBUG oslo_concurrency.lockutils [None req-bf8f4ec5-63a0-4ddc-bc14-7d6705c06bed tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1123.929398] env[61806]: DEBUG oslo_concurrency.lockutils [None req-bf8f4ec5-63a0-4ddc-bc14-7d6705c06bed tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.000s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1123.932826] env[61806]: INFO nova.compute.claims [None req-bf8f4ec5-63a0-4ddc-bc14-7d6705c06bed tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] [instance: c66fe2e7-7af0-48fa-8509-fcc24c1a437b] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1125.022018] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7e8b8028-a194-4770-ada0-532e673da986 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1125.028693] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-37024ade-df63-484a-b744-d2bd623e70f7 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1125.063158] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c0e579f8-b119-436f-ae6c-7c25f70211aa {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1125.070681] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-97c732ac-c62c-4613-a3c7-776aba15957c {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1125.089435] env[61806]: DEBUG nova.compute.provider_tree [None req-bf8f4ec5-63a0-4ddc-bc14-7d6705c06bed tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Updating inventory in ProviderTree for provider a2858be1-fd22-4e08-979e-87ad25293407 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 139, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61806) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1125.600769] env[61806]: DEBUG oslo_concurrency.lockutils [None req-a41d787d-234d-4a55-b94d-87e11c973c8a tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Acquiring lock "d0669232-6f33-4b30-97c2-2e4239af1ad0" by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1125.601296] env[61806]: DEBUG oslo_concurrency.lockutils [None req-a41d787d-234d-4a55-b94d-87e11c973c8a tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Lock "d0669232-6f33-4b30-97c2-2e4239af1ad0" acquired by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" :: waited 0.000s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1125.601296] env[61806]: DEBUG nova.compute.manager [None req-a41d787d-234d-4a55-b94d-87e11c973c8a tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] [instance: d0669232-6f33-4b30-97c2-2e4239af1ad0] Going to confirm migration 6 {{(pid=61806) do_confirm_resize /opt/stack/nova/nova/compute/manager.py:4783}} [ 1125.612704] env[61806]: ERROR nova.scheduler.client.report [None req-bf8f4ec5-63a0-4ddc-bc14-7d6705c06bed tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] [req-4f693136-dcc9-4af9-ab9a-1d0d202601e0] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 139, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID a2858be1-fd22-4e08-979e-87ad25293407. Got 409: {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-4f693136-dcc9-4af9-ab9a-1d0d202601e0"}]} [ 1125.628530] env[61806]: DEBUG nova.scheduler.client.report [None req-bf8f4ec5-63a0-4ddc-bc14-7d6705c06bed tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Refreshing inventories for resource provider a2858be1-fd22-4e08-979e-87ad25293407 {{(pid=61806) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:818}} [ 1125.642683] env[61806]: DEBUG nova.scheduler.client.report [None req-bf8f4ec5-63a0-4ddc-bc14-7d6705c06bed tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Updating ProviderTree inventory for provider a2858be1-fd22-4e08-979e-87ad25293407 from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 138, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61806) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:782}} [ 1125.642904] env[61806]: DEBUG nova.compute.provider_tree [None req-bf8f4ec5-63a0-4ddc-bc14-7d6705c06bed tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Updating inventory in ProviderTree for provider a2858be1-fd22-4e08-979e-87ad25293407 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 138, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61806) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1125.653435] env[61806]: DEBUG nova.scheduler.client.report [None req-bf8f4ec5-63a0-4ddc-bc14-7d6705c06bed tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Refreshing aggregate associations for resource provider a2858be1-fd22-4e08-979e-87ad25293407, aggregates: None {{(pid=61806) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:827}} [ 1125.670943] env[61806]: DEBUG nova.scheduler.client.report [None req-bf8f4ec5-63a0-4ddc-bc14-7d6705c06bed tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Refreshing trait associations for resource provider a2858be1-fd22-4e08-979e-87ad25293407, traits: COMPUTE_IMAGE_TYPE_ISO,COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_SAME_HOST_COLD_MIGRATE,COMPUTE_NET_ATTACH_INTERFACE,COMPUTE_NODE {{(pid=61806) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:839}} [ 1125.721613] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b2021249-ba62-4141-8e49-11344bb490d7 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1125.729246] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-709eb1a0-609e-478e-878c-e9234980ed73 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1125.757968] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e017d8a6-754b-4e20-b83f-e3955dab2065 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1125.764838] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-08634102-0b68-41c1-b2ca-784545d6d070 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1125.777246] env[61806]: DEBUG nova.compute.provider_tree [None req-bf8f4ec5-63a0-4ddc-bc14-7d6705c06bed tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Updating inventory in ProviderTree for provider a2858be1-fd22-4e08-979e-87ad25293407 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 139, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61806) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1126.137592] env[61806]: DEBUG oslo_concurrency.lockutils [None req-a41d787d-234d-4a55-b94d-87e11c973c8a tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Acquiring lock "refresh_cache-d0669232-6f33-4b30-97c2-2e4239af1ad0" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1126.137847] env[61806]: DEBUG oslo_concurrency.lockutils [None req-a41d787d-234d-4a55-b94d-87e11c973c8a tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Acquired lock "refresh_cache-d0669232-6f33-4b30-97c2-2e4239af1ad0" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1126.137969] env[61806]: DEBUG nova.network.neutron [None req-a41d787d-234d-4a55-b94d-87e11c973c8a tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] [instance: d0669232-6f33-4b30-97c2-2e4239af1ad0] Building network info cache for instance {{(pid=61806) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1126.138181] env[61806]: DEBUG nova.objects.instance [None req-a41d787d-234d-4a55-b94d-87e11c973c8a tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Lazy-loading 'info_cache' on Instance uuid d0669232-6f33-4b30-97c2-2e4239af1ad0 {{(pid=61806) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1126.306582] env[61806]: DEBUG nova.scheduler.client.report [None req-bf8f4ec5-63a0-4ddc-bc14-7d6705c06bed tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Updated inventory for provider a2858be1-fd22-4e08-979e-87ad25293407 with generation 142 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 139, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61806) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:971}} [ 1126.306861] env[61806]: DEBUG nova.compute.provider_tree [None req-bf8f4ec5-63a0-4ddc-bc14-7d6705c06bed tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Updating resource provider a2858be1-fd22-4e08-979e-87ad25293407 generation from 142 to 143 during operation: update_inventory {{(pid=61806) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 1126.307078] env[61806]: DEBUG nova.compute.provider_tree [None req-bf8f4ec5-63a0-4ddc-bc14-7d6705c06bed tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Updating inventory in ProviderTree for provider a2858be1-fd22-4e08-979e-87ad25293407 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 139, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61806) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1126.814211] env[61806]: DEBUG oslo_concurrency.lockutils [None req-bf8f4ec5-63a0-4ddc-bc14-7d6705c06bed tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.885s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1126.814808] env[61806]: DEBUG nova.compute.manager [None req-bf8f4ec5-63a0-4ddc-bc14-7d6705c06bed tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] [instance: c66fe2e7-7af0-48fa-8509-fcc24c1a437b] Start building networks asynchronously for instance. {{(pid=61806) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1127.319891] env[61806]: DEBUG nova.compute.utils [None req-bf8f4ec5-63a0-4ddc-bc14-7d6705c06bed tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Using /dev/sd instead of None {{(pid=61806) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1127.321309] env[61806]: DEBUG nova.compute.manager [None req-bf8f4ec5-63a0-4ddc-bc14-7d6705c06bed tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] [instance: c66fe2e7-7af0-48fa-8509-fcc24c1a437b] Allocating IP information in the background. {{(pid=61806) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1127.321487] env[61806]: DEBUG nova.network.neutron [None req-bf8f4ec5-63a0-4ddc-bc14-7d6705c06bed tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] [instance: c66fe2e7-7af0-48fa-8509-fcc24c1a437b] allocate_for_instance() {{(pid=61806) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 1127.347151] env[61806]: DEBUG nova.network.neutron [None req-a41d787d-234d-4a55-b94d-87e11c973c8a tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] [instance: d0669232-6f33-4b30-97c2-2e4239af1ad0] Updating instance_info_cache with network_info: [{"id": "f958e9a3-f8d6-4665-a59f-876aab66edfb", "address": "fa:16:3e:fc:93:23", "network": {"id": "33bf5b21-82b0-432e-ae3d-9c2ca75161d0", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-1294159336-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.187", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "1c55cc0455584fbb8cda88d33b8f8b7e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "5b0fa7a2-ebd9-4788-8904-7bf250ce466c", "external-id": "nsx-vlan-transportzone-669", "segmentation_id": 669, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf958e9a3-f8", "ovs_interfaceid": "f958e9a3-f8d6-4665-a59f-876aab66edfb", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61806) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1127.367800] env[61806]: DEBUG nova.policy [None req-bf8f4ec5-63a0-4ddc-bc14-7d6705c06bed tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '921aeb4e90e64f89888a9c65f873d66b', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'c1f2d161c50f4b70987b90e6364c3779', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61806) authorize /opt/stack/nova/nova/policy.py:201}} [ 1127.684060] env[61806]: DEBUG nova.network.neutron [None req-bf8f4ec5-63a0-4ddc-bc14-7d6705c06bed tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] [instance: c66fe2e7-7af0-48fa-8509-fcc24c1a437b] Successfully created port: 5bb12b4b-dfcb-416e-8095-91002ba3ca7d {{(pid=61806) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1127.825271] env[61806]: DEBUG nova.compute.manager [None req-bf8f4ec5-63a0-4ddc-bc14-7d6705c06bed tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] [instance: c66fe2e7-7af0-48fa-8509-fcc24c1a437b] Start building block device mappings for instance. {{(pid=61806) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1127.849350] env[61806]: DEBUG oslo_concurrency.lockutils [None req-a41d787d-234d-4a55-b94d-87e11c973c8a tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Releasing lock "refresh_cache-d0669232-6f33-4b30-97c2-2e4239af1ad0" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1127.849657] env[61806]: DEBUG nova.objects.instance [None req-a41d787d-234d-4a55-b94d-87e11c973c8a tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Lazy-loading 'migration_context' on Instance uuid d0669232-6f33-4b30-97c2-2e4239af1ad0 {{(pid=61806) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1128.352690] env[61806]: DEBUG nova.objects.base [None req-a41d787d-234d-4a55-b94d-87e11c973c8a tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Object Instance lazy-loaded attributes: info_cache,migration_context {{(pid=61806) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 1128.353602] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8f71cf55-c01b-47c3-a458-94ed80748d87 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1128.372932] env[61806]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-951bdbe6-8d18-49e7-b464-6fdad81d6df1 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1128.378872] env[61806]: DEBUG oslo_vmware.api [None req-a41d787d-234d-4a55-b94d-87e11c973c8a tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Waiting for the task: (returnval){ [ 1128.378872] env[61806]: value = "session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]52d6cd1b-256c-4f47-7f9b-c3770ba92051" [ 1128.378872] env[61806]: _type = "Task" [ 1128.378872] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1128.386287] env[61806]: DEBUG oslo_vmware.api [None req-a41d787d-234d-4a55-b94d-87e11c973c8a tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Task: {'id': session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]52d6cd1b-256c-4f47-7f9b-c3770ba92051, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1128.834730] env[61806]: DEBUG nova.compute.manager [None req-bf8f4ec5-63a0-4ddc-bc14-7d6705c06bed tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] [instance: c66fe2e7-7af0-48fa-8509-fcc24c1a437b] Start spawning the instance on the hypervisor. {{(pid=61806) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1128.862388] env[61806]: DEBUG nova.virt.hardware [None req-bf8f4ec5-63a0-4ddc-bc14-7d6705c06bed tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-15T18:05:10Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-15T18:04:52Z,direct_url=,disk_format='vmdk',id=ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='84ba83fce2c349a6988173c5d6fc3b07',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-15T18:04:53Z,virtual_size=,visibility=), allow threads: False {{(pid=61806) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1128.862663] env[61806]: DEBUG nova.virt.hardware [None req-bf8f4ec5-63a0-4ddc-bc14-7d6705c06bed tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Flavor limits 0:0:0 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1128.862828] env[61806]: DEBUG nova.virt.hardware [None req-bf8f4ec5-63a0-4ddc-bc14-7d6705c06bed tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Image limits 0:0:0 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1128.863028] env[61806]: DEBUG nova.virt.hardware [None req-bf8f4ec5-63a0-4ddc-bc14-7d6705c06bed tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Flavor pref 0:0:0 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1128.863190] env[61806]: DEBUG nova.virt.hardware [None req-bf8f4ec5-63a0-4ddc-bc14-7d6705c06bed tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Image pref 0:0:0 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1128.863348] env[61806]: DEBUG nova.virt.hardware [None req-bf8f4ec5-63a0-4ddc-bc14-7d6705c06bed tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1128.863562] env[61806]: DEBUG nova.virt.hardware [None req-bf8f4ec5-63a0-4ddc-bc14-7d6705c06bed tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61806) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1128.863874] env[61806]: DEBUG nova.virt.hardware [None req-bf8f4ec5-63a0-4ddc-bc14-7d6705c06bed tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61806) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1128.864136] env[61806]: DEBUG nova.virt.hardware [None req-bf8f4ec5-63a0-4ddc-bc14-7d6705c06bed tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Got 1 possible topologies {{(pid=61806) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1128.864335] env[61806]: DEBUG nova.virt.hardware [None req-bf8f4ec5-63a0-4ddc-bc14-7d6705c06bed tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61806) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1128.864555] env[61806]: DEBUG nova.virt.hardware [None req-bf8f4ec5-63a0-4ddc-bc14-7d6705c06bed tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61806) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1128.865441] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-071b9ab8-c1ec-4253-a473-fbd79fe81e9c {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1128.873109] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-59c461c0-172a-405b-874d-125e5cd5f4f2 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1128.893887] env[61806]: DEBUG oslo_vmware.api [None req-a41d787d-234d-4a55-b94d-87e11c973c8a tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Task: {'id': session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]52d6cd1b-256c-4f47-7f9b-c3770ba92051, 'name': SearchDatastore_Task, 'duration_secs': 0.00791} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1128.894166] env[61806]: DEBUG oslo_concurrency.lockutils [None req-a41d787d-234d-4a55-b94d-87e11c973c8a tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1128.894398] env[61806]: DEBUG oslo_concurrency.lockutils [None req-a41d787d-234d-4a55-b94d-87e11c973c8a tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" :: waited 0.000s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1129.060390] env[61806]: DEBUG nova.compute.manager [req-2327ce61-cbce-4c28-8a01-c97c0a30b501 req-f962e0a9-5be3-4cbd-bbfe-9c7a1b1b6f97 service nova] [instance: c66fe2e7-7af0-48fa-8509-fcc24c1a437b] Received event network-vif-plugged-5bb12b4b-dfcb-416e-8095-91002ba3ca7d {{(pid=61806) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1129.060591] env[61806]: DEBUG oslo_concurrency.lockutils [req-2327ce61-cbce-4c28-8a01-c97c0a30b501 req-f962e0a9-5be3-4cbd-bbfe-9c7a1b1b6f97 service nova] Acquiring lock "c66fe2e7-7af0-48fa-8509-fcc24c1a437b-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1129.060815] env[61806]: DEBUG oslo_concurrency.lockutils [req-2327ce61-cbce-4c28-8a01-c97c0a30b501 req-f962e0a9-5be3-4cbd-bbfe-9c7a1b1b6f97 service nova] Lock "c66fe2e7-7af0-48fa-8509-fcc24c1a437b-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1129.060992] env[61806]: DEBUG oslo_concurrency.lockutils [req-2327ce61-cbce-4c28-8a01-c97c0a30b501 req-f962e0a9-5be3-4cbd-bbfe-9c7a1b1b6f97 service nova] Lock "c66fe2e7-7af0-48fa-8509-fcc24c1a437b-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1129.061190] env[61806]: DEBUG nova.compute.manager [req-2327ce61-cbce-4c28-8a01-c97c0a30b501 req-f962e0a9-5be3-4cbd-bbfe-9c7a1b1b6f97 service nova] [instance: c66fe2e7-7af0-48fa-8509-fcc24c1a437b] No waiting events found dispatching network-vif-plugged-5bb12b4b-dfcb-416e-8095-91002ba3ca7d {{(pid=61806) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1129.061366] env[61806]: WARNING nova.compute.manager [req-2327ce61-cbce-4c28-8a01-c97c0a30b501 req-f962e0a9-5be3-4cbd-bbfe-9c7a1b1b6f97 service nova] [instance: c66fe2e7-7af0-48fa-8509-fcc24c1a437b] Received unexpected event network-vif-plugged-5bb12b4b-dfcb-416e-8095-91002ba3ca7d for instance with vm_state building and task_state spawning. [ 1129.147111] env[61806]: DEBUG nova.network.neutron [None req-bf8f4ec5-63a0-4ddc-bc14-7d6705c06bed tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] [instance: c66fe2e7-7af0-48fa-8509-fcc24c1a437b] Successfully updated port: 5bb12b4b-dfcb-416e-8095-91002ba3ca7d {{(pid=61806) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1129.462576] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bde6aa75-d214-4d9e-90bc-e89c78d6b6c2 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1129.470662] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4b4fbee0-0e3a-4b73-9f6a-7ce8ecd625f3 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1129.500975] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-205804f9-93f3-4f0d-830b-04a94e5e83b8 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1129.508256] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aabaa613-b22d-4260-b37c-bd73a5aaf8fa {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1129.521010] env[61806]: DEBUG nova.compute.provider_tree [None req-a41d787d-234d-4a55-b94d-87e11c973c8a tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Inventory has not changed in ProviderTree for provider: a2858be1-fd22-4e08-979e-87ad25293407 {{(pid=61806) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1129.649767] env[61806]: DEBUG oslo_concurrency.lockutils [None req-bf8f4ec5-63a0-4ddc-bc14-7d6705c06bed tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Acquiring lock "refresh_cache-c66fe2e7-7af0-48fa-8509-fcc24c1a437b" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1129.649959] env[61806]: DEBUG oslo_concurrency.lockutils [None req-bf8f4ec5-63a0-4ddc-bc14-7d6705c06bed tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Acquired lock "refresh_cache-c66fe2e7-7af0-48fa-8509-fcc24c1a437b" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1129.650132] env[61806]: DEBUG nova.network.neutron [None req-bf8f4ec5-63a0-4ddc-bc14-7d6705c06bed tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] [instance: c66fe2e7-7af0-48fa-8509-fcc24c1a437b] Building network info cache for instance {{(pid=61806) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1130.024462] env[61806]: DEBUG nova.scheduler.client.report [None req-a41d787d-234d-4a55-b94d-87e11c973c8a tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Inventory has not changed for provider a2858be1-fd22-4e08-979e-87ad25293407 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 139, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61806) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1130.183797] env[61806]: DEBUG nova.network.neutron [None req-bf8f4ec5-63a0-4ddc-bc14-7d6705c06bed tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] [instance: c66fe2e7-7af0-48fa-8509-fcc24c1a437b] Instance cache missing network info. {{(pid=61806) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 1130.305489] env[61806]: DEBUG nova.network.neutron [None req-bf8f4ec5-63a0-4ddc-bc14-7d6705c06bed tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] [instance: c66fe2e7-7af0-48fa-8509-fcc24c1a437b] Updating instance_info_cache with network_info: [{"id": "5bb12b4b-dfcb-416e-8095-91002ba3ca7d", "address": "fa:16:3e:9d:21:7a", "network": {"id": "746a03b6-c1a6-4d27-98b9-0bee6376634f", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-1971382587-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c1f2d161c50f4b70987b90e6364c3779", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6a31b224-77d7-48a4-af87-312758611b8e", "external-id": "nsx-vlan-transportzone-761", "segmentation_id": 761, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap5bb12b4b-df", "ovs_interfaceid": "5bb12b4b-dfcb-416e-8095-91002ba3ca7d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61806) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1130.535326] env[61806]: DEBUG oslo_service.periodic_task [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=61806) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1130.808411] env[61806]: DEBUG oslo_concurrency.lockutils [None req-bf8f4ec5-63a0-4ddc-bc14-7d6705c06bed tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Releasing lock "refresh_cache-c66fe2e7-7af0-48fa-8509-fcc24c1a437b" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1130.808732] env[61806]: DEBUG nova.compute.manager [None req-bf8f4ec5-63a0-4ddc-bc14-7d6705c06bed tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] [instance: c66fe2e7-7af0-48fa-8509-fcc24c1a437b] Instance network_info: |[{"id": "5bb12b4b-dfcb-416e-8095-91002ba3ca7d", "address": "fa:16:3e:9d:21:7a", "network": {"id": "746a03b6-c1a6-4d27-98b9-0bee6376634f", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-1971382587-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c1f2d161c50f4b70987b90e6364c3779", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6a31b224-77d7-48a4-af87-312758611b8e", "external-id": "nsx-vlan-transportzone-761", "segmentation_id": 761, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap5bb12b4b-df", "ovs_interfaceid": "5bb12b4b-dfcb-416e-8095-91002ba3ca7d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61806) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1130.809184] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-bf8f4ec5-63a0-4ddc-bc14-7d6705c06bed tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] [instance: c66fe2e7-7af0-48fa-8509-fcc24c1a437b] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:9d:21:7a', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '6a31b224-77d7-48a4-af87-312758611b8e', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '5bb12b4b-dfcb-416e-8095-91002ba3ca7d', 'vif_model': 'vmxnet3'}] {{(pid=61806) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1130.816567] env[61806]: DEBUG oslo.service.loopingcall [None req-bf8f4ec5-63a0-4ddc-bc14-7d6705c06bed tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61806) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1130.816768] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: c66fe2e7-7af0-48fa-8509-fcc24c1a437b] Creating VM on the ESX host {{(pid=61806) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 1130.816990] env[61806]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-7e8de74b-1cd9-4475-8c9a-21150c45efd0 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1130.836208] env[61806]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1130.836208] env[61806]: value = "task-1295262" [ 1130.836208] env[61806]: _type = "Task" [ 1130.836208] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1130.843483] env[61806]: DEBUG oslo_vmware.api [-] Task: {'id': task-1295262, 'name': CreateVM_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1131.022185] env[61806]: DEBUG oslo_service.periodic_task [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=61806) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1131.022411] env[61806]: DEBUG nova.compute.manager [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Starting heal instance info cache {{(pid=61806) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9926}} [ 1131.034638] env[61806]: DEBUG oslo_concurrency.lockutils [None req-a41d787d-234d-4a55-b94d-87e11c973c8a tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" :: held 2.140s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1131.034970] env[61806]: DEBUG nova.compute.manager [None req-a41d787d-234d-4a55-b94d-87e11c973c8a tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] [instance: d0669232-6f33-4b30-97c2-2e4239af1ad0] Resized/migrated instance is powered off. Setting vm_state to 'stopped'. {{(pid=61806) _confirm_resize /opt/stack/nova/nova/compute/manager.py:4909}} [ 1131.108601] env[61806]: DEBUG nova.compute.manager [req-1dd02d10-ccc9-4a70-b29c-c6332cd4f932 req-7b752e8e-d4da-4289-8977-5627ed82ff10 service nova] [instance: c66fe2e7-7af0-48fa-8509-fcc24c1a437b] Received event network-changed-5bb12b4b-dfcb-416e-8095-91002ba3ca7d {{(pid=61806) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1131.108869] env[61806]: DEBUG nova.compute.manager [req-1dd02d10-ccc9-4a70-b29c-c6332cd4f932 req-7b752e8e-d4da-4289-8977-5627ed82ff10 service nova] [instance: c66fe2e7-7af0-48fa-8509-fcc24c1a437b] Refreshing instance network info cache due to event network-changed-5bb12b4b-dfcb-416e-8095-91002ba3ca7d. {{(pid=61806) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 1131.109066] env[61806]: DEBUG oslo_concurrency.lockutils [req-1dd02d10-ccc9-4a70-b29c-c6332cd4f932 req-7b752e8e-d4da-4289-8977-5627ed82ff10 service nova] Acquiring lock "refresh_cache-c66fe2e7-7af0-48fa-8509-fcc24c1a437b" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1131.109213] env[61806]: DEBUG oslo_concurrency.lockutils [req-1dd02d10-ccc9-4a70-b29c-c6332cd4f932 req-7b752e8e-d4da-4289-8977-5627ed82ff10 service nova] Acquired lock "refresh_cache-c66fe2e7-7af0-48fa-8509-fcc24c1a437b" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1131.109380] env[61806]: DEBUG nova.network.neutron [req-1dd02d10-ccc9-4a70-b29c-c6332cd4f932 req-7b752e8e-d4da-4289-8977-5627ed82ff10 service nova] [instance: c66fe2e7-7af0-48fa-8509-fcc24c1a437b] Refreshing network info cache for port 5bb12b4b-dfcb-416e-8095-91002ba3ca7d {{(pid=61806) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1131.346116] env[61806]: DEBUG oslo_vmware.api [-] Task: {'id': task-1295262, 'name': CreateVM_Task, 'duration_secs': 0.318281} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1131.346289] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: c66fe2e7-7af0-48fa-8509-fcc24c1a437b] Created VM on the ESX host {{(pid=61806) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 1131.346935] env[61806]: DEBUG oslo_concurrency.lockutils [None req-bf8f4ec5-63a0-4ddc-bc14-7d6705c06bed tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1131.347132] env[61806]: DEBUG oslo_concurrency.lockutils [None req-bf8f4ec5-63a0-4ddc-bc14-7d6705c06bed tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Acquired lock "[datastore1] devstack-image-cache_base/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1131.347458] env[61806]: DEBUG oslo_concurrency.lockutils [None req-bf8f4ec5-63a0-4ddc-bc14-7d6705c06bed tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1131.347703] env[61806]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-16391a74-5417-4c03-8290-6e4e652202fd {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1131.352170] env[61806]: DEBUG oslo_vmware.api [None req-bf8f4ec5-63a0-4ddc-bc14-7d6705c06bed tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Waiting for the task: (returnval){ [ 1131.352170] env[61806]: value = "session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]523739ae-9014-018c-6600-13ba4078fbbb" [ 1131.352170] env[61806]: _type = "Task" [ 1131.352170] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1131.359172] env[61806]: DEBUG oslo_vmware.api [None req-bf8f4ec5-63a0-4ddc-bc14-7d6705c06bed tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Task: {'id': session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]523739ae-9014-018c-6600-13ba4078fbbb, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1131.551707] env[61806]: DEBUG oslo_concurrency.lockutils [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Acquiring lock "refresh_cache-d0669232-6f33-4b30-97c2-2e4239af1ad0" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1131.551942] env[61806]: DEBUG oslo_concurrency.lockutils [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Acquired lock "refresh_cache-d0669232-6f33-4b30-97c2-2e4239af1ad0" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1131.551980] env[61806]: DEBUG nova.network.neutron [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] [instance: d0669232-6f33-4b30-97c2-2e4239af1ad0] Forcefully refreshing network info cache for instance {{(pid=61806) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2004}} [ 1131.591686] env[61806]: INFO nova.scheduler.client.report [None req-a41d787d-234d-4a55-b94d-87e11c973c8a tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Deleted allocation for migration 46ba7ac1-5345-483f-b3ae-3d5aed8a7aec [ 1131.800419] env[61806]: DEBUG nova.network.neutron [req-1dd02d10-ccc9-4a70-b29c-c6332cd4f932 req-7b752e8e-d4da-4289-8977-5627ed82ff10 service nova] [instance: c66fe2e7-7af0-48fa-8509-fcc24c1a437b] Updated VIF entry in instance network info cache for port 5bb12b4b-dfcb-416e-8095-91002ba3ca7d. {{(pid=61806) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1131.800771] env[61806]: DEBUG nova.network.neutron [req-1dd02d10-ccc9-4a70-b29c-c6332cd4f932 req-7b752e8e-d4da-4289-8977-5627ed82ff10 service nova] [instance: c66fe2e7-7af0-48fa-8509-fcc24c1a437b] Updating instance_info_cache with network_info: [{"id": "5bb12b4b-dfcb-416e-8095-91002ba3ca7d", "address": "fa:16:3e:9d:21:7a", "network": {"id": "746a03b6-c1a6-4d27-98b9-0bee6376634f", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-1971382587-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c1f2d161c50f4b70987b90e6364c3779", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6a31b224-77d7-48a4-af87-312758611b8e", "external-id": "nsx-vlan-transportzone-761", "segmentation_id": 761, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap5bb12b4b-df", "ovs_interfaceid": "5bb12b4b-dfcb-416e-8095-91002ba3ca7d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61806) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1131.862576] env[61806]: DEBUG oslo_vmware.api [None req-bf8f4ec5-63a0-4ddc-bc14-7d6705c06bed tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Task: {'id': session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]523739ae-9014-018c-6600-13ba4078fbbb, 'name': SearchDatastore_Task, 'duration_secs': 0.010402} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1131.862884] env[61806]: DEBUG oslo_concurrency.lockutils [None req-bf8f4ec5-63a0-4ddc-bc14-7d6705c06bed tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Releasing lock "[datastore1] devstack-image-cache_base/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1131.863140] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-bf8f4ec5-63a0-4ddc-bc14-7d6705c06bed tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] [instance: c66fe2e7-7af0-48fa-8509-fcc24c1a437b] Processing image ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa {{(pid=61806) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1131.864097] env[61806]: DEBUG oslo_concurrency.lockutils [None req-bf8f4ec5-63a0-4ddc-bc14-7d6705c06bed tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa.vmdk" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1131.864097] env[61806]: DEBUG oslo_concurrency.lockutils [None req-bf8f4ec5-63a0-4ddc-bc14-7d6705c06bed tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Acquired lock "[datastore1] devstack-image-cache_base/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa.vmdk" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1131.864097] env[61806]: DEBUG nova.virt.vmwareapi.ds_util [None req-bf8f4ec5-63a0-4ddc-bc14-7d6705c06bed tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61806) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1131.864097] env[61806]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-ed2e2c89-9aa4-4920-bc68-3a3a4787bbb4 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1131.871809] env[61806]: DEBUG nova.virt.vmwareapi.ds_util [None req-bf8f4ec5-63a0-4ddc-bc14-7d6705c06bed tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61806) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1131.872568] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-bf8f4ec5-63a0-4ddc-bc14-7d6705c06bed tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61806) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 1131.872683] env[61806]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-00e2f5f0-bda3-4008-9285-dc14e336bb75 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1131.878067] env[61806]: DEBUG oslo_vmware.api [None req-bf8f4ec5-63a0-4ddc-bc14-7d6705c06bed tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Waiting for the task: (returnval){ [ 1131.878067] env[61806]: value = "session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]52bcd099-23a4-990d-9a04-ed07482a41ba" [ 1131.878067] env[61806]: _type = "Task" [ 1131.878067] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1131.884708] env[61806]: DEBUG oslo_vmware.api [None req-bf8f4ec5-63a0-4ddc-bc14-7d6705c06bed tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Task: {'id': session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]52bcd099-23a4-990d-9a04-ed07482a41ba, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1132.096737] env[61806]: DEBUG oslo_concurrency.lockutils [None req-a41d787d-234d-4a55-b94d-87e11c973c8a tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Lock "d0669232-6f33-4b30-97c2-2e4239af1ad0" "released" by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" :: held 6.496s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1132.303636] env[61806]: DEBUG oslo_concurrency.lockutils [req-1dd02d10-ccc9-4a70-b29c-c6332cd4f932 req-7b752e8e-d4da-4289-8977-5627ed82ff10 service nova] Releasing lock "refresh_cache-c66fe2e7-7af0-48fa-8509-fcc24c1a437b" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1132.391037] env[61806]: DEBUG oslo_vmware.api [None req-bf8f4ec5-63a0-4ddc-bc14-7d6705c06bed tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Task: {'id': session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]52bcd099-23a4-990d-9a04-ed07482a41ba, 'name': SearchDatastore_Task, 'duration_secs': 0.011835} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1132.391037] env[61806]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c7cfb4d5-eaa5-444d-a279-cbe215325100 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1132.393993] env[61806]: DEBUG oslo_vmware.api [None req-bf8f4ec5-63a0-4ddc-bc14-7d6705c06bed tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Waiting for the task: (returnval){ [ 1132.393993] env[61806]: value = "session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]526e5dd9-74d3-9f45-b13e-b5fa215206b6" [ 1132.393993] env[61806]: _type = "Task" [ 1132.393993] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1132.401245] env[61806]: DEBUG oslo_vmware.api [None req-bf8f4ec5-63a0-4ddc-bc14-7d6705c06bed tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Task: {'id': session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]526e5dd9-74d3-9f45-b13e-b5fa215206b6, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1132.542486] env[61806]: DEBUG nova.objects.instance [None req-6f41d8c4-228e-4028-861e-4b121b76c64a tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Lazy-loading 'flavor' on Instance uuid d0669232-6f33-4b30-97c2-2e4239af1ad0 {{(pid=61806) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1132.766944] env[61806]: DEBUG nova.network.neutron [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] [instance: d0669232-6f33-4b30-97c2-2e4239af1ad0] Updating instance_info_cache with network_info: [{"id": "f958e9a3-f8d6-4665-a59f-876aab66edfb", "address": "fa:16:3e:fc:93:23", "network": {"id": "33bf5b21-82b0-432e-ae3d-9c2ca75161d0", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-1294159336-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.187", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "1c55cc0455584fbb8cda88d33b8f8b7e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "5b0fa7a2-ebd9-4788-8904-7bf250ce466c", "external-id": "nsx-vlan-transportzone-669", "segmentation_id": 669, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf958e9a3-f8", "ovs_interfaceid": "f958e9a3-f8d6-4665-a59f-876aab66edfb", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61806) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1132.903604] env[61806]: DEBUG oslo_vmware.api [None req-bf8f4ec5-63a0-4ddc-bc14-7d6705c06bed tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Task: {'id': session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]526e5dd9-74d3-9f45-b13e-b5fa215206b6, 'name': SearchDatastore_Task, 'duration_secs': 0.012357} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1132.903867] env[61806]: DEBUG oslo_concurrency.lockutils [None req-bf8f4ec5-63a0-4ddc-bc14-7d6705c06bed tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Releasing lock "[datastore1] devstack-image-cache_base/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa.vmdk" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1132.904151] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [None req-bf8f4ec5-63a0-4ddc-bc14-7d6705c06bed tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa.vmdk to [datastore1] c66fe2e7-7af0-48fa-8509-fcc24c1a437b/c66fe2e7-7af0-48fa-8509-fcc24c1a437b.vmdk {{(pid=61806) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 1132.904412] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-e919bbcd-d460-452c-b644-1e3c4f5d1e73 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1132.910641] env[61806]: DEBUG oslo_vmware.api [None req-bf8f4ec5-63a0-4ddc-bc14-7d6705c06bed tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Waiting for the task: (returnval){ [ 1132.910641] env[61806]: value = "task-1295263" [ 1132.910641] env[61806]: _type = "Task" [ 1132.910641] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1132.917741] env[61806]: DEBUG oslo_vmware.api [None req-bf8f4ec5-63a0-4ddc-bc14-7d6705c06bed tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Task: {'id': task-1295263, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1133.047334] env[61806]: DEBUG oslo_concurrency.lockutils [None req-6f41d8c4-228e-4028-861e-4b121b76c64a tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Acquiring lock "refresh_cache-d0669232-6f33-4b30-97c2-2e4239af1ad0" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1133.269794] env[61806]: DEBUG oslo_concurrency.lockutils [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Releasing lock "refresh_cache-d0669232-6f33-4b30-97c2-2e4239af1ad0" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1133.270046] env[61806]: DEBUG nova.compute.manager [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] [instance: d0669232-6f33-4b30-97c2-2e4239af1ad0] Updated the network info_cache for instance {{(pid=61806) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9997}} [ 1133.270384] env[61806]: DEBUG oslo_concurrency.lockutils [None req-6f41d8c4-228e-4028-861e-4b121b76c64a tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Acquired lock "refresh_cache-d0669232-6f33-4b30-97c2-2e4239af1ad0" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1133.270576] env[61806]: DEBUG nova.network.neutron [None req-6f41d8c4-228e-4028-861e-4b121b76c64a tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] [instance: d0669232-6f33-4b30-97c2-2e4239af1ad0] Building network info cache for instance {{(pid=61806) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1133.270769] env[61806]: DEBUG nova.objects.instance [None req-6f41d8c4-228e-4028-861e-4b121b76c64a tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Lazy-loading 'info_cache' on Instance uuid d0669232-6f33-4b30-97c2-2e4239af1ad0 {{(pid=61806) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1133.272134] env[61806]: DEBUG oslo_service.periodic_task [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=61806) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1133.272336] env[61806]: DEBUG oslo_service.periodic_task [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Running periodic task ComputeManager.update_available_resource {{(pid=61806) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1133.421015] env[61806]: DEBUG oslo_vmware.api [None req-bf8f4ec5-63a0-4ddc-bc14-7d6705c06bed tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Task: {'id': task-1295263, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1133.775500] env[61806]: DEBUG nova.objects.base [None req-6f41d8c4-228e-4028-861e-4b121b76c64a tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Object Instance lazy-loaded attributes: flavor,info_cache {{(pid=61806) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 1133.777466] env[61806]: DEBUG oslo_concurrency.lockutils [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1133.777611] env[61806]: DEBUG oslo_concurrency.lockutils [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 0.000s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1133.777694] env[61806]: DEBUG oslo_concurrency.lockutils [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1133.777843] env[61806]: DEBUG nova.compute.resource_tracker [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=61806) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 1133.778902] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2b27df9e-a683-49e1-b923-4c3b292d1220 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1133.787345] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-505092b5-68de-41a0-82d7-098705446956 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1133.800893] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-43fea358-3c27-494a-a52f-147dc20cab7f {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1133.807011] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-93f94a74-4855-47be-b3de-1c3102b251f1 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1133.836296] env[61806]: DEBUG nova.compute.resource_tracker [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=180464MB free_disk=138GB free_vcpus=48 pci_devices=None {{(pid=61806) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 1133.836438] env[61806]: DEBUG oslo_concurrency.lockutils [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1133.836620] env[61806]: DEBUG oslo_concurrency.lockutils [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 0.000s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1133.921371] env[61806]: DEBUG oslo_vmware.api [None req-bf8f4ec5-63a0-4ddc-bc14-7d6705c06bed tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Task: {'id': task-1295263, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.560229} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1133.921901] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [None req-bf8f4ec5-63a0-4ddc-bc14-7d6705c06bed tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa.vmdk to [datastore1] c66fe2e7-7af0-48fa-8509-fcc24c1a437b/c66fe2e7-7af0-48fa-8509-fcc24c1a437b.vmdk {{(pid=61806) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 1133.922138] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-bf8f4ec5-63a0-4ddc-bc14-7d6705c06bed tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] [instance: c66fe2e7-7af0-48fa-8509-fcc24c1a437b] Extending root virtual disk to 1048576 {{(pid=61806) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1133.922392] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-b1e56ff6-80ae-454f-82fe-55650c9d8f62 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1133.928671] env[61806]: DEBUG oslo_vmware.api [None req-bf8f4ec5-63a0-4ddc-bc14-7d6705c06bed tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Waiting for the task: (returnval){ [ 1133.928671] env[61806]: value = "task-1295264" [ 1133.928671] env[61806]: _type = "Task" [ 1133.928671] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1133.937094] env[61806]: DEBUG oslo_vmware.api [None req-bf8f4ec5-63a0-4ddc-bc14-7d6705c06bed tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Task: {'id': task-1295264, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1134.438452] env[61806]: DEBUG oslo_vmware.api [None req-bf8f4ec5-63a0-4ddc-bc14-7d6705c06bed tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Task: {'id': task-1295264, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.087851} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1134.440955] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-bf8f4ec5-63a0-4ddc-bc14-7d6705c06bed tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] [instance: c66fe2e7-7af0-48fa-8509-fcc24c1a437b] Extended root virtual disk {{(pid=61806) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1134.441817] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6b16ce6a-d343-4cd4-afc3-968ff0935f84 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1134.464032] env[61806]: DEBUG nova.virt.vmwareapi.volumeops [None req-bf8f4ec5-63a0-4ddc-bc14-7d6705c06bed tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] [instance: c66fe2e7-7af0-48fa-8509-fcc24c1a437b] Reconfiguring VM instance instance-0000006b to attach disk [datastore1] c66fe2e7-7af0-48fa-8509-fcc24c1a437b/c66fe2e7-7af0-48fa-8509-fcc24c1a437b.vmdk or device None with type sparse {{(pid=61806) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1134.464032] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-5a9a3f2b-5d7d-4486-b61a-46eba8c49e36 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1134.485019] env[61806]: DEBUG oslo_vmware.api [None req-bf8f4ec5-63a0-4ddc-bc14-7d6705c06bed tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Waiting for the task: (returnval){ [ 1134.485019] env[61806]: value = "task-1295265" [ 1134.485019] env[61806]: _type = "Task" [ 1134.485019] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1134.492950] env[61806]: DEBUG oslo_vmware.api [None req-bf8f4ec5-63a0-4ddc-bc14-7d6705c06bed tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Task: {'id': task-1295265, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1134.510201] env[61806]: DEBUG nova.network.neutron [None req-6f41d8c4-228e-4028-861e-4b121b76c64a tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] [instance: d0669232-6f33-4b30-97c2-2e4239af1ad0] Updating instance_info_cache with network_info: [{"id": "f958e9a3-f8d6-4665-a59f-876aab66edfb", "address": "fa:16:3e:fc:93:23", "network": {"id": "33bf5b21-82b0-432e-ae3d-9c2ca75161d0", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-1294159336-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.187", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "1c55cc0455584fbb8cda88d33b8f8b7e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "5b0fa7a2-ebd9-4788-8904-7bf250ce466c", "external-id": "nsx-vlan-transportzone-669", "segmentation_id": 669, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf958e9a3-f8", "ovs_interfaceid": "f958e9a3-f8d6-4665-a59f-876aab66edfb", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61806) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1134.863728] env[61806]: DEBUG nova.compute.resource_tracker [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Instance 29210bcd-0fe1-4ed9-a459-abc8016a4255 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61806) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1134.863977] env[61806]: DEBUG nova.compute.resource_tracker [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Instance d0669232-6f33-4b30-97c2-2e4239af1ad0 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 256, 'VCPU': 1}}. {{(pid=61806) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1134.864038] env[61806]: DEBUG nova.compute.resource_tracker [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Instance c66fe2e7-7af0-48fa-8509-fcc24c1a437b actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61806) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1134.864221] env[61806]: DEBUG nova.compute.resource_tracker [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Total usable vcpus: 48, total allocated vcpus: 3 {{(pid=61806) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 1134.864370] env[61806]: DEBUG nova.compute.resource_tracker [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=1152MB phys_disk=200GB used_disk=3GB total_vcpus=48 used_vcpus=3 pci_stats=[] {{(pid=61806) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 1134.909171] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8da121fa-3f37-4303-88f7-927a771a2153 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1134.916528] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a62ea0e8-94b6-45ec-81eb-60f153492c8e {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1134.945746] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-28b2871b-4e0d-4b15-ac04-2de350cd8733 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1134.953212] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b0b0b51f-42fd-46c4-a866-0a3da3c0b59a {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1134.966220] env[61806]: DEBUG nova.compute.provider_tree [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Updating inventory in ProviderTree for provider a2858be1-fd22-4e08-979e-87ad25293407 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 138, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61806) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1134.993586] env[61806]: DEBUG oslo_vmware.api [None req-bf8f4ec5-63a0-4ddc-bc14-7d6705c06bed tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Task: {'id': task-1295265, 'name': ReconfigVM_Task, 'duration_secs': 0.299802} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1134.993860] env[61806]: DEBUG nova.virt.vmwareapi.volumeops [None req-bf8f4ec5-63a0-4ddc-bc14-7d6705c06bed tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] [instance: c66fe2e7-7af0-48fa-8509-fcc24c1a437b] Reconfigured VM instance instance-0000006b to attach disk [datastore1] c66fe2e7-7af0-48fa-8509-fcc24c1a437b/c66fe2e7-7af0-48fa-8509-fcc24c1a437b.vmdk or device None with type sparse {{(pid=61806) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1134.994493] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-61f29764-6640-4398-8add-f3122ad78672 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1135.000859] env[61806]: DEBUG oslo_vmware.api [None req-bf8f4ec5-63a0-4ddc-bc14-7d6705c06bed tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Waiting for the task: (returnval){ [ 1135.000859] env[61806]: value = "task-1295266" [ 1135.000859] env[61806]: _type = "Task" [ 1135.000859] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1135.008013] env[61806]: DEBUG oslo_vmware.api [None req-bf8f4ec5-63a0-4ddc-bc14-7d6705c06bed tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Task: {'id': task-1295266, 'name': Rename_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1135.013025] env[61806]: DEBUG oslo_concurrency.lockutils [None req-6f41d8c4-228e-4028-861e-4b121b76c64a tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Releasing lock "refresh_cache-d0669232-6f33-4b30-97c2-2e4239af1ad0" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1135.484960] env[61806]: ERROR nova.scheduler.client.report [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] [req-073bbcdb-dfb6-4384-8e5a-1603d684b994] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 138, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID a2858be1-fd22-4e08-979e-87ad25293407. Got 409: {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-073bbcdb-dfb6-4384-8e5a-1603d684b994"}]} [ 1135.500636] env[61806]: DEBUG nova.scheduler.client.report [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Refreshing inventories for resource provider a2858be1-fd22-4e08-979e-87ad25293407 {{(pid=61806) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:818}} [ 1135.510699] env[61806]: DEBUG oslo_vmware.api [None req-bf8f4ec5-63a0-4ddc-bc14-7d6705c06bed tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Task: {'id': task-1295266, 'name': Rename_Task, 'duration_secs': 0.151262} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1135.510947] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [None req-bf8f4ec5-63a0-4ddc-bc14-7d6705c06bed tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] [instance: c66fe2e7-7af0-48fa-8509-fcc24c1a437b] Powering on the VM {{(pid=61806) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1135.511237] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-d1cacd54-05ae-4afc-986c-283fca54081e {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1135.513202] env[61806]: DEBUG nova.scheduler.client.report [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Updating ProviderTree inventory for provider a2858be1-fd22-4e08-979e-87ad25293407 from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 139, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61806) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:782}} [ 1135.513369] env[61806]: DEBUG nova.compute.provider_tree [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Updating inventory in ProviderTree for provider a2858be1-fd22-4e08-979e-87ad25293407 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 139, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61806) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1135.515683] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [None req-6f41d8c4-228e-4028-861e-4b121b76c64a tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] [instance: d0669232-6f33-4b30-97c2-2e4239af1ad0] Powering on the VM {{(pid=61806) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1135.516388] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-24372327-7ed9-42a0-9688-d1d34c7de8b0 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1135.520724] env[61806]: DEBUG oslo_vmware.api [None req-bf8f4ec5-63a0-4ddc-bc14-7d6705c06bed tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Waiting for the task: (returnval){ [ 1135.520724] env[61806]: value = "task-1295267" [ 1135.520724] env[61806]: _type = "Task" [ 1135.520724] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1135.524438] env[61806]: DEBUG oslo_vmware.api [None req-6f41d8c4-228e-4028-861e-4b121b76c64a tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Waiting for the task: (returnval){ [ 1135.524438] env[61806]: value = "task-1295268" [ 1135.524438] env[61806]: _type = "Task" [ 1135.524438] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1135.529625] env[61806]: DEBUG nova.scheduler.client.report [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Refreshing aggregate associations for resource provider a2858be1-fd22-4e08-979e-87ad25293407, aggregates: None {{(pid=61806) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:827}} [ 1135.537209] env[61806]: DEBUG oslo_vmware.api [None req-bf8f4ec5-63a0-4ddc-bc14-7d6705c06bed tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Task: {'id': task-1295267, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1135.540700] env[61806]: DEBUG oslo_vmware.api [None req-6f41d8c4-228e-4028-861e-4b121b76c64a tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Task: {'id': task-1295268, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1135.547371] env[61806]: DEBUG nova.scheduler.client.report [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Refreshing trait associations for resource provider a2858be1-fd22-4e08-979e-87ad25293407, traits: COMPUTE_IMAGE_TYPE_ISO,COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_SAME_HOST_COLD_MIGRATE,COMPUTE_NET_ATTACH_INTERFACE,COMPUTE_NODE {{(pid=61806) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:839}} [ 1135.595536] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e05cefbb-f788-46a6-80b4-4952c022fbcb {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1135.603126] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cb021280-7b76-4d6b-aed1-0b0e4d564a56 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1135.633169] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-82c20fe4-f6bf-4e75-8286-5ddb113af13a {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1135.640566] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9902e1bf-f8cd-43cc-af11-9f7be0f45264 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1135.653847] env[61806]: DEBUG nova.compute.provider_tree [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Updating inventory in ProviderTree for provider a2858be1-fd22-4e08-979e-87ad25293407 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 138, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61806) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1136.033697] env[61806]: DEBUG oslo_vmware.api [None req-6f41d8c4-228e-4028-861e-4b121b76c64a tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Task: {'id': task-1295268, 'name': PowerOnVM_Task, 'duration_secs': 0.386069} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1136.038063] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [None req-6f41d8c4-228e-4028-861e-4b121b76c64a tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] [instance: d0669232-6f33-4b30-97c2-2e4239af1ad0] Powered on the VM {{(pid=61806) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1136.038063] env[61806]: DEBUG nova.compute.manager [None req-6f41d8c4-228e-4028-861e-4b121b76c64a tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] [instance: d0669232-6f33-4b30-97c2-2e4239af1ad0] Checking state {{(pid=61806) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1136.038268] env[61806]: DEBUG oslo_vmware.api [None req-bf8f4ec5-63a0-4ddc-bc14-7d6705c06bed tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Task: {'id': task-1295267, 'name': PowerOnVM_Task} progress is 94%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1136.038958] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-afd35db5-563b-409d-85f8-309e84b33cd3 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1136.182436] env[61806]: DEBUG nova.scheduler.client.report [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Updated inventory for provider a2858be1-fd22-4e08-979e-87ad25293407 with generation 144 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 138, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61806) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:971}} [ 1136.182673] env[61806]: DEBUG nova.compute.provider_tree [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Updating resource provider a2858be1-fd22-4e08-979e-87ad25293407 generation from 144 to 145 during operation: update_inventory {{(pid=61806) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 1136.182825] env[61806]: DEBUG nova.compute.provider_tree [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Updating inventory in ProviderTree for provider a2858be1-fd22-4e08-979e-87ad25293407 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 138, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61806) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1136.531016] env[61806]: DEBUG oslo_vmware.api [None req-bf8f4ec5-63a0-4ddc-bc14-7d6705c06bed tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Task: {'id': task-1295267, 'name': PowerOnVM_Task, 'duration_secs': 0.591942} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1136.531335] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [None req-bf8f4ec5-63a0-4ddc-bc14-7d6705c06bed tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] [instance: c66fe2e7-7af0-48fa-8509-fcc24c1a437b] Powered on the VM {{(pid=61806) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1136.531544] env[61806]: INFO nova.compute.manager [None req-bf8f4ec5-63a0-4ddc-bc14-7d6705c06bed tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] [instance: c66fe2e7-7af0-48fa-8509-fcc24c1a437b] Took 7.70 seconds to spawn the instance on the hypervisor. [ 1136.531730] env[61806]: DEBUG nova.compute.manager [None req-bf8f4ec5-63a0-4ddc-bc14-7d6705c06bed tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] [instance: c66fe2e7-7af0-48fa-8509-fcc24c1a437b] Checking state {{(pid=61806) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1136.532510] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8af955c6-10de-4b41-9b63-f170d0380725 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1136.687705] env[61806]: DEBUG nova.compute.resource_tracker [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=61806) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 1136.687903] env[61806]: DEBUG oslo_concurrency.lockutils [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 2.851s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1137.051897] env[61806]: INFO nova.compute.manager [None req-bf8f4ec5-63a0-4ddc-bc14-7d6705c06bed tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] [instance: c66fe2e7-7af0-48fa-8509-fcc24c1a437b] Took 13.14 seconds to build instance. [ 1137.141917] env[61806]: DEBUG oslo_concurrency.lockutils [None req-868c882b-7772-43ab-9e8d-ec477bb859bf tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Acquiring lock "d0669232-6f33-4b30-97c2-2e4239af1ad0" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1137.142238] env[61806]: DEBUG oslo_concurrency.lockutils [None req-868c882b-7772-43ab-9e8d-ec477bb859bf tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Lock "d0669232-6f33-4b30-97c2-2e4239af1ad0" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1137.142473] env[61806]: DEBUG oslo_concurrency.lockutils [None req-868c882b-7772-43ab-9e8d-ec477bb859bf tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Acquiring lock "d0669232-6f33-4b30-97c2-2e4239af1ad0-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1137.142669] env[61806]: DEBUG oslo_concurrency.lockutils [None req-868c882b-7772-43ab-9e8d-ec477bb859bf tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Lock "d0669232-6f33-4b30-97c2-2e4239af1ad0-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1137.142844] env[61806]: DEBUG oslo_concurrency.lockutils [None req-868c882b-7772-43ab-9e8d-ec477bb859bf tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Lock "d0669232-6f33-4b30-97c2-2e4239af1ad0-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1137.144742] env[61806]: INFO nova.compute.manager [None req-868c882b-7772-43ab-9e8d-ec477bb859bf tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] [instance: d0669232-6f33-4b30-97c2-2e4239af1ad0] Terminating instance [ 1137.146450] env[61806]: DEBUG nova.compute.manager [None req-868c882b-7772-43ab-9e8d-ec477bb859bf tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] [instance: d0669232-6f33-4b30-97c2-2e4239af1ad0] Start destroying the instance on the hypervisor. {{(pid=61806) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1137.146645] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-868c882b-7772-43ab-9e8d-ec477bb859bf tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] [instance: d0669232-6f33-4b30-97c2-2e4239af1ad0] Destroying instance {{(pid=61806) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1137.147496] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-189f8c9c-765f-408b-910c-00015aa2d9c7 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1137.155071] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [None req-868c882b-7772-43ab-9e8d-ec477bb859bf tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] [instance: d0669232-6f33-4b30-97c2-2e4239af1ad0] Powering off the VM {{(pid=61806) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1137.155297] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-65beeb39-e8d9-495b-981e-5b7a70350d85 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1137.161166] env[61806]: DEBUG oslo_vmware.api [None req-868c882b-7772-43ab-9e8d-ec477bb859bf tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Waiting for the task: (returnval){ [ 1137.161166] env[61806]: value = "task-1295269" [ 1137.161166] env[61806]: _type = "Task" [ 1137.161166] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1137.168570] env[61806]: DEBUG oslo_vmware.api [None req-868c882b-7772-43ab-9e8d-ec477bb859bf tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Task: {'id': task-1295269, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1137.438415] env[61806]: DEBUG oslo_service.periodic_task [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=61806) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1137.438679] env[61806]: DEBUG oslo_service.periodic_task [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=61806) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1137.438849] env[61806]: DEBUG oslo_service.periodic_task [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=61806) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1137.439010] env[61806]: DEBUG oslo_service.periodic_task [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=61806) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1137.439165] env[61806]: DEBUG nova.compute.manager [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=61806) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10545}} [ 1137.554190] env[61806]: DEBUG oslo_concurrency.lockutils [None req-bf8f4ec5-63a0-4ddc-bc14-7d6705c06bed tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Lock "c66fe2e7-7af0-48fa-8509-fcc24c1a437b" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 14.648s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1137.672029] env[61806]: DEBUG oslo_vmware.api [None req-868c882b-7772-43ab-9e8d-ec477bb859bf tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Task: {'id': task-1295269, 'name': PowerOffVM_Task, 'duration_secs': 0.180344} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1137.672029] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [None req-868c882b-7772-43ab-9e8d-ec477bb859bf tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] [instance: d0669232-6f33-4b30-97c2-2e4239af1ad0] Powered off the VM {{(pid=61806) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1137.672029] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-868c882b-7772-43ab-9e8d-ec477bb859bf tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] [instance: d0669232-6f33-4b30-97c2-2e4239af1ad0] Unregistering the VM {{(pid=61806) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1137.672250] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-1c8fd374-ac0f-4978-9e42-32d54b3eaff1 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1137.746035] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-868c882b-7772-43ab-9e8d-ec477bb859bf tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] [instance: d0669232-6f33-4b30-97c2-2e4239af1ad0] Unregistered the VM {{(pid=61806) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1137.746248] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-868c882b-7772-43ab-9e8d-ec477bb859bf tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] [instance: d0669232-6f33-4b30-97c2-2e4239af1ad0] Deleting contents of the VM from datastore datastore2 {{(pid=61806) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1137.746410] env[61806]: DEBUG nova.virt.vmwareapi.ds_util [None req-868c882b-7772-43ab-9e8d-ec477bb859bf tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Deleting the datastore file [datastore2] d0669232-6f33-4b30-97c2-2e4239af1ad0 {{(pid=61806) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1137.746687] env[61806]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-63cb03c8-987a-4671-a786-9fee3156f7f6 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1137.753537] env[61806]: DEBUG oslo_vmware.api [None req-868c882b-7772-43ab-9e8d-ec477bb859bf tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Waiting for the task: (returnval){ [ 1137.753537] env[61806]: value = "task-1295271" [ 1137.753537] env[61806]: _type = "Task" [ 1137.753537] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1137.761772] env[61806]: DEBUG oslo_vmware.api [None req-868c882b-7772-43ab-9e8d-ec477bb859bf tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Task: {'id': task-1295271, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1138.022956] env[61806]: DEBUG oslo_service.periodic_task [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=61806) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1138.213514] env[61806]: DEBUG nova.compute.manager [req-608b958e-4d9d-4848-87fe-15a79c19bcdd req-15c06326-31bd-4f82-a03c-7ef621c5c65a service nova] [instance: c66fe2e7-7af0-48fa-8509-fcc24c1a437b] Received event network-changed-5bb12b4b-dfcb-416e-8095-91002ba3ca7d {{(pid=61806) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1138.213749] env[61806]: DEBUG nova.compute.manager [req-608b958e-4d9d-4848-87fe-15a79c19bcdd req-15c06326-31bd-4f82-a03c-7ef621c5c65a service nova] [instance: c66fe2e7-7af0-48fa-8509-fcc24c1a437b] Refreshing instance network info cache due to event network-changed-5bb12b4b-dfcb-416e-8095-91002ba3ca7d. {{(pid=61806) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 1138.214060] env[61806]: DEBUG oslo_concurrency.lockutils [req-608b958e-4d9d-4848-87fe-15a79c19bcdd req-15c06326-31bd-4f82-a03c-7ef621c5c65a service nova] Acquiring lock "refresh_cache-c66fe2e7-7af0-48fa-8509-fcc24c1a437b" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1138.214229] env[61806]: DEBUG oslo_concurrency.lockutils [req-608b958e-4d9d-4848-87fe-15a79c19bcdd req-15c06326-31bd-4f82-a03c-7ef621c5c65a service nova] Acquired lock "refresh_cache-c66fe2e7-7af0-48fa-8509-fcc24c1a437b" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1138.214401] env[61806]: DEBUG nova.network.neutron [req-608b958e-4d9d-4848-87fe-15a79c19bcdd req-15c06326-31bd-4f82-a03c-7ef621c5c65a service nova] [instance: c66fe2e7-7af0-48fa-8509-fcc24c1a437b] Refreshing network info cache for port 5bb12b4b-dfcb-416e-8095-91002ba3ca7d {{(pid=61806) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1138.263753] env[61806]: DEBUG oslo_vmware.api [None req-868c882b-7772-43ab-9e8d-ec477bb859bf tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Task: {'id': task-1295271, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.143022} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1138.264070] env[61806]: DEBUG nova.virt.vmwareapi.ds_util [None req-868c882b-7772-43ab-9e8d-ec477bb859bf tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Deleted the datastore file {{(pid=61806) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1138.264236] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-868c882b-7772-43ab-9e8d-ec477bb859bf tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] [instance: d0669232-6f33-4b30-97c2-2e4239af1ad0] Deleted contents of the VM from datastore datastore2 {{(pid=61806) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1138.264429] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-868c882b-7772-43ab-9e8d-ec477bb859bf tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] [instance: d0669232-6f33-4b30-97c2-2e4239af1ad0] Instance destroyed {{(pid=61806) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1138.264600] env[61806]: INFO nova.compute.manager [None req-868c882b-7772-43ab-9e8d-ec477bb859bf tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] [instance: d0669232-6f33-4b30-97c2-2e4239af1ad0] Took 1.12 seconds to destroy the instance on the hypervisor. [ 1138.264846] env[61806]: DEBUG oslo.service.loopingcall [None req-868c882b-7772-43ab-9e8d-ec477bb859bf tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61806) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1138.265052] env[61806]: DEBUG nova.compute.manager [-] [instance: d0669232-6f33-4b30-97c2-2e4239af1ad0] Deallocating network for instance {{(pid=61806) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1138.265151] env[61806]: DEBUG nova.network.neutron [-] [instance: d0669232-6f33-4b30-97c2-2e4239af1ad0] deallocate_for_instance() {{(pid=61806) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1138.961226] env[61806]: DEBUG nova.network.neutron [req-608b958e-4d9d-4848-87fe-15a79c19bcdd req-15c06326-31bd-4f82-a03c-7ef621c5c65a service nova] [instance: c66fe2e7-7af0-48fa-8509-fcc24c1a437b] Updated VIF entry in instance network info cache for port 5bb12b4b-dfcb-416e-8095-91002ba3ca7d. {{(pid=61806) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1138.961651] env[61806]: DEBUG nova.network.neutron [req-608b958e-4d9d-4848-87fe-15a79c19bcdd req-15c06326-31bd-4f82-a03c-7ef621c5c65a service nova] [instance: c66fe2e7-7af0-48fa-8509-fcc24c1a437b] Updating instance_info_cache with network_info: [{"id": "5bb12b4b-dfcb-416e-8095-91002ba3ca7d", "address": "fa:16:3e:9d:21:7a", "network": {"id": "746a03b6-c1a6-4d27-98b9-0bee6376634f", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-1971382587-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.245", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c1f2d161c50f4b70987b90e6364c3779", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6a31b224-77d7-48a4-af87-312758611b8e", "external-id": "nsx-vlan-transportzone-761", "segmentation_id": 761, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap5bb12b4b-df", "ovs_interfaceid": "5bb12b4b-dfcb-416e-8095-91002ba3ca7d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61806) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1138.975422] env[61806]: DEBUG oslo_concurrency.lockutils [None req-f8795417-655c-4a2b-beb9-4c53a9fb716b tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] Acquiring lock "29210bcd-0fe1-4ed9-a459-abc8016a4255" by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1138.975659] env[61806]: DEBUG oslo_concurrency.lockutils [None req-f8795417-655c-4a2b-beb9-4c53a9fb716b tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] Lock "29210bcd-0fe1-4ed9-a459-abc8016a4255" acquired by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" :: waited 0.000s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1138.975832] env[61806]: INFO nova.compute.manager [None req-f8795417-655c-4a2b-beb9-4c53a9fb716b tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] [instance: 29210bcd-0fe1-4ed9-a459-abc8016a4255] Shelving [ 1139.197337] env[61806]: DEBUG nova.network.neutron [-] [instance: d0669232-6f33-4b30-97c2-2e4239af1ad0] Updating instance_info_cache with network_info: [] {{(pid=61806) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1139.464439] env[61806]: DEBUG oslo_concurrency.lockutils [req-608b958e-4d9d-4848-87fe-15a79c19bcdd req-15c06326-31bd-4f82-a03c-7ef621c5c65a service nova] Releasing lock "refresh_cache-c66fe2e7-7af0-48fa-8509-fcc24c1a437b" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1139.484036] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [None req-f8795417-655c-4a2b-beb9-4c53a9fb716b tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] [instance: 29210bcd-0fe1-4ed9-a459-abc8016a4255] Powering off the VM {{(pid=61806) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1139.484036] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-f2c6a83b-a511-4fb2-a879-c40739a76568 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1139.491826] env[61806]: DEBUG oslo_vmware.api [None req-f8795417-655c-4a2b-beb9-4c53a9fb716b tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] Waiting for the task: (returnval){ [ 1139.491826] env[61806]: value = "task-1295272" [ 1139.491826] env[61806]: _type = "Task" [ 1139.491826] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1139.500742] env[61806]: DEBUG oslo_vmware.api [None req-f8795417-655c-4a2b-beb9-4c53a9fb716b tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] Task: {'id': task-1295272, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1139.700107] env[61806]: INFO nova.compute.manager [-] [instance: d0669232-6f33-4b30-97c2-2e4239af1ad0] Took 1.43 seconds to deallocate network for instance. [ 1140.001354] env[61806]: DEBUG oslo_vmware.api [None req-f8795417-655c-4a2b-beb9-4c53a9fb716b tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] Task: {'id': task-1295272, 'name': PowerOffVM_Task, 'duration_secs': 0.399715} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1140.001640] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [None req-f8795417-655c-4a2b-beb9-4c53a9fb716b tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] [instance: 29210bcd-0fe1-4ed9-a459-abc8016a4255] Powered off the VM {{(pid=61806) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1140.002410] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-32118672-20a6-4e7f-a8b8-809bb713323f {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1140.019767] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d308e260-1e66-407d-bc54-160c3cedfb87 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1140.206659] env[61806]: DEBUG oslo_concurrency.lockutils [None req-868c882b-7772-43ab-9e8d-ec477bb859bf tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1140.206928] env[61806]: DEBUG oslo_concurrency.lockutils [None req-868c882b-7772-43ab-9e8d-ec477bb859bf tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.000s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1140.207172] env[61806]: DEBUG nova.objects.instance [None req-868c882b-7772-43ab-9e8d-ec477bb859bf tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Lazy-loading 'resources' on Instance uuid d0669232-6f33-4b30-97c2-2e4239af1ad0 {{(pid=61806) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1140.238241] env[61806]: DEBUG nova.compute.manager [req-6fbe5ea7-bf24-414e-b4ef-7ec2db07ef52 req-6f779dfd-85eb-49be-b1b6-c9c0e4e1605d service nova] [instance: d0669232-6f33-4b30-97c2-2e4239af1ad0] Received event network-vif-deleted-f958e9a3-f8d6-4665-a59f-876aab66edfb {{(pid=61806) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1140.529287] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-f8795417-655c-4a2b-beb9-4c53a9fb716b tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] [instance: 29210bcd-0fe1-4ed9-a459-abc8016a4255] Creating Snapshot of the VM instance {{(pid=61806) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 1140.529627] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-8aaf2aeb-3522-4865-853f-7b536ce720df {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1140.537048] env[61806]: DEBUG oslo_vmware.api [None req-f8795417-655c-4a2b-beb9-4c53a9fb716b tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] Waiting for the task: (returnval){ [ 1140.537048] env[61806]: value = "task-1295273" [ 1140.537048] env[61806]: _type = "Task" [ 1140.537048] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1140.544558] env[61806]: DEBUG oslo_vmware.api [None req-f8795417-655c-4a2b-beb9-4c53a9fb716b tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] Task: {'id': task-1295273, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1140.767376] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f1fadb44-d26a-42bf-a73b-a1f560296c44 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1140.775235] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-548f5dea-d33b-469c-af68-fcc2909c9a31 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1140.806264] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6e3a5974-535a-48da-923c-bdab6760e421 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1140.813643] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c68a7ee2-7942-4631-a766-070f98523b08 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1140.827859] env[61806]: DEBUG nova.compute.provider_tree [None req-868c882b-7772-43ab-9e8d-ec477bb859bf tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Inventory has not changed in ProviderTree for provider: a2858be1-fd22-4e08-979e-87ad25293407 {{(pid=61806) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1141.048457] env[61806]: DEBUG oslo_vmware.api [None req-f8795417-655c-4a2b-beb9-4c53a9fb716b tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] Task: {'id': task-1295273, 'name': CreateSnapshot_Task, 'duration_secs': 0.442179} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1141.048790] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-f8795417-655c-4a2b-beb9-4c53a9fb716b tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] [instance: 29210bcd-0fe1-4ed9-a459-abc8016a4255] Created Snapshot of the VM instance {{(pid=61806) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 1141.049624] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-87590e38-4121-48ee-9174-f3679357060c {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1141.331099] env[61806]: DEBUG nova.scheduler.client.report [None req-868c882b-7772-43ab-9e8d-ec477bb859bf tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Inventory has not changed for provider a2858be1-fd22-4e08-979e-87ad25293407 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 138, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61806) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1141.567286] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-f8795417-655c-4a2b-beb9-4c53a9fb716b tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] [instance: 29210bcd-0fe1-4ed9-a459-abc8016a4255] Creating linked-clone VM from snapshot {{(pid=61806) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 1141.567714] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-9c94ea14-f57d-46cb-954f-59507d7747e5 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1141.576517] env[61806]: DEBUG oslo_vmware.api [None req-f8795417-655c-4a2b-beb9-4c53a9fb716b tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] Waiting for the task: (returnval){ [ 1141.576517] env[61806]: value = "task-1295274" [ 1141.576517] env[61806]: _type = "Task" [ 1141.576517] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1141.584228] env[61806]: DEBUG oslo_vmware.api [None req-f8795417-655c-4a2b-beb9-4c53a9fb716b tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] Task: {'id': task-1295274, 'name': CloneVM_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1141.836312] env[61806]: DEBUG oslo_concurrency.lockutils [None req-868c882b-7772-43ab-9e8d-ec477bb859bf tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.629s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1141.855588] env[61806]: INFO nova.scheduler.client.report [None req-868c882b-7772-43ab-9e8d-ec477bb859bf tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Deleted allocations for instance d0669232-6f33-4b30-97c2-2e4239af1ad0 [ 1142.086277] env[61806]: DEBUG oslo_vmware.api [None req-f8795417-655c-4a2b-beb9-4c53a9fb716b tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] Task: {'id': task-1295274, 'name': CloneVM_Task} progress is 94%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1142.363095] env[61806]: DEBUG oslo_concurrency.lockutils [None req-868c882b-7772-43ab-9e8d-ec477bb859bf tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Lock "d0669232-6f33-4b30-97c2-2e4239af1ad0" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.221s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1142.587154] env[61806]: DEBUG oslo_vmware.api [None req-f8795417-655c-4a2b-beb9-4c53a9fb716b tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] Task: {'id': task-1295274, 'name': CloneVM_Task} progress is 95%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1143.087537] env[61806]: DEBUG oslo_vmware.api [None req-f8795417-655c-4a2b-beb9-4c53a9fb716b tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] Task: {'id': task-1295274, 'name': CloneVM_Task, 'duration_secs': 1.096508} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1143.088948] env[61806]: INFO nova.virt.vmwareapi.vmops [None req-f8795417-655c-4a2b-beb9-4c53a9fb716b tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] [instance: 29210bcd-0fe1-4ed9-a459-abc8016a4255] Created linked-clone VM from snapshot [ 1143.092233] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e39522e9-2faa-4b51-82c2-95a24980d8d7 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1143.094924] env[61806]: DEBUG oslo_concurrency.lockutils [None req-2bcb8d7f-cd6c-4f73-bd59-e28663212f52 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Acquiring lock "4f26445f-35d0-4f53-9014-ef3e2eb1cbdf" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1143.095164] env[61806]: DEBUG oslo_concurrency.lockutils [None req-2bcb8d7f-cd6c-4f73-bd59-e28663212f52 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Lock "4f26445f-35d0-4f53-9014-ef3e2eb1cbdf" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1143.101730] env[61806]: DEBUG nova.virt.vmwareapi.images [None req-f8795417-655c-4a2b-beb9-4c53a9fb716b tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] [instance: 29210bcd-0fe1-4ed9-a459-abc8016a4255] Uploading image 2ea32304-20d1-4039-a93a-b4c97a5656bf {{(pid=61806) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:442}} [ 1143.126296] env[61806]: DEBUG oslo_vmware.rw_handles [None req-f8795417-655c-4a2b-beb9-4c53a9fb716b tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] Creating HttpNfcLease lease for exporting VM: (result){ [ 1143.126296] env[61806]: value = "vm-277768" [ 1143.126296] env[61806]: _type = "VirtualMachine" [ 1143.126296] env[61806]: }. {{(pid=61806) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:478}} [ 1143.126580] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ExportVm with opID=oslo.vmware-6434a02f-2ccc-4b78-affa-b15bf0abcdc1 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1143.133164] env[61806]: DEBUG oslo_vmware.rw_handles [None req-f8795417-655c-4a2b-beb9-4c53a9fb716b tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] Lease: (returnval){ [ 1143.133164] env[61806]: value = "session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]52964e8c-67cf-5d81-a365-36bbff3312f5" [ 1143.133164] env[61806]: _type = "HttpNfcLease" [ 1143.133164] env[61806]: } obtained for exporting VM: (result){ [ 1143.133164] env[61806]: value = "vm-277768" [ 1143.133164] env[61806]: _type = "VirtualMachine" [ 1143.133164] env[61806]: }. {{(pid=61806) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:481}} [ 1143.133388] env[61806]: DEBUG oslo_vmware.api [None req-f8795417-655c-4a2b-beb9-4c53a9fb716b tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] Waiting for the lease: (returnval){ [ 1143.133388] env[61806]: value = "session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]52964e8c-67cf-5d81-a365-36bbff3312f5" [ 1143.133388] env[61806]: _type = "HttpNfcLease" [ 1143.133388] env[61806]: } to be ready. {{(pid=61806) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 1143.139187] env[61806]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1143.139187] env[61806]: value = "session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]52964e8c-67cf-5d81-a365-36bbff3312f5" [ 1143.139187] env[61806]: _type = "HttpNfcLease" [ 1143.139187] env[61806]: } is initializing. {{(pid=61806) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 1143.597418] env[61806]: DEBUG nova.compute.manager [None req-2bcb8d7f-cd6c-4f73-bd59-e28663212f52 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] [instance: 4f26445f-35d0-4f53-9014-ef3e2eb1cbdf] Starting instance... {{(pid=61806) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1143.642797] env[61806]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1143.642797] env[61806]: value = "session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]52964e8c-67cf-5d81-a365-36bbff3312f5" [ 1143.642797] env[61806]: _type = "HttpNfcLease" [ 1143.642797] env[61806]: } is ready. {{(pid=61806) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 1143.643108] env[61806]: DEBUG oslo_vmware.rw_handles [None req-f8795417-655c-4a2b-beb9-4c53a9fb716b tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 1143.643108] env[61806]: value = "session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]52964e8c-67cf-5d81-a365-36bbff3312f5" [ 1143.643108] env[61806]: _type = "HttpNfcLease" [ 1143.643108] env[61806]: }. {{(pid=61806) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:486}} [ 1143.643829] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b6de62b2-f82f-4300-a4e8-6738bb2d7fd1 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1143.651039] env[61806]: DEBUG oslo_vmware.rw_handles [None req-f8795417-655c-4a2b-beb9-4c53a9fb716b tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] Found VMDK URL: https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52ad2f21-283b-03c3-db4e-9ab0e06ccfeb/disk-0.vmdk from lease info. {{(pid=61806) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 1143.651232] env[61806]: DEBUG oslo_vmware.rw_handles [None req-f8795417-655c-4a2b-beb9-4c53a9fb716b tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] Opening URL: https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52ad2f21-283b-03c3-db4e-9ab0e06ccfeb/disk-0.vmdk for reading. {{(pid=61806) _create_read_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:99}} [ 1143.832131] env[61806]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-12a8229a-d2f7-4edd-8754-37dfd9c70aa0 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1144.118297] env[61806]: DEBUG oslo_concurrency.lockutils [None req-2bcb8d7f-cd6c-4f73-bd59-e28663212f52 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1144.118592] env[61806]: DEBUG oslo_concurrency.lockutils [None req-2bcb8d7f-cd6c-4f73-bd59-e28663212f52 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.001s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1144.120292] env[61806]: INFO nova.compute.claims [None req-2bcb8d7f-cd6c-4f73-bd59-e28663212f52 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] [instance: 4f26445f-35d0-4f53-9014-ef3e2eb1cbdf] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1145.182028] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fdcf8f9c-064d-4924-ba8e-b9878e932592 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1145.190072] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9ffe7cf2-8f54-46a4-a753-116011c6ee54 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1145.222478] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-70c7cb23-83b3-45b5-8768-3835f3f7e283 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1145.230358] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-14df155e-b485-42d1-944a-8e1b518c17d1 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1145.244567] env[61806]: DEBUG nova.compute.provider_tree [None req-2bcb8d7f-cd6c-4f73-bd59-e28663212f52 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Inventory has not changed in ProviderTree for provider: a2858be1-fd22-4e08-979e-87ad25293407 {{(pid=61806) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1145.747877] env[61806]: DEBUG nova.scheduler.client.report [None req-2bcb8d7f-cd6c-4f73-bd59-e28663212f52 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Inventory has not changed for provider a2858be1-fd22-4e08-979e-87ad25293407 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 138, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61806) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1146.253389] env[61806]: DEBUG oslo_concurrency.lockutils [None req-2bcb8d7f-cd6c-4f73-bd59-e28663212f52 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.135s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1146.253943] env[61806]: DEBUG nova.compute.manager [None req-2bcb8d7f-cd6c-4f73-bd59-e28663212f52 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] [instance: 4f26445f-35d0-4f53-9014-ef3e2eb1cbdf] Start building networks asynchronously for instance. {{(pid=61806) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1146.759095] env[61806]: DEBUG nova.compute.utils [None req-2bcb8d7f-cd6c-4f73-bd59-e28663212f52 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Using /dev/sd instead of None {{(pid=61806) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1146.760723] env[61806]: DEBUG nova.compute.manager [None req-2bcb8d7f-cd6c-4f73-bd59-e28663212f52 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] [instance: 4f26445f-35d0-4f53-9014-ef3e2eb1cbdf] Allocating IP information in the background. {{(pid=61806) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1146.760900] env[61806]: DEBUG nova.network.neutron [None req-2bcb8d7f-cd6c-4f73-bd59-e28663212f52 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] [instance: 4f26445f-35d0-4f53-9014-ef3e2eb1cbdf] allocate_for_instance() {{(pid=61806) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 1146.810846] env[61806]: DEBUG nova.policy [None req-2bcb8d7f-cd6c-4f73-bd59-e28663212f52 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '772f4c6d731c4d5daea11b37a0f6718b', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '1c55cc0455584fbb8cda88d33b8f8b7e', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61806) authorize /opt/stack/nova/nova/policy.py:201}} [ 1147.082594] env[61806]: DEBUG nova.network.neutron [None req-2bcb8d7f-cd6c-4f73-bd59-e28663212f52 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] [instance: 4f26445f-35d0-4f53-9014-ef3e2eb1cbdf] Successfully created port: a71b2870-bc50-47f4-840d-57e956c3f86d {{(pid=61806) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1147.263774] env[61806]: DEBUG nova.compute.manager [None req-2bcb8d7f-cd6c-4f73-bd59-e28663212f52 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] [instance: 4f26445f-35d0-4f53-9014-ef3e2eb1cbdf] Start building block device mappings for instance. {{(pid=61806) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1148.274567] env[61806]: DEBUG nova.compute.manager [None req-2bcb8d7f-cd6c-4f73-bd59-e28663212f52 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] [instance: 4f26445f-35d0-4f53-9014-ef3e2eb1cbdf] Start spawning the instance on the hypervisor. {{(pid=61806) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1148.303563] env[61806]: DEBUG nova.virt.hardware [None req-2bcb8d7f-cd6c-4f73-bd59-e28663212f52 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-15T18:05:10Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-15T18:04:52Z,direct_url=,disk_format='vmdk',id=ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='84ba83fce2c349a6988173c5d6fc3b07',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-15T18:04:53Z,virtual_size=,visibility=), allow threads: False {{(pid=61806) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1148.303879] env[61806]: DEBUG nova.virt.hardware [None req-2bcb8d7f-cd6c-4f73-bd59-e28663212f52 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Flavor limits 0:0:0 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1148.304097] env[61806]: DEBUG nova.virt.hardware [None req-2bcb8d7f-cd6c-4f73-bd59-e28663212f52 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Image limits 0:0:0 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1148.304310] env[61806]: DEBUG nova.virt.hardware [None req-2bcb8d7f-cd6c-4f73-bd59-e28663212f52 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Flavor pref 0:0:0 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1148.304487] env[61806]: DEBUG nova.virt.hardware [None req-2bcb8d7f-cd6c-4f73-bd59-e28663212f52 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Image pref 0:0:0 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1148.304642] env[61806]: DEBUG nova.virt.hardware [None req-2bcb8d7f-cd6c-4f73-bd59-e28663212f52 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1148.304896] env[61806]: DEBUG nova.virt.hardware [None req-2bcb8d7f-cd6c-4f73-bd59-e28663212f52 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61806) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1148.305093] env[61806]: DEBUG nova.virt.hardware [None req-2bcb8d7f-cd6c-4f73-bd59-e28663212f52 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61806) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1148.305347] env[61806]: DEBUG nova.virt.hardware [None req-2bcb8d7f-cd6c-4f73-bd59-e28663212f52 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Got 1 possible topologies {{(pid=61806) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1148.305543] env[61806]: DEBUG nova.virt.hardware [None req-2bcb8d7f-cd6c-4f73-bd59-e28663212f52 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61806) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1148.305740] env[61806]: DEBUG nova.virt.hardware [None req-2bcb8d7f-cd6c-4f73-bd59-e28663212f52 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61806) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1148.306716] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-34d33dab-90d7-4287-acce-537691a7c4a5 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1148.314890] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7eb7c40a-8269-4f7e-b44c-862cc521f8f2 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1148.480895] env[61806]: DEBUG nova.compute.manager [req-bc9d242d-9b9f-4017-b56a-d231af431b70 req-e7dad048-c8f4-4131-8d02-2ca7c9e51cf6 service nova] [instance: 4f26445f-35d0-4f53-9014-ef3e2eb1cbdf] Received event network-vif-plugged-a71b2870-bc50-47f4-840d-57e956c3f86d {{(pid=61806) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1148.480895] env[61806]: DEBUG oslo_concurrency.lockutils [req-bc9d242d-9b9f-4017-b56a-d231af431b70 req-e7dad048-c8f4-4131-8d02-2ca7c9e51cf6 service nova] Acquiring lock "4f26445f-35d0-4f53-9014-ef3e2eb1cbdf-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1148.481176] env[61806]: DEBUG oslo_concurrency.lockutils [req-bc9d242d-9b9f-4017-b56a-d231af431b70 req-e7dad048-c8f4-4131-8d02-2ca7c9e51cf6 service nova] Lock "4f26445f-35d0-4f53-9014-ef3e2eb1cbdf-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1148.481363] env[61806]: DEBUG oslo_concurrency.lockutils [req-bc9d242d-9b9f-4017-b56a-d231af431b70 req-e7dad048-c8f4-4131-8d02-2ca7c9e51cf6 service nova] Lock "4f26445f-35d0-4f53-9014-ef3e2eb1cbdf-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1148.481552] env[61806]: DEBUG nova.compute.manager [req-bc9d242d-9b9f-4017-b56a-d231af431b70 req-e7dad048-c8f4-4131-8d02-2ca7c9e51cf6 service nova] [instance: 4f26445f-35d0-4f53-9014-ef3e2eb1cbdf] No waiting events found dispatching network-vif-plugged-a71b2870-bc50-47f4-840d-57e956c3f86d {{(pid=61806) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1148.482429] env[61806]: WARNING nova.compute.manager [req-bc9d242d-9b9f-4017-b56a-d231af431b70 req-e7dad048-c8f4-4131-8d02-2ca7c9e51cf6 service nova] [instance: 4f26445f-35d0-4f53-9014-ef3e2eb1cbdf] Received unexpected event network-vif-plugged-a71b2870-bc50-47f4-840d-57e956c3f86d for instance with vm_state building and task_state spawning. [ 1149.009449] env[61806]: DEBUG nova.network.neutron [None req-2bcb8d7f-cd6c-4f73-bd59-e28663212f52 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] [instance: 4f26445f-35d0-4f53-9014-ef3e2eb1cbdf] Successfully updated port: a71b2870-bc50-47f4-840d-57e956c3f86d {{(pid=61806) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1149.031340] env[61806]: DEBUG nova.compute.manager [req-a4457966-cc2c-4288-8b1f-8a8fbc012d93 req-0f9ac021-c4d4-441d-8393-724130a2b605 service nova] [instance: 4f26445f-35d0-4f53-9014-ef3e2eb1cbdf] Received event network-changed-a71b2870-bc50-47f4-840d-57e956c3f86d {{(pid=61806) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1149.031697] env[61806]: DEBUG nova.compute.manager [req-a4457966-cc2c-4288-8b1f-8a8fbc012d93 req-0f9ac021-c4d4-441d-8393-724130a2b605 service nova] [instance: 4f26445f-35d0-4f53-9014-ef3e2eb1cbdf] Refreshing instance network info cache due to event network-changed-a71b2870-bc50-47f4-840d-57e956c3f86d. {{(pid=61806) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 1149.031905] env[61806]: DEBUG oslo_concurrency.lockutils [req-a4457966-cc2c-4288-8b1f-8a8fbc012d93 req-0f9ac021-c4d4-441d-8393-724130a2b605 service nova] Acquiring lock "refresh_cache-4f26445f-35d0-4f53-9014-ef3e2eb1cbdf" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1149.031991] env[61806]: DEBUG oslo_concurrency.lockutils [req-a4457966-cc2c-4288-8b1f-8a8fbc012d93 req-0f9ac021-c4d4-441d-8393-724130a2b605 service nova] Acquired lock "refresh_cache-4f26445f-35d0-4f53-9014-ef3e2eb1cbdf" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1149.032178] env[61806]: DEBUG nova.network.neutron [req-a4457966-cc2c-4288-8b1f-8a8fbc012d93 req-0f9ac021-c4d4-441d-8393-724130a2b605 service nova] [instance: 4f26445f-35d0-4f53-9014-ef3e2eb1cbdf] Refreshing network info cache for port a71b2870-bc50-47f4-840d-57e956c3f86d {{(pid=61806) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1149.513262] env[61806]: DEBUG oslo_concurrency.lockutils [None req-2bcb8d7f-cd6c-4f73-bd59-e28663212f52 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Acquiring lock "refresh_cache-4f26445f-35d0-4f53-9014-ef3e2eb1cbdf" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1149.566143] env[61806]: DEBUG nova.network.neutron [req-a4457966-cc2c-4288-8b1f-8a8fbc012d93 req-0f9ac021-c4d4-441d-8393-724130a2b605 service nova] [instance: 4f26445f-35d0-4f53-9014-ef3e2eb1cbdf] Instance cache missing network info. {{(pid=61806) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 1149.640173] env[61806]: DEBUG nova.network.neutron [req-a4457966-cc2c-4288-8b1f-8a8fbc012d93 req-0f9ac021-c4d4-441d-8393-724130a2b605 service nova] [instance: 4f26445f-35d0-4f53-9014-ef3e2eb1cbdf] Updating instance_info_cache with network_info: [] {{(pid=61806) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1150.146635] env[61806]: DEBUG oslo_concurrency.lockutils [req-a4457966-cc2c-4288-8b1f-8a8fbc012d93 req-0f9ac021-c4d4-441d-8393-724130a2b605 service nova] Releasing lock "refresh_cache-4f26445f-35d0-4f53-9014-ef3e2eb1cbdf" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1150.147149] env[61806]: DEBUG oslo_concurrency.lockutils [None req-2bcb8d7f-cd6c-4f73-bd59-e28663212f52 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Acquired lock "refresh_cache-4f26445f-35d0-4f53-9014-ef3e2eb1cbdf" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1150.147340] env[61806]: DEBUG nova.network.neutron [None req-2bcb8d7f-cd6c-4f73-bd59-e28663212f52 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] [instance: 4f26445f-35d0-4f53-9014-ef3e2eb1cbdf] Building network info cache for instance {{(pid=61806) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1150.689186] env[61806]: DEBUG nova.network.neutron [None req-2bcb8d7f-cd6c-4f73-bd59-e28663212f52 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] [instance: 4f26445f-35d0-4f53-9014-ef3e2eb1cbdf] Instance cache missing network info. {{(pid=61806) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 1150.860471] env[61806]: DEBUG nova.network.neutron [None req-2bcb8d7f-cd6c-4f73-bd59-e28663212f52 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] [instance: 4f26445f-35d0-4f53-9014-ef3e2eb1cbdf] Updating instance_info_cache with network_info: [{"id": "a71b2870-bc50-47f4-840d-57e956c3f86d", "address": "fa:16:3e:74:22:b8", "network": {"id": "33bf5b21-82b0-432e-ae3d-9c2ca75161d0", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-1294159336-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "1c55cc0455584fbb8cda88d33b8f8b7e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "5b0fa7a2-ebd9-4788-8904-7bf250ce466c", "external-id": "nsx-vlan-transportzone-669", "segmentation_id": 669, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa71b2870-bc", "ovs_interfaceid": "a71b2870-bc50-47f4-840d-57e956c3f86d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61806) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1151.363197] env[61806]: DEBUG oslo_concurrency.lockutils [None req-2bcb8d7f-cd6c-4f73-bd59-e28663212f52 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Releasing lock "refresh_cache-4f26445f-35d0-4f53-9014-ef3e2eb1cbdf" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1151.363550] env[61806]: DEBUG nova.compute.manager [None req-2bcb8d7f-cd6c-4f73-bd59-e28663212f52 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] [instance: 4f26445f-35d0-4f53-9014-ef3e2eb1cbdf] Instance network_info: |[{"id": "a71b2870-bc50-47f4-840d-57e956c3f86d", "address": "fa:16:3e:74:22:b8", "network": {"id": "33bf5b21-82b0-432e-ae3d-9c2ca75161d0", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-1294159336-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "1c55cc0455584fbb8cda88d33b8f8b7e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "5b0fa7a2-ebd9-4788-8904-7bf250ce466c", "external-id": "nsx-vlan-transportzone-669", "segmentation_id": 669, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa71b2870-bc", "ovs_interfaceid": "a71b2870-bc50-47f4-840d-57e956c3f86d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61806) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1151.364072] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-2bcb8d7f-cd6c-4f73-bd59-e28663212f52 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] [instance: 4f26445f-35d0-4f53-9014-ef3e2eb1cbdf] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:74:22:b8', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '5b0fa7a2-ebd9-4788-8904-7bf250ce466c', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'a71b2870-bc50-47f4-840d-57e956c3f86d', 'vif_model': 'vmxnet3'}] {{(pid=61806) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1151.371733] env[61806]: DEBUG oslo.service.loopingcall [None req-2bcb8d7f-cd6c-4f73-bd59-e28663212f52 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61806) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1151.371972] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 4f26445f-35d0-4f53-9014-ef3e2eb1cbdf] Creating VM on the ESX host {{(pid=61806) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 1151.372226] env[61806]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-63f9d9d3-1c82-4b64-9044-8dcf383ff1a2 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1151.392330] env[61806]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1151.392330] env[61806]: value = "task-1295276" [ 1151.392330] env[61806]: _type = "Task" [ 1151.392330] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1151.399904] env[61806]: DEBUG oslo_vmware.api [-] Task: {'id': task-1295276, 'name': CreateVM_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1151.902130] env[61806]: DEBUG oslo_vmware.api [-] Task: {'id': task-1295276, 'name': CreateVM_Task, 'duration_secs': 0.380636} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1151.902430] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 4f26445f-35d0-4f53-9014-ef3e2eb1cbdf] Created VM on the ESX host {{(pid=61806) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 1151.902997] env[61806]: DEBUG oslo_concurrency.lockutils [None req-2bcb8d7f-cd6c-4f73-bd59-e28663212f52 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1151.903190] env[61806]: DEBUG oslo_concurrency.lockutils [None req-2bcb8d7f-cd6c-4f73-bd59-e28663212f52 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Acquired lock "[datastore1] devstack-image-cache_base/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1151.903534] env[61806]: DEBUG oslo_concurrency.lockutils [None req-2bcb8d7f-cd6c-4f73-bd59-e28663212f52 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1151.904120] env[61806]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ab9b78de-6e07-4cf2-84d9-80dfb5a6903f {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1151.908669] env[61806]: DEBUG oslo_vmware.api [None req-2bcb8d7f-cd6c-4f73-bd59-e28663212f52 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Waiting for the task: (returnval){ [ 1151.908669] env[61806]: value = "session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]526c988c-0a4f-3686-20e1-2c022719843b" [ 1151.908669] env[61806]: _type = "Task" [ 1151.908669] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1151.916248] env[61806]: DEBUG oslo_vmware.api [None req-2bcb8d7f-cd6c-4f73-bd59-e28663212f52 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Task: {'id': session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]526c988c-0a4f-3686-20e1-2c022719843b, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1152.418956] env[61806]: DEBUG oslo_vmware.api [None req-2bcb8d7f-cd6c-4f73-bd59-e28663212f52 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Task: {'id': session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]526c988c-0a4f-3686-20e1-2c022719843b, 'name': SearchDatastore_Task, 'duration_secs': 0.038661} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1152.419307] env[61806]: DEBUG oslo_concurrency.lockutils [None req-2bcb8d7f-cd6c-4f73-bd59-e28663212f52 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Releasing lock "[datastore1] devstack-image-cache_base/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1152.419552] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-2bcb8d7f-cd6c-4f73-bd59-e28663212f52 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] [instance: 4f26445f-35d0-4f53-9014-ef3e2eb1cbdf] Processing image ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa {{(pid=61806) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1152.419793] env[61806]: DEBUG oslo_concurrency.lockutils [None req-2bcb8d7f-cd6c-4f73-bd59-e28663212f52 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa.vmdk" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1152.419949] env[61806]: DEBUG oslo_concurrency.lockutils [None req-2bcb8d7f-cd6c-4f73-bd59-e28663212f52 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Acquired lock "[datastore1] devstack-image-cache_base/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa.vmdk" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1152.420158] env[61806]: DEBUG nova.virt.vmwareapi.ds_util [None req-2bcb8d7f-cd6c-4f73-bd59-e28663212f52 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61806) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1152.420428] env[61806]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-d2ef2aec-bf12-452e-9f54-8ef23fc7585e {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1152.428902] env[61806]: DEBUG nova.virt.vmwareapi.ds_util [None req-2bcb8d7f-cd6c-4f73-bd59-e28663212f52 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61806) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1152.429100] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-2bcb8d7f-cd6c-4f73-bd59-e28663212f52 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61806) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 1152.429845] env[61806]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-39159b9f-65e2-4492-87fe-3d37a90b3aa1 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1152.436577] env[61806]: DEBUG oslo_vmware.api [None req-2bcb8d7f-cd6c-4f73-bd59-e28663212f52 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Waiting for the task: (returnval){ [ 1152.436577] env[61806]: value = "session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]52a67362-e851-be99-772f-1d4f259d0c38" [ 1152.436577] env[61806]: _type = "Task" [ 1152.436577] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1152.443788] env[61806]: DEBUG oslo_vmware.api [None req-2bcb8d7f-cd6c-4f73-bd59-e28663212f52 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Task: {'id': session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]52a67362-e851-be99-772f-1d4f259d0c38, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1152.761284] env[61806]: DEBUG oslo_vmware.rw_handles [None req-f8795417-655c-4a2b-beb9-4c53a9fb716b tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] Getting lease state for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52ad2f21-283b-03c3-db4e-9ab0e06ccfeb/disk-0.vmdk. {{(pid=61806) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 1152.762245] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dba53c53-2e93-4ff9-8b3b-4a18047fe8ce {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1152.768810] env[61806]: DEBUG oslo_vmware.rw_handles [None req-f8795417-655c-4a2b-beb9-4c53a9fb716b tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] Lease for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52ad2f21-283b-03c3-db4e-9ab0e06ccfeb/disk-0.vmdk is in state: ready. {{(pid=61806) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 1152.768987] env[61806]: ERROR oslo_vmware.rw_handles [None req-f8795417-655c-4a2b-beb9-4c53a9fb716b tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] Aborting lease for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52ad2f21-283b-03c3-db4e-9ab0e06ccfeb/disk-0.vmdk due to incomplete transfer. [ 1152.769255] env[61806]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseAbort with opID=oslo.vmware-326df72b-1ce6-4b0b-bffb-2f85bc17e5cb {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1152.775360] env[61806]: DEBUG oslo_vmware.rw_handles [None req-f8795417-655c-4a2b-beb9-4c53a9fb716b tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] Closed VMDK read handle for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52ad2f21-283b-03c3-db4e-9ab0e06ccfeb/disk-0.vmdk. {{(pid=61806) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:735}} [ 1152.775563] env[61806]: DEBUG nova.virt.vmwareapi.images [None req-f8795417-655c-4a2b-beb9-4c53a9fb716b tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] [instance: 29210bcd-0fe1-4ed9-a459-abc8016a4255] Uploaded image 2ea32304-20d1-4039-a93a-b4c97a5656bf to the Glance image server {{(pid=61806) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:473}} [ 1152.778050] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [None req-f8795417-655c-4a2b-beb9-4c53a9fb716b tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] [instance: 29210bcd-0fe1-4ed9-a459-abc8016a4255] Destroying the VM {{(pid=61806) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1358}} [ 1152.778292] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-e4ec227d-a486-4be8-ad48-e5d535bea959 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1152.783385] env[61806]: DEBUG oslo_vmware.api [None req-f8795417-655c-4a2b-beb9-4c53a9fb716b tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] Waiting for the task: (returnval){ [ 1152.783385] env[61806]: value = "task-1295277" [ 1152.783385] env[61806]: _type = "Task" [ 1152.783385] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1152.790676] env[61806]: DEBUG oslo_vmware.api [None req-f8795417-655c-4a2b-beb9-4c53a9fb716b tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] Task: {'id': task-1295277, 'name': Destroy_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1152.947209] env[61806]: DEBUG oslo_vmware.api [None req-2bcb8d7f-cd6c-4f73-bd59-e28663212f52 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Task: {'id': session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]52a67362-e851-be99-772f-1d4f259d0c38, 'name': SearchDatastore_Task, 'duration_secs': 0.01166} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1152.948168] env[61806]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ef3706c5-0dc4-4017-aeeb-15f0758ccd13 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1152.953080] env[61806]: DEBUG oslo_vmware.api [None req-2bcb8d7f-cd6c-4f73-bd59-e28663212f52 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Waiting for the task: (returnval){ [ 1152.953080] env[61806]: value = "session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]523b5610-b0d4-1cf5-6a39-ee113c2870a7" [ 1152.953080] env[61806]: _type = "Task" [ 1152.953080] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1152.960362] env[61806]: DEBUG oslo_vmware.api [None req-2bcb8d7f-cd6c-4f73-bd59-e28663212f52 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Task: {'id': session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]523b5610-b0d4-1cf5-6a39-ee113c2870a7, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1153.292075] env[61806]: DEBUG oslo_vmware.api [None req-f8795417-655c-4a2b-beb9-4c53a9fb716b tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] Task: {'id': task-1295277, 'name': Destroy_Task, 'duration_secs': 0.387162} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1153.292344] env[61806]: INFO nova.virt.vmwareapi.vm_util [None req-f8795417-655c-4a2b-beb9-4c53a9fb716b tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] [instance: 29210bcd-0fe1-4ed9-a459-abc8016a4255] Destroyed the VM [ 1153.292586] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-f8795417-655c-4a2b-beb9-4c53a9fb716b tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] [instance: 29210bcd-0fe1-4ed9-a459-abc8016a4255] Deleting Snapshot of the VM instance {{(pid=61806) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 1153.292833] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-c284727a-972d-4bda-b86b-452c4285452a {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1153.298827] env[61806]: DEBUG oslo_vmware.api [None req-f8795417-655c-4a2b-beb9-4c53a9fb716b tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] Waiting for the task: (returnval){ [ 1153.298827] env[61806]: value = "task-1295278" [ 1153.298827] env[61806]: _type = "Task" [ 1153.298827] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1153.305716] env[61806]: DEBUG oslo_vmware.api [None req-f8795417-655c-4a2b-beb9-4c53a9fb716b tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] Task: {'id': task-1295278, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1153.463632] env[61806]: DEBUG oslo_vmware.api [None req-2bcb8d7f-cd6c-4f73-bd59-e28663212f52 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Task: {'id': session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]523b5610-b0d4-1cf5-6a39-ee113c2870a7, 'name': SearchDatastore_Task, 'duration_secs': 0.035146} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1153.463877] env[61806]: DEBUG oslo_concurrency.lockutils [None req-2bcb8d7f-cd6c-4f73-bd59-e28663212f52 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Releasing lock "[datastore1] devstack-image-cache_base/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa.vmdk" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1153.464152] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [None req-2bcb8d7f-cd6c-4f73-bd59-e28663212f52 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa.vmdk to [datastore1] 4f26445f-35d0-4f53-9014-ef3e2eb1cbdf/4f26445f-35d0-4f53-9014-ef3e2eb1cbdf.vmdk {{(pid=61806) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 1153.464414] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-857bcf19-6ee7-42bf-99b7-b80dafc18802 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1153.470013] env[61806]: DEBUG oslo_vmware.api [None req-2bcb8d7f-cd6c-4f73-bd59-e28663212f52 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Waiting for the task: (returnval){ [ 1153.470013] env[61806]: value = "task-1295279" [ 1153.470013] env[61806]: _type = "Task" [ 1153.470013] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1153.477031] env[61806]: DEBUG oslo_vmware.api [None req-2bcb8d7f-cd6c-4f73-bd59-e28663212f52 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Task: {'id': task-1295279, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1153.809485] env[61806]: DEBUG oslo_vmware.api [None req-f8795417-655c-4a2b-beb9-4c53a9fb716b tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] Task: {'id': task-1295278, 'name': RemoveSnapshot_Task, 'duration_secs': 0.471803} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1153.809749] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-f8795417-655c-4a2b-beb9-4c53a9fb716b tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] [instance: 29210bcd-0fe1-4ed9-a459-abc8016a4255] Deleted Snapshot of the VM instance {{(pid=61806) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 1153.810107] env[61806]: DEBUG nova.compute.manager [None req-f8795417-655c-4a2b-beb9-4c53a9fb716b tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] [instance: 29210bcd-0fe1-4ed9-a459-abc8016a4255] Checking state {{(pid=61806) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1153.810953] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-de955923-1490-4e97-af1e-21b261171b93 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1153.979119] env[61806]: DEBUG oslo_vmware.api [None req-2bcb8d7f-cd6c-4f73-bd59-e28663212f52 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Task: {'id': task-1295279, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.486318} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1153.979540] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [None req-2bcb8d7f-cd6c-4f73-bd59-e28663212f52 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa.vmdk to [datastore1] 4f26445f-35d0-4f53-9014-ef3e2eb1cbdf/4f26445f-35d0-4f53-9014-ef3e2eb1cbdf.vmdk {{(pid=61806) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 1153.979650] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-2bcb8d7f-cd6c-4f73-bd59-e28663212f52 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] [instance: 4f26445f-35d0-4f53-9014-ef3e2eb1cbdf] Extending root virtual disk to 1048576 {{(pid=61806) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1153.979954] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-5abc4315-b949-4ca4-b635-bc20d0c2e6d3 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1153.985904] env[61806]: DEBUG oslo_vmware.api [None req-2bcb8d7f-cd6c-4f73-bd59-e28663212f52 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Waiting for the task: (returnval){ [ 1153.985904] env[61806]: value = "task-1295280" [ 1153.985904] env[61806]: _type = "Task" [ 1153.985904] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1153.993044] env[61806]: DEBUG oslo_vmware.api [None req-2bcb8d7f-cd6c-4f73-bd59-e28663212f52 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Task: {'id': task-1295280, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1154.323568] env[61806]: INFO nova.compute.manager [None req-f8795417-655c-4a2b-beb9-4c53a9fb716b tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] [instance: 29210bcd-0fe1-4ed9-a459-abc8016a4255] Shelve offloading [ 1154.325233] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [None req-f8795417-655c-4a2b-beb9-4c53a9fb716b tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] [instance: 29210bcd-0fe1-4ed9-a459-abc8016a4255] Powering off the VM {{(pid=61806) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1154.325488] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-6f0c6f18-7204-4c45-97d1-c42f52c4c93c {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1154.332453] env[61806]: DEBUG oslo_vmware.api [None req-f8795417-655c-4a2b-beb9-4c53a9fb716b tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] Waiting for the task: (returnval){ [ 1154.332453] env[61806]: value = "task-1295281" [ 1154.332453] env[61806]: _type = "Task" [ 1154.332453] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1154.342750] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [None req-f8795417-655c-4a2b-beb9-4c53a9fb716b tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] [instance: 29210bcd-0fe1-4ed9-a459-abc8016a4255] VM already powered off {{(pid=61806) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1509}} [ 1154.342986] env[61806]: DEBUG nova.compute.manager [None req-f8795417-655c-4a2b-beb9-4c53a9fb716b tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] [instance: 29210bcd-0fe1-4ed9-a459-abc8016a4255] Checking state {{(pid=61806) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1154.343692] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f3a702ef-ef10-456b-bf1f-ed4070ebc12a {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1154.349214] env[61806]: DEBUG oslo_concurrency.lockutils [None req-f8795417-655c-4a2b-beb9-4c53a9fb716b tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] Acquiring lock "refresh_cache-29210bcd-0fe1-4ed9-a459-abc8016a4255" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1154.349393] env[61806]: DEBUG oslo_concurrency.lockutils [None req-f8795417-655c-4a2b-beb9-4c53a9fb716b tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] Acquired lock "refresh_cache-29210bcd-0fe1-4ed9-a459-abc8016a4255" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1154.349569] env[61806]: DEBUG nova.network.neutron [None req-f8795417-655c-4a2b-beb9-4c53a9fb716b tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] [instance: 29210bcd-0fe1-4ed9-a459-abc8016a4255] Building network info cache for instance {{(pid=61806) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1154.495826] env[61806]: DEBUG oslo_vmware.api [None req-2bcb8d7f-cd6c-4f73-bd59-e28663212f52 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Task: {'id': task-1295280, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.069382} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1154.496130] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-2bcb8d7f-cd6c-4f73-bd59-e28663212f52 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] [instance: 4f26445f-35d0-4f53-9014-ef3e2eb1cbdf] Extended root virtual disk {{(pid=61806) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1154.496955] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-601ba778-d9ac-4199-96cb-04a3a8098f6e {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1154.518628] env[61806]: DEBUG nova.virt.vmwareapi.volumeops [None req-2bcb8d7f-cd6c-4f73-bd59-e28663212f52 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] [instance: 4f26445f-35d0-4f53-9014-ef3e2eb1cbdf] Reconfiguring VM instance instance-0000006c to attach disk [datastore1] 4f26445f-35d0-4f53-9014-ef3e2eb1cbdf/4f26445f-35d0-4f53-9014-ef3e2eb1cbdf.vmdk or device None with type sparse {{(pid=61806) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1154.518893] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-533be39b-f786-4d44-802e-17f272f645ab {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1154.538111] env[61806]: DEBUG oslo_vmware.api [None req-2bcb8d7f-cd6c-4f73-bd59-e28663212f52 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Waiting for the task: (returnval){ [ 1154.538111] env[61806]: value = "task-1295282" [ 1154.538111] env[61806]: _type = "Task" [ 1154.538111] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1154.546484] env[61806]: DEBUG oslo_vmware.api [None req-2bcb8d7f-cd6c-4f73-bd59-e28663212f52 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Task: {'id': task-1295282, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1155.047323] env[61806]: DEBUG oslo_vmware.api [None req-2bcb8d7f-cd6c-4f73-bd59-e28663212f52 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Task: {'id': task-1295282, 'name': ReconfigVM_Task, 'duration_secs': 0.311861} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1155.047685] env[61806]: DEBUG nova.virt.vmwareapi.volumeops [None req-2bcb8d7f-cd6c-4f73-bd59-e28663212f52 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] [instance: 4f26445f-35d0-4f53-9014-ef3e2eb1cbdf] Reconfigured VM instance instance-0000006c to attach disk [datastore1] 4f26445f-35d0-4f53-9014-ef3e2eb1cbdf/4f26445f-35d0-4f53-9014-ef3e2eb1cbdf.vmdk or device None with type sparse {{(pid=61806) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1155.048293] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-713c86a3-a195-42f8-822c-bf8b511ee999 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1155.054337] env[61806]: DEBUG oslo_vmware.api [None req-2bcb8d7f-cd6c-4f73-bd59-e28663212f52 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Waiting for the task: (returnval){ [ 1155.054337] env[61806]: value = "task-1295283" [ 1155.054337] env[61806]: _type = "Task" [ 1155.054337] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1155.057833] env[61806]: DEBUG nova.network.neutron [None req-f8795417-655c-4a2b-beb9-4c53a9fb716b tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] [instance: 29210bcd-0fe1-4ed9-a459-abc8016a4255] Updating instance_info_cache with network_info: [{"id": "8aad34b7-4309-44b6-b88d-ae5961da22c2", "address": "fa:16:3e:c0:a7:c5", "network": {"id": "a9d993bc-0817-4ee6-ab6e-1ef9432a110f", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-939246736-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.196", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "bbbb09a49203421d807f733616c854df", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ad4c015b-4a8b-46ca-9556-74bad8db9fb3", "external-id": "nsx-vlan-transportzone-246", "segmentation_id": 246, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap8aad34b7-43", "ovs_interfaceid": "8aad34b7-4309-44b6-b88d-ae5961da22c2", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61806) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1155.063893] env[61806]: DEBUG oslo_vmware.api [None req-2bcb8d7f-cd6c-4f73-bd59-e28663212f52 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Task: {'id': task-1295283, 'name': Rename_Task} progress is 5%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1155.562058] env[61806]: DEBUG oslo_concurrency.lockutils [None req-f8795417-655c-4a2b-beb9-4c53a9fb716b tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] Releasing lock "refresh_cache-29210bcd-0fe1-4ed9-a459-abc8016a4255" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1155.565689] env[61806]: DEBUG oslo_vmware.api [None req-2bcb8d7f-cd6c-4f73-bd59-e28663212f52 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Task: {'id': task-1295283, 'name': Rename_Task, 'duration_secs': 0.144468} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1155.566232] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [None req-2bcb8d7f-cd6c-4f73-bd59-e28663212f52 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] [instance: 4f26445f-35d0-4f53-9014-ef3e2eb1cbdf] Powering on the VM {{(pid=61806) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1155.566512] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-7e3fb8f6-92c3-46b5-9b96-bc245ad6a739 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1155.572079] env[61806]: DEBUG oslo_vmware.api [None req-2bcb8d7f-cd6c-4f73-bd59-e28663212f52 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Waiting for the task: (returnval){ [ 1155.572079] env[61806]: value = "task-1295284" [ 1155.572079] env[61806]: _type = "Task" [ 1155.572079] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1155.579097] env[61806]: DEBUG oslo_vmware.api [None req-2bcb8d7f-cd6c-4f73-bd59-e28663212f52 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Task: {'id': task-1295284, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1155.861073] env[61806]: DEBUG nova.compute.manager [req-79b84ffa-b968-45b8-a0d8-1b36dfbc22cf req-7a0a0acb-8196-4971-809e-3d8e69594ed3 service nova] [instance: 29210bcd-0fe1-4ed9-a459-abc8016a4255] Received event network-vif-unplugged-8aad34b7-4309-44b6-b88d-ae5961da22c2 {{(pid=61806) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1155.861313] env[61806]: DEBUG oslo_concurrency.lockutils [req-79b84ffa-b968-45b8-a0d8-1b36dfbc22cf req-7a0a0acb-8196-4971-809e-3d8e69594ed3 service nova] Acquiring lock "29210bcd-0fe1-4ed9-a459-abc8016a4255-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1155.861526] env[61806]: DEBUG oslo_concurrency.lockutils [req-79b84ffa-b968-45b8-a0d8-1b36dfbc22cf req-7a0a0acb-8196-4971-809e-3d8e69594ed3 service nova] Lock "29210bcd-0fe1-4ed9-a459-abc8016a4255-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1155.861701] env[61806]: DEBUG oslo_concurrency.lockutils [req-79b84ffa-b968-45b8-a0d8-1b36dfbc22cf req-7a0a0acb-8196-4971-809e-3d8e69594ed3 service nova] Lock "29210bcd-0fe1-4ed9-a459-abc8016a4255-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1155.861879] env[61806]: DEBUG nova.compute.manager [req-79b84ffa-b968-45b8-a0d8-1b36dfbc22cf req-7a0a0acb-8196-4971-809e-3d8e69594ed3 service nova] [instance: 29210bcd-0fe1-4ed9-a459-abc8016a4255] No waiting events found dispatching network-vif-unplugged-8aad34b7-4309-44b6-b88d-ae5961da22c2 {{(pid=61806) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1155.862334] env[61806]: WARNING nova.compute.manager [req-79b84ffa-b968-45b8-a0d8-1b36dfbc22cf req-7a0a0acb-8196-4971-809e-3d8e69594ed3 service nova] [instance: 29210bcd-0fe1-4ed9-a459-abc8016a4255] Received unexpected event network-vif-unplugged-8aad34b7-4309-44b6-b88d-ae5961da22c2 for instance with vm_state shelved and task_state shelving_offloading. [ 1156.003215] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-f8795417-655c-4a2b-beb9-4c53a9fb716b tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] [instance: 29210bcd-0fe1-4ed9-a459-abc8016a4255] Destroying instance {{(pid=61806) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1156.004124] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d00ed34b-0daf-4cfc-a609-c264a0733133 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1156.011920] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-f8795417-655c-4a2b-beb9-4c53a9fb716b tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] [instance: 29210bcd-0fe1-4ed9-a459-abc8016a4255] Unregistering the VM {{(pid=61806) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1156.012198] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-abe95dec-1167-4734-8b80-dad835437456 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1156.081843] env[61806]: DEBUG oslo_vmware.api [None req-2bcb8d7f-cd6c-4f73-bd59-e28663212f52 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Task: {'id': task-1295284, 'name': PowerOnVM_Task} progress is 88%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1156.268050] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-f8795417-655c-4a2b-beb9-4c53a9fb716b tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] [instance: 29210bcd-0fe1-4ed9-a459-abc8016a4255] Unregistered the VM {{(pid=61806) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1156.268302] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-f8795417-655c-4a2b-beb9-4c53a9fb716b tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] [instance: 29210bcd-0fe1-4ed9-a459-abc8016a4255] Deleting contents of the VM from datastore datastore1 {{(pid=61806) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1156.268495] env[61806]: DEBUG nova.virt.vmwareapi.ds_util [None req-f8795417-655c-4a2b-beb9-4c53a9fb716b tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] Deleting the datastore file [datastore1] 29210bcd-0fe1-4ed9-a459-abc8016a4255 {{(pid=61806) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1156.268779] env[61806]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-f7e94eb5-ebc9-4318-b47d-fd8523cc1b51 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1156.274674] env[61806]: DEBUG oslo_vmware.api [None req-f8795417-655c-4a2b-beb9-4c53a9fb716b tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] Waiting for the task: (returnval){ [ 1156.274674] env[61806]: value = "task-1295286" [ 1156.274674] env[61806]: _type = "Task" [ 1156.274674] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1156.283451] env[61806]: DEBUG oslo_vmware.api [None req-f8795417-655c-4a2b-beb9-4c53a9fb716b tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] Task: {'id': task-1295286, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1156.582425] env[61806]: DEBUG oslo_vmware.api [None req-2bcb8d7f-cd6c-4f73-bd59-e28663212f52 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Task: {'id': task-1295284, 'name': PowerOnVM_Task, 'duration_secs': 0.68223} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1156.582689] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [None req-2bcb8d7f-cd6c-4f73-bd59-e28663212f52 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] [instance: 4f26445f-35d0-4f53-9014-ef3e2eb1cbdf] Powered on the VM {{(pid=61806) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1156.582900] env[61806]: INFO nova.compute.manager [None req-2bcb8d7f-cd6c-4f73-bd59-e28663212f52 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] [instance: 4f26445f-35d0-4f53-9014-ef3e2eb1cbdf] Took 8.31 seconds to spawn the instance on the hypervisor. [ 1156.583101] env[61806]: DEBUG nova.compute.manager [None req-2bcb8d7f-cd6c-4f73-bd59-e28663212f52 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] [instance: 4f26445f-35d0-4f53-9014-ef3e2eb1cbdf] Checking state {{(pid=61806) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1156.583880] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-37a9e991-b959-41cc-84ef-ddcb016460ce {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1156.786114] env[61806]: DEBUG oslo_vmware.api [None req-f8795417-655c-4a2b-beb9-4c53a9fb716b tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] Task: {'id': task-1295286, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.140204} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1156.786408] env[61806]: DEBUG nova.virt.vmwareapi.ds_util [None req-f8795417-655c-4a2b-beb9-4c53a9fb716b tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] Deleted the datastore file {{(pid=61806) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1156.786607] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-f8795417-655c-4a2b-beb9-4c53a9fb716b tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] [instance: 29210bcd-0fe1-4ed9-a459-abc8016a4255] Deleted contents of the VM from datastore datastore1 {{(pid=61806) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1156.786792] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-f8795417-655c-4a2b-beb9-4c53a9fb716b tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] [instance: 29210bcd-0fe1-4ed9-a459-abc8016a4255] Instance destroyed {{(pid=61806) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1156.811551] env[61806]: INFO nova.scheduler.client.report [None req-f8795417-655c-4a2b-beb9-4c53a9fb716b tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] Deleted allocations for instance 29210bcd-0fe1-4ed9-a459-abc8016a4255 [ 1157.100414] env[61806]: INFO nova.compute.manager [None req-2bcb8d7f-cd6c-4f73-bd59-e28663212f52 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] [instance: 4f26445f-35d0-4f53-9014-ef3e2eb1cbdf] Took 13.00 seconds to build instance. [ 1157.316068] env[61806]: DEBUG oslo_concurrency.lockutils [None req-f8795417-655c-4a2b-beb9-4c53a9fb716b tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1157.316568] env[61806]: DEBUG oslo_concurrency.lockutils [None req-f8795417-655c-4a2b-beb9-4c53a9fb716b tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.001s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1157.316893] env[61806]: DEBUG nova.objects.instance [None req-f8795417-655c-4a2b-beb9-4c53a9fb716b tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] Lazy-loading 'resources' on Instance uuid 29210bcd-0fe1-4ed9-a459-abc8016a4255 {{(pid=61806) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1157.605855] env[61806]: DEBUG oslo_concurrency.lockutils [None req-2bcb8d7f-cd6c-4f73-bd59-e28663212f52 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Lock "4f26445f-35d0-4f53-9014-ef3e2eb1cbdf" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 14.510s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1157.820288] env[61806]: DEBUG nova.objects.instance [None req-f8795417-655c-4a2b-beb9-4c53a9fb716b tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] Lazy-loading 'numa_topology' on Instance uuid 29210bcd-0fe1-4ed9-a459-abc8016a4255 {{(pid=61806) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1157.896530] env[61806]: DEBUG nova.compute.manager [req-29e2f6d8-4c86-413b-a11d-1c1d999acc16 req-0ae483f6-d5c9-4baa-bfec-f595806d438b service nova] [instance: 29210bcd-0fe1-4ed9-a459-abc8016a4255] Received event network-changed-8aad34b7-4309-44b6-b88d-ae5961da22c2 {{(pid=61806) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1157.896771] env[61806]: DEBUG nova.compute.manager [req-29e2f6d8-4c86-413b-a11d-1c1d999acc16 req-0ae483f6-d5c9-4baa-bfec-f595806d438b service nova] [instance: 29210bcd-0fe1-4ed9-a459-abc8016a4255] Refreshing instance network info cache due to event network-changed-8aad34b7-4309-44b6-b88d-ae5961da22c2. {{(pid=61806) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 1157.897357] env[61806]: DEBUG oslo_concurrency.lockutils [req-29e2f6d8-4c86-413b-a11d-1c1d999acc16 req-0ae483f6-d5c9-4baa-bfec-f595806d438b service nova] Acquiring lock "refresh_cache-29210bcd-0fe1-4ed9-a459-abc8016a4255" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1157.897546] env[61806]: DEBUG oslo_concurrency.lockutils [req-29e2f6d8-4c86-413b-a11d-1c1d999acc16 req-0ae483f6-d5c9-4baa-bfec-f595806d438b service nova] Acquired lock "refresh_cache-29210bcd-0fe1-4ed9-a459-abc8016a4255" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1157.897735] env[61806]: DEBUG nova.network.neutron [req-29e2f6d8-4c86-413b-a11d-1c1d999acc16 req-0ae483f6-d5c9-4baa-bfec-f595806d438b service nova] [instance: 29210bcd-0fe1-4ed9-a459-abc8016a4255] Refreshing network info cache for port 8aad34b7-4309-44b6-b88d-ae5961da22c2 {{(pid=61806) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1158.322699] env[61806]: DEBUG nova.objects.base [None req-f8795417-655c-4a2b-beb9-4c53a9fb716b tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] Object Instance<29210bcd-0fe1-4ed9-a459-abc8016a4255> lazy-loaded attributes: resources,numa_topology {{(pid=61806) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 1158.358341] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1558a05c-03bd-46a7-8a53-aeb6c06e0555 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1158.366049] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5f99f20c-9f17-474a-bce3-c037ea0d8262 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1158.394603] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f75519dd-2c88-488f-b0c0-fdc8b47c26d0 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1158.403102] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-14aaf990-7375-4194-a787-7ff6dcadbbbf {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1158.417767] env[61806]: DEBUG nova.compute.provider_tree [None req-f8795417-655c-4a2b-beb9-4c53a9fb716b tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] Inventory has not changed in ProviderTree for provider: a2858be1-fd22-4e08-979e-87ad25293407 {{(pid=61806) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1158.641802] env[61806]: DEBUG nova.network.neutron [req-29e2f6d8-4c86-413b-a11d-1c1d999acc16 req-0ae483f6-d5c9-4baa-bfec-f595806d438b service nova] [instance: 29210bcd-0fe1-4ed9-a459-abc8016a4255] Updated VIF entry in instance network info cache for port 8aad34b7-4309-44b6-b88d-ae5961da22c2. {{(pid=61806) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1158.642202] env[61806]: DEBUG nova.network.neutron [req-29e2f6d8-4c86-413b-a11d-1c1d999acc16 req-0ae483f6-d5c9-4baa-bfec-f595806d438b service nova] [instance: 29210bcd-0fe1-4ed9-a459-abc8016a4255] Updating instance_info_cache with network_info: [{"id": "8aad34b7-4309-44b6-b88d-ae5961da22c2", "address": "fa:16:3e:c0:a7:c5", "network": {"id": "a9d993bc-0817-4ee6-ab6e-1ef9432a110f", "bridge": null, "label": "tempest-AttachVolumeShelveTestJSON-939246736-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.196", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "bbbb09a49203421d807f733616c854df", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "unbound", "details": {}, "devname": "tap8aad34b7-43", "ovs_interfaceid": null, "qbh_params": null, "qbg_params": null, "active": false, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61806) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1158.920652] env[61806]: DEBUG nova.scheduler.client.report [None req-f8795417-655c-4a2b-beb9-4c53a9fb716b tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] Inventory has not changed for provider a2858be1-fd22-4e08-979e-87ad25293407 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 138, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61806) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1159.144631] env[61806]: DEBUG oslo_concurrency.lockutils [req-29e2f6d8-4c86-413b-a11d-1c1d999acc16 req-0ae483f6-d5c9-4baa-bfec-f595806d438b service nova] Releasing lock "refresh_cache-29210bcd-0fe1-4ed9-a459-abc8016a4255" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1159.144915] env[61806]: DEBUG nova.compute.manager [req-29e2f6d8-4c86-413b-a11d-1c1d999acc16 req-0ae483f6-d5c9-4baa-bfec-f595806d438b service nova] [instance: 4f26445f-35d0-4f53-9014-ef3e2eb1cbdf] Received event network-changed-a71b2870-bc50-47f4-840d-57e956c3f86d {{(pid=61806) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1159.145091] env[61806]: DEBUG nova.compute.manager [req-29e2f6d8-4c86-413b-a11d-1c1d999acc16 req-0ae483f6-d5c9-4baa-bfec-f595806d438b service nova] [instance: 4f26445f-35d0-4f53-9014-ef3e2eb1cbdf] Refreshing instance network info cache due to event network-changed-a71b2870-bc50-47f4-840d-57e956c3f86d. {{(pid=61806) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 1159.145313] env[61806]: DEBUG oslo_concurrency.lockutils [req-29e2f6d8-4c86-413b-a11d-1c1d999acc16 req-0ae483f6-d5c9-4baa-bfec-f595806d438b service nova] Acquiring lock "refresh_cache-4f26445f-35d0-4f53-9014-ef3e2eb1cbdf" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1159.145514] env[61806]: DEBUG oslo_concurrency.lockutils [req-29e2f6d8-4c86-413b-a11d-1c1d999acc16 req-0ae483f6-d5c9-4baa-bfec-f595806d438b service nova] Acquired lock "refresh_cache-4f26445f-35d0-4f53-9014-ef3e2eb1cbdf" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1159.145630] env[61806]: DEBUG nova.network.neutron [req-29e2f6d8-4c86-413b-a11d-1c1d999acc16 req-0ae483f6-d5c9-4baa-bfec-f595806d438b service nova] [instance: 4f26445f-35d0-4f53-9014-ef3e2eb1cbdf] Refreshing network info cache for port a71b2870-bc50-47f4-840d-57e956c3f86d {{(pid=61806) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1159.426906] env[61806]: DEBUG oslo_concurrency.lockutils [None req-f8795417-655c-4a2b-beb9-4c53a9fb716b tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.110s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1159.634661] env[61806]: DEBUG oslo_concurrency.lockutils [None req-055516b6-e34c-4ddb-87ba-434948f7c77d tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] Acquiring lock "29210bcd-0fe1-4ed9-a459-abc8016a4255" by "nova.compute.manager.ComputeManager.unshelve_instance..do_unshelve_instance" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1159.873090] env[61806]: DEBUG nova.network.neutron [req-29e2f6d8-4c86-413b-a11d-1c1d999acc16 req-0ae483f6-d5c9-4baa-bfec-f595806d438b service nova] [instance: 4f26445f-35d0-4f53-9014-ef3e2eb1cbdf] Updated VIF entry in instance network info cache for port a71b2870-bc50-47f4-840d-57e956c3f86d. {{(pid=61806) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1159.873481] env[61806]: DEBUG nova.network.neutron [req-29e2f6d8-4c86-413b-a11d-1c1d999acc16 req-0ae483f6-d5c9-4baa-bfec-f595806d438b service nova] [instance: 4f26445f-35d0-4f53-9014-ef3e2eb1cbdf] Updating instance_info_cache with network_info: [{"id": "a71b2870-bc50-47f4-840d-57e956c3f86d", "address": "fa:16:3e:74:22:b8", "network": {"id": "33bf5b21-82b0-432e-ae3d-9c2ca75161d0", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-1294159336-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.187", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "1c55cc0455584fbb8cda88d33b8f8b7e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "5b0fa7a2-ebd9-4788-8904-7bf250ce466c", "external-id": "nsx-vlan-transportzone-669", "segmentation_id": 669, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa71b2870-bc", "ovs_interfaceid": "a71b2870-bc50-47f4-840d-57e956c3f86d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61806) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1159.934499] env[61806]: DEBUG oslo_concurrency.lockutils [None req-f8795417-655c-4a2b-beb9-4c53a9fb716b tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] Lock "29210bcd-0fe1-4ed9-a459-abc8016a4255" "released" by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" :: held 20.958s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1159.935375] env[61806]: DEBUG oslo_concurrency.lockutils [None req-055516b6-e34c-4ddb-87ba-434948f7c77d tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] Lock "29210bcd-0fe1-4ed9-a459-abc8016a4255" acquired by "nova.compute.manager.ComputeManager.unshelve_instance..do_unshelve_instance" :: waited 0.301s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1159.935375] env[61806]: INFO nova.compute.manager [None req-055516b6-e34c-4ddb-87ba-434948f7c77d tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] [instance: 29210bcd-0fe1-4ed9-a459-abc8016a4255] Unshelving [ 1160.376711] env[61806]: DEBUG oslo_concurrency.lockutils [req-29e2f6d8-4c86-413b-a11d-1c1d999acc16 req-0ae483f6-d5c9-4baa-bfec-f595806d438b service nova] Releasing lock "refresh_cache-4f26445f-35d0-4f53-9014-ef3e2eb1cbdf" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1160.960950] env[61806]: DEBUG oslo_concurrency.lockutils [None req-055516b6-e34c-4ddb-87ba-434948f7c77d tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1160.961247] env[61806]: DEBUG oslo_concurrency.lockutils [None req-055516b6-e34c-4ddb-87ba-434948f7c77d tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.000s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1160.961515] env[61806]: DEBUG nova.objects.instance [None req-055516b6-e34c-4ddb-87ba-434948f7c77d tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] Lazy-loading 'pci_requests' on Instance uuid 29210bcd-0fe1-4ed9-a459-abc8016a4255 {{(pid=61806) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1161.465287] env[61806]: DEBUG nova.objects.instance [None req-055516b6-e34c-4ddb-87ba-434948f7c77d tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] Lazy-loading 'numa_topology' on Instance uuid 29210bcd-0fe1-4ed9-a459-abc8016a4255 {{(pid=61806) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1161.967472] env[61806]: INFO nova.compute.claims [None req-055516b6-e34c-4ddb-87ba-434948f7c77d tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] [instance: 29210bcd-0fe1-4ed9-a459-abc8016a4255] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1163.021593] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cee904ab-0c50-40e2-a6ba-6078f25be572 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1163.029374] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6faa92f1-8255-42c8-b995-65b537462aff {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1163.059302] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f25828d0-840a-49ff-a456-97225dcdffc8 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1163.066363] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b740bacf-d628-4c18-8593-22c11128c43d {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1163.079567] env[61806]: DEBUG nova.compute.provider_tree [None req-055516b6-e34c-4ddb-87ba-434948f7c77d tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] Inventory has not changed in ProviderTree for provider: a2858be1-fd22-4e08-979e-87ad25293407 {{(pid=61806) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1163.582419] env[61806]: DEBUG nova.scheduler.client.report [None req-055516b6-e34c-4ddb-87ba-434948f7c77d tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] Inventory has not changed for provider a2858be1-fd22-4e08-979e-87ad25293407 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 138, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61806) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1164.087511] env[61806]: DEBUG oslo_concurrency.lockutils [None req-055516b6-e34c-4ddb-87ba-434948f7c77d tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 3.126s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1164.140753] env[61806]: INFO nova.network.neutron [None req-055516b6-e34c-4ddb-87ba-434948f7c77d tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] [instance: 29210bcd-0fe1-4ed9-a459-abc8016a4255] Updating port 8aad34b7-4309-44b6-b88d-ae5961da22c2 with attributes {'binding:host_id': 'cpu-1', 'device_owner': 'compute:nova'} [ 1165.557260] env[61806]: DEBUG nova.compute.manager [req-388c65f4-3a8d-43c5-b806-0e55fd269023 req-4ccbe21e-4e87-4d99-ab82-9460b82005e5 service nova] [instance: 29210bcd-0fe1-4ed9-a459-abc8016a4255] Received event network-vif-plugged-8aad34b7-4309-44b6-b88d-ae5961da22c2 {{(pid=61806) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1165.557488] env[61806]: DEBUG oslo_concurrency.lockutils [req-388c65f4-3a8d-43c5-b806-0e55fd269023 req-4ccbe21e-4e87-4d99-ab82-9460b82005e5 service nova] Acquiring lock "29210bcd-0fe1-4ed9-a459-abc8016a4255-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1165.557810] env[61806]: DEBUG oslo_concurrency.lockutils [req-388c65f4-3a8d-43c5-b806-0e55fd269023 req-4ccbe21e-4e87-4d99-ab82-9460b82005e5 service nova] Lock "29210bcd-0fe1-4ed9-a459-abc8016a4255-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1165.558069] env[61806]: DEBUG oslo_concurrency.lockutils [req-388c65f4-3a8d-43c5-b806-0e55fd269023 req-4ccbe21e-4e87-4d99-ab82-9460b82005e5 service nova] Lock "29210bcd-0fe1-4ed9-a459-abc8016a4255-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1165.558298] env[61806]: DEBUG nova.compute.manager [req-388c65f4-3a8d-43c5-b806-0e55fd269023 req-4ccbe21e-4e87-4d99-ab82-9460b82005e5 service nova] [instance: 29210bcd-0fe1-4ed9-a459-abc8016a4255] No waiting events found dispatching network-vif-plugged-8aad34b7-4309-44b6-b88d-ae5961da22c2 {{(pid=61806) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1165.558484] env[61806]: WARNING nova.compute.manager [req-388c65f4-3a8d-43c5-b806-0e55fd269023 req-4ccbe21e-4e87-4d99-ab82-9460b82005e5 service nova] [instance: 29210bcd-0fe1-4ed9-a459-abc8016a4255] Received unexpected event network-vif-plugged-8aad34b7-4309-44b6-b88d-ae5961da22c2 for instance with vm_state shelved_offloaded and task_state spawning. [ 1165.648336] env[61806]: DEBUG oslo_concurrency.lockutils [None req-055516b6-e34c-4ddb-87ba-434948f7c77d tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] Acquiring lock "refresh_cache-29210bcd-0fe1-4ed9-a459-abc8016a4255" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1165.648494] env[61806]: DEBUG oslo_concurrency.lockutils [None req-055516b6-e34c-4ddb-87ba-434948f7c77d tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] Acquired lock "refresh_cache-29210bcd-0fe1-4ed9-a459-abc8016a4255" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1165.648681] env[61806]: DEBUG nova.network.neutron [None req-055516b6-e34c-4ddb-87ba-434948f7c77d tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] [instance: 29210bcd-0fe1-4ed9-a459-abc8016a4255] Building network info cache for instance {{(pid=61806) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1166.347114] env[61806]: DEBUG nova.network.neutron [None req-055516b6-e34c-4ddb-87ba-434948f7c77d tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] [instance: 29210bcd-0fe1-4ed9-a459-abc8016a4255] Updating instance_info_cache with network_info: [{"id": "8aad34b7-4309-44b6-b88d-ae5961da22c2", "address": "fa:16:3e:c0:a7:c5", "network": {"id": "a9d993bc-0817-4ee6-ab6e-1ef9432a110f", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-939246736-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.196", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "bbbb09a49203421d807f733616c854df", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ad4c015b-4a8b-46ca-9556-74bad8db9fb3", "external-id": "nsx-vlan-transportzone-246", "segmentation_id": 246, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap8aad34b7-43", "ovs_interfaceid": "8aad34b7-4309-44b6-b88d-ae5961da22c2", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61806) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1166.852450] env[61806]: DEBUG oslo_concurrency.lockutils [None req-055516b6-e34c-4ddb-87ba-434948f7c77d tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] Releasing lock "refresh_cache-29210bcd-0fe1-4ed9-a459-abc8016a4255" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1166.878442] env[61806]: DEBUG nova.virt.hardware [None req-055516b6-e34c-4ddb-87ba-434948f7c77d tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-15T18:05:10Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='a8eed7c8d6cbd52628d05ae651e61036',container_format='bare',created_at=2024-10-15T18:17:19Z,direct_url=,disk_format='vmdk',id=2ea32304-20d1-4039-a93a-b4c97a5656bf,min_disk=1,min_ram=0,name='tempest-AttachVolumeShelveTestJSON-server-1589379504-shelved',owner='bbbb09a49203421d807f733616c854df',properties=ImageMetaProps,protected=,size=31669760,status='active',tags=,updated_at=2024-10-15T18:17:33Z,virtual_size=,visibility=), allow threads: False {{(pid=61806) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1166.878730] env[61806]: DEBUG nova.virt.hardware [None req-055516b6-e34c-4ddb-87ba-434948f7c77d tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] Flavor limits 0:0:0 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1166.878903] env[61806]: DEBUG nova.virt.hardware [None req-055516b6-e34c-4ddb-87ba-434948f7c77d tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] Image limits 0:0:0 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1166.879110] env[61806]: DEBUG nova.virt.hardware [None req-055516b6-e34c-4ddb-87ba-434948f7c77d tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] Flavor pref 0:0:0 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1166.879265] env[61806]: DEBUG nova.virt.hardware [None req-055516b6-e34c-4ddb-87ba-434948f7c77d tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] Image pref 0:0:0 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1166.879420] env[61806]: DEBUG nova.virt.hardware [None req-055516b6-e34c-4ddb-87ba-434948f7c77d tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1166.879635] env[61806]: DEBUG nova.virt.hardware [None req-055516b6-e34c-4ddb-87ba-434948f7c77d tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61806) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1166.879799] env[61806]: DEBUG nova.virt.hardware [None req-055516b6-e34c-4ddb-87ba-434948f7c77d tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61806) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1166.879977] env[61806]: DEBUG nova.virt.hardware [None req-055516b6-e34c-4ddb-87ba-434948f7c77d tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] Got 1 possible topologies {{(pid=61806) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1166.880166] env[61806]: DEBUG nova.virt.hardware [None req-055516b6-e34c-4ddb-87ba-434948f7c77d tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61806) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1166.880343] env[61806]: DEBUG nova.virt.hardware [None req-055516b6-e34c-4ddb-87ba-434948f7c77d tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61806) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1166.881197] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1e644b36-2be3-4588-a0bd-6b09c456646c {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1166.888926] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e7e31e56-801c-4576-a509-50f2a7d7ba45 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1166.901903] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-055516b6-e34c-4ddb-87ba-434948f7c77d tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] [instance: 29210bcd-0fe1-4ed9-a459-abc8016a4255] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:c0:a7:c5', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'ad4c015b-4a8b-46ca-9556-74bad8db9fb3', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '8aad34b7-4309-44b6-b88d-ae5961da22c2', 'vif_model': 'vmxnet3'}] {{(pid=61806) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1166.909353] env[61806]: DEBUG oslo.service.loopingcall [None req-055516b6-e34c-4ddb-87ba-434948f7c77d tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61806) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1166.909612] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 29210bcd-0fe1-4ed9-a459-abc8016a4255] Creating VM on the ESX host {{(pid=61806) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 1166.909848] env[61806]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-2d313fb1-90bb-4b1b-9c5f-ab168bd61895 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1166.929920] env[61806]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1166.929920] env[61806]: value = "task-1295287" [ 1166.929920] env[61806]: _type = "Task" [ 1166.929920] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1166.939429] env[61806]: DEBUG oslo_vmware.api [-] Task: {'id': task-1295287, 'name': CreateVM_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1167.440259] env[61806]: DEBUG oslo_vmware.api [-] Task: {'id': task-1295287, 'name': CreateVM_Task, 'duration_secs': 0.272495} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1167.440471] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 29210bcd-0fe1-4ed9-a459-abc8016a4255] Created VM on the ESX host {{(pid=61806) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 1167.441102] env[61806]: DEBUG oslo_concurrency.lockutils [None req-055516b6-e34c-4ddb-87ba-434948f7c77d tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/2ea32304-20d1-4039-a93a-b4c97a5656bf" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1167.441293] env[61806]: DEBUG oslo_concurrency.lockutils [None req-055516b6-e34c-4ddb-87ba-434948f7c77d tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] Acquired lock "[datastore2] devstack-image-cache_base/2ea32304-20d1-4039-a93a-b4c97a5656bf" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1167.441700] env[61806]: DEBUG oslo_concurrency.lockutils [None req-055516b6-e34c-4ddb-87ba-434948f7c77d tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/2ea32304-20d1-4039-a93a-b4c97a5656bf" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1167.441982] env[61806]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b125e0a1-1960-4227-a21c-360ce621367b {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1167.446504] env[61806]: DEBUG oslo_vmware.api [None req-055516b6-e34c-4ddb-87ba-434948f7c77d tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] Waiting for the task: (returnval){ [ 1167.446504] env[61806]: value = "session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]52658b5f-ecbc-3e42-d67c-618689471e0f" [ 1167.446504] env[61806]: _type = "Task" [ 1167.446504] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1167.458861] env[61806]: DEBUG oslo_vmware.api [None req-055516b6-e34c-4ddb-87ba-434948f7c77d tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] Task: {'id': session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]52658b5f-ecbc-3e42-d67c-618689471e0f, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1167.580630] env[61806]: DEBUG nova.compute.manager [req-13b1c5c8-d7ae-4da8-951d-c3f2ecb47b07 req-16e60915-6020-4e33-bba0-4f9c47386446 service nova] [instance: 29210bcd-0fe1-4ed9-a459-abc8016a4255] Received event network-changed-8aad34b7-4309-44b6-b88d-ae5961da22c2 {{(pid=61806) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1167.580846] env[61806]: DEBUG nova.compute.manager [req-13b1c5c8-d7ae-4da8-951d-c3f2ecb47b07 req-16e60915-6020-4e33-bba0-4f9c47386446 service nova] [instance: 29210bcd-0fe1-4ed9-a459-abc8016a4255] Refreshing instance network info cache due to event network-changed-8aad34b7-4309-44b6-b88d-ae5961da22c2. {{(pid=61806) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 1167.581077] env[61806]: DEBUG oslo_concurrency.lockutils [req-13b1c5c8-d7ae-4da8-951d-c3f2ecb47b07 req-16e60915-6020-4e33-bba0-4f9c47386446 service nova] Acquiring lock "refresh_cache-29210bcd-0fe1-4ed9-a459-abc8016a4255" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1167.581231] env[61806]: DEBUG oslo_concurrency.lockutils [req-13b1c5c8-d7ae-4da8-951d-c3f2ecb47b07 req-16e60915-6020-4e33-bba0-4f9c47386446 service nova] Acquired lock "refresh_cache-29210bcd-0fe1-4ed9-a459-abc8016a4255" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1167.581398] env[61806]: DEBUG nova.network.neutron [req-13b1c5c8-d7ae-4da8-951d-c3f2ecb47b07 req-16e60915-6020-4e33-bba0-4f9c47386446 service nova] [instance: 29210bcd-0fe1-4ed9-a459-abc8016a4255] Refreshing network info cache for port 8aad34b7-4309-44b6-b88d-ae5961da22c2 {{(pid=61806) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1167.956520] env[61806]: DEBUG oslo_concurrency.lockutils [None req-055516b6-e34c-4ddb-87ba-434948f7c77d tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] Releasing lock "[datastore2] devstack-image-cache_base/2ea32304-20d1-4039-a93a-b4c97a5656bf" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1167.956831] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-055516b6-e34c-4ddb-87ba-434948f7c77d tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] [instance: 29210bcd-0fe1-4ed9-a459-abc8016a4255] Processing image 2ea32304-20d1-4039-a93a-b4c97a5656bf {{(pid=61806) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1167.957056] env[61806]: DEBUG oslo_concurrency.lockutils [None req-055516b6-e34c-4ddb-87ba-434948f7c77d tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/2ea32304-20d1-4039-a93a-b4c97a5656bf/2ea32304-20d1-4039-a93a-b4c97a5656bf.vmdk" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1167.957220] env[61806]: DEBUG oslo_concurrency.lockutils [None req-055516b6-e34c-4ddb-87ba-434948f7c77d tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] Acquired lock "[datastore2] devstack-image-cache_base/2ea32304-20d1-4039-a93a-b4c97a5656bf/2ea32304-20d1-4039-a93a-b4c97a5656bf.vmdk" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1167.957404] env[61806]: DEBUG nova.virt.vmwareapi.ds_util [None req-055516b6-e34c-4ddb-87ba-434948f7c77d tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61806) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1167.957656] env[61806]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-835adb20-67b8-4f2d-af1f-bf6dc58b4562 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1167.965676] env[61806]: DEBUG nova.virt.vmwareapi.ds_util [None req-055516b6-e34c-4ddb-87ba-434948f7c77d tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61806) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1167.965855] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-055516b6-e34c-4ddb-87ba-434948f7c77d tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61806) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 1167.966608] env[61806]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-9d68f1b0-2d7d-4ad5-81d7-784cfed48335 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1167.971135] env[61806]: DEBUG oslo_vmware.api [None req-055516b6-e34c-4ddb-87ba-434948f7c77d tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] Waiting for the task: (returnval){ [ 1167.971135] env[61806]: value = "session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]52fbdec1-f662-0571-f561-b144ab9e6932" [ 1167.971135] env[61806]: _type = "Task" [ 1167.971135] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1167.977885] env[61806]: DEBUG oslo_vmware.api [None req-055516b6-e34c-4ddb-87ba-434948f7c77d tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] Task: {'id': session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]52fbdec1-f662-0571-f561-b144ab9e6932, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1168.273566] env[61806]: DEBUG nova.network.neutron [req-13b1c5c8-d7ae-4da8-951d-c3f2ecb47b07 req-16e60915-6020-4e33-bba0-4f9c47386446 service nova] [instance: 29210bcd-0fe1-4ed9-a459-abc8016a4255] Updated VIF entry in instance network info cache for port 8aad34b7-4309-44b6-b88d-ae5961da22c2. {{(pid=61806) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1168.273969] env[61806]: DEBUG nova.network.neutron [req-13b1c5c8-d7ae-4da8-951d-c3f2ecb47b07 req-16e60915-6020-4e33-bba0-4f9c47386446 service nova] [instance: 29210bcd-0fe1-4ed9-a459-abc8016a4255] Updating instance_info_cache with network_info: [{"id": "8aad34b7-4309-44b6-b88d-ae5961da22c2", "address": "fa:16:3e:c0:a7:c5", "network": {"id": "a9d993bc-0817-4ee6-ab6e-1ef9432a110f", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-939246736-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.196", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "bbbb09a49203421d807f733616c854df", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ad4c015b-4a8b-46ca-9556-74bad8db9fb3", "external-id": "nsx-vlan-transportzone-246", "segmentation_id": 246, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap8aad34b7-43", "ovs_interfaceid": "8aad34b7-4309-44b6-b88d-ae5961da22c2", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61806) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1168.481103] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-055516b6-e34c-4ddb-87ba-434948f7c77d tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] [instance: 29210bcd-0fe1-4ed9-a459-abc8016a4255] Preparing fetch location {{(pid=61806) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:633}} [ 1168.481382] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-055516b6-e34c-4ddb-87ba-434948f7c77d tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] [instance: 29210bcd-0fe1-4ed9-a459-abc8016a4255] Fetch image to [datastore2] OSTACK_IMG_520967df-fd29-4a4c-ab10-c405db541a06/OSTACK_IMG_520967df-fd29-4a4c-ab10-c405db541a06.vmdk {{(pid=61806) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:635}} [ 1168.481603] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-055516b6-e34c-4ddb-87ba-434948f7c77d tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] [instance: 29210bcd-0fe1-4ed9-a459-abc8016a4255] Downloading stream optimized image 2ea32304-20d1-4039-a93a-b4c97a5656bf to [datastore2] OSTACK_IMG_520967df-fd29-4a4c-ab10-c405db541a06/OSTACK_IMG_520967df-fd29-4a4c-ab10-c405db541a06.vmdk on the data store datastore2 as vApp {{(pid=61806) _fetch_image_as_vapp /opt/stack/nova/nova/virt/vmwareapi/vmops.py:437}} [ 1168.481802] env[61806]: DEBUG nova.virt.vmwareapi.images [None req-055516b6-e34c-4ddb-87ba-434948f7c77d tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] [instance: 29210bcd-0fe1-4ed9-a459-abc8016a4255] Downloading image file data 2ea32304-20d1-4039-a93a-b4c97a5656bf to the ESX as VM named 'OSTACK_IMG_520967df-fd29-4a4c-ab10-c405db541a06' {{(pid=61806) fetch_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:336}} [ 1168.547604] env[61806]: DEBUG oslo_vmware.rw_handles [None req-055516b6-e34c-4ddb-87ba-434948f7c77d tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] Creating HttpNfcLease lease for vApp import into resource pool: (val){ [ 1168.547604] env[61806]: value = "resgroup-9" [ 1168.547604] env[61806]: _type = "ResourcePool" [ 1168.547604] env[61806]: }. {{(pid=61806) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:453}} [ 1168.547915] env[61806]: DEBUG oslo_vmware.service [-] Invoking ResourcePool.ImportVApp with opID=oslo.vmware-33f33673-015e-4ddd-a323-7a78ff2a5517 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1168.567607] env[61806]: DEBUG oslo_vmware.rw_handles [None req-055516b6-e34c-4ddb-87ba-434948f7c77d tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] Lease: (returnval){ [ 1168.567607] env[61806]: value = "session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]524c7bc1-9321-0c97-ea87-ae5eefdff497" [ 1168.567607] env[61806]: _type = "HttpNfcLease" [ 1168.567607] env[61806]: } obtained for vApp import into resource pool (val){ [ 1168.567607] env[61806]: value = "resgroup-9" [ 1168.567607] env[61806]: _type = "ResourcePool" [ 1168.567607] env[61806]: }. {{(pid=61806) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:461}} [ 1168.568027] env[61806]: DEBUG oslo_vmware.api [None req-055516b6-e34c-4ddb-87ba-434948f7c77d tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] Waiting for the lease: (returnval){ [ 1168.568027] env[61806]: value = "session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]524c7bc1-9321-0c97-ea87-ae5eefdff497" [ 1168.568027] env[61806]: _type = "HttpNfcLease" [ 1168.568027] env[61806]: } to be ready. {{(pid=61806) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 1168.573855] env[61806]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1168.573855] env[61806]: value = "session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]524c7bc1-9321-0c97-ea87-ae5eefdff497" [ 1168.573855] env[61806]: _type = "HttpNfcLease" [ 1168.573855] env[61806]: } is initializing. {{(pid=61806) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 1168.776672] env[61806]: DEBUG oslo_concurrency.lockutils [req-13b1c5c8-d7ae-4da8-951d-c3f2ecb47b07 req-16e60915-6020-4e33-bba0-4f9c47386446 service nova] Releasing lock "refresh_cache-29210bcd-0fe1-4ed9-a459-abc8016a4255" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1169.075646] env[61806]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1169.075646] env[61806]: value = "session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]524c7bc1-9321-0c97-ea87-ae5eefdff497" [ 1169.075646] env[61806]: _type = "HttpNfcLease" [ 1169.075646] env[61806]: } is ready. {{(pid=61806) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 1169.076201] env[61806]: DEBUG oslo_vmware.rw_handles [None req-055516b6-e34c-4ddb-87ba-434948f7c77d tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 1169.076201] env[61806]: value = "session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]524c7bc1-9321-0c97-ea87-ae5eefdff497" [ 1169.076201] env[61806]: _type = "HttpNfcLease" [ 1169.076201] env[61806]: }. {{(pid=61806) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:467}} [ 1169.076667] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e2b4b14d-9fea-4c94-836c-ecf23c98b1d0 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1169.083893] env[61806]: DEBUG oslo_vmware.rw_handles [None req-055516b6-e34c-4ddb-87ba-434948f7c77d tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] Found VMDK URL: https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/521cc0b4-43ae-cf7d-116c-8c3d8ad4d73e/disk-0.vmdk from lease info. {{(pid=61806) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 1169.084078] env[61806]: DEBUG oslo_vmware.rw_handles [None req-055516b6-e34c-4ddb-87ba-434948f7c77d tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] Creating HTTP connection to write to file with size = 31669760 and URL = https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/521cc0b4-43ae-cf7d-116c-8c3d8ad4d73e/disk-0.vmdk. {{(pid=61806) _create_write_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:122}} [ 1169.147381] env[61806]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-de96a4b3-4a44-4606-b333-3c4d051eac99 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1170.279221] env[61806]: DEBUG oslo_vmware.rw_handles [None req-055516b6-e34c-4ddb-87ba-434948f7c77d tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] Completed reading data from the image iterator. {{(pid=61806) read /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:765}} [ 1170.279613] env[61806]: DEBUG oslo_vmware.rw_handles [None req-055516b6-e34c-4ddb-87ba-434948f7c77d tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] Getting lease state for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/521cc0b4-43ae-cf7d-116c-8c3d8ad4d73e/disk-0.vmdk. {{(pid=61806) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 1170.280519] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2ddab6a1-533d-49a0-aaf3-acab1bc0c2cf {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1170.287789] env[61806]: DEBUG oslo_vmware.rw_handles [None req-055516b6-e34c-4ddb-87ba-434948f7c77d tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] Lease for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/521cc0b4-43ae-cf7d-116c-8c3d8ad4d73e/disk-0.vmdk is in state: ready. {{(pid=61806) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 1170.287965] env[61806]: DEBUG oslo_vmware.rw_handles [None req-055516b6-e34c-4ddb-87ba-434948f7c77d tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] Releasing lease for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/521cc0b4-43ae-cf7d-116c-8c3d8ad4d73e/disk-0.vmdk. {{(pid=61806) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:440}} [ 1170.288214] env[61806]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseComplete with opID=oslo.vmware-1f90fb9a-b78a-4e1e-9f43-ecb2b69b106e {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1170.660710] env[61806]: DEBUG oslo_vmware.rw_handles [None req-055516b6-e34c-4ddb-87ba-434948f7c77d tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] Closed VMDK write handle for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/521cc0b4-43ae-cf7d-116c-8c3d8ad4d73e/disk-0.vmdk. {{(pid=61806) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:646}} [ 1170.660939] env[61806]: INFO nova.virt.vmwareapi.images [None req-055516b6-e34c-4ddb-87ba-434948f7c77d tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] [instance: 29210bcd-0fe1-4ed9-a459-abc8016a4255] Downloaded image file data 2ea32304-20d1-4039-a93a-b4c97a5656bf [ 1170.661822] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3e92a80d-1972-4647-8b46-3b655bc0bd38 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1170.677147] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-1a2ce4c2-bd2a-4f17-ac8c-b0bec79da655 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1170.701379] env[61806]: INFO nova.virt.vmwareapi.images [None req-055516b6-e34c-4ddb-87ba-434948f7c77d tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] [instance: 29210bcd-0fe1-4ed9-a459-abc8016a4255] The imported VM was unregistered [ 1170.704011] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-055516b6-e34c-4ddb-87ba-434948f7c77d tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] [instance: 29210bcd-0fe1-4ed9-a459-abc8016a4255] Caching image {{(pid=61806) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:638}} [ 1170.704320] env[61806]: DEBUG nova.virt.vmwareapi.ds_util [None req-055516b6-e34c-4ddb-87ba-434948f7c77d tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] Creating directory with path [datastore2] devstack-image-cache_base/2ea32304-20d1-4039-a93a-b4c97a5656bf {{(pid=61806) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1170.704595] env[61806]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-5e36c4b9-0686-4610-b330-f46b65771017 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1170.714219] env[61806]: DEBUG nova.virt.vmwareapi.ds_util [None req-055516b6-e34c-4ddb-87ba-434948f7c77d tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] Created directory with path [datastore2] devstack-image-cache_base/2ea32304-20d1-4039-a93a-b4c97a5656bf {{(pid=61806) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1170.714401] env[61806]: DEBUG nova.virt.vmwareapi.ds_util [None req-055516b6-e34c-4ddb-87ba-434948f7c77d tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] Moving virtual disk from [datastore2] OSTACK_IMG_520967df-fd29-4a4c-ab10-c405db541a06/OSTACK_IMG_520967df-fd29-4a4c-ab10-c405db541a06.vmdk to [datastore2] devstack-image-cache_base/2ea32304-20d1-4039-a93a-b4c97a5656bf/2ea32304-20d1-4039-a93a-b4c97a5656bf.vmdk. {{(pid=61806) disk_move /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:263}} [ 1170.714649] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.MoveVirtualDisk_Task with opID=oslo.vmware-f435116e-19f3-452f-a104-a5fa916f48e8 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1170.720219] env[61806]: DEBUG oslo_vmware.api [None req-055516b6-e34c-4ddb-87ba-434948f7c77d tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] Waiting for the task: (returnval){ [ 1170.720219] env[61806]: value = "task-1295290" [ 1170.720219] env[61806]: _type = "Task" [ 1170.720219] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1170.727535] env[61806]: DEBUG oslo_vmware.api [None req-055516b6-e34c-4ddb-87ba-434948f7c77d tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] Task: {'id': task-1295290, 'name': MoveVirtualDisk_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1171.235146] env[61806]: DEBUG oslo_vmware.api [None req-055516b6-e34c-4ddb-87ba-434948f7c77d tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] Task: {'id': task-1295290, 'name': MoveVirtualDisk_Task} progress is 21%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1171.733085] env[61806]: DEBUG oslo_vmware.api [None req-055516b6-e34c-4ddb-87ba-434948f7c77d tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] Task: {'id': task-1295290, 'name': MoveVirtualDisk_Task} progress is 46%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1172.235872] env[61806]: DEBUG oslo_vmware.api [None req-055516b6-e34c-4ddb-87ba-434948f7c77d tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] Task: {'id': task-1295290, 'name': MoveVirtualDisk_Task} progress is 66%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1172.736895] env[61806]: DEBUG oslo_vmware.api [None req-055516b6-e34c-4ddb-87ba-434948f7c77d tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] Task: {'id': task-1295290, 'name': MoveVirtualDisk_Task} progress is 88%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1173.233045] env[61806]: DEBUG oslo_vmware.api [None req-055516b6-e34c-4ddb-87ba-434948f7c77d tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] Task: {'id': task-1295290, 'name': MoveVirtualDisk_Task, 'duration_secs': 2.280388} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1173.233212] env[61806]: INFO nova.virt.vmwareapi.ds_util [None req-055516b6-e34c-4ddb-87ba-434948f7c77d tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] Moved virtual disk from [datastore2] OSTACK_IMG_520967df-fd29-4a4c-ab10-c405db541a06/OSTACK_IMG_520967df-fd29-4a4c-ab10-c405db541a06.vmdk to [datastore2] devstack-image-cache_base/2ea32304-20d1-4039-a93a-b4c97a5656bf/2ea32304-20d1-4039-a93a-b4c97a5656bf.vmdk. [ 1173.233255] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-055516b6-e34c-4ddb-87ba-434948f7c77d tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] [instance: 29210bcd-0fe1-4ed9-a459-abc8016a4255] Cleaning up location [datastore2] OSTACK_IMG_520967df-fd29-4a4c-ab10-c405db541a06 {{(pid=61806) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:640}} [ 1173.233419] env[61806]: DEBUG nova.virt.vmwareapi.ds_util [None req-055516b6-e34c-4ddb-87ba-434948f7c77d tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] Deleting the datastore file [datastore2] OSTACK_IMG_520967df-fd29-4a4c-ab10-c405db541a06 {{(pid=61806) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1173.233713] env[61806]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-ae82e0b9-8038-4315-a049-2d844ba4b78f {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1173.240246] env[61806]: DEBUG oslo_vmware.api [None req-055516b6-e34c-4ddb-87ba-434948f7c77d tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] Waiting for the task: (returnval){ [ 1173.240246] env[61806]: value = "task-1295291" [ 1173.240246] env[61806]: _type = "Task" [ 1173.240246] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1173.247637] env[61806]: DEBUG oslo_vmware.api [None req-055516b6-e34c-4ddb-87ba-434948f7c77d tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] Task: {'id': task-1295291, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1173.749344] env[61806]: DEBUG oslo_vmware.api [None req-055516b6-e34c-4ddb-87ba-434948f7c77d tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] Task: {'id': task-1295291, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.033805} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1173.749776] env[61806]: DEBUG nova.virt.vmwareapi.ds_util [None req-055516b6-e34c-4ddb-87ba-434948f7c77d tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] Deleted the datastore file {{(pid=61806) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1173.749776] env[61806]: DEBUG oslo_concurrency.lockutils [None req-055516b6-e34c-4ddb-87ba-434948f7c77d tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] Releasing lock "[datastore2] devstack-image-cache_base/2ea32304-20d1-4039-a93a-b4c97a5656bf/2ea32304-20d1-4039-a93a-b4c97a5656bf.vmdk" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1173.749985] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [None req-055516b6-e34c-4ddb-87ba-434948f7c77d tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/2ea32304-20d1-4039-a93a-b4c97a5656bf/2ea32304-20d1-4039-a93a-b4c97a5656bf.vmdk to [datastore2] 29210bcd-0fe1-4ed9-a459-abc8016a4255/29210bcd-0fe1-4ed9-a459-abc8016a4255.vmdk {{(pid=61806) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 1173.750254] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-84ef77d3-5a45-44ef-b07d-650ac4cd04a4 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1173.756866] env[61806]: DEBUG oslo_vmware.api [None req-055516b6-e34c-4ddb-87ba-434948f7c77d tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] Waiting for the task: (returnval){ [ 1173.756866] env[61806]: value = "task-1295292" [ 1173.756866] env[61806]: _type = "Task" [ 1173.756866] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1173.763605] env[61806]: DEBUG oslo_vmware.api [None req-055516b6-e34c-4ddb-87ba-434948f7c77d tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] Task: {'id': task-1295292, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1174.272789] env[61806]: DEBUG oslo_vmware.api [None req-055516b6-e34c-4ddb-87ba-434948f7c77d tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] Task: {'id': task-1295292, 'name': CopyVirtualDisk_Task} progress is 24%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1174.772245] env[61806]: DEBUG oslo_vmware.api [None req-055516b6-e34c-4ddb-87ba-434948f7c77d tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] Task: {'id': task-1295292, 'name': CopyVirtualDisk_Task} progress is 46%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1175.271776] env[61806]: DEBUG oslo_vmware.api [None req-055516b6-e34c-4ddb-87ba-434948f7c77d tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] Task: {'id': task-1295292, 'name': CopyVirtualDisk_Task} progress is 69%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1175.772169] env[61806]: DEBUG oslo_vmware.api [None req-055516b6-e34c-4ddb-87ba-434948f7c77d tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] Task: {'id': task-1295292, 'name': CopyVirtualDisk_Task} progress is 94%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1176.269964] env[61806]: DEBUG oslo_vmware.api [None req-055516b6-e34c-4ddb-87ba-434948f7c77d tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] Task: {'id': task-1295292, 'name': CopyVirtualDisk_Task, 'duration_secs': 2.134776} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1176.270208] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [None req-055516b6-e34c-4ddb-87ba-434948f7c77d tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/2ea32304-20d1-4039-a93a-b4c97a5656bf/2ea32304-20d1-4039-a93a-b4c97a5656bf.vmdk to [datastore2] 29210bcd-0fe1-4ed9-a459-abc8016a4255/29210bcd-0fe1-4ed9-a459-abc8016a4255.vmdk {{(pid=61806) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 1176.271115] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f33533bd-6030-4737-bc2a-15f00a15a7e9 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1176.293493] env[61806]: DEBUG nova.virt.vmwareapi.volumeops [None req-055516b6-e34c-4ddb-87ba-434948f7c77d tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] [instance: 29210bcd-0fe1-4ed9-a459-abc8016a4255] Reconfiguring VM instance instance-0000006a to attach disk [datastore2] 29210bcd-0fe1-4ed9-a459-abc8016a4255/29210bcd-0fe1-4ed9-a459-abc8016a4255.vmdk or device None with type streamOptimized {{(pid=61806) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1176.293838] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-c2a0b22c-f552-41ef-bd11-2b0ce214bc0f {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1176.313300] env[61806]: DEBUG oslo_vmware.api [None req-055516b6-e34c-4ddb-87ba-434948f7c77d tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] Waiting for the task: (returnval){ [ 1176.313300] env[61806]: value = "task-1295293" [ 1176.313300] env[61806]: _type = "Task" [ 1176.313300] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1176.320544] env[61806]: DEBUG oslo_vmware.api [None req-055516b6-e34c-4ddb-87ba-434948f7c77d tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] Task: {'id': task-1295293, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1176.452489] env[61806]: DEBUG oslo_concurrency.lockutils [None req-189c9291-3e04-4e60-8c40-ab22f37b23a0 tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Acquiring lock "c66fe2e7-7af0-48fa-8509-fcc24c1a437b" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1176.452802] env[61806]: DEBUG oslo_concurrency.lockutils [None req-189c9291-3e04-4e60-8c40-ab22f37b23a0 tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Lock "c66fe2e7-7af0-48fa-8509-fcc24c1a437b" acquired by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: waited 0.000s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1176.453021] env[61806]: DEBUG nova.compute.manager [None req-189c9291-3e04-4e60-8c40-ab22f37b23a0 tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] [instance: c66fe2e7-7af0-48fa-8509-fcc24c1a437b] Checking state {{(pid=61806) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1176.453941] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-57ed18d4-85b7-4291-b83e-7eac3766d317 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1176.460327] env[61806]: DEBUG nova.compute.manager [None req-189c9291-3e04-4e60-8c40-ab22f37b23a0 tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] [instance: c66fe2e7-7af0-48fa-8509-fcc24c1a437b] Stopping instance; current vm_state: active, current task_state: powering-off, current DB power_state: 1, current VM power_state: 1 {{(pid=61806) do_stop_instance /opt/stack/nova/nova/compute/manager.py:3368}} [ 1176.460899] env[61806]: DEBUG nova.objects.instance [None req-189c9291-3e04-4e60-8c40-ab22f37b23a0 tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Lazy-loading 'flavor' on Instance uuid c66fe2e7-7af0-48fa-8509-fcc24c1a437b {{(pid=61806) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1176.822245] env[61806]: DEBUG oslo_vmware.api [None req-055516b6-e34c-4ddb-87ba-434948f7c77d tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] Task: {'id': task-1295293, 'name': ReconfigVM_Task, 'duration_secs': 0.26923} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1176.822744] env[61806]: DEBUG nova.virt.vmwareapi.volumeops [None req-055516b6-e34c-4ddb-87ba-434948f7c77d tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] [instance: 29210bcd-0fe1-4ed9-a459-abc8016a4255] Reconfigured VM instance instance-0000006a to attach disk [datastore2] 29210bcd-0fe1-4ed9-a459-abc8016a4255/29210bcd-0fe1-4ed9-a459-abc8016a4255.vmdk or device None with type streamOptimized {{(pid=61806) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1176.823129] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-89d38143-cd02-4ae6-81da-717fdfc4a512 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1176.829288] env[61806]: DEBUG oslo_vmware.api [None req-055516b6-e34c-4ddb-87ba-434948f7c77d tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] Waiting for the task: (returnval){ [ 1176.829288] env[61806]: value = "task-1295294" [ 1176.829288] env[61806]: _type = "Task" [ 1176.829288] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1176.836248] env[61806]: DEBUG oslo_vmware.api [None req-055516b6-e34c-4ddb-87ba-434948f7c77d tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] Task: {'id': task-1295294, 'name': Rename_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1176.966634] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [None req-189c9291-3e04-4e60-8c40-ab22f37b23a0 tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] [instance: c66fe2e7-7af0-48fa-8509-fcc24c1a437b] Powering off the VM {{(pid=61806) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1176.966941] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-fad79515-29b4-4f63-b06f-e46e95d2c123 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1176.972603] env[61806]: DEBUG oslo_vmware.api [None req-189c9291-3e04-4e60-8c40-ab22f37b23a0 tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Waiting for the task: (returnval){ [ 1176.972603] env[61806]: value = "task-1295295" [ 1176.972603] env[61806]: _type = "Task" [ 1176.972603] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1176.981640] env[61806]: DEBUG oslo_vmware.api [None req-189c9291-3e04-4e60-8c40-ab22f37b23a0 tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Task: {'id': task-1295295, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1177.339069] env[61806]: DEBUG oslo_vmware.api [None req-055516b6-e34c-4ddb-87ba-434948f7c77d tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] Task: {'id': task-1295294, 'name': Rename_Task, 'duration_secs': 0.131004} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1177.339382] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [None req-055516b6-e34c-4ddb-87ba-434948f7c77d tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] [instance: 29210bcd-0fe1-4ed9-a459-abc8016a4255] Powering on the VM {{(pid=61806) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1177.339570] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-d1b3397a-a4b3-475c-8b17-de74b1fc80c2 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1177.345484] env[61806]: DEBUG oslo_vmware.api [None req-055516b6-e34c-4ddb-87ba-434948f7c77d tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] Waiting for the task: (returnval){ [ 1177.345484] env[61806]: value = "task-1295296" [ 1177.345484] env[61806]: _type = "Task" [ 1177.345484] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1177.352530] env[61806]: DEBUG oslo_vmware.api [None req-055516b6-e34c-4ddb-87ba-434948f7c77d tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] Task: {'id': task-1295296, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1177.482986] env[61806]: DEBUG oslo_vmware.api [None req-189c9291-3e04-4e60-8c40-ab22f37b23a0 tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Task: {'id': task-1295295, 'name': PowerOffVM_Task, 'duration_secs': 0.168189} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1177.483258] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [None req-189c9291-3e04-4e60-8c40-ab22f37b23a0 tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] [instance: c66fe2e7-7af0-48fa-8509-fcc24c1a437b] Powered off the VM {{(pid=61806) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1177.483443] env[61806]: DEBUG nova.compute.manager [None req-189c9291-3e04-4e60-8c40-ab22f37b23a0 tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] [instance: c66fe2e7-7af0-48fa-8509-fcc24c1a437b] Checking state {{(pid=61806) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1177.484256] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-80ce4f1e-e3f3-42f8-9d03-b23ca96e51c4 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1177.855224] env[61806]: DEBUG oslo_vmware.api [None req-055516b6-e34c-4ddb-87ba-434948f7c77d tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] Task: {'id': task-1295296, 'name': PowerOnVM_Task, 'duration_secs': 0.42057} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1177.855505] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [None req-055516b6-e34c-4ddb-87ba-434948f7c77d tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] [instance: 29210bcd-0fe1-4ed9-a459-abc8016a4255] Powered on the VM {{(pid=61806) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1177.959351] env[61806]: DEBUG nova.compute.manager [None req-055516b6-e34c-4ddb-87ba-434948f7c77d tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] [instance: 29210bcd-0fe1-4ed9-a459-abc8016a4255] Checking state {{(pid=61806) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1177.960268] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-32393e52-8b15-4786-a260-690f975ad972 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1177.994996] env[61806]: DEBUG oslo_concurrency.lockutils [None req-189c9291-3e04-4e60-8c40-ab22f37b23a0 tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Lock "c66fe2e7-7af0-48fa-8509-fcc24c1a437b" "released" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: held 1.542s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1178.475920] env[61806]: DEBUG oslo_concurrency.lockutils [None req-055516b6-e34c-4ddb-87ba-434948f7c77d tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] Lock "29210bcd-0fe1-4ed9-a459-abc8016a4255" "released" by "nova.compute.manager.ComputeManager.unshelve_instance..do_unshelve_instance" :: held 18.541s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1178.813721] env[61806]: DEBUG nova.objects.instance [None req-35dc2799-1c81-4beb-8cb3-dbeb5e07d5df tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Lazy-loading 'flavor' on Instance uuid c66fe2e7-7af0-48fa-8509-fcc24c1a437b {{(pid=61806) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1179.318854] env[61806]: DEBUG oslo_concurrency.lockutils [None req-35dc2799-1c81-4beb-8cb3-dbeb5e07d5df tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Acquiring lock "refresh_cache-c66fe2e7-7af0-48fa-8509-fcc24c1a437b" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1179.319066] env[61806]: DEBUG oslo_concurrency.lockutils [None req-35dc2799-1c81-4beb-8cb3-dbeb5e07d5df tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Acquired lock "refresh_cache-c66fe2e7-7af0-48fa-8509-fcc24c1a437b" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1179.319283] env[61806]: DEBUG nova.network.neutron [None req-35dc2799-1c81-4beb-8cb3-dbeb5e07d5df tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] [instance: c66fe2e7-7af0-48fa-8509-fcc24c1a437b] Building network info cache for instance {{(pid=61806) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1179.319499] env[61806]: DEBUG nova.objects.instance [None req-35dc2799-1c81-4beb-8cb3-dbeb5e07d5df tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Lazy-loading 'info_cache' on Instance uuid c66fe2e7-7af0-48fa-8509-fcc24c1a437b {{(pid=61806) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1179.824610] env[61806]: DEBUG nova.objects.base [None req-35dc2799-1c81-4beb-8cb3-dbeb5e07d5df tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Object Instance lazy-loaded attributes: flavor,info_cache {{(pid=61806) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 1180.552371] env[61806]: DEBUG nova.network.neutron [None req-35dc2799-1c81-4beb-8cb3-dbeb5e07d5df tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] [instance: c66fe2e7-7af0-48fa-8509-fcc24c1a437b] Updating instance_info_cache with network_info: [{"id": "5bb12b4b-dfcb-416e-8095-91002ba3ca7d", "address": "fa:16:3e:9d:21:7a", "network": {"id": "746a03b6-c1a6-4d27-98b9-0bee6376634f", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-1971382587-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.245", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c1f2d161c50f4b70987b90e6364c3779", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6a31b224-77d7-48a4-af87-312758611b8e", "external-id": "nsx-vlan-transportzone-761", "segmentation_id": 761, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap5bb12b4b-df", "ovs_interfaceid": "5bb12b4b-dfcb-416e-8095-91002ba3ca7d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61806) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1181.055599] env[61806]: DEBUG oslo_concurrency.lockutils [None req-35dc2799-1c81-4beb-8cb3-dbeb5e07d5df tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Releasing lock "refresh_cache-c66fe2e7-7af0-48fa-8509-fcc24c1a437b" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1181.558800] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [None req-35dc2799-1c81-4beb-8cb3-dbeb5e07d5df tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] [instance: c66fe2e7-7af0-48fa-8509-fcc24c1a437b] Powering on the VM {{(pid=61806) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1181.559141] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-7ab08ddc-f9b4-4bf3-b60b-0b90c3c69c1d {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1181.566366] env[61806]: DEBUG oslo_vmware.api [None req-35dc2799-1c81-4beb-8cb3-dbeb5e07d5df tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Waiting for the task: (returnval){ [ 1181.566366] env[61806]: value = "task-1295297" [ 1181.566366] env[61806]: _type = "Task" [ 1181.566366] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1181.573577] env[61806]: DEBUG oslo_vmware.api [None req-35dc2799-1c81-4beb-8cb3-dbeb5e07d5df tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Task: {'id': task-1295297, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1182.079155] env[61806]: DEBUG oslo_vmware.api [None req-35dc2799-1c81-4beb-8cb3-dbeb5e07d5df tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Task: {'id': task-1295297, 'name': PowerOnVM_Task, 'duration_secs': 0.40526} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1182.079619] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [None req-35dc2799-1c81-4beb-8cb3-dbeb5e07d5df tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] [instance: c66fe2e7-7af0-48fa-8509-fcc24c1a437b] Powered on the VM {{(pid=61806) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1182.079929] env[61806]: DEBUG nova.compute.manager [None req-35dc2799-1c81-4beb-8cb3-dbeb5e07d5df tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] [instance: c66fe2e7-7af0-48fa-8509-fcc24c1a437b] Checking state {{(pid=61806) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1182.081104] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2f89b57c-67b8-48a8-9ff3-e486e0f945b6 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1183.747349] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1f9e6e0f-a366-4d43-9692-6da2897d3130 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1183.754148] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-5099ec65-c3b4-44b7-848b-cae8d4c863a6 tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] [instance: c66fe2e7-7af0-48fa-8509-fcc24c1a437b] Suspending the VM {{(pid=61806) suspend /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1163}} [ 1183.754396] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.SuspendVM_Task with opID=oslo.vmware-eebeeee0-e981-4948-a836-64bed9ac669b {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1183.760561] env[61806]: DEBUG oslo_vmware.api [None req-5099ec65-c3b4-44b7-848b-cae8d4c863a6 tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Waiting for the task: (returnval){ [ 1183.760561] env[61806]: value = "task-1295298" [ 1183.760561] env[61806]: _type = "Task" [ 1183.760561] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1183.768373] env[61806]: DEBUG oslo_vmware.api [None req-5099ec65-c3b4-44b7-848b-cae8d4c863a6 tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Task: {'id': task-1295298, 'name': SuspendVM_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1184.270897] env[61806]: DEBUG oslo_vmware.api [None req-5099ec65-c3b4-44b7-848b-cae8d4c863a6 tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Task: {'id': task-1295298, 'name': SuspendVM_Task} progress is 66%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1184.771808] env[61806]: DEBUG oslo_vmware.api [None req-5099ec65-c3b4-44b7-848b-cae8d4c863a6 tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Task: {'id': task-1295298, 'name': SuspendVM_Task, 'duration_secs': 0.736692} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1184.772206] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-5099ec65-c3b4-44b7-848b-cae8d4c863a6 tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] [instance: c66fe2e7-7af0-48fa-8509-fcc24c1a437b] Suspended the VM {{(pid=61806) suspend /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1167}} [ 1184.772274] env[61806]: DEBUG nova.compute.manager [None req-5099ec65-c3b4-44b7-848b-cae8d4c863a6 tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] [instance: c66fe2e7-7af0-48fa-8509-fcc24c1a437b] Checking state {{(pid=61806) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1184.773029] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9665bed7-7733-4b15-ade6-7bceb3bf37e3 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1186.090453] env[61806]: INFO nova.compute.manager [None req-32b70d3b-a56e-411f-944a-b7a594a008e8 tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] [instance: c66fe2e7-7af0-48fa-8509-fcc24c1a437b] Resuming [ 1186.091045] env[61806]: DEBUG nova.objects.instance [None req-32b70d3b-a56e-411f-944a-b7a594a008e8 tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Lazy-loading 'flavor' on Instance uuid c66fe2e7-7af0-48fa-8509-fcc24c1a437b {{(pid=61806) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1187.099210] env[61806]: DEBUG oslo_concurrency.lockutils [None req-32b70d3b-a56e-411f-944a-b7a594a008e8 tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Acquiring lock "refresh_cache-c66fe2e7-7af0-48fa-8509-fcc24c1a437b" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1187.099610] env[61806]: DEBUG oslo_concurrency.lockutils [None req-32b70d3b-a56e-411f-944a-b7a594a008e8 tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Acquired lock "refresh_cache-c66fe2e7-7af0-48fa-8509-fcc24c1a437b" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1187.099610] env[61806]: DEBUG nova.network.neutron [None req-32b70d3b-a56e-411f-944a-b7a594a008e8 tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] [instance: c66fe2e7-7af0-48fa-8509-fcc24c1a437b] Building network info cache for instance {{(pid=61806) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1187.985340] env[61806]: DEBUG nova.network.neutron [None req-32b70d3b-a56e-411f-944a-b7a594a008e8 tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] [instance: c66fe2e7-7af0-48fa-8509-fcc24c1a437b] Updating instance_info_cache with network_info: [{"id": "5bb12b4b-dfcb-416e-8095-91002ba3ca7d", "address": "fa:16:3e:9d:21:7a", "network": {"id": "746a03b6-c1a6-4d27-98b9-0bee6376634f", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-1971382587-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.245", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c1f2d161c50f4b70987b90e6364c3779", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6a31b224-77d7-48a4-af87-312758611b8e", "external-id": "nsx-vlan-transportzone-761", "segmentation_id": 761, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap5bb12b4b-df", "ovs_interfaceid": "5bb12b4b-dfcb-416e-8095-91002ba3ca7d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61806) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1188.487985] env[61806]: DEBUG oslo_concurrency.lockutils [None req-32b70d3b-a56e-411f-944a-b7a594a008e8 tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Releasing lock "refresh_cache-c66fe2e7-7af0-48fa-8509-fcc24c1a437b" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1188.489398] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-61b39181-9edb-4983-8bfd-fe531ac0d920 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1188.496165] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-32b70d3b-a56e-411f-944a-b7a594a008e8 tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] [instance: c66fe2e7-7af0-48fa-8509-fcc24c1a437b] Resuming the VM {{(pid=61806) resume /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1184}} [ 1188.496395] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-046a077b-ce31-4ae9-ac54-ebdef92996ac {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1188.503789] env[61806]: DEBUG oslo_vmware.api [None req-32b70d3b-a56e-411f-944a-b7a594a008e8 tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Waiting for the task: (returnval){ [ 1188.503789] env[61806]: value = "task-1295299" [ 1188.503789] env[61806]: _type = "Task" [ 1188.503789] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1188.510780] env[61806]: DEBUG oslo_vmware.api [None req-32b70d3b-a56e-411f-944a-b7a594a008e8 tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Task: {'id': task-1295299, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1189.015737] env[61806]: DEBUG oslo_vmware.api [None req-32b70d3b-a56e-411f-944a-b7a594a008e8 tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Task: {'id': task-1295299, 'name': PowerOnVM_Task} progress is 100%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1189.517856] env[61806]: DEBUG oslo_vmware.api [None req-32b70d3b-a56e-411f-944a-b7a594a008e8 tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Task: {'id': task-1295299, 'name': PowerOnVM_Task, 'duration_secs': 0.532198} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1189.518417] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-32b70d3b-a56e-411f-944a-b7a594a008e8 tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] [instance: c66fe2e7-7af0-48fa-8509-fcc24c1a437b] Resumed the VM {{(pid=61806) resume /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1189}} [ 1189.518716] env[61806]: DEBUG nova.compute.manager [None req-32b70d3b-a56e-411f-944a-b7a594a008e8 tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] [instance: c66fe2e7-7af0-48fa-8509-fcc24c1a437b] Checking state {{(pid=61806) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1189.519849] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5d65bd00-b7df-44fc-836b-1e92d5fb32cf {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1190.018121] env[61806]: DEBUG oslo_service.periodic_task [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=61806) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1190.775155] env[61806]: DEBUG oslo_concurrency.lockutils [None req-5c62cd58-8080-43f6-b08a-0e362578d740 tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Acquiring lock "c66fe2e7-7af0-48fa-8509-fcc24c1a437b" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1190.775582] env[61806]: DEBUG oslo_concurrency.lockutils [None req-5c62cd58-8080-43f6-b08a-0e362578d740 tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Lock "c66fe2e7-7af0-48fa-8509-fcc24c1a437b" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1190.775684] env[61806]: DEBUG oslo_concurrency.lockutils [None req-5c62cd58-8080-43f6-b08a-0e362578d740 tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Acquiring lock "c66fe2e7-7af0-48fa-8509-fcc24c1a437b-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1190.775881] env[61806]: DEBUG oslo_concurrency.lockutils [None req-5c62cd58-8080-43f6-b08a-0e362578d740 tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Lock "c66fe2e7-7af0-48fa-8509-fcc24c1a437b-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1190.776151] env[61806]: DEBUG oslo_concurrency.lockutils [None req-5c62cd58-8080-43f6-b08a-0e362578d740 tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Lock "c66fe2e7-7af0-48fa-8509-fcc24c1a437b-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1190.778465] env[61806]: INFO nova.compute.manager [None req-5c62cd58-8080-43f6-b08a-0e362578d740 tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] [instance: c66fe2e7-7af0-48fa-8509-fcc24c1a437b] Terminating instance [ 1190.780216] env[61806]: DEBUG nova.compute.manager [None req-5c62cd58-8080-43f6-b08a-0e362578d740 tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] [instance: c66fe2e7-7af0-48fa-8509-fcc24c1a437b] Start destroying the instance on the hypervisor. {{(pid=61806) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1190.780432] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-5c62cd58-8080-43f6-b08a-0e362578d740 tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] [instance: c66fe2e7-7af0-48fa-8509-fcc24c1a437b] Destroying instance {{(pid=61806) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1190.781301] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2efc27ce-dd66-469a-adb3-78d80a2eb681 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1190.789211] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [None req-5c62cd58-8080-43f6-b08a-0e362578d740 tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] [instance: c66fe2e7-7af0-48fa-8509-fcc24c1a437b] Powering off the VM {{(pid=61806) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1190.789438] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-aeac70a8-b339-4638-ad5f-f919bd5e3f32 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1190.796399] env[61806]: DEBUG oslo_vmware.api [None req-5c62cd58-8080-43f6-b08a-0e362578d740 tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Waiting for the task: (returnval){ [ 1190.796399] env[61806]: value = "task-1295300" [ 1190.796399] env[61806]: _type = "Task" [ 1190.796399] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1190.804268] env[61806]: DEBUG oslo_vmware.api [None req-5c62cd58-8080-43f6-b08a-0e362578d740 tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Task: {'id': task-1295300, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1191.023156] env[61806]: DEBUG oslo_service.periodic_task [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=61806) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1191.023360] env[61806]: DEBUG nova.compute.manager [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Starting heal instance info cache {{(pid=61806) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9926}} [ 1191.307247] env[61806]: DEBUG oslo_vmware.api [None req-5c62cd58-8080-43f6-b08a-0e362578d740 tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Task: {'id': task-1295300, 'name': PowerOffVM_Task, 'duration_secs': 0.182745} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1191.307480] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [None req-5c62cd58-8080-43f6-b08a-0e362578d740 tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] [instance: c66fe2e7-7af0-48fa-8509-fcc24c1a437b] Powered off the VM {{(pid=61806) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1191.307652] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-5c62cd58-8080-43f6-b08a-0e362578d740 tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] [instance: c66fe2e7-7af0-48fa-8509-fcc24c1a437b] Unregistering the VM {{(pid=61806) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1191.307936] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-cbe07244-a43d-4da5-8eac-ce7bcc3a51bd {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1191.815700] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-5c62cd58-8080-43f6-b08a-0e362578d740 tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] [instance: c66fe2e7-7af0-48fa-8509-fcc24c1a437b] Unregistered the VM {{(pid=61806) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1191.816072] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-5c62cd58-8080-43f6-b08a-0e362578d740 tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] [instance: c66fe2e7-7af0-48fa-8509-fcc24c1a437b] Deleting contents of the VM from datastore datastore1 {{(pid=61806) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1191.816181] env[61806]: DEBUG nova.virt.vmwareapi.ds_util [None req-5c62cd58-8080-43f6-b08a-0e362578d740 tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Deleting the datastore file [datastore1] c66fe2e7-7af0-48fa-8509-fcc24c1a437b {{(pid=61806) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1191.816441] env[61806]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-823874c5-10f4-4503-b0be-2119e1bdf6e0 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1191.822821] env[61806]: DEBUG oslo_vmware.api [None req-5c62cd58-8080-43f6-b08a-0e362578d740 tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Waiting for the task: (returnval){ [ 1191.822821] env[61806]: value = "task-1295302" [ 1191.822821] env[61806]: _type = "Task" [ 1191.822821] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1191.830362] env[61806]: DEBUG oslo_vmware.api [None req-5c62cd58-8080-43f6-b08a-0e362578d740 tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Task: {'id': task-1295302, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1192.029266] env[61806]: DEBUG nova.compute.manager [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Didn't find any instances for network info cache update. {{(pid=61806) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10012}} [ 1192.332966] env[61806]: DEBUG oslo_vmware.api [None req-5c62cd58-8080-43f6-b08a-0e362578d740 tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Task: {'id': task-1295302, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.152984} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1192.333225] env[61806]: DEBUG nova.virt.vmwareapi.ds_util [None req-5c62cd58-8080-43f6-b08a-0e362578d740 tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Deleted the datastore file {{(pid=61806) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1192.333397] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-5c62cd58-8080-43f6-b08a-0e362578d740 tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] [instance: c66fe2e7-7af0-48fa-8509-fcc24c1a437b] Deleted contents of the VM from datastore datastore1 {{(pid=61806) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1192.333578] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-5c62cd58-8080-43f6-b08a-0e362578d740 tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] [instance: c66fe2e7-7af0-48fa-8509-fcc24c1a437b] Instance destroyed {{(pid=61806) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1192.333756] env[61806]: INFO nova.compute.manager [None req-5c62cd58-8080-43f6-b08a-0e362578d740 tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] [instance: c66fe2e7-7af0-48fa-8509-fcc24c1a437b] Took 1.55 seconds to destroy the instance on the hypervisor. [ 1192.334017] env[61806]: DEBUG oslo.service.loopingcall [None req-5c62cd58-8080-43f6-b08a-0e362578d740 tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61806) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1192.334218] env[61806]: DEBUG nova.compute.manager [-] [instance: c66fe2e7-7af0-48fa-8509-fcc24c1a437b] Deallocating network for instance {{(pid=61806) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1192.334313] env[61806]: DEBUG nova.network.neutron [-] [instance: c66fe2e7-7af0-48fa-8509-fcc24c1a437b] deallocate_for_instance() {{(pid=61806) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1192.823667] env[61806]: DEBUG nova.compute.manager [req-5be7572c-9816-4398-b361-b9f115abdd57 req-6fe2a8ae-f2a7-417a-b417-74b615c3edc6 service nova] [instance: c66fe2e7-7af0-48fa-8509-fcc24c1a437b] Received event network-vif-deleted-5bb12b4b-dfcb-416e-8095-91002ba3ca7d {{(pid=61806) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1192.823910] env[61806]: INFO nova.compute.manager [req-5be7572c-9816-4398-b361-b9f115abdd57 req-6fe2a8ae-f2a7-417a-b417-74b615c3edc6 service nova] [instance: c66fe2e7-7af0-48fa-8509-fcc24c1a437b] Neutron deleted interface 5bb12b4b-dfcb-416e-8095-91002ba3ca7d; detaching it from the instance and deleting it from the info cache [ 1192.824075] env[61806]: DEBUG nova.network.neutron [req-5be7572c-9816-4398-b361-b9f115abdd57 req-6fe2a8ae-f2a7-417a-b417-74b615c3edc6 service nova] [instance: c66fe2e7-7af0-48fa-8509-fcc24c1a437b] Updating instance_info_cache with network_info: [] {{(pid=61806) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1193.023213] env[61806]: DEBUG oslo_service.periodic_task [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Running periodic task ComputeManager.update_available_resource {{(pid=61806) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1193.300159] env[61806]: DEBUG nova.network.neutron [-] [instance: c66fe2e7-7af0-48fa-8509-fcc24c1a437b] Updating instance_info_cache with network_info: [] {{(pid=61806) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1193.326649] env[61806]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-978713b8-16c9-444a-aba6-becb7e9530c1 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1193.336598] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9aedff89-0e73-44ad-bd2a-3dd3d6ffb000 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1193.360350] env[61806]: DEBUG nova.compute.manager [req-5be7572c-9816-4398-b361-b9f115abdd57 req-6fe2a8ae-f2a7-417a-b417-74b615c3edc6 service nova] [instance: c66fe2e7-7af0-48fa-8509-fcc24c1a437b] Detach interface failed, port_id=5bb12b4b-dfcb-416e-8095-91002ba3ca7d, reason: Instance c66fe2e7-7af0-48fa-8509-fcc24c1a437b could not be found. {{(pid=61806) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10963}} [ 1193.527704] env[61806]: DEBUG oslo_concurrency.lockutils [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1193.527998] env[61806]: DEBUG oslo_concurrency.lockutils [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 0.000s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1193.528201] env[61806]: DEBUG oslo_concurrency.lockutils [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1193.528356] env[61806]: DEBUG nova.compute.resource_tracker [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=61806) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 1193.529333] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-80677dcb-f98e-4218-bb13-f24143ee6355 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1193.537359] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fb7bad96-e8bf-4c8b-8a50-e834158d48f8 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1193.550946] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7efde558-19ba-4a68-9352-503ab94ea12b {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1193.557165] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0f56568a-4f0e-487e-9bc9-391930f0ad4f {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1193.586736] env[61806]: DEBUG nova.compute.resource_tracker [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=180942MB free_disk=138GB free_vcpus=48 pci_devices=None {{(pid=61806) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 1193.586917] env[61806]: DEBUG oslo_concurrency.lockutils [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1193.587146] env[61806]: DEBUG oslo_concurrency.lockutils [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 0.000s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1193.802998] env[61806]: INFO nova.compute.manager [-] [instance: c66fe2e7-7af0-48fa-8509-fcc24c1a437b] Took 1.47 seconds to deallocate network for instance. [ 1194.310122] env[61806]: DEBUG oslo_concurrency.lockutils [None req-5c62cd58-8080-43f6-b08a-0e362578d740 tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1194.441668] env[61806]: DEBUG oslo_concurrency.lockutils [None req-ae894a1e-25be-4869-843d-5c8733810fbb tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Acquiring lock "4f26445f-35d0-4f53-9014-ef3e2eb1cbdf" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1194.441917] env[61806]: DEBUG oslo_concurrency.lockutils [None req-ae894a1e-25be-4869-843d-5c8733810fbb tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Lock "4f26445f-35d0-4f53-9014-ef3e2eb1cbdf" acquired by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: waited 0.000s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1194.614196] env[61806]: DEBUG nova.compute.resource_tracker [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Instance c66fe2e7-7af0-48fa-8509-fcc24c1a437b actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61806) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1194.614408] env[61806]: DEBUG nova.compute.resource_tracker [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Instance 4f26445f-35d0-4f53-9014-ef3e2eb1cbdf actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61806) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1194.614569] env[61806]: DEBUG nova.compute.resource_tracker [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Instance 29210bcd-0fe1-4ed9-a459-abc8016a4255 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61806) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1194.614779] env[61806]: DEBUG nova.compute.resource_tracker [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Total usable vcpus: 48, total allocated vcpus: 3 {{(pid=61806) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 1194.614993] env[61806]: DEBUG nova.compute.resource_tracker [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=1088MB phys_disk=200GB used_disk=3GB total_vcpus=48 used_vcpus=3 pci_stats=[] {{(pid=61806) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 1194.671952] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-683165e9-760f-4b83-9fa7-d4ef9740c98b {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1194.680359] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7d704f55-88da-4bef-8ea5-ce883344b551 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1194.711876] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e79bb06a-e513-4df2-802b-243f916cf09e {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1194.718783] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-36dff0f6-d7fd-488f-a920-8edeeeac2ce1 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1194.731777] env[61806]: DEBUG nova.compute.provider_tree [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Inventory has not changed in ProviderTree for provider: a2858be1-fd22-4e08-979e-87ad25293407 {{(pid=61806) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1194.945569] env[61806]: DEBUG nova.compute.utils [None req-ae894a1e-25be-4869-843d-5c8733810fbb tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Using /dev/sd instead of None {{(pid=61806) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1195.234448] env[61806]: DEBUG nova.scheduler.client.report [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Inventory has not changed for provider a2858be1-fd22-4e08-979e-87ad25293407 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 138, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61806) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1195.448358] env[61806]: DEBUG oslo_concurrency.lockutils [None req-ae894a1e-25be-4869-843d-5c8733810fbb tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Lock "4f26445f-35d0-4f53-9014-ef3e2eb1cbdf" "released" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: held 1.006s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1195.742429] env[61806]: DEBUG nova.compute.resource_tracker [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=61806) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 1195.742429] env[61806]: DEBUG oslo_concurrency.lockutils [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 2.155s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1195.742571] env[61806]: DEBUG oslo_concurrency.lockutils [None req-5c62cd58-8080-43f6-b08a-0e362578d740 tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 1.433s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1195.742672] env[61806]: DEBUG nova.objects.instance [None req-5c62cd58-8080-43f6-b08a-0e362578d740 tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Lazy-loading 'resources' on Instance uuid c66fe2e7-7af0-48fa-8509-fcc24c1a437b {{(pid=61806) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1196.299491] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ed9ab874-f9e6-4289-8948-0e2a71c9d72c {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1196.306597] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-da5e92b1-144c-4525-8bd0-0e5aca5c2938 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1196.336010] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5a777456-5714-4cf3-b1cc-482877873156 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1196.342506] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6ee96576-d854-4892-a85c-932351924d1c {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1196.354726] env[61806]: DEBUG nova.compute.provider_tree [None req-5c62cd58-8080-43f6-b08a-0e362578d740 tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Inventory has not changed in ProviderTree for provider: a2858be1-fd22-4e08-979e-87ad25293407 {{(pid=61806) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1196.508467] env[61806]: DEBUG oslo_concurrency.lockutils [None req-ae894a1e-25be-4869-843d-5c8733810fbb tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Acquiring lock "4f26445f-35d0-4f53-9014-ef3e2eb1cbdf" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1196.508804] env[61806]: DEBUG oslo_concurrency.lockutils [None req-ae894a1e-25be-4869-843d-5c8733810fbb tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Lock "4f26445f-35d0-4f53-9014-ef3e2eb1cbdf" acquired by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: waited 0.000s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1196.508899] env[61806]: INFO nova.compute.manager [None req-ae894a1e-25be-4869-843d-5c8733810fbb tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] [instance: 4f26445f-35d0-4f53-9014-ef3e2eb1cbdf] Attaching volume ba848344-b950-4c4f-ba25-0fd09d2bdd70 to /dev/sdb [ 1196.538284] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-97b81679-cbd0-487d-9cb6-34dcb007b92f {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1196.544848] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4fa88f44-1f01-44fa-b6c4-58d3099ffa8e {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1196.557873] env[61806]: DEBUG nova.virt.block_device [None req-ae894a1e-25be-4869-843d-5c8733810fbb tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] [instance: 4f26445f-35d0-4f53-9014-ef3e2eb1cbdf] Updating existing volume attachment record: cb623ac5-ea1d-491e-b2b4-f102594c2d80 {{(pid=61806) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 1196.743881] env[61806]: DEBUG oslo_service.periodic_task [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=61806) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1196.744134] env[61806]: DEBUG oslo_service.periodic_task [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=61806) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1196.857545] env[61806]: DEBUG nova.scheduler.client.report [None req-5c62cd58-8080-43f6-b08a-0e362578d740 tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Inventory has not changed for provider a2858be1-fd22-4e08-979e-87ad25293407 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 138, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61806) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1197.022528] env[61806]: DEBUG oslo_service.periodic_task [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=61806) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1197.362875] env[61806]: DEBUG oslo_concurrency.lockutils [None req-5c62cd58-8080-43f6-b08a-0e362578d740 tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.620s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1197.382082] env[61806]: INFO nova.scheduler.client.report [None req-5c62cd58-8080-43f6-b08a-0e362578d740 tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Deleted allocations for instance c66fe2e7-7af0-48fa-8509-fcc24c1a437b [ 1197.891032] env[61806]: DEBUG oslo_concurrency.lockutils [None req-5c62cd58-8080-43f6-b08a-0e362578d740 tempest-ServerActionsTestJSON-910105209 tempest-ServerActionsTestJSON-910105209-project-member] Lock "c66fe2e7-7af0-48fa-8509-fcc24c1a437b" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 7.115s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1198.023056] env[61806]: DEBUG oslo_service.periodic_task [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=61806) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1198.023186] env[61806]: DEBUG oslo_service.periodic_task [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=61806) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1198.023238] env[61806]: DEBUG nova.compute.manager [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=61806) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10545}} [ 1199.019957] env[61806]: DEBUG oslo_service.periodic_task [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Running periodic task ComputeManager._sync_scheduler_instance_info {{(pid=61806) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1200.023145] env[61806]: DEBUG oslo_service.periodic_task [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=61806) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1201.099925] env[61806]: DEBUG nova.virt.vmwareapi.volumeops [None req-ae894a1e-25be-4869-843d-5c8733810fbb tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] [instance: 4f26445f-35d0-4f53-9014-ef3e2eb1cbdf] Volume attach. Driver type: vmdk {{(pid=61806) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 1201.100210] env[61806]: DEBUG nova.virt.vmwareapi.volumeops [None req-ae894a1e-25be-4869-843d-5c8733810fbb tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] [instance: 4f26445f-35d0-4f53-9014-ef3e2eb1cbdf] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-277773', 'volume_id': 'ba848344-b950-4c4f-ba25-0fd09d2bdd70', 'name': 'volume-ba848344-b950-4c4f-ba25-0fd09d2bdd70', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '4f26445f-35d0-4f53-9014-ef3e2eb1cbdf', 'attached_at': '', 'detached_at': '', 'volume_id': 'ba848344-b950-4c4f-ba25-0fd09d2bdd70', 'serial': 'ba848344-b950-4c4f-ba25-0fd09d2bdd70'} {{(pid=61806) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 1201.101204] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f0db5987-36ab-4380-9b6a-9e1f6fd4757f {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1201.117108] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c8aecd7b-aa5e-4f50-bc33-b9ce398441c1 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1201.141706] env[61806]: DEBUG nova.virt.vmwareapi.volumeops [None req-ae894a1e-25be-4869-843d-5c8733810fbb tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] [instance: 4f26445f-35d0-4f53-9014-ef3e2eb1cbdf] Reconfiguring VM instance instance-0000006c to attach disk [datastore1] volume-ba848344-b950-4c4f-ba25-0fd09d2bdd70/volume-ba848344-b950-4c4f-ba25-0fd09d2bdd70.vmdk or device None with type thin {{(pid=61806) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1201.141948] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-c7e73d19-cae9-4ca8-966c-f520f1e02723 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1201.158675] env[61806]: DEBUG oslo_vmware.api [None req-ae894a1e-25be-4869-843d-5c8733810fbb tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Waiting for the task: (returnval){ [ 1201.158675] env[61806]: value = "task-1295307" [ 1201.158675] env[61806]: _type = "Task" [ 1201.158675] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1201.165696] env[61806]: DEBUG oslo_vmware.api [None req-ae894a1e-25be-4869-843d-5c8733810fbb tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Task: {'id': task-1295307, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1201.668020] env[61806]: DEBUG oslo_vmware.api [None req-ae894a1e-25be-4869-843d-5c8733810fbb tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Task: {'id': task-1295307, 'name': ReconfigVM_Task, 'duration_secs': 0.33188} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1201.668325] env[61806]: DEBUG nova.virt.vmwareapi.volumeops [None req-ae894a1e-25be-4869-843d-5c8733810fbb tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] [instance: 4f26445f-35d0-4f53-9014-ef3e2eb1cbdf] Reconfigured VM instance instance-0000006c to attach disk [datastore1] volume-ba848344-b950-4c4f-ba25-0fd09d2bdd70/volume-ba848344-b950-4c4f-ba25-0fd09d2bdd70.vmdk or device None with type thin {{(pid=61806) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1201.672948] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-6763b6fe-4981-4bc4-a56d-970f056efdf4 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1201.687123] env[61806]: DEBUG oslo_vmware.api [None req-ae894a1e-25be-4869-843d-5c8733810fbb tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Waiting for the task: (returnval){ [ 1201.687123] env[61806]: value = "task-1295308" [ 1201.687123] env[61806]: _type = "Task" [ 1201.687123] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1201.694948] env[61806]: DEBUG oslo_vmware.api [None req-ae894a1e-25be-4869-843d-5c8733810fbb tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Task: {'id': task-1295308, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1202.197593] env[61806]: DEBUG oslo_vmware.api [None req-ae894a1e-25be-4869-843d-5c8733810fbb tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Task: {'id': task-1295308, 'name': ReconfigVM_Task, 'duration_secs': 0.136063} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1202.198070] env[61806]: DEBUG nova.virt.vmwareapi.volumeops [None req-ae894a1e-25be-4869-843d-5c8733810fbb tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] [instance: 4f26445f-35d0-4f53-9014-ef3e2eb1cbdf] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-277773', 'volume_id': 'ba848344-b950-4c4f-ba25-0fd09d2bdd70', 'name': 'volume-ba848344-b950-4c4f-ba25-0fd09d2bdd70', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '4f26445f-35d0-4f53-9014-ef3e2eb1cbdf', 'attached_at': '', 'detached_at': '', 'volume_id': 'ba848344-b950-4c4f-ba25-0fd09d2bdd70', 'serial': 'ba848344-b950-4c4f-ba25-0fd09d2bdd70'} {{(pid=61806) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 1203.233395] env[61806]: DEBUG nova.objects.instance [None req-ae894a1e-25be-4869-843d-5c8733810fbb tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Lazy-loading 'flavor' on Instance uuid 4f26445f-35d0-4f53-9014-ef3e2eb1cbdf {{(pid=61806) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1203.738849] env[61806]: DEBUG oslo_concurrency.lockutils [None req-ae894a1e-25be-4869-843d-5c8733810fbb tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Lock "4f26445f-35d0-4f53-9014-ef3e2eb1cbdf" "released" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: held 7.230s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1204.699771] env[61806]: DEBUG nova.compute.manager [None req-f0ae362a-a43f-4e32-a197-e9215d619ba1 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] [instance: 4f26445f-35d0-4f53-9014-ef3e2eb1cbdf] Stashing vm_state: active {{(pid=61806) _prep_resize /opt/stack/nova/nova/compute/manager.py:5624}} [ 1205.218230] env[61806]: DEBUG oslo_concurrency.lockutils [None req-f0ae362a-a43f-4e32-a197-e9215d619ba1 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1205.218496] env[61806]: DEBUG oslo_concurrency.lockutils [None req-f0ae362a-a43f-4e32-a197-e9215d619ba1 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: waited 0.000s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1205.724513] env[61806]: INFO nova.compute.claims [None req-f0ae362a-a43f-4e32-a197-e9215d619ba1 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] [instance: 4f26445f-35d0-4f53-9014-ef3e2eb1cbdf] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1206.687463] env[61806]: INFO nova.compute.resource_tracker [None req-f0ae362a-a43f-4e32-a197-e9215d619ba1 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] [instance: 4f26445f-35d0-4f53-9014-ef3e2eb1cbdf] Updating resource usage from migration bf3eae4b-1865-4dba-9161-9d137e96b682 [ 1206.732764] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d0ac8515-2a48-4df1-bd8d-4e235701ead3 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1206.740129] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7d897d64-0048-4e6d-ac26-cefe625ad02b {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1206.768563] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1c002204-2b97-4fc9-8e0e-968d1f595b88 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1206.774939] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7e73f622-f0d0-4a9a-8488-0c98f83e5317 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1206.787442] env[61806]: DEBUG nova.compute.provider_tree [None req-f0ae362a-a43f-4e32-a197-e9215d619ba1 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Inventory has not changed in ProviderTree for provider: a2858be1-fd22-4e08-979e-87ad25293407 {{(pid=61806) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1207.290186] env[61806]: DEBUG nova.scheduler.client.report [None req-f0ae362a-a43f-4e32-a197-e9215d619ba1 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Inventory has not changed for provider a2858be1-fd22-4e08-979e-87ad25293407 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 138, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61806) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1207.794748] env[61806]: DEBUG oslo_concurrency.lockutils [None req-f0ae362a-a43f-4e32-a197-e9215d619ba1 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: held 2.576s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1207.795117] env[61806]: INFO nova.compute.manager [None req-f0ae362a-a43f-4e32-a197-e9215d619ba1 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] [instance: 4f26445f-35d0-4f53-9014-ef3e2eb1cbdf] Migrating [ 1208.309906] env[61806]: DEBUG oslo_concurrency.lockutils [None req-f0ae362a-a43f-4e32-a197-e9215d619ba1 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Acquiring lock "refresh_cache-4f26445f-35d0-4f53-9014-ef3e2eb1cbdf" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1208.310153] env[61806]: DEBUG oslo_concurrency.lockutils [None req-f0ae362a-a43f-4e32-a197-e9215d619ba1 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Acquired lock "refresh_cache-4f26445f-35d0-4f53-9014-ef3e2eb1cbdf" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1208.310392] env[61806]: DEBUG nova.network.neutron [None req-f0ae362a-a43f-4e32-a197-e9215d619ba1 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] [instance: 4f26445f-35d0-4f53-9014-ef3e2eb1cbdf] Building network info cache for instance {{(pid=61806) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1209.020428] env[61806]: DEBUG nova.network.neutron [None req-f0ae362a-a43f-4e32-a197-e9215d619ba1 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] [instance: 4f26445f-35d0-4f53-9014-ef3e2eb1cbdf] Updating instance_info_cache with network_info: [{"id": "a71b2870-bc50-47f4-840d-57e956c3f86d", "address": "fa:16:3e:74:22:b8", "network": {"id": "33bf5b21-82b0-432e-ae3d-9c2ca75161d0", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-1294159336-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.187", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "1c55cc0455584fbb8cda88d33b8f8b7e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "5b0fa7a2-ebd9-4788-8904-7bf250ce466c", "external-id": "nsx-vlan-transportzone-669", "segmentation_id": 669, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa71b2870-bc", "ovs_interfaceid": "a71b2870-bc50-47f4-840d-57e956c3f86d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61806) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1209.523822] env[61806]: DEBUG oslo_concurrency.lockutils [None req-f0ae362a-a43f-4e32-a197-e9215d619ba1 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Releasing lock "refresh_cache-4f26445f-35d0-4f53-9014-ef3e2eb1cbdf" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1211.049055] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e978f011-3516-4f50-8b57-ded3372dad1d {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1211.070762] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-f0ae362a-a43f-4e32-a197-e9215d619ba1 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] [instance: 4f26445f-35d0-4f53-9014-ef3e2eb1cbdf] Updating instance '4f26445f-35d0-4f53-9014-ef3e2eb1cbdf' progress to 0 {{(pid=61806) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 1211.578120] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [None req-f0ae362a-a43f-4e32-a197-e9215d619ba1 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] [instance: 4f26445f-35d0-4f53-9014-ef3e2eb1cbdf] Powering off the VM {{(pid=61806) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1211.578120] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-a4be96ed-8cf4-469f-b689-8bddef432afa {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1211.585469] env[61806]: DEBUG oslo_vmware.api [None req-f0ae362a-a43f-4e32-a197-e9215d619ba1 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Waiting for the task: (returnval){ [ 1211.585469] env[61806]: value = "task-1295309" [ 1211.585469] env[61806]: _type = "Task" [ 1211.585469] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1211.592905] env[61806]: DEBUG oslo_vmware.api [None req-f0ae362a-a43f-4e32-a197-e9215d619ba1 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Task: {'id': task-1295309, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1212.095694] env[61806]: DEBUG oslo_vmware.api [None req-f0ae362a-a43f-4e32-a197-e9215d619ba1 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Task: {'id': task-1295309, 'name': PowerOffVM_Task, 'duration_secs': 0.211606} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1212.096188] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [None req-f0ae362a-a43f-4e32-a197-e9215d619ba1 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] [instance: 4f26445f-35d0-4f53-9014-ef3e2eb1cbdf] Powered off the VM {{(pid=61806) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1212.096188] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-f0ae362a-a43f-4e32-a197-e9215d619ba1 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] [instance: 4f26445f-35d0-4f53-9014-ef3e2eb1cbdf] Updating instance '4f26445f-35d0-4f53-9014-ef3e2eb1cbdf' progress to 17 {{(pid=61806) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 1212.602534] env[61806]: DEBUG nova.virt.hardware [None req-f0ae362a-a43f-4e32-a197-e9215d619ba1 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-15T18:05:12Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='84',id=12,is_public=True,memory_mb=256,name='m1.micro',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=61806) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1212.602805] env[61806]: DEBUG nova.virt.hardware [None req-f0ae362a-a43f-4e32-a197-e9215d619ba1 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Flavor limits 0:0:0 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1212.602972] env[61806]: DEBUG nova.virt.hardware [None req-f0ae362a-a43f-4e32-a197-e9215d619ba1 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Image limits 0:0:0 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1212.603185] env[61806]: DEBUG nova.virt.hardware [None req-f0ae362a-a43f-4e32-a197-e9215d619ba1 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Flavor pref 0:0:0 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1212.603372] env[61806]: DEBUG nova.virt.hardware [None req-f0ae362a-a43f-4e32-a197-e9215d619ba1 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Image pref 0:0:0 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1212.603530] env[61806]: DEBUG nova.virt.hardware [None req-f0ae362a-a43f-4e32-a197-e9215d619ba1 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1212.603741] env[61806]: DEBUG nova.virt.hardware [None req-f0ae362a-a43f-4e32-a197-e9215d619ba1 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61806) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1212.603908] env[61806]: DEBUG nova.virt.hardware [None req-f0ae362a-a43f-4e32-a197-e9215d619ba1 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61806) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1212.604104] env[61806]: DEBUG nova.virt.hardware [None req-f0ae362a-a43f-4e32-a197-e9215d619ba1 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Got 1 possible topologies {{(pid=61806) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1212.604279] env[61806]: DEBUG nova.virt.hardware [None req-f0ae362a-a43f-4e32-a197-e9215d619ba1 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61806) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1212.604460] env[61806]: DEBUG nova.virt.hardware [None req-f0ae362a-a43f-4e32-a197-e9215d619ba1 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61806) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1212.609678] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-f5a79c1a-b258-45ad-9b5d-fd4cacc5ba23 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1212.626202] env[61806]: DEBUG oslo_vmware.api [None req-f0ae362a-a43f-4e32-a197-e9215d619ba1 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Waiting for the task: (returnval){ [ 1212.626202] env[61806]: value = "task-1295310" [ 1212.626202] env[61806]: _type = "Task" [ 1212.626202] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1212.633883] env[61806]: DEBUG oslo_vmware.api [None req-f0ae362a-a43f-4e32-a197-e9215d619ba1 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Task: {'id': task-1295310, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1213.135329] env[61806]: DEBUG oslo_vmware.api [None req-f0ae362a-a43f-4e32-a197-e9215d619ba1 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Task: {'id': task-1295310, 'name': ReconfigVM_Task, 'duration_secs': 0.160862} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1213.135715] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-f0ae362a-a43f-4e32-a197-e9215d619ba1 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] [instance: 4f26445f-35d0-4f53-9014-ef3e2eb1cbdf] Updating instance '4f26445f-35d0-4f53-9014-ef3e2eb1cbdf' progress to 33 {{(pid=61806) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 1213.641912] env[61806]: DEBUG nova.virt.hardware [None req-f0ae362a-a43f-4e32-a197-e9215d619ba1 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-15T18:05:10Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=61806) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1213.642206] env[61806]: DEBUG nova.virt.hardware [None req-f0ae362a-a43f-4e32-a197-e9215d619ba1 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Flavor limits 0:0:0 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1213.642389] env[61806]: DEBUG nova.virt.hardware [None req-f0ae362a-a43f-4e32-a197-e9215d619ba1 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Image limits 0:0:0 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1213.642581] env[61806]: DEBUG nova.virt.hardware [None req-f0ae362a-a43f-4e32-a197-e9215d619ba1 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Flavor pref 0:0:0 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1213.642732] env[61806]: DEBUG nova.virt.hardware [None req-f0ae362a-a43f-4e32-a197-e9215d619ba1 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Image pref 0:0:0 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1213.642882] env[61806]: DEBUG nova.virt.hardware [None req-f0ae362a-a43f-4e32-a197-e9215d619ba1 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1213.643104] env[61806]: DEBUG nova.virt.hardware [None req-f0ae362a-a43f-4e32-a197-e9215d619ba1 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61806) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1213.643277] env[61806]: DEBUG nova.virt.hardware [None req-f0ae362a-a43f-4e32-a197-e9215d619ba1 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61806) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1213.643448] env[61806]: DEBUG nova.virt.hardware [None req-f0ae362a-a43f-4e32-a197-e9215d619ba1 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Got 1 possible topologies {{(pid=61806) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1213.643617] env[61806]: DEBUG nova.virt.hardware [None req-f0ae362a-a43f-4e32-a197-e9215d619ba1 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61806) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1213.643791] env[61806]: DEBUG nova.virt.hardware [None req-f0ae362a-a43f-4e32-a197-e9215d619ba1 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61806) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1213.649164] env[61806]: DEBUG nova.virt.vmwareapi.volumeops [None req-f0ae362a-a43f-4e32-a197-e9215d619ba1 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] [instance: 4f26445f-35d0-4f53-9014-ef3e2eb1cbdf] Reconfiguring VM instance instance-0000006c to detach disk 2000 {{(pid=61806) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1213.649457] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-1d673ea9-fa23-4f1a-913e-67c47e2dff12 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1213.668651] env[61806]: DEBUG oslo_vmware.api [None req-f0ae362a-a43f-4e32-a197-e9215d619ba1 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Waiting for the task: (returnval){ [ 1213.668651] env[61806]: value = "task-1295311" [ 1213.668651] env[61806]: _type = "Task" [ 1213.668651] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1213.676017] env[61806]: DEBUG oslo_vmware.api [None req-f0ae362a-a43f-4e32-a197-e9215d619ba1 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Task: {'id': task-1295311, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1214.179046] env[61806]: DEBUG oslo_vmware.api [None req-f0ae362a-a43f-4e32-a197-e9215d619ba1 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Task: {'id': task-1295311, 'name': ReconfigVM_Task, 'duration_secs': 0.178705} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1214.179394] env[61806]: DEBUG nova.virt.vmwareapi.volumeops [None req-f0ae362a-a43f-4e32-a197-e9215d619ba1 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] [instance: 4f26445f-35d0-4f53-9014-ef3e2eb1cbdf] Reconfigured VM instance instance-0000006c to detach disk 2000 {{(pid=61806) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1214.180100] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-03fb43d2-d66b-4659-bb99-371e38f76fa9 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1214.203215] env[61806]: DEBUG nova.virt.vmwareapi.volumeops [None req-f0ae362a-a43f-4e32-a197-e9215d619ba1 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] [instance: 4f26445f-35d0-4f53-9014-ef3e2eb1cbdf] Reconfiguring VM instance instance-0000006c to attach disk [datastore1] 4f26445f-35d0-4f53-9014-ef3e2eb1cbdf/4f26445f-35d0-4f53-9014-ef3e2eb1cbdf.vmdk or device None with type thin {{(pid=61806) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1214.203460] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-474a1bb0-511a-41bb-810f-ed0798aeb22c {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1214.220288] env[61806]: DEBUG oslo_vmware.api [None req-f0ae362a-a43f-4e32-a197-e9215d619ba1 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Waiting for the task: (returnval){ [ 1214.220288] env[61806]: value = "task-1295312" [ 1214.220288] env[61806]: _type = "Task" [ 1214.220288] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1214.227289] env[61806]: DEBUG oslo_vmware.api [None req-f0ae362a-a43f-4e32-a197-e9215d619ba1 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Task: {'id': task-1295312, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1214.407382] env[61806]: DEBUG oslo_concurrency.lockutils [None req-295351ed-4dc9-4e64-b701-ee00f0fc6686 tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] Acquiring lock "29210bcd-0fe1-4ed9-a459-abc8016a4255" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1214.407662] env[61806]: DEBUG oslo_concurrency.lockutils [None req-295351ed-4dc9-4e64-b701-ee00f0fc6686 tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] Lock "29210bcd-0fe1-4ed9-a459-abc8016a4255" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1214.407884] env[61806]: DEBUG oslo_concurrency.lockutils [None req-295351ed-4dc9-4e64-b701-ee00f0fc6686 tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] Acquiring lock "29210bcd-0fe1-4ed9-a459-abc8016a4255-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1214.408090] env[61806]: DEBUG oslo_concurrency.lockutils [None req-295351ed-4dc9-4e64-b701-ee00f0fc6686 tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] Lock "29210bcd-0fe1-4ed9-a459-abc8016a4255-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1214.408268] env[61806]: DEBUG oslo_concurrency.lockutils [None req-295351ed-4dc9-4e64-b701-ee00f0fc6686 tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] Lock "29210bcd-0fe1-4ed9-a459-abc8016a4255-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1214.410530] env[61806]: INFO nova.compute.manager [None req-295351ed-4dc9-4e64-b701-ee00f0fc6686 tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] [instance: 29210bcd-0fe1-4ed9-a459-abc8016a4255] Terminating instance [ 1214.412306] env[61806]: DEBUG nova.compute.manager [None req-295351ed-4dc9-4e64-b701-ee00f0fc6686 tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] [instance: 29210bcd-0fe1-4ed9-a459-abc8016a4255] Start destroying the instance on the hypervisor. {{(pid=61806) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1214.412518] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-295351ed-4dc9-4e64-b701-ee00f0fc6686 tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] [instance: 29210bcd-0fe1-4ed9-a459-abc8016a4255] Destroying instance {{(pid=61806) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1214.413329] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-514473f4-28f3-4d2c-9691-5ac4d7ec79da {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1214.420359] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [None req-295351ed-4dc9-4e64-b701-ee00f0fc6686 tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] [instance: 29210bcd-0fe1-4ed9-a459-abc8016a4255] Powering off the VM {{(pid=61806) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1214.420580] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-35f9245b-bdfe-424b-8b2a-9e7f8a847b05 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1214.427071] env[61806]: DEBUG oslo_vmware.api [None req-295351ed-4dc9-4e64-b701-ee00f0fc6686 tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] Waiting for the task: (returnval){ [ 1214.427071] env[61806]: value = "task-1295313" [ 1214.427071] env[61806]: _type = "Task" [ 1214.427071] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1214.434238] env[61806]: DEBUG oslo_vmware.api [None req-295351ed-4dc9-4e64-b701-ee00f0fc6686 tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] Task: {'id': task-1295313, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1214.731474] env[61806]: DEBUG oslo_vmware.api [None req-f0ae362a-a43f-4e32-a197-e9215d619ba1 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Task: {'id': task-1295312, 'name': ReconfigVM_Task, 'duration_secs': 0.289994} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1214.731701] env[61806]: DEBUG nova.virt.vmwareapi.volumeops [None req-f0ae362a-a43f-4e32-a197-e9215d619ba1 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] [instance: 4f26445f-35d0-4f53-9014-ef3e2eb1cbdf] Reconfigured VM instance instance-0000006c to attach disk [datastore1] 4f26445f-35d0-4f53-9014-ef3e2eb1cbdf/4f26445f-35d0-4f53-9014-ef3e2eb1cbdf.vmdk or device None with type thin {{(pid=61806) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1214.731978] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-f0ae362a-a43f-4e32-a197-e9215d619ba1 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] [instance: 4f26445f-35d0-4f53-9014-ef3e2eb1cbdf] Updating instance '4f26445f-35d0-4f53-9014-ef3e2eb1cbdf' progress to 50 {{(pid=61806) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 1214.937445] env[61806]: DEBUG oslo_vmware.api [None req-295351ed-4dc9-4e64-b701-ee00f0fc6686 tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] Task: {'id': task-1295313, 'name': PowerOffVM_Task, 'duration_secs': 0.147794} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1214.937743] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [None req-295351ed-4dc9-4e64-b701-ee00f0fc6686 tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] [instance: 29210bcd-0fe1-4ed9-a459-abc8016a4255] Powered off the VM {{(pid=61806) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1214.937923] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-295351ed-4dc9-4e64-b701-ee00f0fc6686 tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] [instance: 29210bcd-0fe1-4ed9-a459-abc8016a4255] Unregistering the VM {{(pid=61806) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1214.938231] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-4c3c71c5-6775-483b-a03f-e9cbcb4ea2ca {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1215.001309] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-295351ed-4dc9-4e64-b701-ee00f0fc6686 tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] [instance: 29210bcd-0fe1-4ed9-a459-abc8016a4255] Unregistered the VM {{(pid=61806) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1215.001535] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-295351ed-4dc9-4e64-b701-ee00f0fc6686 tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] [instance: 29210bcd-0fe1-4ed9-a459-abc8016a4255] Deleting contents of the VM from datastore datastore2 {{(pid=61806) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1215.001700] env[61806]: DEBUG nova.virt.vmwareapi.ds_util [None req-295351ed-4dc9-4e64-b701-ee00f0fc6686 tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] Deleting the datastore file [datastore2] 29210bcd-0fe1-4ed9-a459-abc8016a4255 {{(pid=61806) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1215.001984] env[61806]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-e3849daa-ee19-43b8-8105-f14ecefc6d0f {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1215.010849] env[61806]: DEBUG oslo_vmware.api [None req-295351ed-4dc9-4e64-b701-ee00f0fc6686 tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] Waiting for the task: (returnval){ [ 1215.010849] env[61806]: value = "task-1295315" [ 1215.010849] env[61806]: _type = "Task" [ 1215.010849] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1215.019524] env[61806]: DEBUG oslo_vmware.api [None req-295351ed-4dc9-4e64-b701-ee00f0fc6686 tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] Task: {'id': task-1295315, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1215.238915] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c54d4ec3-0cbd-4978-8d36-ade182689632 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1215.259255] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ceb64488-0a0f-4ef5-8de5-cad6b21cb399 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1215.277901] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-f0ae362a-a43f-4e32-a197-e9215d619ba1 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] [instance: 4f26445f-35d0-4f53-9014-ef3e2eb1cbdf] Updating instance '4f26445f-35d0-4f53-9014-ef3e2eb1cbdf' progress to 67 {{(pid=61806) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 1215.519949] env[61806]: DEBUG oslo_vmware.api [None req-295351ed-4dc9-4e64-b701-ee00f0fc6686 tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] Task: {'id': task-1295315, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.119428} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1215.520195] env[61806]: DEBUG nova.virt.vmwareapi.ds_util [None req-295351ed-4dc9-4e64-b701-ee00f0fc6686 tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] Deleted the datastore file {{(pid=61806) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1215.520410] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-295351ed-4dc9-4e64-b701-ee00f0fc6686 tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] [instance: 29210bcd-0fe1-4ed9-a459-abc8016a4255] Deleted contents of the VM from datastore datastore2 {{(pid=61806) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1215.520596] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-295351ed-4dc9-4e64-b701-ee00f0fc6686 tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] [instance: 29210bcd-0fe1-4ed9-a459-abc8016a4255] Instance destroyed {{(pid=61806) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1215.520772] env[61806]: INFO nova.compute.manager [None req-295351ed-4dc9-4e64-b701-ee00f0fc6686 tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] [instance: 29210bcd-0fe1-4ed9-a459-abc8016a4255] Took 1.11 seconds to destroy the instance on the hypervisor. [ 1215.521019] env[61806]: DEBUG oslo.service.loopingcall [None req-295351ed-4dc9-4e64-b701-ee00f0fc6686 tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61806) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1215.521216] env[61806]: DEBUG nova.compute.manager [-] [instance: 29210bcd-0fe1-4ed9-a459-abc8016a4255] Deallocating network for instance {{(pid=61806) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1215.521312] env[61806]: DEBUG nova.network.neutron [-] [instance: 29210bcd-0fe1-4ed9-a459-abc8016a4255] deallocate_for_instance() {{(pid=61806) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1216.017561] env[61806]: DEBUG nova.compute.manager [req-171f3931-215e-4041-ba12-7be57b7086f6 req-a0eac6b0-3d05-4256-bda1-6495be0c73bf service nova] [instance: 29210bcd-0fe1-4ed9-a459-abc8016a4255] Received event network-vif-deleted-8aad34b7-4309-44b6-b88d-ae5961da22c2 {{(pid=61806) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1216.017733] env[61806]: INFO nova.compute.manager [req-171f3931-215e-4041-ba12-7be57b7086f6 req-a0eac6b0-3d05-4256-bda1-6495be0c73bf service nova] [instance: 29210bcd-0fe1-4ed9-a459-abc8016a4255] Neutron deleted interface 8aad34b7-4309-44b6-b88d-ae5961da22c2; detaching it from the instance and deleting it from the info cache [ 1216.017912] env[61806]: DEBUG nova.network.neutron [req-171f3931-215e-4041-ba12-7be57b7086f6 req-a0eac6b0-3d05-4256-bda1-6495be0c73bf service nova] [instance: 29210bcd-0fe1-4ed9-a459-abc8016a4255] Updating instance_info_cache with network_info: [] {{(pid=61806) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1216.492417] env[61806]: DEBUG nova.network.neutron [-] [instance: 29210bcd-0fe1-4ed9-a459-abc8016a4255] Updating instance_info_cache with network_info: [] {{(pid=61806) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1216.520103] env[61806]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-4a2ec988-1f9c-486c-bfdf-b53057f7d683 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1216.533524] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-04e5618d-fc78-4024-a062-834ebf729038 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1216.564203] env[61806]: DEBUG nova.compute.manager [req-171f3931-215e-4041-ba12-7be57b7086f6 req-a0eac6b0-3d05-4256-bda1-6495be0c73bf service nova] [instance: 29210bcd-0fe1-4ed9-a459-abc8016a4255] Detach interface failed, port_id=8aad34b7-4309-44b6-b88d-ae5961da22c2, reason: Instance 29210bcd-0fe1-4ed9-a459-abc8016a4255 could not be found. {{(pid=61806) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10963}} [ 1216.913189] env[61806]: DEBUG nova.network.neutron [None req-f0ae362a-a43f-4e32-a197-e9215d619ba1 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] [instance: 4f26445f-35d0-4f53-9014-ef3e2eb1cbdf] Port a71b2870-bc50-47f4-840d-57e956c3f86d binding to destination host cpu-1 is already ACTIVE {{(pid=61806) migrate_instance_start /opt/stack/nova/nova/network/neutron.py:3171}} [ 1216.994582] env[61806]: INFO nova.compute.manager [-] [instance: 29210bcd-0fe1-4ed9-a459-abc8016a4255] Took 1.47 seconds to deallocate network for instance. [ 1217.500855] env[61806]: DEBUG oslo_concurrency.lockutils [None req-295351ed-4dc9-4e64-b701-ee00f0fc6686 tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1217.501256] env[61806]: DEBUG oslo_concurrency.lockutils [None req-295351ed-4dc9-4e64-b701-ee00f0fc6686 tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.000s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1217.501381] env[61806]: DEBUG nova.objects.instance [None req-295351ed-4dc9-4e64-b701-ee00f0fc6686 tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] Lazy-loading 'resources' on Instance uuid 29210bcd-0fe1-4ed9-a459-abc8016a4255 {{(pid=61806) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1217.934764] env[61806]: DEBUG oslo_concurrency.lockutils [None req-f0ae362a-a43f-4e32-a197-e9215d619ba1 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Acquiring lock "4f26445f-35d0-4f53-9014-ef3e2eb1cbdf-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1217.935029] env[61806]: DEBUG oslo_concurrency.lockutils [None req-f0ae362a-a43f-4e32-a197-e9215d619ba1 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Lock "4f26445f-35d0-4f53-9014-ef3e2eb1cbdf-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1217.935203] env[61806]: DEBUG oslo_concurrency.lockutils [None req-f0ae362a-a43f-4e32-a197-e9215d619ba1 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Lock "4f26445f-35d0-4f53-9014-ef3e2eb1cbdf-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1218.054475] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-daee0d65-4738-4ccb-a3db-6c33fd38ab2a {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1218.062088] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1d197cf1-028b-4fd9-b546-d2d6e75be144 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1218.092379] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eaff930b-aec1-45de-9e55-a6cb31cb5673 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1218.099161] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bce61345-2533-410a-8fa9-0f9cd2ac9ca0 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1218.111511] env[61806]: DEBUG nova.compute.provider_tree [None req-295351ed-4dc9-4e64-b701-ee00f0fc6686 tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] Inventory has not changed in ProviderTree for provider: a2858be1-fd22-4e08-979e-87ad25293407 {{(pid=61806) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1218.614964] env[61806]: DEBUG nova.scheduler.client.report [None req-295351ed-4dc9-4e64-b701-ee00f0fc6686 tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] Inventory has not changed for provider a2858be1-fd22-4e08-979e-87ad25293407 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 138, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61806) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1218.971295] env[61806]: DEBUG oslo_concurrency.lockutils [None req-f0ae362a-a43f-4e32-a197-e9215d619ba1 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Acquiring lock "refresh_cache-4f26445f-35d0-4f53-9014-ef3e2eb1cbdf" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1218.971515] env[61806]: DEBUG oslo_concurrency.lockutils [None req-f0ae362a-a43f-4e32-a197-e9215d619ba1 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Acquired lock "refresh_cache-4f26445f-35d0-4f53-9014-ef3e2eb1cbdf" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1218.971699] env[61806]: DEBUG nova.network.neutron [None req-f0ae362a-a43f-4e32-a197-e9215d619ba1 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] [instance: 4f26445f-35d0-4f53-9014-ef3e2eb1cbdf] Building network info cache for instance {{(pid=61806) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1219.120326] env[61806]: DEBUG oslo_concurrency.lockutils [None req-295351ed-4dc9-4e64-b701-ee00f0fc6686 tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.619s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1219.140470] env[61806]: INFO nova.scheduler.client.report [None req-295351ed-4dc9-4e64-b701-ee00f0fc6686 tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] Deleted allocations for instance 29210bcd-0fe1-4ed9-a459-abc8016a4255 [ 1219.649269] env[61806]: DEBUG oslo_concurrency.lockutils [None req-295351ed-4dc9-4e64-b701-ee00f0fc6686 tempest-AttachVolumeShelveTestJSON-1685583356 tempest-AttachVolumeShelveTestJSON-1685583356-project-member] Lock "29210bcd-0fe1-4ed9-a459-abc8016a4255" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.241s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1219.776216] env[61806]: DEBUG nova.network.neutron [None req-f0ae362a-a43f-4e32-a197-e9215d619ba1 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] [instance: 4f26445f-35d0-4f53-9014-ef3e2eb1cbdf] Updating instance_info_cache with network_info: [{"id": "a71b2870-bc50-47f4-840d-57e956c3f86d", "address": "fa:16:3e:74:22:b8", "network": {"id": "33bf5b21-82b0-432e-ae3d-9c2ca75161d0", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-1294159336-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.187", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "1c55cc0455584fbb8cda88d33b8f8b7e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "5b0fa7a2-ebd9-4788-8904-7bf250ce466c", "external-id": "nsx-vlan-transportzone-669", "segmentation_id": 669, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa71b2870-bc", "ovs_interfaceid": "a71b2870-bc50-47f4-840d-57e956c3f86d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61806) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1220.278311] env[61806]: DEBUG oslo_concurrency.lockutils [None req-f0ae362a-a43f-4e32-a197-e9215d619ba1 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Releasing lock "refresh_cache-4f26445f-35d0-4f53-9014-ef3e2eb1cbdf" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1220.787878] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-89bc37b5-35bf-45bd-81a1-baadecca6d40 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1220.794955] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c12df8ee-e69a-4ae8-a6f1-1d05dd9abf9d {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1221.892754] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3a63178e-d97d-4c47-a925-e3b927f6b6e9 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1221.914909] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0adf29dc-170a-4397-aef8-875d93361cd6 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1221.921618] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-f0ae362a-a43f-4e32-a197-e9215d619ba1 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] [instance: 4f26445f-35d0-4f53-9014-ef3e2eb1cbdf] Updating instance '4f26445f-35d0-4f53-9014-ef3e2eb1cbdf' progress to 83 {{(pid=61806) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 1222.427974] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [None req-f0ae362a-a43f-4e32-a197-e9215d619ba1 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] [instance: 4f26445f-35d0-4f53-9014-ef3e2eb1cbdf] Powering on the VM {{(pid=61806) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1222.428333] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-ee062bbd-4c50-4b4e-af83-6cb2c478a845 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1222.435572] env[61806]: DEBUG oslo_vmware.api [None req-f0ae362a-a43f-4e32-a197-e9215d619ba1 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Waiting for the task: (returnval){ [ 1222.435572] env[61806]: value = "task-1295317" [ 1222.435572] env[61806]: _type = "Task" [ 1222.435572] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1222.443061] env[61806]: DEBUG oslo_vmware.api [None req-f0ae362a-a43f-4e32-a197-e9215d619ba1 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Task: {'id': task-1295317, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1222.945435] env[61806]: DEBUG oslo_vmware.api [None req-f0ae362a-a43f-4e32-a197-e9215d619ba1 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Task: {'id': task-1295317, 'name': PowerOnVM_Task, 'duration_secs': 0.395091} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1222.945869] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [None req-f0ae362a-a43f-4e32-a197-e9215d619ba1 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] [instance: 4f26445f-35d0-4f53-9014-ef3e2eb1cbdf] Powered on the VM {{(pid=61806) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1222.945914] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-f0ae362a-a43f-4e32-a197-e9215d619ba1 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] [instance: 4f26445f-35d0-4f53-9014-ef3e2eb1cbdf] Updating instance '4f26445f-35d0-4f53-9014-ef3e2eb1cbdf' progress to 100 {{(pid=61806) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 1225.355415] env[61806]: DEBUG nova.network.neutron [None req-3aec3f2d-4ed5-4066-a721-5ba416343c9c tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] [instance: 4f26445f-35d0-4f53-9014-ef3e2eb1cbdf] Port a71b2870-bc50-47f4-840d-57e956c3f86d binding to destination host cpu-1 is already ACTIVE {{(pid=61806) migrate_instance_start /opt/stack/nova/nova/network/neutron.py:3171}} [ 1225.355725] env[61806]: DEBUG oslo_concurrency.lockutils [None req-3aec3f2d-4ed5-4066-a721-5ba416343c9c tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Acquiring lock "refresh_cache-4f26445f-35d0-4f53-9014-ef3e2eb1cbdf" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1225.355900] env[61806]: DEBUG oslo_concurrency.lockutils [None req-3aec3f2d-4ed5-4066-a721-5ba416343c9c tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Acquired lock "refresh_cache-4f26445f-35d0-4f53-9014-ef3e2eb1cbdf" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1225.356122] env[61806]: DEBUG nova.network.neutron [None req-3aec3f2d-4ed5-4066-a721-5ba416343c9c tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] [instance: 4f26445f-35d0-4f53-9014-ef3e2eb1cbdf] Building network info cache for instance {{(pid=61806) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1226.060144] env[61806]: DEBUG nova.network.neutron [None req-3aec3f2d-4ed5-4066-a721-5ba416343c9c tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] [instance: 4f26445f-35d0-4f53-9014-ef3e2eb1cbdf] Updating instance_info_cache with network_info: [{"id": "a71b2870-bc50-47f4-840d-57e956c3f86d", "address": "fa:16:3e:74:22:b8", "network": {"id": "33bf5b21-82b0-432e-ae3d-9c2ca75161d0", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-1294159336-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.187", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "1c55cc0455584fbb8cda88d33b8f8b7e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "5b0fa7a2-ebd9-4788-8904-7bf250ce466c", "external-id": "nsx-vlan-transportzone-669", "segmentation_id": 669, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa71b2870-bc", "ovs_interfaceid": "a71b2870-bc50-47f4-840d-57e956c3f86d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61806) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1226.562802] env[61806]: DEBUG oslo_concurrency.lockutils [None req-3aec3f2d-4ed5-4066-a721-5ba416343c9c tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Releasing lock "refresh_cache-4f26445f-35d0-4f53-9014-ef3e2eb1cbdf" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1227.066494] env[61806]: DEBUG nova.compute.manager [None req-3aec3f2d-4ed5-4066-a721-5ba416343c9c tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] [instance: 4f26445f-35d0-4f53-9014-ef3e2eb1cbdf] Hypervisor driver does not support instance shared storage check, assuming it's not on shared storage {{(pid=61806) _is_instance_storage_shared /opt/stack/nova/nova/compute/manager.py:897}} [ 1228.160138] env[61806]: DEBUG oslo_concurrency.lockutils [None req-3aec3f2d-4ed5-4066-a721-5ba416343c9c tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_dest" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1228.160439] env[61806]: DEBUG oslo_concurrency.lockutils [None req-3aec3f2d-4ed5-4066-a721-5ba416343c9c tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_dest" :: waited 0.001s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1228.663232] env[61806]: DEBUG nova.objects.instance [None req-3aec3f2d-4ed5-4066-a721-5ba416343c9c tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Lazy-loading 'migration_context' on Instance uuid 4f26445f-35d0-4f53-9014-ef3e2eb1cbdf {{(pid=61806) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1229.208049] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-52cbfada-957e-4d46-ba84-0b61b33917e7 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1229.215794] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d7ca1aaa-b378-4936-abcc-36f86a3aa8ee {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1229.245025] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0406335b-9cc3-44dd-8ef9-64bed2e29623 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1229.251384] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9436b879-2b2e-474f-8508-29d6120adf0a {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1229.264784] env[61806]: DEBUG nova.compute.provider_tree [None req-3aec3f2d-4ed5-4066-a721-5ba416343c9c tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Inventory has not changed in ProviderTree for provider: a2858be1-fd22-4e08-979e-87ad25293407 {{(pid=61806) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1229.767859] env[61806]: DEBUG nova.scheduler.client.report [None req-3aec3f2d-4ed5-4066-a721-5ba416343c9c tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Inventory has not changed for provider a2858be1-fd22-4e08-979e-87ad25293407 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 138, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61806) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1230.779104] env[61806]: DEBUG oslo_concurrency.lockutils [None req-3aec3f2d-4ed5-4066-a721-5ba416343c9c tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_dest" :: held 2.618s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1232.313303] env[61806]: INFO nova.compute.manager [None req-3aec3f2d-4ed5-4066-a721-5ba416343c9c tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] [instance: 4f26445f-35d0-4f53-9014-ef3e2eb1cbdf] Swapping old allocation on dict_keys(['a2858be1-fd22-4e08-979e-87ad25293407']) held by migration bf3eae4b-1865-4dba-9161-9d137e96b682 for instance [ 1232.335221] env[61806]: DEBUG nova.scheduler.client.report [None req-3aec3f2d-4ed5-4066-a721-5ba416343c9c tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Overwriting current allocation {'allocations': {'a2858be1-fd22-4e08-979e-87ad25293407': {'resources': {'DISK_GB': 1, 'MEMORY_MB': 256, 'VCPU': 1}, 'generation': 149}}, 'project_id': '1c55cc0455584fbb8cda88d33b8f8b7e', 'user_id': '772f4c6d731c4d5daea11b37a0f6718b', 'consumer_generation': 1} on consumer 4f26445f-35d0-4f53-9014-ef3e2eb1cbdf {{(pid=61806) move_allocations /opt/stack/nova/nova/scheduler/client/report.py:2032}} [ 1232.408343] env[61806]: DEBUG oslo_concurrency.lockutils [None req-3aec3f2d-4ed5-4066-a721-5ba416343c9c tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Acquiring lock "refresh_cache-4f26445f-35d0-4f53-9014-ef3e2eb1cbdf" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1232.408574] env[61806]: DEBUG oslo_concurrency.lockutils [None req-3aec3f2d-4ed5-4066-a721-5ba416343c9c tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Acquired lock "refresh_cache-4f26445f-35d0-4f53-9014-ef3e2eb1cbdf" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1232.408759] env[61806]: DEBUG nova.network.neutron [None req-3aec3f2d-4ed5-4066-a721-5ba416343c9c tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] [instance: 4f26445f-35d0-4f53-9014-ef3e2eb1cbdf] Building network info cache for instance {{(pid=61806) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1233.117085] env[61806]: DEBUG nova.network.neutron [None req-3aec3f2d-4ed5-4066-a721-5ba416343c9c tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] [instance: 4f26445f-35d0-4f53-9014-ef3e2eb1cbdf] Updating instance_info_cache with network_info: [{"id": "a71b2870-bc50-47f4-840d-57e956c3f86d", "address": "fa:16:3e:74:22:b8", "network": {"id": "33bf5b21-82b0-432e-ae3d-9c2ca75161d0", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-1294159336-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.187", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "1c55cc0455584fbb8cda88d33b8f8b7e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "5b0fa7a2-ebd9-4788-8904-7bf250ce466c", "external-id": "nsx-vlan-transportzone-669", "segmentation_id": 669, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa71b2870-bc", "ovs_interfaceid": "a71b2870-bc50-47f4-840d-57e956c3f86d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61806) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1233.619915] env[61806]: DEBUG oslo_concurrency.lockutils [None req-3aec3f2d-4ed5-4066-a721-5ba416343c9c tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Releasing lock "refresh_cache-4f26445f-35d0-4f53-9014-ef3e2eb1cbdf" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1233.621060] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-67c4fd11-75e4-4afe-931c-4cd926c71e2d {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1233.629032] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-872749a1-2892-491b-a43a-81e80bf1fcfe {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1234.709803] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [None req-3aec3f2d-4ed5-4066-a721-5ba416343c9c tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] [instance: 4f26445f-35d0-4f53-9014-ef3e2eb1cbdf] Powering off the VM {{(pid=61806) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1234.710168] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-7eebc9f6-ced7-4715-aa57-045319600de8 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1234.717103] env[61806]: DEBUG oslo_vmware.api [None req-3aec3f2d-4ed5-4066-a721-5ba416343c9c tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Waiting for the task: (returnval){ [ 1234.717103] env[61806]: value = "task-1295318" [ 1234.717103] env[61806]: _type = "Task" [ 1234.717103] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1234.725144] env[61806]: DEBUG oslo_vmware.api [None req-3aec3f2d-4ed5-4066-a721-5ba416343c9c tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Task: {'id': task-1295318, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1235.226715] env[61806]: DEBUG oslo_vmware.api [None req-3aec3f2d-4ed5-4066-a721-5ba416343c9c tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Task: {'id': task-1295318, 'name': PowerOffVM_Task, 'duration_secs': 0.200408} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1235.226970] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [None req-3aec3f2d-4ed5-4066-a721-5ba416343c9c tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] [instance: 4f26445f-35d0-4f53-9014-ef3e2eb1cbdf] Powered off the VM {{(pid=61806) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1235.227692] env[61806]: DEBUG nova.virt.hardware [None req-3aec3f2d-4ed5-4066-a721-5ba416343c9c tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-15T18:05:10Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=61806) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1235.227917] env[61806]: DEBUG nova.virt.hardware [None req-3aec3f2d-4ed5-4066-a721-5ba416343c9c tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Flavor limits 0:0:0 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1235.228094] env[61806]: DEBUG nova.virt.hardware [None req-3aec3f2d-4ed5-4066-a721-5ba416343c9c tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Image limits 0:0:0 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1235.228294] env[61806]: DEBUG nova.virt.hardware [None req-3aec3f2d-4ed5-4066-a721-5ba416343c9c tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Flavor pref 0:0:0 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1235.228447] env[61806]: DEBUG nova.virt.hardware [None req-3aec3f2d-4ed5-4066-a721-5ba416343c9c tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Image pref 0:0:0 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1235.228599] env[61806]: DEBUG nova.virt.hardware [None req-3aec3f2d-4ed5-4066-a721-5ba416343c9c tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1235.228802] env[61806]: DEBUG nova.virt.hardware [None req-3aec3f2d-4ed5-4066-a721-5ba416343c9c tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61806) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1235.228966] env[61806]: DEBUG nova.virt.hardware [None req-3aec3f2d-4ed5-4066-a721-5ba416343c9c tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61806) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1235.229154] env[61806]: DEBUG nova.virt.hardware [None req-3aec3f2d-4ed5-4066-a721-5ba416343c9c tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Got 1 possible topologies {{(pid=61806) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1235.229321] env[61806]: DEBUG nova.virt.hardware [None req-3aec3f2d-4ed5-4066-a721-5ba416343c9c tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61806) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1235.229501] env[61806]: DEBUG nova.virt.hardware [None req-3aec3f2d-4ed5-4066-a721-5ba416343c9c tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61806) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1235.234273] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-5c7bdb34-339e-4ce8-a401-de57c3bf487d {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1235.248419] env[61806]: DEBUG oslo_vmware.api [None req-3aec3f2d-4ed5-4066-a721-5ba416343c9c tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Waiting for the task: (returnval){ [ 1235.248419] env[61806]: value = "task-1295319" [ 1235.248419] env[61806]: _type = "Task" [ 1235.248419] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1235.255676] env[61806]: DEBUG oslo_vmware.api [None req-3aec3f2d-4ed5-4066-a721-5ba416343c9c tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Task: {'id': task-1295319, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1235.758122] env[61806]: DEBUG oslo_vmware.api [None req-3aec3f2d-4ed5-4066-a721-5ba416343c9c tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Task: {'id': task-1295319, 'name': ReconfigVM_Task, 'duration_secs': 0.136092} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1235.758955] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f59a12ff-69c1-4729-8eb1-c15aa5c35529 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1235.779155] env[61806]: DEBUG nova.virt.hardware [None req-3aec3f2d-4ed5-4066-a721-5ba416343c9c tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-15T18:05:10Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=61806) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1235.779409] env[61806]: DEBUG nova.virt.hardware [None req-3aec3f2d-4ed5-4066-a721-5ba416343c9c tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Flavor limits 0:0:0 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1235.779573] env[61806]: DEBUG nova.virt.hardware [None req-3aec3f2d-4ed5-4066-a721-5ba416343c9c tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Image limits 0:0:0 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1235.779770] env[61806]: DEBUG nova.virt.hardware [None req-3aec3f2d-4ed5-4066-a721-5ba416343c9c tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Flavor pref 0:0:0 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1235.779947] env[61806]: DEBUG nova.virt.hardware [None req-3aec3f2d-4ed5-4066-a721-5ba416343c9c tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Image pref 0:0:0 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1235.780153] env[61806]: DEBUG nova.virt.hardware [None req-3aec3f2d-4ed5-4066-a721-5ba416343c9c tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1235.780369] env[61806]: DEBUG nova.virt.hardware [None req-3aec3f2d-4ed5-4066-a721-5ba416343c9c tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61806) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1235.780535] env[61806]: DEBUG nova.virt.hardware [None req-3aec3f2d-4ed5-4066-a721-5ba416343c9c tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61806) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1235.780708] env[61806]: DEBUG nova.virt.hardware [None req-3aec3f2d-4ed5-4066-a721-5ba416343c9c tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Got 1 possible topologies {{(pid=61806) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1235.780875] env[61806]: DEBUG nova.virt.hardware [None req-3aec3f2d-4ed5-4066-a721-5ba416343c9c tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61806) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1235.781064] env[61806]: DEBUG nova.virt.hardware [None req-3aec3f2d-4ed5-4066-a721-5ba416343c9c tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61806) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1235.781840] env[61806]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-62aed6c0-57e3-40e9-a135-180eb0798d04 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1235.786813] env[61806]: DEBUG oslo_vmware.api [None req-3aec3f2d-4ed5-4066-a721-5ba416343c9c tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Waiting for the task: (returnval){ [ 1235.786813] env[61806]: value = "session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]5273c423-3034-273e-7b7a-7e3b6236b74f" [ 1235.786813] env[61806]: _type = "Task" [ 1235.786813] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1235.794653] env[61806]: DEBUG oslo_vmware.api [None req-3aec3f2d-4ed5-4066-a721-5ba416343c9c tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Task: {'id': session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]5273c423-3034-273e-7b7a-7e3b6236b74f, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1236.296586] env[61806]: DEBUG oslo_vmware.api [None req-3aec3f2d-4ed5-4066-a721-5ba416343c9c tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Task: {'id': session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]5273c423-3034-273e-7b7a-7e3b6236b74f, 'name': SearchDatastore_Task, 'duration_secs': 0.008917} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1236.301968] env[61806]: DEBUG nova.virt.vmwareapi.volumeops [None req-3aec3f2d-4ed5-4066-a721-5ba416343c9c tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] [instance: 4f26445f-35d0-4f53-9014-ef3e2eb1cbdf] Reconfiguring VM instance instance-0000006c to detach disk 2000 {{(pid=61806) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1236.302254] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-0d5ae9c7-24da-443e-8e19-28b2a5b324b5 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1236.319641] env[61806]: DEBUG oslo_vmware.api [None req-3aec3f2d-4ed5-4066-a721-5ba416343c9c tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Waiting for the task: (returnval){ [ 1236.319641] env[61806]: value = "task-1295320" [ 1236.319641] env[61806]: _type = "Task" [ 1236.319641] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1236.326643] env[61806]: DEBUG oslo_vmware.api [None req-3aec3f2d-4ed5-4066-a721-5ba416343c9c tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Task: {'id': task-1295320, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1236.829988] env[61806]: DEBUG oslo_vmware.api [None req-3aec3f2d-4ed5-4066-a721-5ba416343c9c tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Task: {'id': task-1295320, 'name': ReconfigVM_Task, 'duration_secs': 0.205429} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1236.830359] env[61806]: DEBUG nova.virt.vmwareapi.volumeops [None req-3aec3f2d-4ed5-4066-a721-5ba416343c9c tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] [instance: 4f26445f-35d0-4f53-9014-ef3e2eb1cbdf] Reconfigured VM instance instance-0000006c to detach disk 2000 {{(pid=61806) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1236.831179] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2b77ac45-5ac5-48aa-bedb-a97890fd48c5 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1236.857587] env[61806]: DEBUG nova.virt.vmwareapi.volumeops [None req-3aec3f2d-4ed5-4066-a721-5ba416343c9c tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] [instance: 4f26445f-35d0-4f53-9014-ef3e2eb1cbdf] Reconfiguring VM instance instance-0000006c to attach disk [datastore1] 4f26445f-35d0-4f53-9014-ef3e2eb1cbdf/4f26445f-35d0-4f53-9014-ef3e2eb1cbdf.vmdk or device None with type thin {{(pid=61806) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1236.857864] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-09866ecf-2b33-497d-963a-eb35d0532132 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1236.877904] env[61806]: DEBUG oslo_vmware.api [None req-3aec3f2d-4ed5-4066-a721-5ba416343c9c tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Waiting for the task: (returnval){ [ 1236.877904] env[61806]: value = "task-1295321" [ 1236.877904] env[61806]: _type = "Task" [ 1236.877904] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1236.885494] env[61806]: DEBUG oslo_vmware.api [None req-3aec3f2d-4ed5-4066-a721-5ba416343c9c tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Task: {'id': task-1295321, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1237.387720] env[61806]: DEBUG oslo_vmware.api [None req-3aec3f2d-4ed5-4066-a721-5ba416343c9c tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Task: {'id': task-1295321, 'name': ReconfigVM_Task, 'duration_secs': 0.254293} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1237.388052] env[61806]: DEBUG nova.virt.vmwareapi.volumeops [None req-3aec3f2d-4ed5-4066-a721-5ba416343c9c tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] [instance: 4f26445f-35d0-4f53-9014-ef3e2eb1cbdf] Reconfigured VM instance instance-0000006c to attach disk [datastore1] 4f26445f-35d0-4f53-9014-ef3e2eb1cbdf/4f26445f-35d0-4f53-9014-ef3e2eb1cbdf.vmdk or device None with type thin {{(pid=61806) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1237.388861] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6d8c8880-2e8e-4d13-b7eb-aa8689ae5af5 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1237.410967] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-01832567-c149-4510-bd81-8b5f25004b5e {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1237.430407] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dde99730-501e-4371-864f-5a94acc2062e {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1237.449415] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8ac43d57-5470-44c8-9993-8b5d01e11d44 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1237.455383] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [None req-3aec3f2d-4ed5-4066-a721-5ba416343c9c tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] [instance: 4f26445f-35d0-4f53-9014-ef3e2eb1cbdf] Powering on the VM {{(pid=61806) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1237.455593] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-03873c3f-4ca3-422d-aa50-9bf0d89c8790 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1237.461436] env[61806]: DEBUG oslo_vmware.api [None req-3aec3f2d-4ed5-4066-a721-5ba416343c9c tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Waiting for the task: (returnval){ [ 1237.461436] env[61806]: value = "task-1295322" [ 1237.461436] env[61806]: _type = "Task" [ 1237.461436] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1237.468157] env[61806]: DEBUG oslo_vmware.api [None req-3aec3f2d-4ed5-4066-a721-5ba416343c9c tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Task: {'id': task-1295322, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1237.971740] env[61806]: DEBUG oslo_vmware.api [None req-3aec3f2d-4ed5-4066-a721-5ba416343c9c tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Task: {'id': task-1295322, 'name': PowerOnVM_Task, 'duration_secs': 0.34937} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1237.972122] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [None req-3aec3f2d-4ed5-4066-a721-5ba416343c9c tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] [instance: 4f26445f-35d0-4f53-9014-ef3e2eb1cbdf] Powered on the VM {{(pid=61806) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1239.012524] env[61806]: INFO nova.compute.manager [None req-3aec3f2d-4ed5-4066-a721-5ba416343c9c tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] [instance: 4f26445f-35d0-4f53-9014-ef3e2eb1cbdf] Updating instance to original state: 'active' [ 1239.969387] env[61806]: DEBUG oslo_concurrency.lockutils [None req-e4cde284-1fdd-4253-b776-2c10375280e0 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Acquiring lock "4f26445f-35d0-4f53-9014-ef3e2eb1cbdf" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1239.969698] env[61806]: DEBUG oslo_concurrency.lockutils [None req-e4cde284-1fdd-4253-b776-2c10375280e0 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Lock "4f26445f-35d0-4f53-9014-ef3e2eb1cbdf" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1239.969925] env[61806]: DEBUG oslo_concurrency.lockutils [None req-e4cde284-1fdd-4253-b776-2c10375280e0 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Acquiring lock "4f26445f-35d0-4f53-9014-ef3e2eb1cbdf-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1239.970129] env[61806]: DEBUG oslo_concurrency.lockutils [None req-e4cde284-1fdd-4253-b776-2c10375280e0 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Lock "4f26445f-35d0-4f53-9014-ef3e2eb1cbdf-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1239.970317] env[61806]: DEBUG oslo_concurrency.lockutils [None req-e4cde284-1fdd-4253-b776-2c10375280e0 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Lock "4f26445f-35d0-4f53-9014-ef3e2eb1cbdf-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1239.972544] env[61806]: INFO nova.compute.manager [None req-e4cde284-1fdd-4253-b776-2c10375280e0 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] [instance: 4f26445f-35d0-4f53-9014-ef3e2eb1cbdf] Terminating instance [ 1239.974498] env[61806]: DEBUG nova.compute.manager [None req-e4cde284-1fdd-4253-b776-2c10375280e0 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] [instance: 4f26445f-35d0-4f53-9014-ef3e2eb1cbdf] Start destroying the instance on the hypervisor. {{(pid=61806) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1239.974718] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [None req-e4cde284-1fdd-4253-b776-2c10375280e0 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] [instance: 4f26445f-35d0-4f53-9014-ef3e2eb1cbdf] Powering off the VM {{(pid=61806) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1239.974961] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-ff26ab3b-e2fe-44ee-a2d6-e4a2ae95badb {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1239.981421] env[61806]: DEBUG oslo_vmware.api [None req-e4cde284-1fdd-4253-b776-2c10375280e0 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Waiting for the task: (returnval){ [ 1239.981421] env[61806]: value = "task-1295323" [ 1239.981421] env[61806]: _type = "Task" [ 1239.981421] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1239.989469] env[61806]: DEBUG oslo_vmware.api [None req-e4cde284-1fdd-4253-b776-2c10375280e0 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Task: {'id': task-1295323, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1240.492025] env[61806]: DEBUG oslo_vmware.api [None req-e4cde284-1fdd-4253-b776-2c10375280e0 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Task: {'id': task-1295323, 'name': PowerOffVM_Task, 'duration_secs': 0.189917} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1240.492396] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [None req-e4cde284-1fdd-4253-b776-2c10375280e0 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] [instance: 4f26445f-35d0-4f53-9014-ef3e2eb1cbdf] Powered off the VM {{(pid=61806) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1240.492511] env[61806]: DEBUG nova.virt.vmwareapi.volumeops [None req-e4cde284-1fdd-4253-b776-2c10375280e0 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] [instance: 4f26445f-35d0-4f53-9014-ef3e2eb1cbdf] Volume detach. Driver type: vmdk {{(pid=61806) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 1240.492711] env[61806]: DEBUG nova.virt.vmwareapi.volumeops [None req-e4cde284-1fdd-4253-b776-2c10375280e0 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] [instance: 4f26445f-35d0-4f53-9014-ef3e2eb1cbdf] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-277773', 'volume_id': 'ba848344-b950-4c4f-ba25-0fd09d2bdd70', 'name': 'volume-ba848344-b950-4c4f-ba25-0fd09d2bdd70', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'attaching', 'instance': '4f26445f-35d0-4f53-9014-ef3e2eb1cbdf', 'attached_at': '2024-10-15T18:18:54.000000', 'detached_at': '', 'volume_id': 'ba848344-b950-4c4f-ba25-0fd09d2bdd70', 'serial': 'ba848344-b950-4c4f-ba25-0fd09d2bdd70'} {{(pid=61806) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 1240.493460] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3c10e521-7e99-4f73-87d1-e355c66c88d3 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1240.512882] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-935284bd-a68e-4ebf-859a-a2060e1f415d {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1240.518701] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d1f7501a-83c5-4ca1-a846-bf559a1459e5 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1240.537661] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4da54cfa-1c15-475c-a45f-046b7185f465 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1240.551481] env[61806]: DEBUG nova.virt.vmwareapi.volumeops [None req-e4cde284-1fdd-4253-b776-2c10375280e0 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] The volume has not been displaced from its original location: [datastore1] volume-ba848344-b950-4c4f-ba25-0fd09d2bdd70/volume-ba848344-b950-4c4f-ba25-0fd09d2bdd70.vmdk. No consolidation needed. {{(pid=61806) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 1240.556625] env[61806]: DEBUG nova.virt.vmwareapi.volumeops [None req-e4cde284-1fdd-4253-b776-2c10375280e0 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] [instance: 4f26445f-35d0-4f53-9014-ef3e2eb1cbdf] Reconfiguring VM instance instance-0000006c to detach disk 2001 {{(pid=61806) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1240.556882] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-e0ea732c-504a-48e1-af5d-720bfd20207a {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1240.573301] env[61806]: DEBUG oslo_vmware.api [None req-e4cde284-1fdd-4253-b776-2c10375280e0 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Waiting for the task: (returnval){ [ 1240.573301] env[61806]: value = "task-1295324" [ 1240.573301] env[61806]: _type = "Task" [ 1240.573301] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1240.581788] env[61806]: DEBUG oslo_vmware.api [None req-e4cde284-1fdd-4253-b776-2c10375280e0 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Task: {'id': task-1295324, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1241.082993] env[61806]: DEBUG oslo_vmware.api [None req-e4cde284-1fdd-4253-b776-2c10375280e0 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Task: {'id': task-1295324, 'name': ReconfigVM_Task, 'duration_secs': 0.195751} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1241.083314] env[61806]: DEBUG nova.virt.vmwareapi.volumeops [None req-e4cde284-1fdd-4253-b776-2c10375280e0 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] [instance: 4f26445f-35d0-4f53-9014-ef3e2eb1cbdf] Reconfigured VM instance instance-0000006c to detach disk 2001 {{(pid=61806) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1241.087824] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-6137297d-0da4-4dcf-84d1-b84ecf7b96fe {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1241.101427] env[61806]: DEBUG oslo_vmware.api [None req-e4cde284-1fdd-4253-b776-2c10375280e0 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Waiting for the task: (returnval){ [ 1241.101427] env[61806]: value = "task-1295325" [ 1241.101427] env[61806]: _type = "Task" [ 1241.101427] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1241.108348] env[61806]: DEBUG oslo_vmware.api [None req-e4cde284-1fdd-4253-b776-2c10375280e0 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Task: {'id': task-1295325, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1241.611794] env[61806]: DEBUG oslo_vmware.api [None req-e4cde284-1fdd-4253-b776-2c10375280e0 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Task: {'id': task-1295325, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1242.111819] env[61806]: DEBUG oslo_vmware.api [None req-e4cde284-1fdd-4253-b776-2c10375280e0 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Task: {'id': task-1295325, 'name': ReconfigVM_Task} progress is 99%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1242.612953] env[61806]: DEBUG oslo_vmware.api [None req-e4cde284-1fdd-4253-b776-2c10375280e0 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Task: {'id': task-1295325, 'name': ReconfigVM_Task} progress is 99%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1243.112996] env[61806]: DEBUG oslo_vmware.api [None req-e4cde284-1fdd-4253-b776-2c10375280e0 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Task: {'id': task-1295325, 'name': ReconfigVM_Task, 'duration_secs': 1.735782} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1243.113318] env[61806]: DEBUG nova.virt.vmwareapi.volumeops [None req-e4cde284-1fdd-4253-b776-2c10375280e0 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] [instance: 4f26445f-35d0-4f53-9014-ef3e2eb1cbdf] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-277773', 'volume_id': 'ba848344-b950-4c4f-ba25-0fd09d2bdd70', 'name': 'volume-ba848344-b950-4c4f-ba25-0fd09d2bdd70', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'attaching', 'instance': '4f26445f-35d0-4f53-9014-ef3e2eb1cbdf', 'attached_at': '2024-10-15T18:18:54.000000', 'detached_at': '', 'volume_id': 'ba848344-b950-4c4f-ba25-0fd09d2bdd70', 'serial': 'ba848344-b950-4c4f-ba25-0fd09d2bdd70'} {{(pid=61806) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 1243.113603] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-e4cde284-1fdd-4253-b776-2c10375280e0 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] [instance: 4f26445f-35d0-4f53-9014-ef3e2eb1cbdf] Destroying instance {{(pid=61806) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1243.114388] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-29e6d4a1-4fe4-4fa6-8265-4b21b794d60e {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1243.120576] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-e4cde284-1fdd-4253-b776-2c10375280e0 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] [instance: 4f26445f-35d0-4f53-9014-ef3e2eb1cbdf] Unregistering the VM {{(pid=61806) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1243.120791] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-afd69ec9-faf4-4beb-b39d-715331166304 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1243.178944] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-e4cde284-1fdd-4253-b776-2c10375280e0 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] [instance: 4f26445f-35d0-4f53-9014-ef3e2eb1cbdf] Unregistered the VM {{(pid=61806) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1243.179173] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-e4cde284-1fdd-4253-b776-2c10375280e0 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] [instance: 4f26445f-35d0-4f53-9014-ef3e2eb1cbdf] Deleting contents of the VM from datastore datastore1 {{(pid=61806) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1243.179357] env[61806]: DEBUG nova.virt.vmwareapi.ds_util [None req-e4cde284-1fdd-4253-b776-2c10375280e0 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Deleting the datastore file [datastore1] 4f26445f-35d0-4f53-9014-ef3e2eb1cbdf {{(pid=61806) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1243.179619] env[61806]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-61c9cdeb-c768-40df-b4e4-3767802105c4 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1243.185594] env[61806]: DEBUG oslo_vmware.api [None req-e4cde284-1fdd-4253-b776-2c10375280e0 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Waiting for the task: (returnval){ [ 1243.185594] env[61806]: value = "task-1295327" [ 1243.185594] env[61806]: _type = "Task" [ 1243.185594] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1243.192687] env[61806]: DEBUG oslo_vmware.api [None req-e4cde284-1fdd-4253-b776-2c10375280e0 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Task: {'id': task-1295327, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1243.694765] env[61806]: DEBUG oslo_vmware.api [None req-e4cde284-1fdd-4253-b776-2c10375280e0 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Task: {'id': task-1295327, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.152659} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1243.695234] env[61806]: DEBUG nova.virt.vmwareapi.ds_util [None req-e4cde284-1fdd-4253-b776-2c10375280e0 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Deleted the datastore file {{(pid=61806) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1243.695234] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-e4cde284-1fdd-4253-b776-2c10375280e0 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] [instance: 4f26445f-35d0-4f53-9014-ef3e2eb1cbdf] Deleted contents of the VM from datastore datastore1 {{(pid=61806) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1243.695409] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-e4cde284-1fdd-4253-b776-2c10375280e0 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] [instance: 4f26445f-35d0-4f53-9014-ef3e2eb1cbdf] Instance destroyed {{(pid=61806) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1243.695611] env[61806]: INFO nova.compute.manager [None req-e4cde284-1fdd-4253-b776-2c10375280e0 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] [instance: 4f26445f-35d0-4f53-9014-ef3e2eb1cbdf] Took 3.72 seconds to destroy the instance on the hypervisor. [ 1243.695863] env[61806]: DEBUG oslo.service.loopingcall [None req-e4cde284-1fdd-4253-b776-2c10375280e0 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61806) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1243.696070] env[61806]: DEBUG nova.compute.manager [-] [instance: 4f26445f-35d0-4f53-9014-ef3e2eb1cbdf] Deallocating network for instance {{(pid=61806) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1243.696168] env[61806]: DEBUG nova.network.neutron [-] [instance: 4f26445f-35d0-4f53-9014-ef3e2eb1cbdf] deallocate_for_instance() {{(pid=61806) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1244.314876] env[61806]: DEBUG nova.compute.manager [req-33b8fd34-20a5-4b6a-aed6-4b9c90702a23 req-6ae44e88-5e20-4ed6-8597-934b8e3ae0bc service nova] [instance: 4f26445f-35d0-4f53-9014-ef3e2eb1cbdf] Received event network-vif-deleted-a71b2870-bc50-47f4-840d-57e956c3f86d {{(pid=61806) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1244.315019] env[61806]: INFO nova.compute.manager [req-33b8fd34-20a5-4b6a-aed6-4b9c90702a23 req-6ae44e88-5e20-4ed6-8597-934b8e3ae0bc service nova] [instance: 4f26445f-35d0-4f53-9014-ef3e2eb1cbdf] Neutron deleted interface a71b2870-bc50-47f4-840d-57e956c3f86d; detaching it from the instance and deleting it from the info cache [ 1244.315207] env[61806]: DEBUG nova.network.neutron [req-33b8fd34-20a5-4b6a-aed6-4b9c90702a23 req-6ae44e88-5e20-4ed6-8597-934b8e3ae0bc service nova] [instance: 4f26445f-35d0-4f53-9014-ef3e2eb1cbdf] Updating instance_info_cache with network_info: [] {{(pid=61806) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1244.792010] env[61806]: DEBUG nova.network.neutron [-] [instance: 4f26445f-35d0-4f53-9014-ef3e2eb1cbdf] Updating instance_info_cache with network_info: [] {{(pid=61806) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1244.818033] env[61806]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-09070b16-8a76-4bd6-a319-adabb3cbfe01 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1244.827014] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e8c9bdd5-3c9e-4934-922a-6c933017d9f7 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1244.848471] env[61806]: DEBUG nova.compute.manager [req-33b8fd34-20a5-4b6a-aed6-4b9c90702a23 req-6ae44e88-5e20-4ed6-8597-934b8e3ae0bc service nova] [instance: 4f26445f-35d0-4f53-9014-ef3e2eb1cbdf] Detach interface failed, port_id=a71b2870-bc50-47f4-840d-57e956c3f86d, reason: Instance 4f26445f-35d0-4f53-9014-ef3e2eb1cbdf could not be found. {{(pid=61806) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10963}} [ 1245.295384] env[61806]: INFO nova.compute.manager [-] [instance: 4f26445f-35d0-4f53-9014-ef3e2eb1cbdf] Took 1.60 seconds to deallocate network for instance. [ 1245.838647] env[61806]: INFO nova.compute.manager [None req-e4cde284-1fdd-4253-b776-2c10375280e0 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] [instance: 4f26445f-35d0-4f53-9014-ef3e2eb1cbdf] Took 0.54 seconds to detach 1 volumes for instance. [ 1246.346380] env[61806]: DEBUG oslo_concurrency.lockutils [None req-e4cde284-1fdd-4253-b776-2c10375280e0 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1246.346555] env[61806]: DEBUG oslo_concurrency.lockutils [None req-e4cde284-1fdd-4253-b776-2c10375280e0 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.000s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1246.346725] env[61806]: DEBUG oslo_concurrency.lockutils [None req-e4cde284-1fdd-4253-b776-2c10375280e0 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1246.372835] env[61806]: INFO nova.scheduler.client.report [None req-e4cde284-1fdd-4253-b776-2c10375280e0 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Deleted allocations for instance 4f26445f-35d0-4f53-9014-ef3e2eb1cbdf [ 1246.880874] env[61806]: DEBUG oslo_concurrency.lockutils [None req-e4cde284-1fdd-4253-b776-2c10375280e0 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Lock "4f26445f-35d0-4f53-9014-ef3e2eb1cbdf" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 6.911s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1248.306368] env[61806]: DEBUG oslo_concurrency.lockutils [None req-28105e44-4916-46ed-961c-1d7952a7ef4d tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Acquiring lock "7afc6260-7569-44c4-9764-7aa08c662c1c" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1248.306712] env[61806]: DEBUG oslo_concurrency.lockutils [None req-28105e44-4916-46ed-961c-1d7952a7ef4d tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Lock "7afc6260-7569-44c4-9764-7aa08c662c1c" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1248.809385] env[61806]: DEBUG nova.compute.manager [None req-28105e44-4916-46ed-961c-1d7952a7ef4d tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] [instance: 7afc6260-7569-44c4-9764-7aa08c662c1c] Starting instance... {{(pid=61806) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1249.329227] env[61806]: DEBUG oslo_concurrency.lockutils [None req-28105e44-4916-46ed-961c-1d7952a7ef4d tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1249.329483] env[61806]: DEBUG oslo_concurrency.lockutils [None req-28105e44-4916-46ed-961c-1d7952a7ef4d tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.000s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1249.330949] env[61806]: INFO nova.compute.claims [None req-28105e44-4916-46ed-961c-1d7952a7ef4d tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] [instance: 7afc6260-7569-44c4-9764-7aa08c662c1c] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1250.018152] env[61806]: DEBUG oslo_service.periodic_task [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=61806) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1250.366139] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c5ebee1b-4cbb-4e31-9226-728f9df0cc6c {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1250.373786] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-86fbc385-ac19-46dc-97b5-fea8509918ba {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1250.403108] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5037939d-5f41-4319-bbe4-f34c7f56bfda {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1250.409508] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8eb57e0b-34fd-4c98-9d33-7e8502b94159 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1250.421955] env[61806]: DEBUG nova.compute.provider_tree [None req-28105e44-4916-46ed-961c-1d7952a7ef4d tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Updating inventory in ProviderTree for provider a2858be1-fd22-4e08-979e-87ad25293407 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 139, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61806) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1250.941935] env[61806]: ERROR nova.scheduler.client.report [None req-28105e44-4916-46ed-961c-1d7952a7ef4d tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] [req-8aae2fe9-c401-4427-a690-41183b87009a] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 139, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID a2858be1-fd22-4e08-979e-87ad25293407. Got 409: {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-8aae2fe9-c401-4427-a690-41183b87009a"}]} [ 1250.958356] env[61806]: DEBUG nova.scheduler.client.report [None req-28105e44-4916-46ed-961c-1d7952a7ef4d tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Refreshing inventories for resource provider a2858be1-fd22-4e08-979e-87ad25293407 {{(pid=61806) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:818}} [ 1250.971315] env[61806]: DEBUG nova.scheduler.client.report [None req-28105e44-4916-46ed-961c-1d7952a7ef4d tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Updating ProviderTree inventory for provider a2858be1-fd22-4e08-979e-87ad25293407 from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 138, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61806) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:782}} [ 1250.971532] env[61806]: DEBUG nova.compute.provider_tree [None req-28105e44-4916-46ed-961c-1d7952a7ef4d tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Updating inventory in ProviderTree for provider a2858be1-fd22-4e08-979e-87ad25293407 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 138, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61806) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1250.981214] env[61806]: DEBUG nova.scheduler.client.report [None req-28105e44-4916-46ed-961c-1d7952a7ef4d tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Refreshing aggregate associations for resource provider a2858be1-fd22-4e08-979e-87ad25293407, aggregates: None {{(pid=61806) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:827}} [ 1250.997735] env[61806]: DEBUG nova.scheduler.client.report [None req-28105e44-4916-46ed-961c-1d7952a7ef4d tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Refreshing trait associations for resource provider a2858be1-fd22-4e08-979e-87ad25293407, traits: COMPUTE_IMAGE_TYPE_ISO,COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_SAME_HOST_COLD_MIGRATE,COMPUTE_NET_ATTACH_INTERFACE,COMPUTE_NODE {{(pid=61806) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:839}} [ 1251.019093] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c8e625fd-8526-4f5a-8e1e-5ee0f52b4e1d {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1251.026228] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bc80f765-d3fc-445b-8f89-7e4ba4e5cf9e {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1251.054677] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9f5b4e31-a494-4a97-a773-f980a684aa51 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1251.061204] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c5ff26e6-9c7b-4013-819c-b27a35a8702f {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1251.073387] env[61806]: DEBUG nova.compute.provider_tree [None req-28105e44-4916-46ed-961c-1d7952a7ef4d tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Updating inventory in ProviderTree for provider a2858be1-fd22-4e08-979e-87ad25293407 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 139, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61806) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1251.605020] env[61806]: DEBUG nova.scheduler.client.report [None req-28105e44-4916-46ed-961c-1d7952a7ef4d tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Updated inventory for provider a2858be1-fd22-4e08-979e-87ad25293407 with generation 151 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 139, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61806) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:971}} [ 1251.605020] env[61806]: DEBUG nova.compute.provider_tree [None req-28105e44-4916-46ed-961c-1d7952a7ef4d tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Updating resource provider a2858be1-fd22-4e08-979e-87ad25293407 generation from 151 to 152 during operation: update_inventory {{(pid=61806) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 1251.605020] env[61806]: DEBUG nova.compute.provider_tree [None req-28105e44-4916-46ed-961c-1d7952a7ef4d tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Updating inventory in ProviderTree for provider a2858be1-fd22-4e08-979e-87ad25293407 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 139, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61806) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1252.108824] env[61806]: DEBUG oslo_concurrency.lockutils [None req-28105e44-4916-46ed-961c-1d7952a7ef4d tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.779s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1252.109152] env[61806]: DEBUG nova.compute.manager [None req-28105e44-4916-46ed-961c-1d7952a7ef4d tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] [instance: 7afc6260-7569-44c4-9764-7aa08c662c1c] Start building networks asynchronously for instance. {{(pid=61806) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1252.615417] env[61806]: DEBUG nova.compute.utils [None req-28105e44-4916-46ed-961c-1d7952a7ef4d tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Using /dev/sd instead of None {{(pid=61806) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1252.616977] env[61806]: DEBUG nova.compute.manager [None req-28105e44-4916-46ed-961c-1d7952a7ef4d tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] [instance: 7afc6260-7569-44c4-9764-7aa08c662c1c] Allocating IP information in the background. {{(pid=61806) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1252.616977] env[61806]: DEBUG nova.network.neutron [None req-28105e44-4916-46ed-961c-1d7952a7ef4d tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] [instance: 7afc6260-7569-44c4-9764-7aa08c662c1c] allocate_for_instance() {{(pid=61806) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 1252.666077] env[61806]: DEBUG nova.policy [None req-28105e44-4916-46ed-961c-1d7952a7ef4d tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '772f4c6d731c4d5daea11b37a0f6718b', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '1c55cc0455584fbb8cda88d33b8f8b7e', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61806) authorize /opt/stack/nova/nova/policy.py:201}} [ 1252.942312] env[61806]: DEBUG nova.network.neutron [None req-28105e44-4916-46ed-961c-1d7952a7ef4d tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] [instance: 7afc6260-7569-44c4-9764-7aa08c662c1c] Successfully created port: ee859b12-498b-43fe-97f1-47c250ada436 {{(pid=61806) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1253.023092] env[61806]: DEBUG oslo_service.periodic_task [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=61806) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1253.023275] env[61806]: DEBUG nova.compute.manager [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Starting heal instance info cache {{(pid=61806) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9926}} [ 1253.023399] env[61806]: DEBUG nova.compute.manager [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Rebuilding the list of instances to heal {{(pid=61806) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9930}} [ 1253.120262] env[61806]: DEBUG nova.compute.manager [None req-28105e44-4916-46ed-961c-1d7952a7ef4d tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] [instance: 7afc6260-7569-44c4-9764-7aa08c662c1c] Start building block device mappings for instance. {{(pid=61806) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1253.526699] env[61806]: DEBUG nova.compute.manager [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] [instance: 7afc6260-7569-44c4-9764-7aa08c662c1c] Skipping network cache update for instance because it is Building. {{(pid=61806) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9939}} [ 1253.526860] env[61806]: DEBUG nova.compute.manager [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Didn't find any instances for network info cache update. {{(pid=61806) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10012}} [ 1253.527363] env[61806]: DEBUG oslo_service.periodic_task [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Running periodic task ComputeManager.update_available_resource {{(pid=61806) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1254.030642] env[61806]: DEBUG oslo_concurrency.lockutils [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1254.031037] env[61806]: DEBUG oslo_concurrency.lockutils [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 0.000s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1254.031116] env[61806]: DEBUG oslo_concurrency.lockutils [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1254.031282] env[61806]: DEBUG nova.compute.resource_tracker [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=61806) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 1254.032314] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b9682adb-30da-4c5e-91c9-10421848dc11 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1254.041247] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-025b80d7-54b0-4ad1-9a76-73dee281bd10 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1254.056313] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2bb0e8a8-b8fc-478e-84bf-e8b3fae9aec0 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1254.062888] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7cc73f9c-019d-4fe2-8601-623ba4c9f441 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1254.094869] env[61806]: DEBUG nova.compute.resource_tracker [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=180878MB free_disk=139GB free_vcpus=48 pci_devices=None {{(pid=61806) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 1254.095062] env[61806]: DEBUG oslo_concurrency.lockutils [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1254.095285] env[61806]: DEBUG oslo_concurrency.lockutils [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 0.000s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1254.130765] env[61806]: DEBUG nova.compute.manager [None req-28105e44-4916-46ed-961c-1d7952a7ef4d tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] [instance: 7afc6260-7569-44c4-9764-7aa08c662c1c] Start spawning the instance on the hypervisor. {{(pid=61806) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1254.156785] env[61806]: DEBUG nova.virt.hardware [None req-28105e44-4916-46ed-961c-1d7952a7ef4d tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-15T18:05:10Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-15T18:04:52Z,direct_url=,disk_format='vmdk',id=ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='84ba83fce2c349a6988173c5d6fc3b07',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-15T18:04:53Z,virtual_size=,visibility=), allow threads: False {{(pid=61806) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1254.157054] env[61806]: DEBUG nova.virt.hardware [None req-28105e44-4916-46ed-961c-1d7952a7ef4d tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Flavor limits 0:0:0 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1254.157222] env[61806]: DEBUG nova.virt.hardware [None req-28105e44-4916-46ed-961c-1d7952a7ef4d tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Image limits 0:0:0 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1254.157411] env[61806]: DEBUG nova.virt.hardware [None req-28105e44-4916-46ed-961c-1d7952a7ef4d tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Flavor pref 0:0:0 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1254.157585] env[61806]: DEBUG nova.virt.hardware [None req-28105e44-4916-46ed-961c-1d7952a7ef4d tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Image pref 0:0:0 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1254.157765] env[61806]: DEBUG nova.virt.hardware [None req-28105e44-4916-46ed-961c-1d7952a7ef4d tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1254.157979] env[61806]: DEBUG nova.virt.hardware [None req-28105e44-4916-46ed-961c-1d7952a7ef4d tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61806) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1254.158195] env[61806]: DEBUG nova.virt.hardware [None req-28105e44-4916-46ed-961c-1d7952a7ef4d tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61806) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1254.158372] env[61806]: DEBUG nova.virt.hardware [None req-28105e44-4916-46ed-961c-1d7952a7ef4d tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Got 1 possible topologies {{(pid=61806) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1254.158549] env[61806]: DEBUG nova.virt.hardware [None req-28105e44-4916-46ed-961c-1d7952a7ef4d tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61806) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1254.158779] env[61806]: DEBUG nova.virt.hardware [None req-28105e44-4916-46ed-961c-1d7952a7ef4d tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61806) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1254.159648] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e4b0bfda-49b5-47b8-923c-4f11e07fb1c6 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1254.167801] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2b61c5a3-7c4e-4f45-aa8c-dc2c5ce3c868 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1254.364633] env[61806]: DEBUG nova.compute.manager [req-8edac2af-55a7-423a-9630-affeb48b97af req-3c2b89b8-386c-47ff-89cd-a74c1a776d02 service nova] [instance: 7afc6260-7569-44c4-9764-7aa08c662c1c] Received event network-vif-plugged-ee859b12-498b-43fe-97f1-47c250ada436 {{(pid=61806) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1254.364859] env[61806]: DEBUG oslo_concurrency.lockutils [req-8edac2af-55a7-423a-9630-affeb48b97af req-3c2b89b8-386c-47ff-89cd-a74c1a776d02 service nova] Acquiring lock "7afc6260-7569-44c4-9764-7aa08c662c1c-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1254.365093] env[61806]: DEBUG oslo_concurrency.lockutils [req-8edac2af-55a7-423a-9630-affeb48b97af req-3c2b89b8-386c-47ff-89cd-a74c1a776d02 service nova] Lock "7afc6260-7569-44c4-9764-7aa08c662c1c-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1254.365273] env[61806]: DEBUG oslo_concurrency.lockutils [req-8edac2af-55a7-423a-9630-affeb48b97af req-3c2b89b8-386c-47ff-89cd-a74c1a776d02 service nova] Lock "7afc6260-7569-44c4-9764-7aa08c662c1c-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1254.365446] env[61806]: DEBUG nova.compute.manager [req-8edac2af-55a7-423a-9630-affeb48b97af req-3c2b89b8-386c-47ff-89cd-a74c1a776d02 service nova] [instance: 7afc6260-7569-44c4-9764-7aa08c662c1c] No waiting events found dispatching network-vif-plugged-ee859b12-498b-43fe-97f1-47c250ada436 {{(pid=61806) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1254.365615] env[61806]: WARNING nova.compute.manager [req-8edac2af-55a7-423a-9630-affeb48b97af req-3c2b89b8-386c-47ff-89cd-a74c1a776d02 service nova] [instance: 7afc6260-7569-44c4-9764-7aa08c662c1c] Received unexpected event network-vif-plugged-ee859b12-498b-43fe-97f1-47c250ada436 for instance with vm_state building and task_state spawning. [ 1254.919495] env[61806]: DEBUG nova.network.neutron [None req-28105e44-4916-46ed-961c-1d7952a7ef4d tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] [instance: 7afc6260-7569-44c4-9764-7aa08c662c1c] Successfully updated port: ee859b12-498b-43fe-97f1-47c250ada436 {{(pid=61806) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1254.942954] env[61806]: DEBUG nova.compute.manager [req-c9b91dd7-bdd0-4f43-be39-87e32ac085b7 req-26b0260d-2148-4824-a64a-85b9fa6641b3 service nova] [instance: 7afc6260-7569-44c4-9764-7aa08c662c1c] Received event network-changed-ee859b12-498b-43fe-97f1-47c250ada436 {{(pid=61806) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1254.943323] env[61806]: DEBUG nova.compute.manager [req-c9b91dd7-bdd0-4f43-be39-87e32ac085b7 req-26b0260d-2148-4824-a64a-85b9fa6641b3 service nova] [instance: 7afc6260-7569-44c4-9764-7aa08c662c1c] Refreshing instance network info cache due to event network-changed-ee859b12-498b-43fe-97f1-47c250ada436. {{(pid=61806) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 1254.943417] env[61806]: DEBUG oslo_concurrency.lockutils [req-c9b91dd7-bdd0-4f43-be39-87e32ac085b7 req-26b0260d-2148-4824-a64a-85b9fa6641b3 service nova] Acquiring lock "refresh_cache-7afc6260-7569-44c4-9764-7aa08c662c1c" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1254.943549] env[61806]: DEBUG oslo_concurrency.lockutils [req-c9b91dd7-bdd0-4f43-be39-87e32ac085b7 req-26b0260d-2148-4824-a64a-85b9fa6641b3 service nova] Acquired lock "refresh_cache-7afc6260-7569-44c4-9764-7aa08c662c1c" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1254.943720] env[61806]: DEBUG nova.network.neutron [req-c9b91dd7-bdd0-4f43-be39-87e32ac085b7 req-26b0260d-2148-4824-a64a-85b9fa6641b3 service nova] [instance: 7afc6260-7569-44c4-9764-7aa08c662c1c] Refreshing network info cache for port ee859b12-498b-43fe-97f1-47c250ada436 {{(pid=61806) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1255.118551] env[61806]: DEBUG nova.compute.resource_tracker [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Instance 7afc6260-7569-44c4-9764-7aa08c662c1c actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61806) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1255.118824] env[61806]: DEBUG nova.compute.resource_tracker [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Total usable vcpus: 48, total allocated vcpus: 1 {{(pid=61806) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 1255.118875] env[61806]: DEBUG nova.compute.resource_tracker [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=704MB phys_disk=200GB used_disk=1GB total_vcpus=48 used_vcpus=1 pci_stats=[] {{(pid=61806) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 1255.144073] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1fd81b59-a521-49c6-b16d-3e79cf7db0b3 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1255.151514] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-37a569dc-ff00-43f0-9e7b-26e839167566 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1255.180780] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-931adfb5-0f77-4e5c-87a1-44e41a57882e {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1255.187297] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b04f9436-90b6-4963-a4a7-201e9c1c2583 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1255.199582] env[61806]: DEBUG nova.compute.provider_tree [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Inventory has not changed in ProviderTree for provider: a2858be1-fd22-4e08-979e-87ad25293407 {{(pid=61806) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1255.421992] env[61806]: DEBUG oslo_concurrency.lockutils [None req-28105e44-4916-46ed-961c-1d7952a7ef4d tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Acquiring lock "refresh_cache-7afc6260-7569-44c4-9764-7aa08c662c1c" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1255.475264] env[61806]: DEBUG nova.network.neutron [req-c9b91dd7-bdd0-4f43-be39-87e32ac085b7 req-26b0260d-2148-4824-a64a-85b9fa6641b3 service nova] [instance: 7afc6260-7569-44c4-9764-7aa08c662c1c] Instance cache missing network info. {{(pid=61806) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 1255.544175] env[61806]: DEBUG nova.network.neutron [req-c9b91dd7-bdd0-4f43-be39-87e32ac085b7 req-26b0260d-2148-4824-a64a-85b9fa6641b3 service nova] [instance: 7afc6260-7569-44c4-9764-7aa08c662c1c] Updating instance_info_cache with network_info: [] {{(pid=61806) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1255.701965] env[61806]: DEBUG nova.scheduler.client.report [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Inventory has not changed for provider a2858be1-fd22-4e08-979e-87ad25293407 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 139, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61806) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1256.047376] env[61806]: DEBUG oslo_concurrency.lockutils [req-c9b91dd7-bdd0-4f43-be39-87e32ac085b7 req-26b0260d-2148-4824-a64a-85b9fa6641b3 service nova] Releasing lock "refresh_cache-7afc6260-7569-44c4-9764-7aa08c662c1c" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1256.047775] env[61806]: DEBUG oslo_concurrency.lockutils [None req-28105e44-4916-46ed-961c-1d7952a7ef4d tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Acquired lock "refresh_cache-7afc6260-7569-44c4-9764-7aa08c662c1c" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1256.047939] env[61806]: DEBUG nova.network.neutron [None req-28105e44-4916-46ed-961c-1d7952a7ef4d tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] [instance: 7afc6260-7569-44c4-9764-7aa08c662c1c] Building network info cache for instance {{(pid=61806) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1256.206332] env[61806]: DEBUG nova.compute.resource_tracker [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=61806) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 1256.206655] env[61806]: DEBUG oslo_concurrency.lockutils [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 2.111s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1256.585263] env[61806]: DEBUG nova.network.neutron [None req-28105e44-4916-46ed-961c-1d7952a7ef4d tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] [instance: 7afc6260-7569-44c4-9764-7aa08c662c1c] Instance cache missing network info. {{(pid=61806) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 1256.702355] env[61806]: DEBUG oslo_service.periodic_task [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=61806) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1256.708622] env[61806]: DEBUG nova.network.neutron [None req-28105e44-4916-46ed-961c-1d7952a7ef4d tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] [instance: 7afc6260-7569-44c4-9764-7aa08c662c1c] Updating instance_info_cache with network_info: [{"id": "ee859b12-498b-43fe-97f1-47c250ada436", "address": "fa:16:3e:6d:6f:15", "network": {"id": "33bf5b21-82b0-432e-ae3d-9c2ca75161d0", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-1294159336-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "1c55cc0455584fbb8cda88d33b8f8b7e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "5b0fa7a2-ebd9-4788-8904-7bf250ce466c", "external-id": "nsx-vlan-transportzone-669", "segmentation_id": 669, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapee859b12-49", "ovs_interfaceid": "ee859b12-498b-43fe-97f1-47c250ada436", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61806) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1257.023447] env[61806]: DEBUG oslo_service.periodic_task [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=61806) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1257.211527] env[61806]: DEBUG oslo_concurrency.lockutils [None req-28105e44-4916-46ed-961c-1d7952a7ef4d tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Releasing lock "refresh_cache-7afc6260-7569-44c4-9764-7aa08c662c1c" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1257.211862] env[61806]: DEBUG nova.compute.manager [None req-28105e44-4916-46ed-961c-1d7952a7ef4d tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] [instance: 7afc6260-7569-44c4-9764-7aa08c662c1c] Instance network_info: |[{"id": "ee859b12-498b-43fe-97f1-47c250ada436", "address": "fa:16:3e:6d:6f:15", "network": {"id": "33bf5b21-82b0-432e-ae3d-9c2ca75161d0", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-1294159336-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "1c55cc0455584fbb8cda88d33b8f8b7e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "5b0fa7a2-ebd9-4788-8904-7bf250ce466c", "external-id": "nsx-vlan-transportzone-669", "segmentation_id": 669, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapee859b12-49", "ovs_interfaceid": "ee859b12-498b-43fe-97f1-47c250ada436", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61806) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1257.212336] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-28105e44-4916-46ed-961c-1d7952a7ef4d tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] [instance: 7afc6260-7569-44c4-9764-7aa08c662c1c] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:6d:6f:15', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '5b0fa7a2-ebd9-4788-8904-7bf250ce466c', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'ee859b12-498b-43fe-97f1-47c250ada436', 'vif_model': 'vmxnet3'}] {{(pid=61806) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1257.219784] env[61806]: DEBUG oslo.service.loopingcall [None req-28105e44-4916-46ed-961c-1d7952a7ef4d tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61806) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1257.220009] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 7afc6260-7569-44c4-9764-7aa08c662c1c] Creating VM on the ESX host {{(pid=61806) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 1257.220254] env[61806]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-17b425d9-0121-4f83-8ed1-5467dc4a7c85 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1257.239683] env[61806]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1257.239683] env[61806]: value = "task-1295328" [ 1257.239683] env[61806]: _type = "Task" [ 1257.239683] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1257.250770] env[61806]: DEBUG oslo_vmware.api [-] Task: {'id': task-1295328, 'name': CreateVM_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1257.750033] env[61806]: DEBUG oslo_vmware.api [-] Task: {'id': task-1295328, 'name': CreateVM_Task, 'duration_secs': 0.279815} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1257.750033] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 7afc6260-7569-44c4-9764-7aa08c662c1c] Created VM on the ESX host {{(pid=61806) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 1257.750725] env[61806]: DEBUG oslo_concurrency.lockutils [None req-28105e44-4916-46ed-961c-1d7952a7ef4d tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1257.750813] env[61806]: DEBUG oslo_concurrency.lockutils [None req-28105e44-4916-46ed-961c-1d7952a7ef4d tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Acquired lock "[datastore1] devstack-image-cache_base/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1257.751116] env[61806]: DEBUG oslo_concurrency.lockutils [None req-28105e44-4916-46ed-961c-1d7952a7ef4d tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1257.751371] env[61806]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-91f4421b-725a-4747-aa22-7e5ff7c31b02 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1257.755482] env[61806]: DEBUG oslo_vmware.api [None req-28105e44-4916-46ed-961c-1d7952a7ef4d tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Waiting for the task: (returnval){ [ 1257.755482] env[61806]: value = "session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]52d2e33d-884b-81b2-c5d7-ab43acf6fba5" [ 1257.755482] env[61806]: _type = "Task" [ 1257.755482] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1257.762705] env[61806]: DEBUG oslo_vmware.api [None req-28105e44-4916-46ed-961c-1d7952a7ef4d tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Task: {'id': session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]52d2e33d-884b-81b2-c5d7-ab43acf6fba5, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1258.267290] env[61806]: DEBUG oslo_vmware.api [None req-28105e44-4916-46ed-961c-1d7952a7ef4d tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Task: {'id': session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]52d2e33d-884b-81b2-c5d7-ab43acf6fba5, 'name': SearchDatastore_Task, 'duration_secs': 0.012124} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1258.267620] env[61806]: DEBUG oslo_concurrency.lockutils [None req-28105e44-4916-46ed-961c-1d7952a7ef4d tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Releasing lock "[datastore1] devstack-image-cache_base/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1258.267854] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-28105e44-4916-46ed-961c-1d7952a7ef4d tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] [instance: 7afc6260-7569-44c4-9764-7aa08c662c1c] Processing image ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa {{(pid=61806) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1258.268106] env[61806]: DEBUG oslo_concurrency.lockutils [None req-28105e44-4916-46ed-961c-1d7952a7ef4d tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa.vmdk" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1258.268264] env[61806]: DEBUG oslo_concurrency.lockutils [None req-28105e44-4916-46ed-961c-1d7952a7ef4d tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Acquired lock "[datastore1] devstack-image-cache_base/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa.vmdk" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1258.268449] env[61806]: DEBUG nova.virt.vmwareapi.ds_util [None req-28105e44-4916-46ed-961c-1d7952a7ef4d tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61806) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1258.268738] env[61806]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-4c06cf24-5d66-4c12-81f4-c1a1d39ec09f {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1258.276507] env[61806]: DEBUG nova.virt.vmwareapi.ds_util [None req-28105e44-4916-46ed-961c-1d7952a7ef4d tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61806) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1258.276686] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-28105e44-4916-46ed-961c-1d7952a7ef4d tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61806) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 1258.277373] env[61806]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f2a14191-0ced-4de4-9b5b-6af0d7ddad58 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1258.281916] env[61806]: DEBUG oslo_vmware.api [None req-28105e44-4916-46ed-961c-1d7952a7ef4d tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Waiting for the task: (returnval){ [ 1258.281916] env[61806]: value = "session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]52e9021e-5262-7bcd-ca2c-ed0e68c8f226" [ 1258.281916] env[61806]: _type = "Task" [ 1258.281916] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1258.289034] env[61806]: DEBUG oslo_vmware.api [None req-28105e44-4916-46ed-961c-1d7952a7ef4d tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Task: {'id': session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]52e9021e-5262-7bcd-ca2c-ed0e68c8f226, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1258.791961] env[61806]: DEBUG oslo_vmware.api [None req-28105e44-4916-46ed-961c-1d7952a7ef4d tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Task: {'id': session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]52e9021e-5262-7bcd-ca2c-ed0e68c8f226, 'name': SearchDatastore_Task, 'duration_secs': 0.007624} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1258.792785] env[61806]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-97a4ccdb-686f-4ee5-8611-66f4f52d0d20 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1258.797746] env[61806]: DEBUG oslo_vmware.api [None req-28105e44-4916-46ed-961c-1d7952a7ef4d tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Waiting for the task: (returnval){ [ 1258.797746] env[61806]: value = "session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]52b8cc50-7e92-6415-16cd-9d6cb2dad709" [ 1258.797746] env[61806]: _type = "Task" [ 1258.797746] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1258.805180] env[61806]: DEBUG oslo_vmware.api [None req-28105e44-4916-46ed-961c-1d7952a7ef4d tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Task: {'id': session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]52b8cc50-7e92-6415-16cd-9d6cb2dad709, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1259.022732] env[61806]: DEBUG oslo_service.periodic_task [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=61806) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1259.022965] env[61806]: DEBUG oslo_service.periodic_task [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=61806) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1259.023153] env[61806]: DEBUG oslo_service.periodic_task [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=61806) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1259.023299] env[61806]: DEBUG nova.compute.manager [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=61806) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10545}} [ 1259.307492] env[61806]: DEBUG oslo_vmware.api [None req-28105e44-4916-46ed-961c-1d7952a7ef4d tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Task: {'id': session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]52b8cc50-7e92-6415-16cd-9d6cb2dad709, 'name': SearchDatastore_Task, 'duration_secs': 0.00933} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1259.307878] env[61806]: DEBUG oslo_concurrency.lockutils [None req-28105e44-4916-46ed-961c-1d7952a7ef4d tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Releasing lock "[datastore1] devstack-image-cache_base/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa.vmdk" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1259.308055] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [None req-28105e44-4916-46ed-961c-1d7952a7ef4d tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa.vmdk to [datastore1] 7afc6260-7569-44c4-9764-7aa08c662c1c/7afc6260-7569-44c4-9764-7aa08c662c1c.vmdk {{(pid=61806) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 1259.308317] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-f5f2a902-c799-4aec-b2c8-4080b07d9c45 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1259.314810] env[61806]: DEBUG oslo_vmware.api [None req-28105e44-4916-46ed-961c-1d7952a7ef4d tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Waiting for the task: (returnval){ [ 1259.314810] env[61806]: value = "task-1295329" [ 1259.314810] env[61806]: _type = "Task" [ 1259.314810] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1259.321610] env[61806]: DEBUG oslo_vmware.api [None req-28105e44-4916-46ed-961c-1d7952a7ef4d tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Task: {'id': task-1295329, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1259.823771] env[61806]: DEBUG oslo_vmware.api [None req-28105e44-4916-46ed-961c-1d7952a7ef4d tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Task: {'id': task-1295329, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.488563} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1259.824064] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [None req-28105e44-4916-46ed-961c-1d7952a7ef4d tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa.vmdk to [datastore1] 7afc6260-7569-44c4-9764-7aa08c662c1c/7afc6260-7569-44c4-9764-7aa08c662c1c.vmdk {{(pid=61806) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 1259.824289] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-28105e44-4916-46ed-961c-1d7952a7ef4d tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] [instance: 7afc6260-7569-44c4-9764-7aa08c662c1c] Extending root virtual disk to 1048576 {{(pid=61806) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1259.824545] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-dce2b847-9839-4f25-a218-04587de07a7c {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1259.830767] env[61806]: DEBUG oslo_vmware.api [None req-28105e44-4916-46ed-961c-1d7952a7ef4d tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Waiting for the task: (returnval){ [ 1259.830767] env[61806]: value = "task-1295330" [ 1259.830767] env[61806]: _type = "Task" [ 1259.830767] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1259.837422] env[61806]: DEBUG oslo_vmware.api [None req-28105e44-4916-46ed-961c-1d7952a7ef4d tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Task: {'id': task-1295330, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1260.341141] env[61806]: DEBUG oslo_vmware.api [None req-28105e44-4916-46ed-961c-1d7952a7ef4d tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Task: {'id': task-1295330, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.084399} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1260.341512] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-28105e44-4916-46ed-961c-1d7952a7ef4d tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] [instance: 7afc6260-7569-44c4-9764-7aa08c662c1c] Extended root virtual disk {{(pid=61806) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1260.342191] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3714693a-a1fc-49dd-a741-39b9ac1be785 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1260.363138] env[61806]: DEBUG nova.virt.vmwareapi.volumeops [None req-28105e44-4916-46ed-961c-1d7952a7ef4d tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] [instance: 7afc6260-7569-44c4-9764-7aa08c662c1c] Reconfiguring VM instance instance-0000006d to attach disk [datastore1] 7afc6260-7569-44c4-9764-7aa08c662c1c/7afc6260-7569-44c4-9764-7aa08c662c1c.vmdk or device None with type sparse {{(pid=61806) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1260.363363] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-8603f54b-0e8e-4694-8d94-89ea98844235 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1260.382803] env[61806]: DEBUG oslo_vmware.api [None req-28105e44-4916-46ed-961c-1d7952a7ef4d tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Waiting for the task: (returnval){ [ 1260.382803] env[61806]: value = "task-1295331" [ 1260.382803] env[61806]: _type = "Task" [ 1260.382803] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1260.389938] env[61806]: DEBUG oslo_vmware.api [None req-28105e44-4916-46ed-961c-1d7952a7ef4d tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Task: {'id': task-1295331, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1260.892251] env[61806]: DEBUG oslo_vmware.api [None req-28105e44-4916-46ed-961c-1d7952a7ef4d tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Task: {'id': task-1295331, 'name': ReconfigVM_Task, 'duration_secs': 0.269411} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1260.892532] env[61806]: DEBUG nova.virt.vmwareapi.volumeops [None req-28105e44-4916-46ed-961c-1d7952a7ef4d tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] [instance: 7afc6260-7569-44c4-9764-7aa08c662c1c] Reconfigured VM instance instance-0000006d to attach disk [datastore1] 7afc6260-7569-44c4-9764-7aa08c662c1c/7afc6260-7569-44c4-9764-7aa08c662c1c.vmdk or device None with type sparse {{(pid=61806) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1260.893143] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-e3652370-a382-4ef8-b487-99f2fd3d7fa2 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1260.899588] env[61806]: DEBUG oslo_vmware.api [None req-28105e44-4916-46ed-961c-1d7952a7ef4d tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Waiting for the task: (returnval){ [ 1260.899588] env[61806]: value = "task-1295332" [ 1260.899588] env[61806]: _type = "Task" [ 1260.899588] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1260.906809] env[61806]: DEBUG oslo_vmware.api [None req-28105e44-4916-46ed-961c-1d7952a7ef4d tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Task: {'id': task-1295332, 'name': Rename_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1261.409834] env[61806]: DEBUG oslo_vmware.api [None req-28105e44-4916-46ed-961c-1d7952a7ef4d tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Task: {'id': task-1295332, 'name': Rename_Task, 'duration_secs': 0.136302} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1261.410292] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [None req-28105e44-4916-46ed-961c-1d7952a7ef4d tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] [instance: 7afc6260-7569-44c4-9764-7aa08c662c1c] Powering on the VM {{(pid=61806) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1261.410388] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-19b30c28-ddc1-464e-b46a-fd59d14e3e9c {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1261.416140] env[61806]: DEBUG oslo_vmware.api [None req-28105e44-4916-46ed-961c-1d7952a7ef4d tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Waiting for the task: (returnval){ [ 1261.416140] env[61806]: value = "task-1295333" [ 1261.416140] env[61806]: _type = "Task" [ 1261.416140] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1261.424270] env[61806]: DEBUG oslo_vmware.api [None req-28105e44-4916-46ed-961c-1d7952a7ef4d tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Task: {'id': task-1295333, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1261.925703] env[61806]: DEBUG oslo_vmware.api [None req-28105e44-4916-46ed-961c-1d7952a7ef4d tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Task: {'id': task-1295333, 'name': PowerOnVM_Task, 'duration_secs': 0.430477} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1261.925965] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [None req-28105e44-4916-46ed-961c-1d7952a7ef4d tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] [instance: 7afc6260-7569-44c4-9764-7aa08c662c1c] Powered on the VM {{(pid=61806) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1261.926197] env[61806]: INFO nova.compute.manager [None req-28105e44-4916-46ed-961c-1d7952a7ef4d tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] [instance: 7afc6260-7569-44c4-9764-7aa08c662c1c] Took 7.80 seconds to spawn the instance on the hypervisor. [ 1261.926395] env[61806]: DEBUG nova.compute.manager [None req-28105e44-4916-46ed-961c-1d7952a7ef4d tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] [instance: 7afc6260-7569-44c4-9764-7aa08c662c1c] Checking state {{(pid=61806) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1261.927133] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bc1af336-80bc-4634-8fa0-b8fc3794b2c8 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1262.022727] env[61806]: DEBUG oslo_service.periodic_task [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=61806) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1262.447685] env[61806]: INFO nova.compute.manager [None req-28105e44-4916-46ed-961c-1d7952a7ef4d tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] [instance: 7afc6260-7569-44c4-9764-7aa08c662c1c] Took 13.13 seconds to build instance. [ 1262.700521] env[61806]: DEBUG nova.compute.manager [req-81eb11b9-ccab-46da-a0cc-2dd65111b8e1 req-58a60922-fd5f-4b76-b8c6-8344d2f13373 service nova] [instance: 7afc6260-7569-44c4-9764-7aa08c662c1c] Received event network-changed-ee859b12-498b-43fe-97f1-47c250ada436 {{(pid=61806) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1262.700735] env[61806]: DEBUG nova.compute.manager [req-81eb11b9-ccab-46da-a0cc-2dd65111b8e1 req-58a60922-fd5f-4b76-b8c6-8344d2f13373 service nova] [instance: 7afc6260-7569-44c4-9764-7aa08c662c1c] Refreshing instance network info cache due to event network-changed-ee859b12-498b-43fe-97f1-47c250ada436. {{(pid=61806) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 1262.700986] env[61806]: DEBUG oslo_concurrency.lockutils [req-81eb11b9-ccab-46da-a0cc-2dd65111b8e1 req-58a60922-fd5f-4b76-b8c6-8344d2f13373 service nova] Acquiring lock "refresh_cache-7afc6260-7569-44c4-9764-7aa08c662c1c" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1262.701502] env[61806]: DEBUG oslo_concurrency.lockutils [req-81eb11b9-ccab-46da-a0cc-2dd65111b8e1 req-58a60922-fd5f-4b76-b8c6-8344d2f13373 service nova] Acquired lock "refresh_cache-7afc6260-7569-44c4-9764-7aa08c662c1c" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1262.701502] env[61806]: DEBUG nova.network.neutron [req-81eb11b9-ccab-46da-a0cc-2dd65111b8e1 req-58a60922-fd5f-4b76-b8c6-8344d2f13373 service nova] [instance: 7afc6260-7569-44c4-9764-7aa08c662c1c] Refreshing network info cache for port ee859b12-498b-43fe-97f1-47c250ada436 {{(pid=61806) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1262.950150] env[61806]: DEBUG oslo_concurrency.lockutils [None req-28105e44-4916-46ed-961c-1d7952a7ef4d tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Lock "7afc6260-7569-44c4-9764-7aa08c662c1c" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 14.643s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1263.415044] env[61806]: DEBUG nova.network.neutron [req-81eb11b9-ccab-46da-a0cc-2dd65111b8e1 req-58a60922-fd5f-4b76-b8c6-8344d2f13373 service nova] [instance: 7afc6260-7569-44c4-9764-7aa08c662c1c] Updated VIF entry in instance network info cache for port ee859b12-498b-43fe-97f1-47c250ada436. {{(pid=61806) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1263.415445] env[61806]: DEBUG nova.network.neutron [req-81eb11b9-ccab-46da-a0cc-2dd65111b8e1 req-58a60922-fd5f-4b76-b8c6-8344d2f13373 service nova] [instance: 7afc6260-7569-44c4-9764-7aa08c662c1c] Updating instance_info_cache with network_info: [{"id": "ee859b12-498b-43fe-97f1-47c250ada436", "address": "fa:16:3e:6d:6f:15", "network": {"id": "33bf5b21-82b0-432e-ae3d-9c2ca75161d0", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-1294159336-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.187", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "1c55cc0455584fbb8cda88d33b8f8b7e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "5b0fa7a2-ebd9-4788-8904-7bf250ce466c", "external-id": "nsx-vlan-transportzone-669", "segmentation_id": 669, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapee859b12-49", "ovs_interfaceid": "ee859b12-498b-43fe-97f1-47c250ada436", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61806) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1263.918117] env[61806]: DEBUG oslo_concurrency.lockutils [req-81eb11b9-ccab-46da-a0cc-2dd65111b8e1 req-58a60922-fd5f-4b76-b8c6-8344d2f13373 service nova] Releasing lock "refresh_cache-7afc6260-7569-44c4-9764-7aa08c662c1c" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1300.421557] env[61806]: DEBUG oslo_concurrency.lockutils [None req-f381d6bb-88c6-4f40-b94c-c38f50132080 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Acquiring lock "7afc6260-7569-44c4-9764-7aa08c662c1c" by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1300.421973] env[61806]: DEBUG oslo_concurrency.lockutils [None req-f381d6bb-88c6-4f40-b94c-c38f50132080 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Lock "7afc6260-7569-44c4-9764-7aa08c662c1c" acquired by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" :: waited 0.001s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1300.422047] env[61806]: INFO nova.compute.manager [None req-f381d6bb-88c6-4f40-b94c-c38f50132080 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] [instance: 7afc6260-7569-44c4-9764-7aa08c662c1c] Shelving [ 1300.930957] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [None req-f381d6bb-88c6-4f40-b94c-c38f50132080 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] [instance: 7afc6260-7569-44c4-9764-7aa08c662c1c] Powering off the VM {{(pid=61806) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1300.931340] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-4e637118-7f97-436c-a7fb-892b7c9d16c2 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1300.939289] env[61806]: DEBUG oslo_vmware.api [None req-f381d6bb-88c6-4f40-b94c-c38f50132080 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Waiting for the task: (returnval){ [ 1300.939289] env[61806]: value = "task-1295334" [ 1300.939289] env[61806]: _type = "Task" [ 1300.939289] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1300.948756] env[61806]: DEBUG oslo_vmware.api [None req-f381d6bb-88c6-4f40-b94c-c38f50132080 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Task: {'id': task-1295334, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1301.448872] env[61806]: DEBUG oslo_vmware.api [None req-f381d6bb-88c6-4f40-b94c-c38f50132080 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Task: {'id': task-1295334, 'name': PowerOffVM_Task, 'duration_secs': 0.180787} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1301.449184] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [None req-f381d6bb-88c6-4f40-b94c-c38f50132080 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] [instance: 7afc6260-7569-44c4-9764-7aa08c662c1c] Powered off the VM {{(pid=61806) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1301.449914] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ed2aab11-deba-4906-9332-f5b3343322a3 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1301.467917] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fc8ca5dc-594d-44ed-8ccf-95c9dac803e7 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1301.977442] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-f381d6bb-88c6-4f40-b94c-c38f50132080 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] [instance: 7afc6260-7569-44c4-9764-7aa08c662c1c] Creating Snapshot of the VM instance {{(pid=61806) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 1301.977767] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-cd37e001-7d91-4c43-b46c-538b8e4daa80 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1301.985020] env[61806]: DEBUG oslo_vmware.api [None req-f381d6bb-88c6-4f40-b94c-c38f50132080 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Waiting for the task: (returnval){ [ 1301.985020] env[61806]: value = "task-1295335" [ 1301.985020] env[61806]: _type = "Task" [ 1301.985020] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1301.993175] env[61806]: DEBUG oslo_vmware.api [None req-f381d6bb-88c6-4f40-b94c-c38f50132080 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Task: {'id': task-1295335, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1302.495472] env[61806]: DEBUG oslo_vmware.api [None req-f381d6bb-88c6-4f40-b94c-c38f50132080 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Task: {'id': task-1295335, 'name': CreateSnapshot_Task, 'duration_secs': 0.444402} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1302.495922] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-f381d6bb-88c6-4f40-b94c-c38f50132080 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] [instance: 7afc6260-7569-44c4-9764-7aa08c662c1c] Created Snapshot of the VM instance {{(pid=61806) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 1302.496434] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d822dc98-5cae-4e05-91aa-e454de127e18 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1303.013061] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-f381d6bb-88c6-4f40-b94c-c38f50132080 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] [instance: 7afc6260-7569-44c4-9764-7aa08c662c1c] Creating linked-clone VM from snapshot {{(pid=61806) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 1303.013435] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-ae669ff0-01ba-4ef0-92f8-c5122e7f40d3 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1303.024724] env[61806]: DEBUG oslo_vmware.api [None req-f381d6bb-88c6-4f40-b94c-c38f50132080 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Waiting for the task: (returnval){ [ 1303.024724] env[61806]: value = "task-1295336" [ 1303.024724] env[61806]: _type = "Task" [ 1303.024724] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1303.032772] env[61806]: DEBUG oslo_vmware.api [None req-f381d6bb-88c6-4f40-b94c-c38f50132080 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Task: {'id': task-1295336, 'name': CloneVM_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1303.535047] env[61806]: DEBUG oslo_vmware.api [None req-f381d6bb-88c6-4f40-b94c-c38f50132080 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Task: {'id': task-1295336, 'name': CloneVM_Task} progress is 94%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1304.035589] env[61806]: DEBUG oslo_vmware.api [None req-f381d6bb-88c6-4f40-b94c-c38f50132080 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Task: {'id': task-1295336, 'name': CloneVM_Task, 'duration_secs': 0.822047} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1304.035859] env[61806]: INFO nova.virt.vmwareapi.vmops [None req-f381d6bb-88c6-4f40-b94c-c38f50132080 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] [instance: 7afc6260-7569-44c4-9764-7aa08c662c1c] Created linked-clone VM from snapshot [ 1304.036585] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eef60f1a-109e-41e1-a575-46f6948a5d7f {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1304.043438] env[61806]: DEBUG nova.virt.vmwareapi.images [None req-f381d6bb-88c6-4f40-b94c-c38f50132080 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] [instance: 7afc6260-7569-44c4-9764-7aa08c662c1c] Uploading image 23d238dd-6fe7-4d3d-a822-8104bb3bc291 {{(pid=61806) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:442}} [ 1304.064180] env[61806]: DEBUG oslo_vmware.rw_handles [None req-f381d6bb-88c6-4f40-b94c-c38f50132080 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Creating HttpNfcLease lease for exporting VM: (result){ [ 1304.064180] env[61806]: value = "vm-277776" [ 1304.064180] env[61806]: _type = "VirtualMachine" [ 1304.064180] env[61806]: }. {{(pid=61806) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:478}} [ 1304.064412] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ExportVm with opID=oslo.vmware-291b1f9f-affb-43c5-a3d1-1cfd89e5991f {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1304.071006] env[61806]: DEBUG oslo_vmware.rw_handles [None req-f381d6bb-88c6-4f40-b94c-c38f50132080 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Lease: (returnval){ [ 1304.071006] env[61806]: value = "session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]52885486-3ba8-f2c3-578f-d423797a2a98" [ 1304.071006] env[61806]: _type = "HttpNfcLease" [ 1304.071006] env[61806]: } obtained for exporting VM: (result){ [ 1304.071006] env[61806]: value = "vm-277776" [ 1304.071006] env[61806]: _type = "VirtualMachine" [ 1304.071006] env[61806]: }. {{(pid=61806) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:481}} [ 1304.071237] env[61806]: DEBUG oslo_vmware.api [None req-f381d6bb-88c6-4f40-b94c-c38f50132080 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Waiting for the lease: (returnval){ [ 1304.071237] env[61806]: value = "session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]52885486-3ba8-f2c3-578f-d423797a2a98" [ 1304.071237] env[61806]: _type = "HttpNfcLease" [ 1304.071237] env[61806]: } to be ready. {{(pid=61806) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 1304.076901] env[61806]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1304.076901] env[61806]: value = "session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]52885486-3ba8-f2c3-578f-d423797a2a98" [ 1304.076901] env[61806]: _type = "HttpNfcLease" [ 1304.076901] env[61806]: } is initializing. {{(pid=61806) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 1304.579105] env[61806]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1304.579105] env[61806]: value = "session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]52885486-3ba8-f2c3-578f-d423797a2a98" [ 1304.579105] env[61806]: _type = "HttpNfcLease" [ 1304.579105] env[61806]: } is ready. {{(pid=61806) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 1304.579542] env[61806]: DEBUG oslo_vmware.rw_handles [None req-f381d6bb-88c6-4f40-b94c-c38f50132080 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 1304.579542] env[61806]: value = "session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]52885486-3ba8-f2c3-578f-d423797a2a98" [ 1304.579542] env[61806]: _type = "HttpNfcLease" [ 1304.579542] env[61806]: }. {{(pid=61806) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:486}} [ 1304.580127] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8ef2a898-44c5-4921-a973-4d1b9a338545 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1304.588620] env[61806]: DEBUG oslo_vmware.rw_handles [None req-f381d6bb-88c6-4f40-b94c-c38f50132080 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Found VMDK URL: https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52fa3424-fb54-0975-2815-1ebb504df566/disk-0.vmdk from lease info. {{(pid=61806) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 1304.588795] env[61806]: DEBUG oslo_vmware.rw_handles [None req-f381d6bb-88c6-4f40-b94c-c38f50132080 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Opening URL: https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52fa3424-fb54-0975-2815-1ebb504df566/disk-0.vmdk for reading. {{(pid=61806) _create_read_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:99}} [ 1304.677472] env[61806]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-86438673-e724-4e4e-acac-d68b509157a0 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1311.018908] env[61806]: DEBUG oslo_service.periodic_task [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=61806) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1311.448978] env[61806]: DEBUG oslo_vmware.rw_handles [None req-f381d6bb-88c6-4f40-b94c-c38f50132080 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Getting lease state for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52fa3424-fb54-0975-2815-1ebb504df566/disk-0.vmdk. {{(pid=61806) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 1311.449910] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1f634c15-667c-405d-a498-5f2bec4341e6 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1311.456288] env[61806]: DEBUG oslo_vmware.rw_handles [None req-f381d6bb-88c6-4f40-b94c-c38f50132080 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52fa3424-fb54-0975-2815-1ebb504df566/disk-0.vmdk is in state: ready. {{(pid=61806) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 1311.456466] env[61806]: ERROR oslo_vmware.rw_handles [None req-f381d6bb-88c6-4f40-b94c-c38f50132080 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Aborting lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52fa3424-fb54-0975-2815-1ebb504df566/disk-0.vmdk due to incomplete transfer. [ 1311.456676] env[61806]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseAbort with opID=oslo.vmware-13eb388a-c08e-43ef-b532-1f7625eab245 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1311.464192] env[61806]: DEBUG oslo_vmware.rw_handles [None req-f381d6bb-88c6-4f40-b94c-c38f50132080 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Closed VMDK read handle for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52fa3424-fb54-0975-2815-1ebb504df566/disk-0.vmdk. {{(pid=61806) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:735}} [ 1311.464389] env[61806]: DEBUG nova.virt.vmwareapi.images [None req-f381d6bb-88c6-4f40-b94c-c38f50132080 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] [instance: 7afc6260-7569-44c4-9764-7aa08c662c1c] Uploaded image 23d238dd-6fe7-4d3d-a822-8104bb3bc291 to the Glance image server {{(pid=61806) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:473}} [ 1311.466570] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [None req-f381d6bb-88c6-4f40-b94c-c38f50132080 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] [instance: 7afc6260-7569-44c4-9764-7aa08c662c1c] Destroying the VM {{(pid=61806) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1358}} [ 1311.466804] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-260912aa-3363-41e0-8b87-665fa1d2272e {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1311.472596] env[61806]: DEBUG oslo_vmware.api [None req-f381d6bb-88c6-4f40-b94c-c38f50132080 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Waiting for the task: (returnval){ [ 1311.472596] env[61806]: value = "task-1295338" [ 1311.472596] env[61806]: _type = "Task" [ 1311.472596] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1311.479990] env[61806]: DEBUG oslo_vmware.api [None req-f381d6bb-88c6-4f40-b94c-c38f50132080 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Task: {'id': task-1295338, 'name': Destroy_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1311.982356] env[61806]: DEBUG oslo_vmware.api [None req-f381d6bb-88c6-4f40-b94c-c38f50132080 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Task: {'id': task-1295338, 'name': Destroy_Task, 'duration_secs': 0.351493} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1311.982623] env[61806]: INFO nova.virt.vmwareapi.vm_util [None req-f381d6bb-88c6-4f40-b94c-c38f50132080 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] [instance: 7afc6260-7569-44c4-9764-7aa08c662c1c] Destroyed the VM [ 1311.982867] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-f381d6bb-88c6-4f40-b94c-c38f50132080 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] [instance: 7afc6260-7569-44c4-9764-7aa08c662c1c] Deleting Snapshot of the VM instance {{(pid=61806) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 1311.983132] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-8b41dd23-8d51-4ead-9008-b09d8e2b2df3 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1311.988714] env[61806]: DEBUG oslo_vmware.api [None req-f381d6bb-88c6-4f40-b94c-c38f50132080 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Waiting for the task: (returnval){ [ 1311.988714] env[61806]: value = "task-1295339" [ 1311.988714] env[61806]: _type = "Task" [ 1311.988714] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1311.995528] env[61806]: DEBUG oslo_vmware.api [None req-f381d6bb-88c6-4f40-b94c-c38f50132080 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Task: {'id': task-1295339, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1312.498311] env[61806]: DEBUG oslo_vmware.api [None req-f381d6bb-88c6-4f40-b94c-c38f50132080 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Task: {'id': task-1295339, 'name': RemoveSnapshot_Task, 'duration_secs': 0.382833} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1312.498672] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-f381d6bb-88c6-4f40-b94c-c38f50132080 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] [instance: 7afc6260-7569-44c4-9764-7aa08c662c1c] Deleted Snapshot of the VM instance {{(pid=61806) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 1312.498834] env[61806]: DEBUG nova.compute.manager [None req-f381d6bb-88c6-4f40-b94c-c38f50132080 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] [instance: 7afc6260-7569-44c4-9764-7aa08c662c1c] Checking state {{(pid=61806) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1312.499643] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d274e0dd-e766-4306-b0bf-5dc4bb2bd759 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1313.011481] env[61806]: INFO nova.compute.manager [None req-f381d6bb-88c6-4f40-b94c-c38f50132080 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] [instance: 7afc6260-7569-44c4-9764-7aa08c662c1c] Shelve offloading [ 1313.013161] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [None req-f381d6bb-88c6-4f40-b94c-c38f50132080 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] [instance: 7afc6260-7569-44c4-9764-7aa08c662c1c] Powering off the VM {{(pid=61806) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1313.013450] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-e4738cbe-4707-471c-9404-46107706dcad {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1313.020976] env[61806]: DEBUG oslo_vmware.api [None req-f381d6bb-88c6-4f40-b94c-c38f50132080 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Waiting for the task: (returnval){ [ 1313.020976] env[61806]: value = "task-1295340" [ 1313.020976] env[61806]: _type = "Task" [ 1313.020976] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1313.024050] env[61806]: DEBUG oslo_service.periodic_task [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Running periodic task ComputeManager.update_available_resource {{(pid=61806) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1313.031098] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [None req-f381d6bb-88c6-4f40-b94c-c38f50132080 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] [instance: 7afc6260-7569-44c4-9764-7aa08c662c1c] VM already powered off {{(pid=61806) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1509}} [ 1313.031293] env[61806]: DEBUG nova.compute.manager [None req-f381d6bb-88c6-4f40-b94c-c38f50132080 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] [instance: 7afc6260-7569-44c4-9764-7aa08c662c1c] Checking state {{(pid=61806) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1313.032041] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5df13a21-73a8-4a74-b781-72416cc6cbe9 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1313.037658] env[61806]: DEBUG oslo_concurrency.lockutils [None req-f381d6bb-88c6-4f40-b94c-c38f50132080 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Acquiring lock "refresh_cache-7afc6260-7569-44c4-9764-7aa08c662c1c" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1313.037828] env[61806]: DEBUG oslo_concurrency.lockutils [None req-f381d6bb-88c6-4f40-b94c-c38f50132080 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Acquired lock "refresh_cache-7afc6260-7569-44c4-9764-7aa08c662c1c" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1313.038012] env[61806]: DEBUG nova.network.neutron [None req-f381d6bb-88c6-4f40-b94c-c38f50132080 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] [instance: 7afc6260-7569-44c4-9764-7aa08c662c1c] Building network info cache for instance {{(pid=61806) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1313.527876] env[61806]: DEBUG oslo_concurrency.lockutils [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1313.528261] env[61806]: DEBUG oslo_concurrency.lockutils [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 0.000s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1313.528382] env[61806]: DEBUG oslo_concurrency.lockutils [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1313.528535] env[61806]: DEBUG nova.compute.resource_tracker [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=61806) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 1313.529457] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2712e462-f388-4a18-83e0-c89a59acc3f1 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1313.537533] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b3e63396-fa31-4d55-ba99-1760404fe609 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1313.552618] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d082d285-cf89-48f0-a705-01c79f00931b {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1313.558417] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-88954d43-758c-429f-8dfe-a665db9dda13 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1313.587399] env[61806]: DEBUG nova.compute.resource_tracker [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=181311MB free_disk=138GB free_vcpus=48 pci_devices=None {{(pid=61806) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 1313.587520] env[61806]: DEBUG oslo_concurrency.lockutils [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1313.587784] env[61806]: DEBUG oslo_concurrency.lockutils [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 0.000s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1313.756900] env[61806]: DEBUG nova.network.neutron [None req-f381d6bb-88c6-4f40-b94c-c38f50132080 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] [instance: 7afc6260-7569-44c4-9764-7aa08c662c1c] Updating instance_info_cache with network_info: [{"id": "ee859b12-498b-43fe-97f1-47c250ada436", "address": "fa:16:3e:6d:6f:15", "network": {"id": "33bf5b21-82b0-432e-ae3d-9c2ca75161d0", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-1294159336-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.187", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "1c55cc0455584fbb8cda88d33b8f8b7e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "5b0fa7a2-ebd9-4788-8904-7bf250ce466c", "external-id": "nsx-vlan-transportzone-669", "segmentation_id": 669, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapee859b12-49", "ovs_interfaceid": "ee859b12-498b-43fe-97f1-47c250ada436", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61806) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1314.260224] env[61806]: DEBUG oslo_concurrency.lockutils [None req-f381d6bb-88c6-4f40-b94c-c38f50132080 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Releasing lock "refresh_cache-7afc6260-7569-44c4-9764-7aa08c662c1c" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1314.466193] env[61806]: DEBUG nova.compute.manager [req-af90af4e-01d1-4685-b5b8-e24d8c803487 req-ebd95376-c772-486d-9699-945302f642d8 service nova] [instance: 7afc6260-7569-44c4-9764-7aa08c662c1c] Received event network-vif-unplugged-ee859b12-498b-43fe-97f1-47c250ada436 {{(pid=61806) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1314.466446] env[61806]: DEBUG oslo_concurrency.lockutils [req-af90af4e-01d1-4685-b5b8-e24d8c803487 req-ebd95376-c772-486d-9699-945302f642d8 service nova] Acquiring lock "7afc6260-7569-44c4-9764-7aa08c662c1c-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1314.466667] env[61806]: DEBUG oslo_concurrency.lockutils [req-af90af4e-01d1-4685-b5b8-e24d8c803487 req-ebd95376-c772-486d-9699-945302f642d8 service nova] Lock "7afc6260-7569-44c4-9764-7aa08c662c1c-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1314.466841] env[61806]: DEBUG oslo_concurrency.lockutils [req-af90af4e-01d1-4685-b5b8-e24d8c803487 req-ebd95376-c772-486d-9699-945302f642d8 service nova] Lock "7afc6260-7569-44c4-9764-7aa08c662c1c-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1314.467147] env[61806]: DEBUG nova.compute.manager [req-af90af4e-01d1-4685-b5b8-e24d8c803487 req-ebd95376-c772-486d-9699-945302f642d8 service nova] [instance: 7afc6260-7569-44c4-9764-7aa08c662c1c] No waiting events found dispatching network-vif-unplugged-ee859b12-498b-43fe-97f1-47c250ada436 {{(pid=61806) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1314.467373] env[61806]: WARNING nova.compute.manager [req-af90af4e-01d1-4685-b5b8-e24d8c803487 req-ebd95376-c772-486d-9699-945302f642d8 service nova] [instance: 7afc6260-7569-44c4-9764-7aa08c662c1c] Received unexpected event network-vif-unplugged-ee859b12-498b-43fe-97f1-47c250ada436 for instance with vm_state shelved and task_state shelving_offloading. [ 1314.559511] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-f381d6bb-88c6-4f40-b94c-c38f50132080 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] [instance: 7afc6260-7569-44c4-9764-7aa08c662c1c] Destroying instance {{(pid=61806) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1314.560494] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aecf207d-c1a6-4f6f-a6a5-8d46bfe504df {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1314.567896] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-f381d6bb-88c6-4f40-b94c-c38f50132080 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] [instance: 7afc6260-7569-44c4-9764-7aa08c662c1c] Unregistering the VM {{(pid=61806) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1314.568145] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-f914b9fa-3498-43ba-bcc3-df85562eecf1 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1314.613739] env[61806]: DEBUG nova.compute.resource_tracker [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Instance 7afc6260-7569-44c4-9764-7aa08c662c1c actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61806) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1314.614052] env[61806]: DEBUG nova.compute.resource_tracker [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Total usable vcpus: 48, total allocated vcpus: 1 {{(pid=61806) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 1314.614332] env[61806]: DEBUG nova.compute.resource_tracker [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=704MB phys_disk=200GB used_disk=1GB total_vcpus=48 used_vcpus=1 pci_stats=[] {{(pid=61806) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 1314.628738] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-f381d6bb-88c6-4f40-b94c-c38f50132080 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] [instance: 7afc6260-7569-44c4-9764-7aa08c662c1c] Unregistered the VM {{(pid=61806) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1314.628969] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-f381d6bb-88c6-4f40-b94c-c38f50132080 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] [instance: 7afc6260-7569-44c4-9764-7aa08c662c1c] Deleting contents of the VM from datastore datastore1 {{(pid=61806) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1314.629177] env[61806]: DEBUG nova.virt.vmwareapi.ds_util [None req-f381d6bb-88c6-4f40-b94c-c38f50132080 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Deleting the datastore file [datastore1] 7afc6260-7569-44c4-9764-7aa08c662c1c {{(pid=61806) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1314.629448] env[61806]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-585f70b9-ce1d-4138-bd9e-9077c312e50d {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1314.638181] env[61806]: DEBUG oslo_vmware.api [None req-f381d6bb-88c6-4f40-b94c-c38f50132080 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Waiting for the task: (returnval){ [ 1314.638181] env[61806]: value = "task-1295342" [ 1314.638181] env[61806]: _type = "Task" [ 1314.638181] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1314.646586] env[61806]: DEBUG oslo_vmware.api [None req-f381d6bb-88c6-4f40-b94c-c38f50132080 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Task: {'id': task-1295342, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1314.648102] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6c9f0805-2855-4156-9b07-259ca51a685e {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1314.654307] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ed23f1f4-980f-489f-b26b-cb33a44d7220 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1314.684814] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-98483b51-e4c5-43f6-a510-b1ddaf192ce3 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1314.691585] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-25e58963-b1e6-475f-9db0-09b0481d49b6 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1314.704157] env[61806]: DEBUG nova.compute.provider_tree [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Updating inventory in ProviderTree for provider a2858be1-fd22-4e08-979e-87ad25293407 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 138, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61806) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1315.147815] env[61806]: DEBUG oslo_vmware.api [None req-f381d6bb-88c6-4f40-b94c-c38f50132080 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Task: {'id': task-1295342, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.141709} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1315.148086] env[61806]: DEBUG nova.virt.vmwareapi.ds_util [None req-f381d6bb-88c6-4f40-b94c-c38f50132080 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Deleted the datastore file {{(pid=61806) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1315.148311] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-f381d6bb-88c6-4f40-b94c-c38f50132080 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] [instance: 7afc6260-7569-44c4-9764-7aa08c662c1c] Deleted contents of the VM from datastore datastore1 {{(pid=61806) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1315.148522] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-f381d6bb-88c6-4f40-b94c-c38f50132080 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] [instance: 7afc6260-7569-44c4-9764-7aa08c662c1c] Instance destroyed {{(pid=61806) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1315.167830] env[61806]: INFO nova.scheduler.client.report [None req-f381d6bb-88c6-4f40-b94c-c38f50132080 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Deleted allocations for instance 7afc6260-7569-44c4-9764-7aa08c662c1c [ 1315.232375] env[61806]: DEBUG nova.scheduler.client.report [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Updated inventory for provider a2858be1-fd22-4e08-979e-87ad25293407 with generation 152 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 138, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61806) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:971}} [ 1315.232599] env[61806]: DEBUG nova.compute.provider_tree [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Updating resource provider a2858be1-fd22-4e08-979e-87ad25293407 generation from 152 to 153 during operation: update_inventory {{(pid=61806) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 1315.232749] env[61806]: DEBUG nova.compute.provider_tree [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Updating inventory in ProviderTree for provider a2858be1-fd22-4e08-979e-87ad25293407 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 138, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61806) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1315.672829] env[61806]: DEBUG oslo_concurrency.lockutils [None req-f381d6bb-88c6-4f40-b94c-c38f50132080 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1315.737061] env[61806]: DEBUG nova.compute.resource_tracker [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=61806) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 1315.737299] env[61806]: DEBUG oslo_concurrency.lockutils [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 2.150s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1315.737587] env[61806]: DEBUG oslo_concurrency.lockutils [None req-f381d6bb-88c6-4f40-b94c-c38f50132080 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.065s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1315.737855] env[61806]: DEBUG nova.objects.instance [None req-f381d6bb-88c6-4f40-b94c-c38f50132080 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Lazy-loading 'resources' on Instance uuid 7afc6260-7569-44c4-9764-7aa08c662c1c {{(pid=61806) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1316.241084] env[61806]: DEBUG nova.objects.instance [None req-f381d6bb-88c6-4f40-b94c-c38f50132080 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Lazy-loading 'numa_topology' on Instance uuid 7afc6260-7569-44c4-9764-7aa08c662c1c {{(pid=61806) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1316.489524] env[61806]: DEBUG nova.compute.manager [req-6e30349b-4b4c-4db3-b8fe-bb313fa66cac req-d71795ae-842a-4df0-a146-ddbaa664ef1e service nova] [instance: 7afc6260-7569-44c4-9764-7aa08c662c1c] Received event network-changed-ee859b12-498b-43fe-97f1-47c250ada436 {{(pid=61806) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1316.489792] env[61806]: DEBUG nova.compute.manager [req-6e30349b-4b4c-4db3-b8fe-bb313fa66cac req-d71795ae-842a-4df0-a146-ddbaa664ef1e service nova] [instance: 7afc6260-7569-44c4-9764-7aa08c662c1c] Refreshing instance network info cache due to event network-changed-ee859b12-498b-43fe-97f1-47c250ada436. {{(pid=61806) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 1316.489999] env[61806]: DEBUG oslo_concurrency.lockutils [req-6e30349b-4b4c-4db3-b8fe-bb313fa66cac req-d71795ae-842a-4df0-a146-ddbaa664ef1e service nova] Acquiring lock "refresh_cache-7afc6260-7569-44c4-9764-7aa08c662c1c" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1316.490162] env[61806]: DEBUG oslo_concurrency.lockutils [req-6e30349b-4b4c-4db3-b8fe-bb313fa66cac req-d71795ae-842a-4df0-a146-ddbaa664ef1e service nova] Acquired lock "refresh_cache-7afc6260-7569-44c4-9764-7aa08c662c1c" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1316.490334] env[61806]: DEBUG nova.network.neutron [req-6e30349b-4b4c-4db3-b8fe-bb313fa66cac req-d71795ae-842a-4df0-a146-ddbaa664ef1e service nova] [instance: 7afc6260-7569-44c4-9764-7aa08c662c1c] Refreshing network info cache for port ee859b12-498b-43fe-97f1-47c250ada436 {{(pid=61806) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1316.737692] env[61806]: DEBUG oslo_service.periodic_task [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=61806) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1316.738089] env[61806]: DEBUG nova.compute.manager [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Starting heal instance info cache {{(pid=61806) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9926}} [ 1316.738089] env[61806]: DEBUG nova.compute.manager [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Rebuilding the list of instances to heal {{(pid=61806) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9930}} [ 1316.743754] env[61806]: DEBUG nova.objects.base [None req-f381d6bb-88c6-4f40-b94c-c38f50132080 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Object Instance<7afc6260-7569-44c4-9764-7aa08c662c1c> lazy-loaded attributes: resources,numa_topology {{(pid=61806) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 1316.774916] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-709e7e13-a356-4b29-a5b4-42a8e35ec1c7 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1316.783775] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8c0c4bce-e62e-414e-b204-31a61b1e2ea6 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1316.812208] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cd88aaff-7a04-424c-83dc-df01f390627c {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1316.819127] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-82e36173-a054-48d3-9dfd-49f296b5d273 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1316.831620] env[61806]: DEBUG nova.compute.provider_tree [None req-f381d6bb-88c6-4f40-b94c-c38f50132080 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Updating inventory in ProviderTree for provider a2858be1-fd22-4e08-979e-87ad25293407 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 139, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61806) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1316.965109] env[61806]: DEBUG oslo_concurrency.lockutils [None req-aaf896e1-050c-4ea6-8bb3-d9b0acde63e0 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Acquiring lock "7afc6260-7569-44c4-9764-7aa08c662c1c" by "nova.compute.manager.ComputeManager.unshelve_instance..do_unshelve_instance" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1317.190707] env[61806]: DEBUG nova.network.neutron [req-6e30349b-4b4c-4db3-b8fe-bb313fa66cac req-d71795ae-842a-4df0-a146-ddbaa664ef1e service nova] [instance: 7afc6260-7569-44c4-9764-7aa08c662c1c] Updated VIF entry in instance network info cache for port ee859b12-498b-43fe-97f1-47c250ada436. {{(pid=61806) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1317.191076] env[61806]: DEBUG nova.network.neutron [req-6e30349b-4b4c-4db3-b8fe-bb313fa66cac req-d71795ae-842a-4df0-a146-ddbaa664ef1e service nova] [instance: 7afc6260-7569-44c4-9764-7aa08c662c1c] Updating instance_info_cache with network_info: [{"id": "ee859b12-498b-43fe-97f1-47c250ada436", "address": "fa:16:3e:6d:6f:15", "network": {"id": "33bf5b21-82b0-432e-ae3d-9c2ca75161d0", "bridge": null, "label": "tempest-ServerActionsTestOtherB-1294159336-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.187", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "1c55cc0455584fbb8cda88d33b8f8b7e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "unbound", "details": {}, "devname": "tapee859b12-49", "ovs_interfaceid": null, "qbh_params": null, "qbg_params": null, "active": false, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61806) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1317.241441] env[61806]: DEBUG oslo_concurrency.lockutils [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Acquiring lock "refresh_cache-7afc6260-7569-44c4-9764-7aa08c662c1c" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1317.350733] env[61806]: ERROR nova.scheduler.client.report [None req-f381d6bb-88c6-4f40-b94c-c38f50132080 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] [req-accc6695-542c-4257-bb7f-e5bec03ebb48] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 139, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID a2858be1-fd22-4e08-979e-87ad25293407. Got 409: {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-accc6695-542c-4257-bb7f-e5bec03ebb48"}]} [ 1317.365302] env[61806]: DEBUG nova.scheduler.client.report [None req-f381d6bb-88c6-4f40-b94c-c38f50132080 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Refreshing inventories for resource provider a2858be1-fd22-4e08-979e-87ad25293407 {{(pid=61806) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:818}} [ 1317.378930] env[61806]: DEBUG nova.scheduler.client.report [None req-f381d6bb-88c6-4f40-b94c-c38f50132080 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Updating ProviderTree inventory for provider a2858be1-fd22-4e08-979e-87ad25293407 from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 138, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61806) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:782}} [ 1317.379164] env[61806]: DEBUG nova.compute.provider_tree [None req-f381d6bb-88c6-4f40-b94c-c38f50132080 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Updating inventory in ProviderTree for provider a2858be1-fd22-4e08-979e-87ad25293407 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 138, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61806) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1317.389148] env[61806]: DEBUG nova.scheduler.client.report [None req-f381d6bb-88c6-4f40-b94c-c38f50132080 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Refreshing aggregate associations for resource provider a2858be1-fd22-4e08-979e-87ad25293407, aggregates: None {{(pid=61806) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:827}} [ 1317.406269] env[61806]: DEBUG nova.scheduler.client.report [None req-f381d6bb-88c6-4f40-b94c-c38f50132080 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Refreshing trait associations for resource provider a2858be1-fd22-4e08-979e-87ad25293407, traits: COMPUTE_IMAGE_TYPE_ISO,COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_SAME_HOST_COLD_MIGRATE,COMPUTE_NET_ATTACH_INTERFACE,COMPUTE_NODE {{(pid=61806) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:839}} [ 1317.428124] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5a36460e-b912-4c56-9441-3733bdb935d7 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1317.435045] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-646b06f5-dd64-430f-8f4c-128db825bbef {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1317.464666] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7fea96bf-f039-48bb-a11d-a70eae49d2a5 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1317.471117] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fa2ed516-7c1b-49ac-b331-12e7eecf3f8e {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1317.483758] env[61806]: DEBUG nova.compute.provider_tree [None req-f381d6bb-88c6-4f40-b94c-c38f50132080 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Updating inventory in ProviderTree for provider a2858be1-fd22-4e08-979e-87ad25293407 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 139, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61806) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1317.694128] env[61806]: DEBUG oslo_concurrency.lockutils [req-6e30349b-4b4c-4db3-b8fe-bb313fa66cac req-d71795ae-842a-4df0-a146-ddbaa664ef1e service nova] Releasing lock "refresh_cache-7afc6260-7569-44c4-9764-7aa08c662c1c" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1317.694517] env[61806]: DEBUG oslo_concurrency.lockutils [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Acquired lock "refresh_cache-7afc6260-7569-44c4-9764-7aa08c662c1c" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1317.694682] env[61806]: DEBUG nova.network.neutron [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] [instance: 7afc6260-7569-44c4-9764-7aa08c662c1c] Forcefully refreshing network info cache for instance {{(pid=61806) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2004}} [ 1317.694843] env[61806]: DEBUG nova.objects.instance [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Lazy-loading 'info_cache' on Instance uuid 7afc6260-7569-44c4-9764-7aa08c662c1c {{(pid=61806) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1318.012238] env[61806]: DEBUG nova.scheduler.client.report [None req-f381d6bb-88c6-4f40-b94c-c38f50132080 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Updated inventory for provider a2858be1-fd22-4e08-979e-87ad25293407 with generation 154 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 139, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61806) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:971}} [ 1318.012514] env[61806]: DEBUG nova.compute.provider_tree [None req-f381d6bb-88c6-4f40-b94c-c38f50132080 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Updating resource provider a2858be1-fd22-4e08-979e-87ad25293407 generation from 154 to 155 during operation: update_inventory {{(pid=61806) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 1318.012695] env[61806]: DEBUG nova.compute.provider_tree [None req-f381d6bb-88c6-4f40-b94c-c38f50132080 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Updating inventory in ProviderTree for provider a2858be1-fd22-4e08-979e-87ad25293407 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 139, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61806) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1318.518031] env[61806]: DEBUG oslo_concurrency.lockutils [None req-f381d6bb-88c6-4f40-b94c-c38f50132080 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.780s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1319.027835] env[61806]: DEBUG oslo_concurrency.lockutils [None req-f381d6bb-88c6-4f40-b94c-c38f50132080 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Lock "7afc6260-7569-44c4-9764-7aa08c662c1c" "released" by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" :: held 18.606s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1319.028747] env[61806]: DEBUG oslo_concurrency.lockutils [None req-aaf896e1-050c-4ea6-8bb3-d9b0acde63e0 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Lock "7afc6260-7569-44c4-9764-7aa08c662c1c" acquired by "nova.compute.manager.ComputeManager.unshelve_instance..do_unshelve_instance" :: waited 2.064s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1319.028936] env[61806]: INFO nova.compute.manager [None req-aaf896e1-050c-4ea6-8bb3-d9b0acde63e0 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] [instance: 7afc6260-7569-44c4-9764-7aa08c662c1c] Unshelving [ 1319.391039] env[61806]: DEBUG nova.network.neutron [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] [instance: 7afc6260-7569-44c4-9764-7aa08c662c1c] Updating instance_info_cache with network_info: [{"id": "ee859b12-498b-43fe-97f1-47c250ada436", "address": "fa:16:3e:6d:6f:15", "network": {"id": "33bf5b21-82b0-432e-ae3d-9c2ca75161d0", "bridge": null, "label": "tempest-ServerActionsTestOtherB-1294159336-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.187", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "1c55cc0455584fbb8cda88d33b8f8b7e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "unbound", "details": {}, "devname": "tapee859b12-49", "ovs_interfaceid": null, "qbh_params": null, "qbg_params": null, "active": false, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61806) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1319.893240] env[61806]: DEBUG oslo_concurrency.lockutils [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Releasing lock "refresh_cache-7afc6260-7569-44c4-9764-7aa08c662c1c" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1319.893498] env[61806]: DEBUG nova.compute.manager [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] [instance: 7afc6260-7569-44c4-9764-7aa08c662c1c] Updated the network info_cache for instance {{(pid=61806) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9997}} [ 1319.893674] env[61806]: DEBUG oslo_service.periodic_task [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=61806) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1319.893840] env[61806]: DEBUG oslo_service.periodic_task [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=61806) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1319.893999] env[61806]: DEBUG oslo_service.periodic_task [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=61806) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1320.022868] env[61806]: DEBUG oslo_service.periodic_task [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Running periodic task ComputeManager._sync_scheduler_instance_info {{(pid=61806) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1320.052688] env[61806]: DEBUG oslo_concurrency.lockutils [None req-aaf896e1-050c-4ea6-8bb3-d9b0acde63e0 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1320.052965] env[61806]: DEBUG oslo_concurrency.lockutils [None req-aaf896e1-050c-4ea6-8bb3-d9b0acde63e0 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.000s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1320.053197] env[61806]: DEBUG nova.objects.instance [None req-aaf896e1-050c-4ea6-8bb3-d9b0acde63e0 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Lazy-loading 'pci_requests' on Instance uuid 7afc6260-7569-44c4-9764-7aa08c662c1c {{(pid=61806) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1320.527550] env[61806]: DEBUG oslo_service.periodic_task [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=61806) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1320.527765] env[61806]: DEBUG oslo_service.periodic_task [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=61806) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1320.527861] env[61806]: DEBUG nova.compute.manager [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=61806) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10545}} [ 1320.556780] env[61806]: DEBUG nova.objects.instance [None req-aaf896e1-050c-4ea6-8bb3-d9b0acde63e0 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Lazy-loading 'numa_topology' on Instance uuid 7afc6260-7569-44c4-9764-7aa08c662c1c {{(pid=61806) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1321.059057] env[61806]: INFO nova.compute.claims [None req-aaf896e1-050c-4ea6-8bb3-d9b0acde63e0 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] [instance: 7afc6260-7569-44c4-9764-7aa08c662c1c] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1322.094404] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c113aa4d-a53d-4783-9a3a-7119069a3a77 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1322.101759] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0d7f5758-b38c-4d34-8914-a80139ebc931 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1322.130216] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-388b0cf8-d4a3-4866-bf2d-e5f0801fa43f {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1322.136697] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-55d1b5c1-8cf9-487c-8a66-1bd329b84225 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1322.148975] env[61806]: DEBUG nova.compute.provider_tree [None req-aaf896e1-050c-4ea6-8bb3-d9b0acde63e0 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Inventory has not changed in ProviderTree for provider: a2858be1-fd22-4e08-979e-87ad25293407 {{(pid=61806) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1322.652027] env[61806]: DEBUG nova.scheduler.client.report [None req-aaf896e1-050c-4ea6-8bb3-d9b0acde63e0 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Inventory has not changed for provider a2858be1-fd22-4e08-979e-87ad25293407 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 139, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61806) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1323.022950] env[61806]: DEBUG oslo_service.periodic_task [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=61806) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1323.158249] env[61806]: DEBUG oslo_concurrency.lockutils [None req-aaf896e1-050c-4ea6-8bb3-d9b0acde63e0 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 3.105s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1323.185936] env[61806]: INFO nova.network.neutron [None req-aaf896e1-050c-4ea6-8bb3-d9b0acde63e0 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] [instance: 7afc6260-7569-44c4-9764-7aa08c662c1c] Updating port ee859b12-498b-43fe-97f1-47c250ada436 with attributes {'binding:host_id': 'cpu-1', 'device_owner': 'compute:nova'} [ 1324.562758] env[61806]: DEBUG nova.compute.manager [req-f12634f0-4ed2-4ada-8e31-62656132b8fd req-05273bc2-ec07-4106-87aa-29d0422e5474 service nova] [instance: 7afc6260-7569-44c4-9764-7aa08c662c1c] Received event network-vif-plugged-ee859b12-498b-43fe-97f1-47c250ada436 {{(pid=61806) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1324.563032] env[61806]: DEBUG oslo_concurrency.lockutils [req-f12634f0-4ed2-4ada-8e31-62656132b8fd req-05273bc2-ec07-4106-87aa-29d0422e5474 service nova] Acquiring lock "7afc6260-7569-44c4-9764-7aa08c662c1c-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1324.563211] env[61806]: DEBUG oslo_concurrency.lockutils [req-f12634f0-4ed2-4ada-8e31-62656132b8fd req-05273bc2-ec07-4106-87aa-29d0422e5474 service nova] Lock "7afc6260-7569-44c4-9764-7aa08c662c1c-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1324.563408] env[61806]: DEBUG oslo_concurrency.lockutils [req-f12634f0-4ed2-4ada-8e31-62656132b8fd req-05273bc2-ec07-4106-87aa-29d0422e5474 service nova] Lock "7afc6260-7569-44c4-9764-7aa08c662c1c-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1324.563547] env[61806]: DEBUG nova.compute.manager [req-f12634f0-4ed2-4ada-8e31-62656132b8fd req-05273bc2-ec07-4106-87aa-29d0422e5474 service nova] [instance: 7afc6260-7569-44c4-9764-7aa08c662c1c] No waiting events found dispatching network-vif-plugged-ee859b12-498b-43fe-97f1-47c250ada436 {{(pid=61806) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1324.563718] env[61806]: WARNING nova.compute.manager [req-f12634f0-4ed2-4ada-8e31-62656132b8fd req-05273bc2-ec07-4106-87aa-29d0422e5474 service nova] [instance: 7afc6260-7569-44c4-9764-7aa08c662c1c] Received unexpected event network-vif-plugged-ee859b12-498b-43fe-97f1-47c250ada436 for instance with vm_state shelved_offloaded and task_state spawning. [ 1324.644704] env[61806]: DEBUG oslo_concurrency.lockutils [None req-aaf896e1-050c-4ea6-8bb3-d9b0acde63e0 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Acquiring lock "refresh_cache-7afc6260-7569-44c4-9764-7aa08c662c1c" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1324.644990] env[61806]: DEBUG oslo_concurrency.lockutils [None req-aaf896e1-050c-4ea6-8bb3-d9b0acde63e0 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Acquired lock "refresh_cache-7afc6260-7569-44c4-9764-7aa08c662c1c" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1324.645178] env[61806]: DEBUG nova.network.neutron [None req-aaf896e1-050c-4ea6-8bb3-d9b0acde63e0 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] [instance: 7afc6260-7569-44c4-9764-7aa08c662c1c] Building network info cache for instance {{(pid=61806) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1325.337323] env[61806]: DEBUG nova.network.neutron [None req-aaf896e1-050c-4ea6-8bb3-d9b0acde63e0 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] [instance: 7afc6260-7569-44c4-9764-7aa08c662c1c] Updating instance_info_cache with network_info: [{"id": "ee859b12-498b-43fe-97f1-47c250ada436", "address": "fa:16:3e:6d:6f:15", "network": {"id": "33bf5b21-82b0-432e-ae3d-9c2ca75161d0", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-1294159336-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.187", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "1c55cc0455584fbb8cda88d33b8f8b7e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "5b0fa7a2-ebd9-4788-8904-7bf250ce466c", "external-id": "nsx-vlan-transportzone-669", "segmentation_id": 669, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapee859b12-49", "ovs_interfaceid": "ee859b12-498b-43fe-97f1-47c250ada436", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61806) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1325.840050] env[61806]: DEBUG oslo_concurrency.lockutils [None req-aaf896e1-050c-4ea6-8bb3-d9b0acde63e0 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Releasing lock "refresh_cache-7afc6260-7569-44c4-9764-7aa08c662c1c" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1325.865879] env[61806]: DEBUG nova.virt.hardware [None req-aaf896e1-050c-4ea6-8bb3-d9b0acde63e0 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-15T18:05:10Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='2199ffcd1abb30aa9e8967ef005dfd6d',container_format='bare',created_at=2024-10-15T18:20:00Z,direct_url=,disk_format='vmdk',id=23d238dd-6fe7-4d3d-a822-8104bb3bc291,min_disk=1,min_ram=0,name='tempest-ServerActionsTestOtherB-server-2142589452-shelved',owner='1c55cc0455584fbb8cda88d33b8f8b7e',properties=ImageMetaProps,protected=,size=31670272,status='active',tags=,updated_at=2024-10-15T18:20:12Z,virtual_size=,visibility=), allow threads: False {{(pid=61806) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1325.866161] env[61806]: DEBUG nova.virt.hardware [None req-aaf896e1-050c-4ea6-8bb3-d9b0acde63e0 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Flavor limits 0:0:0 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1325.866336] env[61806]: DEBUG nova.virt.hardware [None req-aaf896e1-050c-4ea6-8bb3-d9b0acde63e0 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Image limits 0:0:0 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1325.866551] env[61806]: DEBUG nova.virt.hardware [None req-aaf896e1-050c-4ea6-8bb3-d9b0acde63e0 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Flavor pref 0:0:0 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1325.866709] env[61806]: DEBUG nova.virt.hardware [None req-aaf896e1-050c-4ea6-8bb3-d9b0acde63e0 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Image pref 0:0:0 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1325.866862] env[61806]: DEBUG nova.virt.hardware [None req-aaf896e1-050c-4ea6-8bb3-d9b0acde63e0 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1325.867090] env[61806]: DEBUG nova.virt.hardware [None req-aaf896e1-050c-4ea6-8bb3-d9b0acde63e0 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61806) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1325.867261] env[61806]: DEBUG nova.virt.hardware [None req-aaf896e1-050c-4ea6-8bb3-d9b0acde63e0 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61806) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1325.867433] env[61806]: DEBUG nova.virt.hardware [None req-aaf896e1-050c-4ea6-8bb3-d9b0acde63e0 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Got 1 possible topologies {{(pid=61806) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1325.867602] env[61806]: DEBUG nova.virt.hardware [None req-aaf896e1-050c-4ea6-8bb3-d9b0acde63e0 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61806) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1325.867778] env[61806]: DEBUG nova.virt.hardware [None req-aaf896e1-050c-4ea6-8bb3-d9b0acde63e0 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61806) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1325.868679] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-928b8d98-500b-4b28-8888-fa3e8c0e52e1 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1325.876293] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-67e5e570-455e-42b0-b89a-12ff1aaf862e {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1325.889016] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-aaf896e1-050c-4ea6-8bb3-d9b0acde63e0 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] [instance: 7afc6260-7569-44c4-9764-7aa08c662c1c] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:6d:6f:15', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '5b0fa7a2-ebd9-4788-8904-7bf250ce466c', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'ee859b12-498b-43fe-97f1-47c250ada436', 'vif_model': 'vmxnet3'}] {{(pid=61806) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1325.896179] env[61806]: DEBUG oslo.service.loopingcall [None req-aaf896e1-050c-4ea6-8bb3-d9b0acde63e0 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61806) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1325.896409] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 7afc6260-7569-44c4-9764-7aa08c662c1c] Creating VM on the ESX host {{(pid=61806) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 1325.896608] env[61806]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-49fc96a3-4906-47e6-be0b-f6e63a3580e5 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1325.914599] env[61806]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1325.914599] env[61806]: value = "task-1295343" [ 1325.914599] env[61806]: _type = "Task" [ 1325.914599] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1325.921425] env[61806]: DEBUG oslo_vmware.api [-] Task: {'id': task-1295343, 'name': CreateVM_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1326.424834] env[61806]: DEBUG oslo_vmware.api [-] Task: {'id': task-1295343, 'name': CreateVM_Task, 'duration_secs': 0.276797} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1326.425561] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 7afc6260-7569-44c4-9764-7aa08c662c1c] Created VM on the ESX host {{(pid=61806) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 1326.425968] env[61806]: DEBUG oslo_concurrency.lockutils [None req-aaf896e1-050c-4ea6-8bb3-d9b0acde63e0 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/23d238dd-6fe7-4d3d-a822-8104bb3bc291" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1326.426166] env[61806]: DEBUG oslo_concurrency.lockutils [None req-aaf896e1-050c-4ea6-8bb3-d9b0acde63e0 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Acquired lock "[datastore1] devstack-image-cache_base/23d238dd-6fe7-4d3d-a822-8104bb3bc291" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1326.426568] env[61806]: DEBUG oslo_concurrency.lockutils [None req-aaf896e1-050c-4ea6-8bb3-d9b0acde63e0 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/23d238dd-6fe7-4d3d-a822-8104bb3bc291" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1326.426831] env[61806]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-0be70337-77ae-4bdc-9c16-0b971990659a {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1326.430979] env[61806]: DEBUG oslo_vmware.api [None req-aaf896e1-050c-4ea6-8bb3-d9b0acde63e0 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Waiting for the task: (returnval){ [ 1326.430979] env[61806]: value = "session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]52f4365b-5097-40ac-cfd2-d3dc69e51519" [ 1326.430979] env[61806]: _type = "Task" [ 1326.430979] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1326.437917] env[61806]: DEBUG oslo_vmware.api [None req-aaf896e1-050c-4ea6-8bb3-d9b0acde63e0 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Task: {'id': session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]52f4365b-5097-40ac-cfd2-d3dc69e51519, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1326.587891] env[61806]: DEBUG nova.compute.manager [req-0e2e5a47-d886-4d91-8c47-080a9b0980ef req-742d6334-2a9c-4fe8-9f94-b9121fd78ac5 service nova] [instance: 7afc6260-7569-44c4-9764-7aa08c662c1c] Received event network-changed-ee859b12-498b-43fe-97f1-47c250ada436 {{(pid=61806) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1326.588119] env[61806]: DEBUG nova.compute.manager [req-0e2e5a47-d886-4d91-8c47-080a9b0980ef req-742d6334-2a9c-4fe8-9f94-b9121fd78ac5 service nova] [instance: 7afc6260-7569-44c4-9764-7aa08c662c1c] Refreshing instance network info cache due to event network-changed-ee859b12-498b-43fe-97f1-47c250ada436. {{(pid=61806) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 1326.588339] env[61806]: DEBUG oslo_concurrency.lockutils [req-0e2e5a47-d886-4d91-8c47-080a9b0980ef req-742d6334-2a9c-4fe8-9f94-b9121fd78ac5 service nova] Acquiring lock "refresh_cache-7afc6260-7569-44c4-9764-7aa08c662c1c" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1326.588534] env[61806]: DEBUG oslo_concurrency.lockutils [req-0e2e5a47-d886-4d91-8c47-080a9b0980ef req-742d6334-2a9c-4fe8-9f94-b9121fd78ac5 service nova] Acquired lock "refresh_cache-7afc6260-7569-44c4-9764-7aa08c662c1c" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1326.588711] env[61806]: DEBUG nova.network.neutron [req-0e2e5a47-d886-4d91-8c47-080a9b0980ef req-742d6334-2a9c-4fe8-9f94-b9121fd78ac5 service nova] [instance: 7afc6260-7569-44c4-9764-7aa08c662c1c] Refreshing network info cache for port ee859b12-498b-43fe-97f1-47c250ada436 {{(pid=61806) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1326.940899] env[61806]: DEBUG oslo_concurrency.lockutils [None req-aaf896e1-050c-4ea6-8bb3-d9b0acde63e0 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Releasing lock "[datastore1] devstack-image-cache_base/23d238dd-6fe7-4d3d-a822-8104bb3bc291" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1326.941226] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-aaf896e1-050c-4ea6-8bb3-d9b0acde63e0 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] [instance: 7afc6260-7569-44c4-9764-7aa08c662c1c] Processing image 23d238dd-6fe7-4d3d-a822-8104bb3bc291 {{(pid=61806) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1326.941493] env[61806]: DEBUG oslo_concurrency.lockutils [None req-aaf896e1-050c-4ea6-8bb3-d9b0acde63e0 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/23d238dd-6fe7-4d3d-a822-8104bb3bc291/23d238dd-6fe7-4d3d-a822-8104bb3bc291.vmdk" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1326.941654] env[61806]: DEBUG oslo_concurrency.lockutils [None req-aaf896e1-050c-4ea6-8bb3-d9b0acde63e0 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Acquired lock "[datastore1] devstack-image-cache_base/23d238dd-6fe7-4d3d-a822-8104bb3bc291/23d238dd-6fe7-4d3d-a822-8104bb3bc291.vmdk" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1326.941871] env[61806]: DEBUG nova.virt.vmwareapi.ds_util [None req-aaf896e1-050c-4ea6-8bb3-d9b0acde63e0 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61806) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1326.942157] env[61806]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-72e7c24d-76a5-4452-a2e6-5d4ce38bdc0e {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1326.950333] env[61806]: DEBUG nova.virt.vmwareapi.ds_util [None req-aaf896e1-050c-4ea6-8bb3-d9b0acde63e0 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61806) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1326.950510] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-aaf896e1-050c-4ea6-8bb3-d9b0acde63e0 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61806) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 1326.951225] env[61806]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c083bf77-ec9b-45cb-aa37-6c205be058d1 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1326.955882] env[61806]: DEBUG oslo_vmware.api [None req-aaf896e1-050c-4ea6-8bb3-d9b0acde63e0 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Waiting for the task: (returnval){ [ 1326.955882] env[61806]: value = "session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]52447277-f222-14e1-e43c-e278f6dca58e" [ 1326.955882] env[61806]: _type = "Task" [ 1326.955882] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1326.962845] env[61806]: DEBUG oslo_vmware.api [None req-aaf896e1-050c-4ea6-8bb3-d9b0acde63e0 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Task: {'id': session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]52447277-f222-14e1-e43c-e278f6dca58e, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1327.275251] env[61806]: DEBUG nova.network.neutron [req-0e2e5a47-d886-4d91-8c47-080a9b0980ef req-742d6334-2a9c-4fe8-9f94-b9121fd78ac5 service nova] [instance: 7afc6260-7569-44c4-9764-7aa08c662c1c] Updated VIF entry in instance network info cache for port ee859b12-498b-43fe-97f1-47c250ada436. {{(pid=61806) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1327.275628] env[61806]: DEBUG nova.network.neutron [req-0e2e5a47-d886-4d91-8c47-080a9b0980ef req-742d6334-2a9c-4fe8-9f94-b9121fd78ac5 service nova] [instance: 7afc6260-7569-44c4-9764-7aa08c662c1c] Updating instance_info_cache with network_info: [{"id": "ee859b12-498b-43fe-97f1-47c250ada436", "address": "fa:16:3e:6d:6f:15", "network": {"id": "33bf5b21-82b0-432e-ae3d-9c2ca75161d0", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-1294159336-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.187", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "1c55cc0455584fbb8cda88d33b8f8b7e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "5b0fa7a2-ebd9-4788-8904-7bf250ce466c", "external-id": "nsx-vlan-transportzone-669", "segmentation_id": 669, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapee859b12-49", "ovs_interfaceid": "ee859b12-498b-43fe-97f1-47c250ada436", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61806) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1327.465427] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-aaf896e1-050c-4ea6-8bb3-d9b0acde63e0 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] [instance: 7afc6260-7569-44c4-9764-7aa08c662c1c] Preparing fetch location {{(pid=61806) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:633}} [ 1327.465689] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-aaf896e1-050c-4ea6-8bb3-d9b0acde63e0 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] [instance: 7afc6260-7569-44c4-9764-7aa08c662c1c] Fetch image to [datastore1] OSTACK_IMG_7233a54a-fc8b-4c1c-a4c8-ccaae45e5434/OSTACK_IMG_7233a54a-fc8b-4c1c-a4c8-ccaae45e5434.vmdk {{(pid=61806) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:635}} [ 1327.465935] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-aaf896e1-050c-4ea6-8bb3-d9b0acde63e0 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] [instance: 7afc6260-7569-44c4-9764-7aa08c662c1c] Downloading stream optimized image 23d238dd-6fe7-4d3d-a822-8104bb3bc291 to [datastore1] OSTACK_IMG_7233a54a-fc8b-4c1c-a4c8-ccaae45e5434/OSTACK_IMG_7233a54a-fc8b-4c1c-a4c8-ccaae45e5434.vmdk on the data store datastore1 as vApp {{(pid=61806) _fetch_image_as_vapp /opt/stack/nova/nova/virt/vmwareapi/vmops.py:437}} [ 1327.466142] env[61806]: DEBUG nova.virt.vmwareapi.images [None req-aaf896e1-050c-4ea6-8bb3-d9b0acde63e0 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] [instance: 7afc6260-7569-44c4-9764-7aa08c662c1c] Downloading image file data 23d238dd-6fe7-4d3d-a822-8104bb3bc291 to the ESX as VM named 'OSTACK_IMG_7233a54a-fc8b-4c1c-a4c8-ccaae45e5434' {{(pid=61806) fetch_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:336}} [ 1327.530863] env[61806]: DEBUG oslo_vmware.rw_handles [None req-aaf896e1-050c-4ea6-8bb3-d9b0acde63e0 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Creating HttpNfcLease lease for vApp import into resource pool: (val){ [ 1327.530863] env[61806]: value = "resgroup-9" [ 1327.530863] env[61806]: _type = "ResourcePool" [ 1327.530863] env[61806]: }. {{(pid=61806) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:453}} [ 1327.531179] env[61806]: DEBUG oslo_vmware.service [-] Invoking ResourcePool.ImportVApp with opID=oslo.vmware-c8c67b3f-2dba-454b-9736-4dbacefb7489 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1327.551510] env[61806]: DEBUG oslo_vmware.rw_handles [None req-aaf896e1-050c-4ea6-8bb3-d9b0acde63e0 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Lease: (returnval){ [ 1327.551510] env[61806]: value = "session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]522d494f-4623-c0e3-233a-bd8fa8ca5f00" [ 1327.551510] env[61806]: _type = "HttpNfcLease" [ 1327.551510] env[61806]: } obtained for vApp import into resource pool (val){ [ 1327.551510] env[61806]: value = "resgroup-9" [ 1327.551510] env[61806]: _type = "ResourcePool" [ 1327.551510] env[61806]: }. {{(pid=61806) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:461}} [ 1327.551780] env[61806]: DEBUG oslo_vmware.api [None req-aaf896e1-050c-4ea6-8bb3-d9b0acde63e0 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Waiting for the lease: (returnval){ [ 1327.551780] env[61806]: value = "session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]522d494f-4623-c0e3-233a-bd8fa8ca5f00" [ 1327.551780] env[61806]: _type = "HttpNfcLease" [ 1327.551780] env[61806]: } to be ready. {{(pid=61806) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 1327.559208] env[61806]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1327.559208] env[61806]: value = "session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]522d494f-4623-c0e3-233a-bd8fa8ca5f00" [ 1327.559208] env[61806]: _type = "HttpNfcLease" [ 1327.559208] env[61806]: } is initializing. {{(pid=61806) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 1327.779023] env[61806]: DEBUG oslo_concurrency.lockutils [req-0e2e5a47-d886-4d91-8c47-080a9b0980ef req-742d6334-2a9c-4fe8-9f94-b9121fd78ac5 service nova] Releasing lock "refresh_cache-7afc6260-7569-44c4-9764-7aa08c662c1c" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1328.060739] env[61806]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1328.060739] env[61806]: value = "session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]522d494f-4623-c0e3-233a-bd8fa8ca5f00" [ 1328.060739] env[61806]: _type = "HttpNfcLease" [ 1328.060739] env[61806]: } is ready. {{(pid=61806) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 1328.061120] env[61806]: DEBUG oslo_vmware.rw_handles [None req-aaf896e1-050c-4ea6-8bb3-d9b0acde63e0 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 1328.061120] env[61806]: value = "session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]522d494f-4623-c0e3-233a-bd8fa8ca5f00" [ 1328.061120] env[61806]: _type = "HttpNfcLease" [ 1328.061120] env[61806]: }. {{(pid=61806) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:467}} [ 1328.061779] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ba91e679-dd99-462a-897a-9f536556895f {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1328.068856] env[61806]: DEBUG oslo_vmware.rw_handles [None req-aaf896e1-050c-4ea6-8bb3-d9b0acde63e0 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Found VMDK URL: https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52465b3d-3a88-3b90-6361-6aa6a65bab24/disk-0.vmdk from lease info. {{(pid=61806) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 1328.069052] env[61806]: DEBUG oslo_vmware.rw_handles [None req-aaf896e1-050c-4ea6-8bb3-d9b0acde63e0 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Creating HTTP connection to write to file with size = 31670272 and URL = https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52465b3d-3a88-3b90-6361-6aa6a65bab24/disk-0.vmdk. {{(pid=61806) _create_write_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:122}} [ 1328.131886] env[61806]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-e3f89d03-08a0-46be-95ab-b2e0f48e720a {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1331.183345] env[61806]: DEBUG oslo_vmware.rw_handles [None req-aaf896e1-050c-4ea6-8bb3-d9b0acde63e0 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Completed reading data from the image iterator. {{(pid=61806) read /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:765}} [ 1331.183819] env[61806]: DEBUG oslo_vmware.rw_handles [None req-aaf896e1-050c-4ea6-8bb3-d9b0acde63e0 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Getting lease state for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52465b3d-3a88-3b90-6361-6aa6a65bab24/disk-0.vmdk. {{(pid=61806) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 1331.184570] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cf519262-83e3-406f-9407-7c66dfb068d5 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1331.191380] env[61806]: DEBUG oslo_vmware.rw_handles [None req-aaf896e1-050c-4ea6-8bb3-d9b0acde63e0 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52465b3d-3a88-3b90-6361-6aa6a65bab24/disk-0.vmdk is in state: ready. {{(pid=61806) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 1331.191575] env[61806]: DEBUG oslo_vmware.rw_handles [None req-aaf896e1-050c-4ea6-8bb3-d9b0acde63e0 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Releasing lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52465b3d-3a88-3b90-6361-6aa6a65bab24/disk-0.vmdk. {{(pid=61806) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:440}} [ 1331.191822] env[61806]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseComplete with opID=oslo.vmware-18da7a4d-76a1-408e-9bbd-0178a8c317fb {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1331.403023] env[61806]: DEBUG oslo_vmware.rw_handles [None req-aaf896e1-050c-4ea6-8bb3-d9b0acde63e0 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Closed VMDK write handle for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52465b3d-3a88-3b90-6361-6aa6a65bab24/disk-0.vmdk. {{(pid=61806) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:646}} [ 1331.403342] env[61806]: INFO nova.virt.vmwareapi.images [None req-aaf896e1-050c-4ea6-8bb3-d9b0acde63e0 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] [instance: 7afc6260-7569-44c4-9764-7aa08c662c1c] Downloaded image file data 23d238dd-6fe7-4d3d-a822-8104bb3bc291 [ 1331.404250] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-39599dec-bb4b-42c5-99dd-79da5c3faa7d {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1331.419074] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-631191c0-6e14-4f31-bbf7-36caa47436d8 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1331.454845] env[61806]: INFO nova.virt.vmwareapi.images [None req-aaf896e1-050c-4ea6-8bb3-d9b0acde63e0 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] [instance: 7afc6260-7569-44c4-9764-7aa08c662c1c] The imported VM was unregistered [ 1331.457212] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-aaf896e1-050c-4ea6-8bb3-d9b0acde63e0 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] [instance: 7afc6260-7569-44c4-9764-7aa08c662c1c] Caching image {{(pid=61806) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:638}} [ 1331.457466] env[61806]: DEBUG nova.virt.vmwareapi.ds_util [None req-aaf896e1-050c-4ea6-8bb3-d9b0acde63e0 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Creating directory with path [datastore1] devstack-image-cache_base/23d238dd-6fe7-4d3d-a822-8104bb3bc291 {{(pid=61806) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1331.457735] env[61806]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-8faf8006-9a14-4eb3-b435-c533e54342e0 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1331.469022] env[61806]: DEBUG nova.virt.vmwareapi.ds_util [None req-aaf896e1-050c-4ea6-8bb3-d9b0acde63e0 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Created directory with path [datastore1] devstack-image-cache_base/23d238dd-6fe7-4d3d-a822-8104bb3bc291 {{(pid=61806) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1331.469228] env[61806]: DEBUG nova.virt.vmwareapi.ds_util [None req-aaf896e1-050c-4ea6-8bb3-d9b0acde63e0 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Moving virtual disk from [datastore1] OSTACK_IMG_7233a54a-fc8b-4c1c-a4c8-ccaae45e5434/OSTACK_IMG_7233a54a-fc8b-4c1c-a4c8-ccaae45e5434.vmdk to [datastore1] devstack-image-cache_base/23d238dd-6fe7-4d3d-a822-8104bb3bc291/23d238dd-6fe7-4d3d-a822-8104bb3bc291.vmdk. {{(pid=61806) disk_move /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:263}} [ 1331.469482] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.MoveVirtualDisk_Task with opID=oslo.vmware-83cd688f-4e1b-488f-ac5a-347c9571ba4b {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1331.474889] env[61806]: DEBUG oslo_vmware.api [None req-aaf896e1-050c-4ea6-8bb3-d9b0acde63e0 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Waiting for the task: (returnval){ [ 1331.474889] env[61806]: value = "task-1295346" [ 1331.474889] env[61806]: _type = "Task" [ 1331.474889] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1331.481936] env[61806]: DEBUG oslo_vmware.api [None req-aaf896e1-050c-4ea6-8bb3-d9b0acde63e0 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Task: {'id': task-1295346, 'name': MoveVirtualDisk_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1331.984667] env[61806]: DEBUG oslo_vmware.api [None req-aaf896e1-050c-4ea6-8bb3-d9b0acde63e0 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Task: {'id': task-1295346, 'name': MoveVirtualDisk_Task} progress is 21%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1332.485103] env[61806]: DEBUG oslo_vmware.api [None req-aaf896e1-050c-4ea6-8bb3-d9b0acde63e0 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Task: {'id': task-1295346, 'name': MoveVirtualDisk_Task} progress is 43%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1332.989685] env[61806]: DEBUG oslo_vmware.api [None req-aaf896e1-050c-4ea6-8bb3-d9b0acde63e0 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Task: {'id': task-1295346, 'name': MoveVirtualDisk_Task} progress is 63%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1333.488330] env[61806]: DEBUG oslo_vmware.api [None req-aaf896e1-050c-4ea6-8bb3-d9b0acde63e0 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Task: {'id': task-1295346, 'name': MoveVirtualDisk_Task} progress is 88%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1333.989741] env[61806]: DEBUG oslo_vmware.api [None req-aaf896e1-050c-4ea6-8bb3-d9b0acde63e0 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Task: {'id': task-1295346, 'name': MoveVirtualDisk_Task, 'duration_secs': 2.333727} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1333.990073] env[61806]: INFO nova.virt.vmwareapi.ds_util [None req-aaf896e1-050c-4ea6-8bb3-d9b0acde63e0 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Moved virtual disk from [datastore1] OSTACK_IMG_7233a54a-fc8b-4c1c-a4c8-ccaae45e5434/OSTACK_IMG_7233a54a-fc8b-4c1c-a4c8-ccaae45e5434.vmdk to [datastore1] devstack-image-cache_base/23d238dd-6fe7-4d3d-a822-8104bb3bc291/23d238dd-6fe7-4d3d-a822-8104bb3bc291.vmdk. [ 1333.990310] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-aaf896e1-050c-4ea6-8bb3-d9b0acde63e0 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] [instance: 7afc6260-7569-44c4-9764-7aa08c662c1c] Cleaning up location [datastore1] OSTACK_IMG_7233a54a-fc8b-4c1c-a4c8-ccaae45e5434 {{(pid=61806) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:640}} [ 1333.990531] env[61806]: DEBUG nova.virt.vmwareapi.ds_util [None req-aaf896e1-050c-4ea6-8bb3-d9b0acde63e0 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Deleting the datastore file [datastore1] OSTACK_IMG_7233a54a-fc8b-4c1c-a4c8-ccaae45e5434 {{(pid=61806) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1333.990824] env[61806]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-36e53c6c-0884-439a-a9e6-8227923c65e8 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1333.996837] env[61806]: DEBUG oslo_vmware.api [None req-aaf896e1-050c-4ea6-8bb3-d9b0acde63e0 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Waiting for the task: (returnval){ [ 1333.996837] env[61806]: value = "task-1295347" [ 1333.996837] env[61806]: _type = "Task" [ 1333.996837] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1334.004523] env[61806]: DEBUG oslo_vmware.api [None req-aaf896e1-050c-4ea6-8bb3-d9b0acde63e0 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Task: {'id': task-1295347, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1334.507241] env[61806]: DEBUG oslo_vmware.api [None req-aaf896e1-050c-4ea6-8bb3-d9b0acde63e0 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Task: {'id': task-1295347, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.037624} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1334.507649] env[61806]: DEBUG nova.virt.vmwareapi.ds_util [None req-aaf896e1-050c-4ea6-8bb3-d9b0acde63e0 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Deleted the datastore file {{(pid=61806) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1334.507694] env[61806]: DEBUG oslo_concurrency.lockutils [None req-aaf896e1-050c-4ea6-8bb3-d9b0acde63e0 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Releasing lock "[datastore1] devstack-image-cache_base/23d238dd-6fe7-4d3d-a822-8104bb3bc291/23d238dd-6fe7-4d3d-a822-8104bb3bc291.vmdk" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1334.507908] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [None req-aaf896e1-050c-4ea6-8bb3-d9b0acde63e0 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/23d238dd-6fe7-4d3d-a822-8104bb3bc291/23d238dd-6fe7-4d3d-a822-8104bb3bc291.vmdk to [datastore1] 7afc6260-7569-44c4-9764-7aa08c662c1c/7afc6260-7569-44c4-9764-7aa08c662c1c.vmdk {{(pid=61806) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 1334.508185] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-fb3325e2-7a75-4117-81fb-96ee98078c69 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1334.514940] env[61806]: DEBUG oslo_vmware.api [None req-aaf896e1-050c-4ea6-8bb3-d9b0acde63e0 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Waiting for the task: (returnval){ [ 1334.514940] env[61806]: value = "task-1295348" [ 1334.514940] env[61806]: _type = "Task" [ 1334.514940] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1334.522188] env[61806]: DEBUG oslo_vmware.api [None req-aaf896e1-050c-4ea6-8bb3-d9b0acde63e0 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Task: {'id': task-1295348, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1335.024619] env[61806]: DEBUG oslo_vmware.api [None req-aaf896e1-050c-4ea6-8bb3-d9b0acde63e0 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Task: {'id': task-1295348, 'name': CopyVirtualDisk_Task} progress is 21%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1335.524886] env[61806]: DEBUG oslo_vmware.api [None req-aaf896e1-050c-4ea6-8bb3-d9b0acde63e0 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Task: {'id': task-1295348, 'name': CopyVirtualDisk_Task} progress is 46%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1336.026022] env[61806]: DEBUG oslo_vmware.api [None req-aaf896e1-050c-4ea6-8bb3-d9b0acde63e0 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Task: {'id': task-1295348, 'name': CopyVirtualDisk_Task} progress is 69%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1336.526549] env[61806]: DEBUG oslo_vmware.api [None req-aaf896e1-050c-4ea6-8bb3-d9b0acde63e0 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Task: {'id': task-1295348, 'name': CopyVirtualDisk_Task} progress is 91%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1337.028400] env[61806]: DEBUG oslo_vmware.api [None req-aaf896e1-050c-4ea6-8bb3-d9b0acde63e0 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Task: {'id': task-1295348, 'name': CopyVirtualDisk_Task, 'duration_secs': 2.336463} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1337.028667] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [None req-aaf896e1-050c-4ea6-8bb3-d9b0acde63e0 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/23d238dd-6fe7-4d3d-a822-8104bb3bc291/23d238dd-6fe7-4d3d-a822-8104bb3bc291.vmdk to [datastore1] 7afc6260-7569-44c4-9764-7aa08c662c1c/7afc6260-7569-44c4-9764-7aa08c662c1c.vmdk {{(pid=61806) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 1337.029512] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8f3adf4f-e063-4931-94f8-659b9a41b8db {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1337.050613] env[61806]: DEBUG nova.virt.vmwareapi.volumeops [None req-aaf896e1-050c-4ea6-8bb3-d9b0acde63e0 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] [instance: 7afc6260-7569-44c4-9764-7aa08c662c1c] Reconfiguring VM instance instance-0000006d to attach disk [datastore1] 7afc6260-7569-44c4-9764-7aa08c662c1c/7afc6260-7569-44c4-9764-7aa08c662c1c.vmdk or device None with type streamOptimized {{(pid=61806) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1337.050881] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-b4a2dd19-1627-4623-85f3-f5e99595cda1 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1337.069979] env[61806]: DEBUG oslo_vmware.api [None req-aaf896e1-050c-4ea6-8bb3-d9b0acde63e0 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Waiting for the task: (returnval){ [ 1337.069979] env[61806]: value = "task-1295349" [ 1337.069979] env[61806]: _type = "Task" [ 1337.069979] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1337.077257] env[61806]: DEBUG oslo_vmware.api [None req-aaf896e1-050c-4ea6-8bb3-d9b0acde63e0 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Task: {'id': task-1295349, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1337.580543] env[61806]: DEBUG oslo_vmware.api [None req-aaf896e1-050c-4ea6-8bb3-d9b0acde63e0 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Task: {'id': task-1295349, 'name': ReconfigVM_Task, 'duration_secs': 0.263661} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1337.580876] env[61806]: DEBUG nova.virt.vmwareapi.volumeops [None req-aaf896e1-050c-4ea6-8bb3-d9b0acde63e0 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] [instance: 7afc6260-7569-44c4-9764-7aa08c662c1c] Reconfigured VM instance instance-0000006d to attach disk [datastore1] 7afc6260-7569-44c4-9764-7aa08c662c1c/7afc6260-7569-44c4-9764-7aa08c662c1c.vmdk or device None with type streamOptimized {{(pid=61806) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1337.581461] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-612f13a2-57dc-4295-a234-d49545239b50 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1337.586958] env[61806]: DEBUG oslo_vmware.api [None req-aaf896e1-050c-4ea6-8bb3-d9b0acde63e0 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Waiting for the task: (returnval){ [ 1337.586958] env[61806]: value = "task-1295350" [ 1337.586958] env[61806]: _type = "Task" [ 1337.586958] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1337.594115] env[61806]: DEBUG oslo_vmware.api [None req-aaf896e1-050c-4ea6-8bb3-d9b0acde63e0 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Task: {'id': task-1295350, 'name': Rename_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1338.096442] env[61806]: DEBUG oslo_vmware.api [None req-aaf896e1-050c-4ea6-8bb3-d9b0acde63e0 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Task: {'id': task-1295350, 'name': Rename_Task, 'duration_secs': 0.124925} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1338.096727] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [None req-aaf896e1-050c-4ea6-8bb3-d9b0acde63e0 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] [instance: 7afc6260-7569-44c4-9764-7aa08c662c1c] Powering on the VM {{(pid=61806) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1338.096976] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-14c3d27d-03bc-42ca-a789-57de2c086fc1 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1338.102630] env[61806]: DEBUG oslo_vmware.api [None req-aaf896e1-050c-4ea6-8bb3-d9b0acde63e0 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Waiting for the task: (returnval){ [ 1338.102630] env[61806]: value = "task-1295351" [ 1338.102630] env[61806]: _type = "Task" [ 1338.102630] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1338.109556] env[61806]: DEBUG oslo_vmware.api [None req-aaf896e1-050c-4ea6-8bb3-d9b0acde63e0 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Task: {'id': task-1295351, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1338.613039] env[61806]: DEBUG oslo_vmware.api [None req-aaf896e1-050c-4ea6-8bb3-d9b0acde63e0 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Task: {'id': task-1295351, 'name': PowerOnVM_Task, 'duration_secs': 0.442343} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1338.613410] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [None req-aaf896e1-050c-4ea6-8bb3-d9b0acde63e0 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] [instance: 7afc6260-7569-44c4-9764-7aa08c662c1c] Powered on the VM {{(pid=61806) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1338.711709] env[61806]: DEBUG nova.compute.manager [None req-aaf896e1-050c-4ea6-8bb3-d9b0acde63e0 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] [instance: 7afc6260-7569-44c4-9764-7aa08c662c1c] Checking state {{(pid=61806) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1338.712791] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ee7519b6-00a0-496d-a66d-04d4b3f58369 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1339.232944] env[61806]: DEBUG oslo_concurrency.lockutils [None req-aaf896e1-050c-4ea6-8bb3-d9b0acde63e0 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Lock "7afc6260-7569-44c4-9764-7aa08c662c1c" "released" by "nova.compute.manager.ComputeManager.unshelve_instance..do_unshelve_instance" :: held 20.204s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1339.831974] env[61806]: DEBUG oslo_concurrency.lockutils [None req-3447251d-f389-44d8-920d-c13761e3e5b9 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Acquiring lock "7afc6260-7569-44c4-9764-7aa08c662c1c" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1339.832337] env[61806]: DEBUG oslo_concurrency.lockutils [None req-3447251d-f389-44d8-920d-c13761e3e5b9 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Lock "7afc6260-7569-44c4-9764-7aa08c662c1c" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1339.832500] env[61806]: DEBUG oslo_concurrency.lockutils [None req-3447251d-f389-44d8-920d-c13761e3e5b9 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Acquiring lock "7afc6260-7569-44c4-9764-7aa08c662c1c-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1339.832695] env[61806]: DEBUG oslo_concurrency.lockutils [None req-3447251d-f389-44d8-920d-c13761e3e5b9 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Lock "7afc6260-7569-44c4-9764-7aa08c662c1c-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1339.832875] env[61806]: DEBUG oslo_concurrency.lockutils [None req-3447251d-f389-44d8-920d-c13761e3e5b9 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Lock "7afc6260-7569-44c4-9764-7aa08c662c1c-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1339.835169] env[61806]: INFO nova.compute.manager [None req-3447251d-f389-44d8-920d-c13761e3e5b9 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] [instance: 7afc6260-7569-44c4-9764-7aa08c662c1c] Terminating instance [ 1339.836950] env[61806]: DEBUG nova.compute.manager [None req-3447251d-f389-44d8-920d-c13761e3e5b9 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] [instance: 7afc6260-7569-44c4-9764-7aa08c662c1c] Start destroying the instance on the hypervisor. {{(pid=61806) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1339.837176] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-3447251d-f389-44d8-920d-c13761e3e5b9 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] [instance: 7afc6260-7569-44c4-9764-7aa08c662c1c] Destroying instance {{(pid=61806) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1339.837993] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-467db919-829f-4c64-b7e6-5f6fcd3fa268 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1339.845989] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [None req-3447251d-f389-44d8-920d-c13761e3e5b9 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] [instance: 7afc6260-7569-44c4-9764-7aa08c662c1c] Powering off the VM {{(pid=61806) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1339.846230] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-cf3bc62c-a427-4015-9f6a-ead7a45b9855 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1339.851891] env[61806]: DEBUG oslo_vmware.api [None req-3447251d-f389-44d8-920d-c13761e3e5b9 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Waiting for the task: (returnval){ [ 1339.851891] env[61806]: value = "task-1295352" [ 1339.851891] env[61806]: _type = "Task" [ 1339.851891] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1339.859612] env[61806]: DEBUG oslo_vmware.api [None req-3447251d-f389-44d8-920d-c13761e3e5b9 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Task: {'id': task-1295352, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1340.362324] env[61806]: DEBUG oslo_vmware.api [None req-3447251d-f389-44d8-920d-c13761e3e5b9 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Task: {'id': task-1295352, 'name': PowerOffVM_Task, 'duration_secs': 0.174786} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1340.362534] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [None req-3447251d-f389-44d8-920d-c13761e3e5b9 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] [instance: 7afc6260-7569-44c4-9764-7aa08c662c1c] Powered off the VM {{(pid=61806) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1340.362711] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-3447251d-f389-44d8-920d-c13761e3e5b9 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] [instance: 7afc6260-7569-44c4-9764-7aa08c662c1c] Unregistering the VM {{(pid=61806) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1340.362963] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-1a013daf-0bd7-4512-8e8a-4a6923943f5f {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1340.422060] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-3447251d-f389-44d8-920d-c13761e3e5b9 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] [instance: 7afc6260-7569-44c4-9764-7aa08c662c1c] Unregistered the VM {{(pid=61806) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1340.422280] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-3447251d-f389-44d8-920d-c13761e3e5b9 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] [instance: 7afc6260-7569-44c4-9764-7aa08c662c1c] Deleting contents of the VM from datastore datastore1 {{(pid=61806) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1340.422471] env[61806]: DEBUG nova.virt.vmwareapi.ds_util [None req-3447251d-f389-44d8-920d-c13761e3e5b9 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Deleting the datastore file [datastore1] 7afc6260-7569-44c4-9764-7aa08c662c1c {{(pid=61806) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1340.422735] env[61806]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-d1586bb7-b504-404c-bbdd-8d8b2bc63a70 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1340.429206] env[61806]: DEBUG oslo_vmware.api [None req-3447251d-f389-44d8-920d-c13761e3e5b9 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Waiting for the task: (returnval){ [ 1340.429206] env[61806]: value = "task-1295354" [ 1340.429206] env[61806]: _type = "Task" [ 1340.429206] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1340.436225] env[61806]: DEBUG oslo_vmware.api [None req-3447251d-f389-44d8-920d-c13761e3e5b9 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Task: {'id': task-1295354, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1340.938849] env[61806]: DEBUG oslo_vmware.api [None req-3447251d-f389-44d8-920d-c13761e3e5b9 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Task: {'id': task-1295354, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.149619} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1340.939248] env[61806]: DEBUG nova.virt.vmwareapi.ds_util [None req-3447251d-f389-44d8-920d-c13761e3e5b9 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Deleted the datastore file {{(pid=61806) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1340.939297] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-3447251d-f389-44d8-920d-c13761e3e5b9 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] [instance: 7afc6260-7569-44c4-9764-7aa08c662c1c] Deleted contents of the VM from datastore datastore1 {{(pid=61806) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1340.939476] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-3447251d-f389-44d8-920d-c13761e3e5b9 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] [instance: 7afc6260-7569-44c4-9764-7aa08c662c1c] Instance destroyed {{(pid=61806) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1340.939676] env[61806]: INFO nova.compute.manager [None req-3447251d-f389-44d8-920d-c13761e3e5b9 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] [instance: 7afc6260-7569-44c4-9764-7aa08c662c1c] Took 1.10 seconds to destroy the instance on the hypervisor. [ 1340.939907] env[61806]: DEBUG oslo.service.loopingcall [None req-3447251d-f389-44d8-920d-c13761e3e5b9 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61806) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1340.940134] env[61806]: DEBUG nova.compute.manager [-] [instance: 7afc6260-7569-44c4-9764-7aa08c662c1c] Deallocating network for instance {{(pid=61806) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1340.940233] env[61806]: DEBUG nova.network.neutron [-] [instance: 7afc6260-7569-44c4-9764-7aa08c662c1c] deallocate_for_instance() {{(pid=61806) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1341.340873] env[61806]: DEBUG nova.compute.manager [req-1f68baf7-5f41-4378-89ce-97dee240d93a req-d83ceccc-35a7-4af0-abfd-995f547d76ee service nova] [instance: 7afc6260-7569-44c4-9764-7aa08c662c1c] Received event network-vif-deleted-ee859b12-498b-43fe-97f1-47c250ada436 {{(pid=61806) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1341.341324] env[61806]: INFO nova.compute.manager [req-1f68baf7-5f41-4378-89ce-97dee240d93a req-d83ceccc-35a7-4af0-abfd-995f547d76ee service nova] [instance: 7afc6260-7569-44c4-9764-7aa08c662c1c] Neutron deleted interface ee859b12-498b-43fe-97f1-47c250ada436; detaching it from the instance and deleting it from the info cache [ 1341.341379] env[61806]: DEBUG nova.network.neutron [req-1f68baf7-5f41-4378-89ce-97dee240d93a req-d83ceccc-35a7-4af0-abfd-995f547d76ee service nova] [instance: 7afc6260-7569-44c4-9764-7aa08c662c1c] Updating instance_info_cache with network_info: [] {{(pid=61806) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1341.822030] env[61806]: DEBUG nova.network.neutron [-] [instance: 7afc6260-7569-44c4-9764-7aa08c662c1c] Updating instance_info_cache with network_info: [] {{(pid=61806) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1341.844197] env[61806]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-26eb1925-fbf9-4759-84be-a76cc8db682b {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1341.853865] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f32d7812-ffbf-4d3e-aaa2-5f26627ef5cd {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1341.877039] env[61806]: DEBUG nova.compute.manager [req-1f68baf7-5f41-4378-89ce-97dee240d93a req-d83ceccc-35a7-4af0-abfd-995f547d76ee service nova] [instance: 7afc6260-7569-44c4-9764-7aa08c662c1c] Detach interface failed, port_id=ee859b12-498b-43fe-97f1-47c250ada436, reason: Instance 7afc6260-7569-44c4-9764-7aa08c662c1c could not be found. {{(pid=61806) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10963}} [ 1342.324785] env[61806]: INFO nova.compute.manager [-] [instance: 7afc6260-7569-44c4-9764-7aa08c662c1c] Took 1.38 seconds to deallocate network for instance. [ 1342.831706] env[61806]: DEBUG oslo_concurrency.lockutils [None req-3447251d-f389-44d8-920d-c13761e3e5b9 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1342.831938] env[61806]: DEBUG oslo_concurrency.lockutils [None req-3447251d-f389-44d8-920d-c13761e3e5b9 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.000s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1342.832189] env[61806]: DEBUG nova.objects.instance [None req-3447251d-f389-44d8-920d-c13761e3e5b9 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Lazy-loading 'resources' on Instance uuid 7afc6260-7569-44c4-9764-7aa08c662c1c {{(pid=61806) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1343.373576] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d0abcc76-3c4e-4f35-ac04-8757e3b13799 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1343.380885] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7e801161-d583-4c92-b771-65192b10c0de {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1343.410356] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1cd0a36c-900c-4dab-84ba-135f3a503878 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1343.417150] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-773fd9dc-b179-4fd3-8fc2-2100d914d5d9 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1343.429727] env[61806]: DEBUG nova.compute.provider_tree [None req-3447251d-f389-44d8-920d-c13761e3e5b9 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Updating inventory in ProviderTree for provider a2858be1-fd22-4e08-979e-87ad25293407 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 138, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61806) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1343.961569] env[61806]: DEBUG nova.scheduler.client.report [None req-3447251d-f389-44d8-920d-c13761e3e5b9 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Updated inventory for provider a2858be1-fd22-4e08-979e-87ad25293407 with generation 155 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 138, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61806) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:971}} [ 1343.961862] env[61806]: DEBUG nova.compute.provider_tree [None req-3447251d-f389-44d8-920d-c13761e3e5b9 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Updating resource provider a2858be1-fd22-4e08-979e-87ad25293407 generation from 155 to 156 during operation: update_inventory {{(pid=61806) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 1343.962066] env[61806]: DEBUG nova.compute.provider_tree [None req-3447251d-f389-44d8-920d-c13761e3e5b9 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Updating inventory in ProviderTree for provider a2858be1-fd22-4e08-979e-87ad25293407 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 138, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61806) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1344.466628] env[61806]: DEBUG oslo_concurrency.lockutils [None req-3447251d-f389-44d8-920d-c13761e3e5b9 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.634s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1344.495181] env[61806]: INFO nova.scheduler.client.report [None req-3447251d-f389-44d8-920d-c13761e3e5b9 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Deleted allocations for instance 7afc6260-7569-44c4-9764-7aa08c662c1c [ 1345.003050] env[61806]: DEBUG oslo_concurrency.lockutils [None req-3447251d-f389-44d8-920d-c13761e3e5b9 tempest-ServerActionsTestOtherB-1209037893 tempest-ServerActionsTestOtherB-1209037893-project-member] Lock "7afc6260-7569-44c4-9764-7aa08c662c1c" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.171s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1353.238436] env[61806]: DEBUG oslo_concurrency.lockutils [None req-c1dda7ed-f3be-450a-9e30-07997ea9d73d tempest-ServerGroupTestJSON-1417866451 tempest-ServerGroupTestJSON-1417866451-project-member] Acquiring lock "8fdc4f99-9dd2-4312-9b13-f1e4170ed489" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1353.238999] env[61806]: DEBUG oslo_concurrency.lockutils [None req-c1dda7ed-f3be-450a-9e30-07997ea9d73d tempest-ServerGroupTestJSON-1417866451 tempest-ServerGroupTestJSON-1417866451-project-member] Lock "8fdc4f99-9dd2-4312-9b13-f1e4170ed489" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1353.741246] env[61806]: DEBUG nova.compute.manager [None req-c1dda7ed-f3be-450a-9e30-07997ea9d73d tempest-ServerGroupTestJSON-1417866451 tempest-ServerGroupTestJSON-1417866451-project-member] [instance: 8fdc4f99-9dd2-4312-9b13-f1e4170ed489] Starting instance... {{(pid=61806) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1354.263058] env[61806]: DEBUG oslo_concurrency.lockutils [None req-c1dda7ed-f3be-450a-9e30-07997ea9d73d tempest-ServerGroupTestJSON-1417866451 tempest-ServerGroupTestJSON-1417866451-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1354.263383] env[61806]: DEBUG oslo_concurrency.lockutils [None req-c1dda7ed-f3be-450a-9e30-07997ea9d73d tempest-ServerGroupTestJSON-1417866451 tempest-ServerGroupTestJSON-1417866451-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.000s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1354.264976] env[61806]: INFO nova.compute.claims [None req-c1dda7ed-f3be-450a-9e30-07997ea9d73d tempest-ServerGroupTestJSON-1417866451 tempest-ServerGroupTestJSON-1417866451-project-member] [instance: 8fdc4f99-9dd2-4312-9b13-f1e4170ed489] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1355.302887] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-503518c0-80f1-4996-9446-ccbd92b243bc {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1355.310517] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-505a3559-4c0d-498e-aaf3-4d0009d12ce9 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1355.340658] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a5a36b67-db02-4adf-a8db-91c12ca652a2 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1355.347471] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a42d2f36-b5d6-45db-9292-0742b2423de8 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1355.360060] env[61806]: DEBUG nova.compute.provider_tree [None req-c1dda7ed-f3be-450a-9e30-07997ea9d73d tempest-ServerGroupTestJSON-1417866451 tempest-ServerGroupTestJSON-1417866451-project-member] Inventory has not changed in ProviderTree for provider: a2858be1-fd22-4e08-979e-87ad25293407 {{(pid=61806) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1355.863021] env[61806]: DEBUG nova.scheduler.client.report [None req-c1dda7ed-f3be-450a-9e30-07997ea9d73d tempest-ServerGroupTestJSON-1417866451 tempest-ServerGroupTestJSON-1417866451-project-member] Inventory has not changed for provider a2858be1-fd22-4e08-979e-87ad25293407 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 138, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61806) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1356.368069] env[61806]: DEBUG oslo_concurrency.lockutils [None req-c1dda7ed-f3be-450a-9e30-07997ea9d73d tempest-ServerGroupTestJSON-1417866451 tempest-ServerGroupTestJSON-1417866451-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.104s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1356.871049] env[61806]: DEBUG oslo_concurrency.lockutils [None req-c1dda7ed-f3be-450a-9e30-07997ea9d73d tempest-ServerGroupTestJSON-1417866451 tempest-ServerGroupTestJSON-1417866451-project-member] Acquiring lock "b7594e4a-66ee-40e0-a334-d6846391c983" by "nova.compute.manager.ComputeManager._validate_instance_group_policy.._do_validation" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1356.871204] env[61806]: DEBUG oslo_concurrency.lockutils [None req-c1dda7ed-f3be-450a-9e30-07997ea9d73d tempest-ServerGroupTestJSON-1417866451 tempest-ServerGroupTestJSON-1417866451-project-member] Lock "b7594e4a-66ee-40e0-a334-d6846391c983" acquired by "nova.compute.manager.ComputeManager._validate_instance_group_policy.._do_validation" :: waited 0.000s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1357.375804] env[61806]: DEBUG oslo_concurrency.lockutils [None req-c1dda7ed-f3be-450a-9e30-07997ea9d73d tempest-ServerGroupTestJSON-1417866451 tempest-ServerGroupTestJSON-1417866451-project-member] Lock "b7594e4a-66ee-40e0-a334-d6846391c983" "released" by "nova.compute.manager.ComputeManager._validate_instance_group_policy.._do_validation" :: held 0.504s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1357.376449] env[61806]: DEBUG nova.compute.manager [None req-c1dda7ed-f3be-450a-9e30-07997ea9d73d tempest-ServerGroupTestJSON-1417866451 tempest-ServerGroupTestJSON-1417866451-project-member] [instance: 8fdc4f99-9dd2-4312-9b13-f1e4170ed489] Start building networks asynchronously for instance. {{(pid=61806) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1357.881079] env[61806]: DEBUG nova.compute.utils [None req-c1dda7ed-f3be-450a-9e30-07997ea9d73d tempest-ServerGroupTestJSON-1417866451 tempest-ServerGroupTestJSON-1417866451-project-member] Using /dev/sd instead of None {{(pid=61806) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1357.882550] env[61806]: DEBUG nova.compute.manager [None req-c1dda7ed-f3be-450a-9e30-07997ea9d73d tempest-ServerGroupTestJSON-1417866451 tempest-ServerGroupTestJSON-1417866451-project-member] [instance: 8fdc4f99-9dd2-4312-9b13-f1e4170ed489] Allocating IP information in the background. {{(pid=61806) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1357.882728] env[61806]: DEBUG nova.network.neutron [None req-c1dda7ed-f3be-450a-9e30-07997ea9d73d tempest-ServerGroupTestJSON-1417866451 tempest-ServerGroupTestJSON-1417866451-project-member] [instance: 8fdc4f99-9dd2-4312-9b13-f1e4170ed489] allocate_for_instance() {{(pid=61806) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 1357.931739] env[61806]: DEBUG nova.policy [None req-c1dda7ed-f3be-450a-9e30-07997ea9d73d tempest-ServerGroupTestJSON-1417866451 tempest-ServerGroupTestJSON-1417866451-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'e2674b45b4c146d28143a36111839676', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'dc914b76e41544b8b2ada352cfb5640f', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61806) authorize /opt/stack/nova/nova/policy.py:201}} [ 1358.160605] env[61806]: DEBUG nova.network.neutron [None req-c1dda7ed-f3be-450a-9e30-07997ea9d73d tempest-ServerGroupTestJSON-1417866451 tempest-ServerGroupTestJSON-1417866451-project-member] [instance: 8fdc4f99-9dd2-4312-9b13-f1e4170ed489] Successfully created port: 84c9dc10-8d31-4fb4-aea2-435e8940c4e4 {{(pid=61806) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1358.385851] env[61806]: DEBUG nova.compute.manager [None req-c1dda7ed-f3be-450a-9e30-07997ea9d73d tempest-ServerGroupTestJSON-1417866451 tempest-ServerGroupTestJSON-1417866451-project-member] [instance: 8fdc4f99-9dd2-4312-9b13-f1e4170ed489] Start building block device mappings for instance. {{(pid=61806) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1359.394944] env[61806]: DEBUG nova.compute.manager [None req-c1dda7ed-f3be-450a-9e30-07997ea9d73d tempest-ServerGroupTestJSON-1417866451 tempest-ServerGroupTestJSON-1417866451-project-member] [instance: 8fdc4f99-9dd2-4312-9b13-f1e4170ed489] Start spawning the instance on the hypervisor. {{(pid=61806) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1359.419911] env[61806]: DEBUG nova.virt.hardware [None req-c1dda7ed-f3be-450a-9e30-07997ea9d73d tempest-ServerGroupTestJSON-1417866451 tempest-ServerGroupTestJSON-1417866451-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-15T18:05:10Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-15T18:04:52Z,direct_url=,disk_format='vmdk',id=ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='84ba83fce2c349a6988173c5d6fc3b07',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-15T18:04:53Z,virtual_size=,visibility=), allow threads: False {{(pid=61806) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1359.420105] env[61806]: DEBUG nova.virt.hardware [None req-c1dda7ed-f3be-450a-9e30-07997ea9d73d tempest-ServerGroupTestJSON-1417866451 tempest-ServerGroupTestJSON-1417866451-project-member] Flavor limits 0:0:0 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1359.420282] env[61806]: DEBUG nova.virt.hardware [None req-c1dda7ed-f3be-450a-9e30-07997ea9d73d tempest-ServerGroupTestJSON-1417866451 tempest-ServerGroupTestJSON-1417866451-project-member] Image limits 0:0:0 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1359.420477] env[61806]: DEBUG nova.virt.hardware [None req-c1dda7ed-f3be-450a-9e30-07997ea9d73d tempest-ServerGroupTestJSON-1417866451 tempest-ServerGroupTestJSON-1417866451-project-member] Flavor pref 0:0:0 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1359.420633] env[61806]: DEBUG nova.virt.hardware [None req-c1dda7ed-f3be-450a-9e30-07997ea9d73d tempest-ServerGroupTestJSON-1417866451 tempest-ServerGroupTestJSON-1417866451-project-member] Image pref 0:0:0 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1359.420791] env[61806]: DEBUG nova.virt.hardware [None req-c1dda7ed-f3be-450a-9e30-07997ea9d73d tempest-ServerGroupTestJSON-1417866451 tempest-ServerGroupTestJSON-1417866451-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1359.421018] env[61806]: DEBUG nova.virt.hardware [None req-c1dda7ed-f3be-450a-9e30-07997ea9d73d tempest-ServerGroupTestJSON-1417866451 tempest-ServerGroupTestJSON-1417866451-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61806) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1359.421194] env[61806]: DEBUG nova.virt.hardware [None req-c1dda7ed-f3be-450a-9e30-07997ea9d73d tempest-ServerGroupTestJSON-1417866451 tempest-ServerGroupTestJSON-1417866451-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61806) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1359.421373] env[61806]: DEBUG nova.virt.hardware [None req-c1dda7ed-f3be-450a-9e30-07997ea9d73d tempest-ServerGroupTestJSON-1417866451 tempest-ServerGroupTestJSON-1417866451-project-member] Got 1 possible topologies {{(pid=61806) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1359.421543] env[61806]: DEBUG nova.virt.hardware [None req-c1dda7ed-f3be-450a-9e30-07997ea9d73d tempest-ServerGroupTestJSON-1417866451 tempest-ServerGroupTestJSON-1417866451-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61806) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1359.422299] env[61806]: DEBUG nova.virt.hardware [None req-c1dda7ed-f3be-450a-9e30-07997ea9d73d tempest-ServerGroupTestJSON-1417866451 tempest-ServerGroupTestJSON-1417866451-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61806) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1359.422600] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a8218697-e883-4075-aae8-80cd865a7193 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1359.430670] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7ef8abfc-b73e-4c38-95ff-e9126d85e4c5 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1359.520363] env[61806]: DEBUG nova.compute.manager [req-fe8cf38c-ba6a-4eac-b09d-8f7dfa1a02f8 req-c40393fd-71ab-41d0-a0f1-bb63eb688237 service nova] [instance: 8fdc4f99-9dd2-4312-9b13-f1e4170ed489] Received event network-vif-plugged-84c9dc10-8d31-4fb4-aea2-435e8940c4e4 {{(pid=61806) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1359.520653] env[61806]: DEBUG oslo_concurrency.lockutils [req-fe8cf38c-ba6a-4eac-b09d-8f7dfa1a02f8 req-c40393fd-71ab-41d0-a0f1-bb63eb688237 service nova] Acquiring lock "8fdc4f99-9dd2-4312-9b13-f1e4170ed489-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1359.520835] env[61806]: DEBUG oslo_concurrency.lockutils [req-fe8cf38c-ba6a-4eac-b09d-8f7dfa1a02f8 req-c40393fd-71ab-41d0-a0f1-bb63eb688237 service nova] Lock "8fdc4f99-9dd2-4312-9b13-f1e4170ed489-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1359.520958] env[61806]: DEBUG oslo_concurrency.lockutils [req-fe8cf38c-ba6a-4eac-b09d-8f7dfa1a02f8 req-c40393fd-71ab-41d0-a0f1-bb63eb688237 service nova] Lock "8fdc4f99-9dd2-4312-9b13-f1e4170ed489-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1359.521151] env[61806]: DEBUG nova.compute.manager [req-fe8cf38c-ba6a-4eac-b09d-8f7dfa1a02f8 req-c40393fd-71ab-41d0-a0f1-bb63eb688237 service nova] [instance: 8fdc4f99-9dd2-4312-9b13-f1e4170ed489] No waiting events found dispatching network-vif-plugged-84c9dc10-8d31-4fb4-aea2-435e8940c4e4 {{(pid=61806) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1359.521324] env[61806]: WARNING nova.compute.manager [req-fe8cf38c-ba6a-4eac-b09d-8f7dfa1a02f8 req-c40393fd-71ab-41d0-a0f1-bb63eb688237 service nova] [instance: 8fdc4f99-9dd2-4312-9b13-f1e4170ed489] Received unexpected event network-vif-plugged-84c9dc10-8d31-4fb4-aea2-435e8940c4e4 for instance with vm_state building and task_state spawning. [ 1359.604400] env[61806]: DEBUG nova.network.neutron [None req-c1dda7ed-f3be-450a-9e30-07997ea9d73d tempest-ServerGroupTestJSON-1417866451 tempest-ServerGroupTestJSON-1417866451-project-member] [instance: 8fdc4f99-9dd2-4312-9b13-f1e4170ed489] Successfully updated port: 84c9dc10-8d31-4fb4-aea2-435e8940c4e4 {{(pid=61806) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1360.106901] env[61806]: DEBUG oslo_concurrency.lockutils [None req-c1dda7ed-f3be-450a-9e30-07997ea9d73d tempest-ServerGroupTestJSON-1417866451 tempest-ServerGroupTestJSON-1417866451-project-member] Acquiring lock "refresh_cache-8fdc4f99-9dd2-4312-9b13-f1e4170ed489" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1360.107065] env[61806]: DEBUG oslo_concurrency.lockutils [None req-c1dda7ed-f3be-450a-9e30-07997ea9d73d tempest-ServerGroupTestJSON-1417866451 tempest-ServerGroupTestJSON-1417866451-project-member] Acquired lock "refresh_cache-8fdc4f99-9dd2-4312-9b13-f1e4170ed489" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1360.107217] env[61806]: DEBUG nova.network.neutron [None req-c1dda7ed-f3be-450a-9e30-07997ea9d73d tempest-ServerGroupTestJSON-1417866451 tempest-ServerGroupTestJSON-1417866451-project-member] [instance: 8fdc4f99-9dd2-4312-9b13-f1e4170ed489] Building network info cache for instance {{(pid=61806) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1360.639161] env[61806]: DEBUG nova.network.neutron [None req-c1dda7ed-f3be-450a-9e30-07997ea9d73d tempest-ServerGroupTestJSON-1417866451 tempest-ServerGroupTestJSON-1417866451-project-member] [instance: 8fdc4f99-9dd2-4312-9b13-f1e4170ed489] Instance cache missing network info. {{(pid=61806) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 1360.754607] env[61806]: DEBUG nova.network.neutron [None req-c1dda7ed-f3be-450a-9e30-07997ea9d73d tempest-ServerGroupTestJSON-1417866451 tempest-ServerGroupTestJSON-1417866451-project-member] [instance: 8fdc4f99-9dd2-4312-9b13-f1e4170ed489] Updating instance_info_cache with network_info: [{"id": "84c9dc10-8d31-4fb4-aea2-435e8940c4e4", "address": "fa:16:3e:72:31:dc", "network": {"id": "86065c4a-0172-44b7-aa6e-ab53281d9d74", "bridge": "br-int", "label": "tempest-ServerGroupTestJSON-160946182-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "dc914b76e41544b8b2ada352cfb5640f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "9fa933df-d66f-485e-8cf9-eda7f1a7f283", "external-id": "nsx-vlan-transportzone-87", "segmentation_id": 87, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap84c9dc10-8d", "ovs_interfaceid": "84c9dc10-8d31-4fb4-aea2-435e8940c4e4", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61806) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1361.257412] env[61806]: DEBUG oslo_concurrency.lockutils [None req-c1dda7ed-f3be-450a-9e30-07997ea9d73d tempest-ServerGroupTestJSON-1417866451 tempest-ServerGroupTestJSON-1417866451-project-member] Releasing lock "refresh_cache-8fdc4f99-9dd2-4312-9b13-f1e4170ed489" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1361.257738] env[61806]: DEBUG nova.compute.manager [None req-c1dda7ed-f3be-450a-9e30-07997ea9d73d tempest-ServerGroupTestJSON-1417866451 tempest-ServerGroupTestJSON-1417866451-project-member] [instance: 8fdc4f99-9dd2-4312-9b13-f1e4170ed489] Instance network_info: |[{"id": "84c9dc10-8d31-4fb4-aea2-435e8940c4e4", "address": "fa:16:3e:72:31:dc", "network": {"id": "86065c4a-0172-44b7-aa6e-ab53281d9d74", "bridge": "br-int", "label": "tempest-ServerGroupTestJSON-160946182-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "dc914b76e41544b8b2ada352cfb5640f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "9fa933df-d66f-485e-8cf9-eda7f1a7f283", "external-id": "nsx-vlan-transportzone-87", "segmentation_id": 87, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap84c9dc10-8d", "ovs_interfaceid": "84c9dc10-8d31-4fb4-aea2-435e8940c4e4", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61806) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1361.258210] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-c1dda7ed-f3be-450a-9e30-07997ea9d73d tempest-ServerGroupTestJSON-1417866451 tempest-ServerGroupTestJSON-1417866451-project-member] [instance: 8fdc4f99-9dd2-4312-9b13-f1e4170ed489] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:72:31:dc', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '9fa933df-d66f-485e-8cf9-eda7f1a7f283', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '84c9dc10-8d31-4fb4-aea2-435e8940c4e4', 'vif_model': 'vmxnet3'}] {{(pid=61806) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1361.265908] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [None req-c1dda7ed-f3be-450a-9e30-07997ea9d73d tempest-ServerGroupTestJSON-1417866451 tempest-ServerGroupTestJSON-1417866451-project-member] Creating folder: Project (dc914b76e41544b8b2ada352cfb5640f). Parent ref: group-v277609. {{(pid=61806) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 1361.266211] env[61806]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-03ea638b-9ff3-4e3b-ab5c-a9994a920229 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1361.277805] env[61806]: INFO nova.virt.vmwareapi.vm_util [None req-c1dda7ed-f3be-450a-9e30-07997ea9d73d tempest-ServerGroupTestJSON-1417866451 tempest-ServerGroupTestJSON-1417866451-project-member] Created folder: Project (dc914b76e41544b8b2ada352cfb5640f) in parent group-v277609. [ 1361.277998] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [None req-c1dda7ed-f3be-450a-9e30-07997ea9d73d tempest-ServerGroupTestJSON-1417866451 tempest-ServerGroupTestJSON-1417866451-project-member] Creating folder: Instances. Parent ref: group-v277779. {{(pid=61806) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 1361.278234] env[61806]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-71569b08-da82-48d4-b2a9-07ec1238851d {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1361.286969] env[61806]: INFO nova.virt.vmwareapi.vm_util [None req-c1dda7ed-f3be-450a-9e30-07997ea9d73d tempest-ServerGroupTestJSON-1417866451 tempest-ServerGroupTestJSON-1417866451-project-member] Created folder: Instances in parent group-v277779. [ 1361.287219] env[61806]: DEBUG oslo.service.loopingcall [None req-c1dda7ed-f3be-450a-9e30-07997ea9d73d tempest-ServerGroupTestJSON-1417866451 tempest-ServerGroupTestJSON-1417866451-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61806) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1361.287401] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 8fdc4f99-9dd2-4312-9b13-f1e4170ed489] Creating VM on the ESX host {{(pid=61806) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 1361.287590] env[61806]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-18851994-8568-4f73-b9fe-db9ebf3a0587 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1361.304631] env[61806]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1361.304631] env[61806]: value = "task-1295358" [ 1361.304631] env[61806]: _type = "Task" [ 1361.304631] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1361.311527] env[61806]: DEBUG oslo_vmware.api [-] Task: {'id': task-1295358, 'name': CreateVM_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1361.545713] env[61806]: DEBUG nova.compute.manager [req-32013297-f290-4604-9db2-f91aea4f53d1 req-04a40c6d-accd-43b2-a36b-4055214281ac service nova] [instance: 8fdc4f99-9dd2-4312-9b13-f1e4170ed489] Received event network-changed-84c9dc10-8d31-4fb4-aea2-435e8940c4e4 {{(pid=61806) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1361.545939] env[61806]: DEBUG nova.compute.manager [req-32013297-f290-4604-9db2-f91aea4f53d1 req-04a40c6d-accd-43b2-a36b-4055214281ac service nova] [instance: 8fdc4f99-9dd2-4312-9b13-f1e4170ed489] Refreshing instance network info cache due to event network-changed-84c9dc10-8d31-4fb4-aea2-435e8940c4e4. {{(pid=61806) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 1361.546175] env[61806]: DEBUG oslo_concurrency.lockutils [req-32013297-f290-4604-9db2-f91aea4f53d1 req-04a40c6d-accd-43b2-a36b-4055214281ac service nova] Acquiring lock "refresh_cache-8fdc4f99-9dd2-4312-9b13-f1e4170ed489" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1361.546327] env[61806]: DEBUG oslo_concurrency.lockutils [req-32013297-f290-4604-9db2-f91aea4f53d1 req-04a40c6d-accd-43b2-a36b-4055214281ac service nova] Acquired lock "refresh_cache-8fdc4f99-9dd2-4312-9b13-f1e4170ed489" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1361.546498] env[61806]: DEBUG nova.network.neutron [req-32013297-f290-4604-9db2-f91aea4f53d1 req-04a40c6d-accd-43b2-a36b-4055214281ac service nova] [instance: 8fdc4f99-9dd2-4312-9b13-f1e4170ed489] Refreshing network info cache for port 84c9dc10-8d31-4fb4-aea2-435e8940c4e4 {{(pid=61806) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1361.814787] env[61806]: DEBUG oslo_vmware.api [-] Task: {'id': task-1295358, 'name': CreateVM_Task, 'duration_secs': 0.280689} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1361.815232] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 8fdc4f99-9dd2-4312-9b13-f1e4170ed489] Created VM on the ESX host {{(pid=61806) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 1361.815604] env[61806]: DEBUG oslo_concurrency.lockutils [None req-c1dda7ed-f3be-450a-9e30-07997ea9d73d tempest-ServerGroupTestJSON-1417866451 tempest-ServerGroupTestJSON-1417866451-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1361.815779] env[61806]: DEBUG oslo_concurrency.lockutils [None req-c1dda7ed-f3be-450a-9e30-07997ea9d73d tempest-ServerGroupTestJSON-1417866451 tempest-ServerGroupTestJSON-1417866451-project-member] Acquired lock "[datastore1] devstack-image-cache_base/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1361.816112] env[61806]: DEBUG oslo_concurrency.lockutils [None req-c1dda7ed-f3be-450a-9e30-07997ea9d73d tempest-ServerGroupTestJSON-1417866451 tempest-ServerGroupTestJSON-1417866451-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1361.816370] env[61806]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-403940b7-fd6c-47ea-961c-a1f453764aef {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1361.820786] env[61806]: DEBUG oslo_vmware.api [None req-c1dda7ed-f3be-450a-9e30-07997ea9d73d tempest-ServerGroupTestJSON-1417866451 tempest-ServerGroupTestJSON-1417866451-project-member] Waiting for the task: (returnval){ [ 1361.820786] env[61806]: value = "session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]52d2740d-629b-0fc1-eeae-7d152df21b8f" [ 1361.820786] env[61806]: _type = "Task" [ 1361.820786] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1361.827824] env[61806]: DEBUG oslo_vmware.api [None req-c1dda7ed-f3be-450a-9e30-07997ea9d73d tempest-ServerGroupTestJSON-1417866451 tempest-ServerGroupTestJSON-1417866451-project-member] Task: {'id': session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]52d2740d-629b-0fc1-eeae-7d152df21b8f, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1362.227394] env[61806]: DEBUG nova.network.neutron [req-32013297-f290-4604-9db2-f91aea4f53d1 req-04a40c6d-accd-43b2-a36b-4055214281ac service nova] [instance: 8fdc4f99-9dd2-4312-9b13-f1e4170ed489] Updated VIF entry in instance network info cache for port 84c9dc10-8d31-4fb4-aea2-435e8940c4e4. {{(pid=61806) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1362.227751] env[61806]: DEBUG nova.network.neutron [req-32013297-f290-4604-9db2-f91aea4f53d1 req-04a40c6d-accd-43b2-a36b-4055214281ac service nova] [instance: 8fdc4f99-9dd2-4312-9b13-f1e4170ed489] Updating instance_info_cache with network_info: [{"id": "84c9dc10-8d31-4fb4-aea2-435e8940c4e4", "address": "fa:16:3e:72:31:dc", "network": {"id": "86065c4a-0172-44b7-aa6e-ab53281d9d74", "bridge": "br-int", "label": "tempest-ServerGroupTestJSON-160946182-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "dc914b76e41544b8b2ada352cfb5640f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "9fa933df-d66f-485e-8cf9-eda7f1a7f283", "external-id": "nsx-vlan-transportzone-87", "segmentation_id": 87, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap84c9dc10-8d", "ovs_interfaceid": "84c9dc10-8d31-4fb4-aea2-435e8940c4e4", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61806) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1362.331623] env[61806]: DEBUG oslo_vmware.api [None req-c1dda7ed-f3be-450a-9e30-07997ea9d73d tempest-ServerGroupTestJSON-1417866451 tempest-ServerGroupTestJSON-1417866451-project-member] Task: {'id': session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]52d2740d-629b-0fc1-eeae-7d152df21b8f, 'name': SearchDatastore_Task, 'duration_secs': 0.010583} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1362.331906] env[61806]: DEBUG oslo_concurrency.lockutils [None req-c1dda7ed-f3be-450a-9e30-07997ea9d73d tempest-ServerGroupTestJSON-1417866451 tempest-ServerGroupTestJSON-1417866451-project-member] Releasing lock "[datastore1] devstack-image-cache_base/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1362.332169] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-c1dda7ed-f3be-450a-9e30-07997ea9d73d tempest-ServerGroupTestJSON-1417866451 tempest-ServerGroupTestJSON-1417866451-project-member] [instance: 8fdc4f99-9dd2-4312-9b13-f1e4170ed489] Processing image ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa {{(pid=61806) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1362.332595] env[61806]: DEBUG oslo_concurrency.lockutils [None req-c1dda7ed-f3be-450a-9e30-07997ea9d73d tempest-ServerGroupTestJSON-1417866451 tempest-ServerGroupTestJSON-1417866451-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa.vmdk" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1362.332595] env[61806]: DEBUG oslo_concurrency.lockutils [None req-c1dda7ed-f3be-450a-9e30-07997ea9d73d tempest-ServerGroupTestJSON-1417866451 tempest-ServerGroupTestJSON-1417866451-project-member] Acquired lock "[datastore1] devstack-image-cache_base/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa.vmdk" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1362.332759] env[61806]: DEBUG nova.virt.vmwareapi.ds_util [None req-c1dda7ed-f3be-450a-9e30-07997ea9d73d tempest-ServerGroupTestJSON-1417866451 tempest-ServerGroupTestJSON-1417866451-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61806) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1362.333015] env[61806]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-3cc39490-7a77-464d-8415-62b6eae7d2da {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1362.340699] env[61806]: DEBUG nova.virt.vmwareapi.ds_util [None req-c1dda7ed-f3be-450a-9e30-07997ea9d73d tempest-ServerGroupTestJSON-1417866451 tempest-ServerGroupTestJSON-1417866451-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61806) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1362.340883] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-c1dda7ed-f3be-450a-9e30-07997ea9d73d tempest-ServerGroupTestJSON-1417866451 tempest-ServerGroupTestJSON-1417866451-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61806) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 1362.341587] env[61806]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ec9711c2-8f6e-4052-b4d3-04589e0474c6 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1362.347627] env[61806]: DEBUG oslo_vmware.api [None req-c1dda7ed-f3be-450a-9e30-07997ea9d73d tempest-ServerGroupTestJSON-1417866451 tempest-ServerGroupTestJSON-1417866451-project-member] Waiting for the task: (returnval){ [ 1362.347627] env[61806]: value = "session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]52ebdbe4-4999-af2d-1597-36e0f2c0d8e4" [ 1362.347627] env[61806]: _type = "Task" [ 1362.347627] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1362.354761] env[61806]: DEBUG oslo_vmware.api [None req-c1dda7ed-f3be-450a-9e30-07997ea9d73d tempest-ServerGroupTestJSON-1417866451 tempest-ServerGroupTestJSON-1417866451-project-member] Task: {'id': session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]52ebdbe4-4999-af2d-1597-36e0f2c0d8e4, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1362.730341] env[61806]: DEBUG oslo_concurrency.lockutils [req-32013297-f290-4604-9db2-f91aea4f53d1 req-04a40c6d-accd-43b2-a36b-4055214281ac service nova] Releasing lock "refresh_cache-8fdc4f99-9dd2-4312-9b13-f1e4170ed489" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1362.857859] env[61806]: DEBUG oslo_vmware.api [None req-c1dda7ed-f3be-450a-9e30-07997ea9d73d tempest-ServerGroupTestJSON-1417866451 tempest-ServerGroupTestJSON-1417866451-project-member] Task: {'id': session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]52ebdbe4-4999-af2d-1597-36e0f2c0d8e4, 'name': SearchDatastore_Task, 'duration_secs': 0.007844} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1362.858706] env[61806]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-4c570743-23e4-4e21-8152-3a2a48e21bb0 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1362.863463] env[61806]: DEBUG oslo_vmware.api [None req-c1dda7ed-f3be-450a-9e30-07997ea9d73d tempest-ServerGroupTestJSON-1417866451 tempest-ServerGroupTestJSON-1417866451-project-member] Waiting for the task: (returnval){ [ 1362.863463] env[61806]: value = "session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]520fa988-e280-2435-dae7-cd9d0cdf65bf" [ 1362.863463] env[61806]: _type = "Task" [ 1362.863463] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1362.870579] env[61806]: DEBUG oslo_vmware.api [None req-c1dda7ed-f3be-450a-9e30-07997ea9d73d tempest-ServerGroupTestJSON-1417866451 tempest-ServerGroupTestJSON-1417866451-project-member] Task: {'id': session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]520fa988-e280-2435-dae7-cd9d0cdf65bf, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1363.374200] env[61806]: DEBUG oslo_vmware.api [None req-c1dda7ed-f3be-450a-9e30-07997ea9d73d tempest-ServerGroupTestJSON-1417866451 tempest-ServerGroupTestJSON-1417866451-project-member] Task: {'id': session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]520fa988-e280-2435-dae7-cd9d0cdf65bf, 'name': SearchDatastore_Task, 'duration_secs': 0.008707} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1363.374484] env[61806]: DEBUG oslo_concurrency.lockutils [None req-c1dda7ed-f3be-450a-9e30-07997ea9d73d tempest-ServerGroupTestJSON-1417866451 tempest-ServerGroupTestJSON-1417866451-project-member] Releasing lock "[datastore1] devstack-image-cache_base/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa.vmdk" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1363.374745] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [None req-c1dda7ed-f3be-450a-9e30-07997ea9d73d tempest-ServerGroupTestJSON-1417866451 tempest-ServerGroupTestJSON-1417866451-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa.vmdk to [datastore1] 8fdc4f99-9dd2-4312-9b13-f1e4170ed489/8fdc4f99-9dd2-4312-9b13-f1e4170ed489.vmdk {{(pid=61806) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 1363.375016] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-2880aa7e-da98-4c64-99a0-c126fc5ebf46 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1363.381598] env[61806]: DEBUG oslo_vmware.api [None req-c1dda7ed-f3be-450a-9e30-07997ea9d73d tempest-ServerGroupTestJSON-1417866451 tempest-ServerGroupTestJSON-1417866451-project-member] Waiting for the task: (returnval){ [ 1363.381598] env[61806]: value = "task-1295359" [ 1363.381598] env[61806]: _type = "Task" [ 1363.381598] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1363.389247] env[61806]: DEBUG oslo_vmware.api [None req-c1dda7ed-f3be-450a-9e30-07997ea9d73d tempest-ServerGroupTestJSON-1417866451 tempest-ServerGroupTestJSON-1417866451-project-member] Task: {'id': task-1295359, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1363.891024] env[61806]: DEBUG oslo_vmware.api [None req-c1dda7ed-f3be-450a-9e30-07997ea9d73d tempest-ServerGroupTestJSON-1417866451 tempest-ServerGroupTestJSON-1417866451-project-member] Task: {'id': task-1295359, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.474492} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1363.891417] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [None req-c1dda7ed-f3be-450a-9e30-07997ea9d73d tempest-ServerGroupTestJSON-1417866451 tempest-ServerGroupTestJSON-1417866451-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa.vmdk to [datastore1] 8fdc4f99-9dd2-4312-9b13-f1e4170ed489/8fdc4f99-9dd2-4312-9b13-f1e4170ed489.vmdk {{(pid=61806) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 1363.891621] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-c1dda7ed-f3be-450a-9e30-07997ea9d73d tempest-ServerGroupTestJSON-1417866451 tempest-ServerGroupTestJSON-1417866451-project-member] [instance: 8fdc4f99-9dd2-4312-9b13-f1e4170ed489] Extending root virtual disk to 1048576 {{(pid=61806) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1363.891790] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-a5237452-2fce-4ec9-9ef5-32dbed78ce5f {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1363.897242] env[61806]: DEBUG oslo_vmware.api [None req-c1dda7ed-f3be-450a-9e30-07997ea9d73d tempest-ServerGroupTestJSON-1417866451 tempest-ServerGroupTestJSON-1417866451-project-member] Waiting for the task: (returnval){ [ 1363.897242] env[61806]: value = "task-1295360" [ 1363.897242] env[61806]: _type = "Task" [ 1363.897242] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1363.904468] env[61806]: DEBUG oslo_vmware.api [None req-c1dda7ed-f3be-450a-9e30-07997ea9d73d tempest-ServerGroupTestJSON-1417866451 tempest-ServerGroupTestJSON-1417866451-project-member] Task: {'id': task-1295360, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1364.406427] env[61806]: DEBUG oslo_vmware.api [None req-c1dda7ed-f3be-450a-9e30-07997ea9d73d tempest-ServerGroupTestJSON-1417866451 tempest-ServerGroupTestJSON-1417866451-project-member] Task: {'id': task-1295360, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.070684} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1364.406708] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-c1dda7ed-f3be-450a-9e30-07997ea9d73d tempest-ServerGroupTestJSON-1417866451 tempest-ServerGroupTestJSON-1417866451-project-member] [instance: 8fdc4f99-9dd2-4312-9b13-f1e4170ed489] Extended root virtual disk {{(pid=61806) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1364.407502] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-36bcb848-ab1f-4574-a883-c177b37e1cf5 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1364.428298] env[61806]: DEBUG nova.virt.vmwareapi.volumeops [None req-c1dda7ed-f3be-450a-9e30-07997ea9d73d tempest-ServerGroupTestJSON-1417866451 tempest-ServerGroupTestJSON-1417866451-project-member] [instance: 8fdc4f99-9dd2-4312-9b13-f1e4170ed489] Reconfiguring VM instance instance-0000006e to attach disk [datastore1] 8fdc4f99-9dd2-4312-9b13-f1e4170ed489/8fdc4f99-9dd2-4312-9b13-f1e4170ed489.vmdk or device None with type sparse {{(pid=61806) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1364.428531] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-9b11def2-ae43-4e13-9d2a-c855e38c7c90 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1364.447617] env[61806]: DEBUG oslo_vmware.api [None req-c1dda7ed-f3be-450a-9e30-07997ea9d73d tempest-ServerGroupTestJSON-1417866451 tempest-ServerGroupTestJSON-1417866451-project-member] Waiting for the task: (returnval){ [ 1364.447617] env[61806]: value = "task-1295361" [ 1364.447617] env[61806]: _type = "Task" [ 1364.447617] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1364.454760] env[61806]: DEBUG oslo_vmware.api [None req-c1dda7ed-f3be-450a-9e30-07997ea9d73d tempest-ServerGroupTestJSON-1417866451 tempest-ServerGroupTestJSON-1417866451-project-member] Task: {'id': task-1295361, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1364.957631] env[61806]: DEBUG oslo_vmware.api [None req-c1dda7ed-f3be-450a-9e30-07997ea9d73d tempest-ServerGroupTestJSON-1417866451 tempest-ServerGroupTestJSON-1417866451-project-member] Task: {'id': task-1295361, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1365.458269] env[61806]: DEBUG oslo_vmware.api [None req-c1dda7ed-f3be-450a-9e30-07997ea9d73d tempest-ServerGroupTestJSON-1417866451 tempest-ServerGroupTestJSON-1417866451-project-member] Task: {'id': task-1295361, 'name': ReconfigVM_Task, 'duration_secs': 0.902345} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1365.458596] env[61806]: DEBUG nova.virt.vmwareapi.volumeops [None req-c1dda7ed-f3be-450a-9e30-07997ea9d73d tempest-ServerGroupTestJSON-1417866451 tempest-ServerGroupTestJSON-1417866451-project-member] [instance: 8fdc4f99-9dd2-4312-9b13-f1e4170ed489] Reconfigured VM instance instance-0000006e to attach disk [datastore1] 8fdc4f99-9dd2-4312-9b13-f1e4170ed489/8fdc4f99-9dd2-4312-9b13-f1e4170ed489.vmdk or device None with type sparse {{(pid=61806) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1365.459258] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-71257752-e687-4b9a-b9ee-3bcf4e4c0257 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1365.464926] env[61806]: DEBUG oslo_vmware.api [None req-c1dda7ed-f3be-450a-9e30-07997ea9d73d tempest-ServerGroupTestJSON-1417866451 tempest-ServerGroupTestJSON-1417866451-project-member] Waiting for the task: (returnval){ [ 1365.464926] env[61806]: value = "task-1295362" [ 1365.464926] env[61806]: _type = "Task" [ 1365.464926] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1365.472424] env[61806]: DEBUG oslo_vmware.api [None req-c1dda7ed-f3be-450a-9e30-07997ea9d73d tempest-ServerGroupTestJSON-1417866451 tempest-ServerGroupTestJSON-1417866451-project-member] Task: {'id': task-1295362, 'name': Rename_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1365.974374] env[61806]: DEBUG oslo_vmware.api [None req-c1dda7ed-f3be-450a-9e30-07997ea9d73d tempest-ServerGroupTestJSON-1417866451 tempest-ServerGroupTestJSON-1417866451-project-member] Task: {'id': task-1295362, 'name': Rename_Task, 'duration_secs': 0.139021} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1365.974757] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [None req-c1dda7ed-f3be-450a-9e30-07997ea9d73d tempest-ServerGroupTestJSON-1417866451 tempest-ServerGroupTestJSON-1417866451-project-member] [instance: 8fdc4f99-9dd2-4312-9b13-f1e4170ed489] Powering on the VM {{(pid=61806) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1365.974897] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-0d69084e-186b-4d35-967f-0ec09d512b4a {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1365.980781] env[61806]: DEBUG oslo_vmware.api [None req-c1dda7ed-f3be-450a-9e30-07997ea9d73d tempest-ServerGroupTestJSON-1417866451 tempest-ServerGroupTestJSON-1417866451-project-member] Waiting for the task: (returnval){ [ 1365.980781] env[61806]: value = "task-1295363" [ 1365.980781] env[61806]: _type = "Task" [ 1365.980781] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1365.987640] env[61806]: DEBUG oslo_vmware.api [None req-c1dda7ed-f3be-450a-9e30-07997ea9d73d tempest-ServerGroupTestJSON-1417866451 tempest-ServerGroupTestJSON-1417866451-project-member] Task: {'id': task-1295363, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1366.492376] env[61806]: DEBUG oslo_vmware.api [None req-c1dda7ed-f3be-450a-9e30-07997ea9d73d tempest-ServerGroupTestJSON-1417866451 tempest-ServerGroupTestJSON-1417866451-project-member] Task: {'id': task-1295363, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1366.991947] env[61806]: DEBUG oslo_vmware.api [None req-c1dda7ed-f3be-450a-9e30-07997ea9d73d tempest-ServerGroupTestJSON-1417866451 tempest-ServerGroupTestJSON-1417866451-project-member] Task: {'id': task-1295363, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1367.492680] env[61806]: DEBUG oslo_vmware.api [None req-c1dda7ed-f3be-450a-9e30-07997ea9d73d tempest-ServerGroupTestJSON-1417866451 tempest-ServerGroupTestJSON-1417866451-project-member] Task: {'id': task-1295363, 'name': PowerOnVM_Task, 'duration_secs': 1.433474} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1367.494054] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [None req-c1dda7ed-f3be-450a-9e30-07997ea9d73d tempest-ServerGroupTestJSON-1417866451 tempest-ServerGroupTestJSON-1417866451-project-member] [instance: 8fdc4f99-9dd2-4312-9b13-f1e4170ed489] Powered on the VM {{(pid=61806) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1367.494054] env[61806]: INFO nova.compute.manager [None req-c1dda7ed-f3be-450a-9e30-07997ea9d73d tempest-ServerGroupTestJSON-1417866451 tempest-ServerGroupTestJSON-1417866451-project-member] [instance: 8fdc4f99-9dd2-4312-9b13-f1e4170ed489] Took 8.10 seconds to spawn the instance on the hypervisor. [ 1367.494054] env[61806]: DEBUG nova.compute.manager [None req-c1dda7ed-f3be-450a-9e30-07997ea9d73d tempest-ServerGroupTestJSON-1417866451 tempest-ServerGroupTestJSON-1417866451-project-member] [instance: 8fdc4f99-9dd2-4312-9b13-f1e4170ed489] Checking state {{(pid=61806) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1367.494451] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-147cf16b-ea35-4551-82db-819e8db1b9a1 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1368.010601] env[61806]: INFO nova.compute.manager [None req-c1dda7ed-f3be-450a-9e30-07997ea9d73d tempest-ServerGroupTestJSON-1417866451 tempest-ServerGroupTestJSON-1417866451-project-member] [instance: 8fdc4f99-9dd2-4312-9b13-f1e4170ed489] Took 13.76 seconds to build instance. [ 1368.476480] env[61806]: DEBUG oslo_concurrency.lockutils [None req-8757cee5-da00-4825-a9f1-6f2cb009d066 tempest-ServerGroupTestJSON-1417866451 tempest-ServerGroupTestJSON-1417866451-project-member] Acquiring lock "8fdc4f99-9dd2-4312-9b13-f1e4170ed489" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1368.512745] env[61806]: DEBUG oslo_concurrency.lockutils [None req-c1dda7ed-f3be-450a-9e30-07997ea9d73d tempest-ServerGroupTestJSON-1417866451 tempest-ServerGroupTestJSON-1417866451-project-member] Lock "8fdc4f99-9dd2-4312-9b13-f1e4170ed489" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 15.274s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1368.513052] env[61806]: DEBUG oslo_concurrency.lockutils [None req-8757cee5-da00-4825-a9f1-6f2cb009d066 tempest-ServerGroupTestJSON-1417866451 tempest-ServerGroupTestJSON-1417866451-project-member] Lock "8fdc4f99-9dd2-4312-9b13-f1e4170ed489" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.037s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1368.513293] env[61806]: DEBUG oslo_concurrency.lockutils [None req-8757cee5-da00-4825-a9f1-6f2cb009d066 tempest-ServerGroupTestJSON-1417866451 tempest-ServerGroupTestJSON-1417866451-project-member] Acquiring lock "8fdc4f99-9dd2-4312-9b13-f1e4170ed489-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1368.513510] env[61806]: DEBUG oslo_concurrency.lockutils [None req-8757cee5-da00-4825-a9f1-6f2cb009d066 tempest-ServerGroupTestJSON-1417866451 tempest-ServerGroupTestJSON-1417866451-project-member] Lock "8fdc4f99-9dd2-4312-9b13-f1e4170ed489-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1368.513688] env[61806]: DEBUG oslo_concurrency.lockutils [None req-8757cee5-da00-4825-a9f1-6f2cb009d066 tempest-ServerGroupTestJSON-1417866451 tempest-ServerGroupTestJSON-1417866451-project-member] Lock "8fdc4f99-9dd2-4312-9b13-f1e4170ed489-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1368.515865] env[61806]: INFO nova.compute.manager [None req-8757cee5-da00-4825-a9f1-6f2cb009d066 tempest-ServerGroupTestJSON-1417866451 tempest-ServerGroupTestJSON-1417866451-project-member] [instance: 8fdc4f99-9dd2-4312-9b13-f1e4170ed489] Terminating instance [ 1368.517596] env[61806]: DEBUG nova.compute.manager [None req-8757cee5-da00-4825-a9f1-6f2cb009d066 tempest-ServerGroupTestJSON-1417866451 tempest-ServerGroupTestJSON-1417866451-project-member] [instance: 8fdc4f99-9dd2-4312-9b13-f1e4170ed489] Start destroying the instance on the hypervisor. {{(pid=61806) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1368.517803] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-8757cee5-da00-4825-a9f1-6f2cb009d066 tempest-ServerGroupTestJSON-1417866451 tempest-ServerGroupTestJSON-1417866451-project-member] [instance: 8fdc4f99-9dd2-4312-9b13-f1e4170ed489] Destroying instance {{(pid=61806) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1368.518710] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6b289d7a-7e1a-4d7e-8e3d-48ba1d9949b9 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1368.526595] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [None req-8757cee5-da00-4825-a9f1-6f2cb009d066 tempest-ServerGroupTestJSON-1417866451 tempest-ServerGroupTestJSON-1417866451-project-member] [instance: 8fdc4f99-9dd2-4312-9b13-f1e4170ed489] Powering off the VM {{(pid=61806) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1368.526827] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-ba48964a-5f58-478a-bbf8-81d231cee8a9 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1368.532459] env[61806]: DEBUG oslo_vmware.api [None req-8757cee5-da00-4825-a9f1-6f2cb009d066 tempest-ServerGroupTestJSON-1417866451 tempest-ServerGroupTestJSON-1417866451-project-member] Waiting for the task: (returnval){ [ 1368.532459] env[61806]: value = "task-1295364" [ 1368.532459] env[61806]: _type = "Task" [ 1368.532459] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1368.539742] env[61806]: DEBUG oslo_vmware.api [None req-8757cee5-da00-4825-a9f1-6f2cb009d066 tempest-ServerGroupTestJSON-1417866451 tempest-ServerGroupTestJSON-1417866451-project-member] Task: {'id': task-1295364, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1369.041779] env[61806]: DEBUG oslo_vmware.api [None req-8757cee5-da00-4825-a9f1-6f2cb009d066 tempest-ServerGroupTestJSON-1417866451 tempest-ServerGroupTestJSON-1417866451-project-member] Task: {'id': task-1295364, 'name': PowerOffVM_Task, 'duration_secs': 0.190513} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1369.042172] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [None req-8757cee5-da00-4825-a9f1-6f2cb009d066 tempest-ServerGroupTestJSON-1417866451 tempest-ServerGroupTestJSON-1417866451-project-member] [instance: 8fdc4f99-9dd2-4312-9b13-f1e4170ed489] Powered off the VM {{(pid=61806) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1369.042296] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-8757cee5-da00-4825-a9f1-6f2cb009d066 tempest-ServerGroupTestJSON-1417866451 tempest-ServerGroupTestJSON-1417866451-project-member] [instance: 8fdc4f99-9dd2-4312-9b13-f1e4170ed489] Unregistering the VM {{(pid=61806) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1369.042497] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-427beefc-f9ae-458e-ab9c-0578bcb2d3b6 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1369.103863] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-8757cee5-da00-4825-a9f1-6f2cb009d066 tempest-ServerGroupTestJSON-1417866451 tempest-ServerGroupTestJSON-1417866451-project-member] [instance: 8fdc4f99-9dd2-4312-9b13-f1e4170ed489] Unregistered the VM {{(pid=61806) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1369.104133] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-8757cee5-da00-4825-a9f1-6f2cb009d066 tempest-ServerGroupTestJSON-1417866451 tempest-ServerGroupTestJSON-1417866451-project-member] [instance: 8fdc4f99-9dd2-4312-9b13-f1e4170ed489] Deleting contents of the VM from datastore datastore1 {{(pid=61806) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1369.104318] env[61806]: DEBUG nova.virt.vmwareapi.ds_util [None req-8757cee5-da00-4825-a9f1-6f2cb009d066 tempest-ServerGroupTestJSON-1417866451 tempest-ServerGroupTestJSON-1417866451-project-member] Deleting the datastore file [datastore1] 8fdc4f99-9dd2-4312-9b13-f1e4170ed489 {{(pid=61806) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1369.104600] env[61806]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-ad1bdb6d-7718-4217-a5d8-3812615dbb15 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1369.111435] env[61806]: DEBUG oslo_vmware.api [None req-8757cee5-da00-4825-a9f1-6f2cb009d066 tempest-ServerGroupTestJSON-1417866451 tempest-ServerGroupTestJSON-1417866451-project-member] Waiting for the task: (returnval){ [ 1369.111435] env[61806]: value = "task-1295366" [ 1369.111435] env[61806]: _type = "Task" [ 1369.111435] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1369.118928] env[61806]: DEBUG oslo_vmware.api [None req-8757cee5-da00-4825-a9f1-6f2cb009d066 tempest-ServerGroupTestJSON-1417866451 tempest-ServerGroupTestJSON-1417866451-project-member] Task: {'id': task-1295366, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1369.621664] env[61806]: DEBUG oslo_vmware.api [None req-8757cee5-da00-4825-a9f1-6f2cb009d066 tempest-ServerGroupTestJSON-1417866451 tempest-ServerGroupTestJSON-1417866451-project-member] Task: {'id': task-1295366, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.134329} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1369.622077] env[61806]: DEBUG nova.virt.vmwareapi.ds_util [None req-8757cee5-da00-4825-a9f1-6f2cb009d066 tempest-ServerGroupTestJSON-1417866451 tempest-ServerGroupTestJSON-1417866451-project-member] Deleted the datastore file {{(pid=61806) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1369.622317] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-8757cee5-da00-4825-a9f1-6f2cb009d066 tempest-ServerGroupTestJSON-1417866451 tempest-ServerGroupTestJSON-1417866451-project-member] [instance: 8fdc4f99-9dd2-4312-9b13-f1e4170ed489] Deleted contents of the VM from datastore datastore1 {{(pid=61806) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1369.622557] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-8757cee5-da00-4825-a9f1-6f2cb009d066 tempest-ServerGroupTestJSON-1417866451 tempest-ServerGroupTestJSON-1417866451-project-member] [instance: 8fdc4f99-9dd2-4312-9b13-f1e4170ed489] Instance destroyed {{(pid=61806) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1369.622788] env[61806]: INFO nova.compute.manager [None req-8757cee5-da00-4825-a9f1-6f2cb009d066 tempest-ServerGroupTestJSON-1417866451 tempest-ServerGroupTestJSON-1417866451-project-member] [instance: 8fdc4f99-9dd2-4312-9b13-f1e4170ed489] Took 1.10 seconds to destroy the instance on the hypervisor. [ 1369.623083] env[61806]: DEBUG oslo.service.loopingcall [None req-8757cee5-da00-4825-a9f1-6f2cb009d066 tempest-ServerGroupTestJSON-1417866451 tempest-ServerGroupTestJSON-1417866451-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61806) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1369.623303] env[61806]: DEBUG nova.compute.manager [-] [instance: 8fdc4f99-9dd2-4312-9b13-f1e4170ed489] Deallocating network for instance {{(pid=61806) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1369.623399] env[61806]: DEBUG nova.network.neutron [-] [instance: 8fdc4f99-9dd2-4312-9b13-f1e4170ed489] deallocate_for_instance() {{(pid=61806) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1369.851267] env[61806]: DEBUG nova.compute.manager [req-b28a4aa8-6934-4fd2-b72d-41389f07e462 req-7734dfe4-3b54-4f10-a91c-f6f837698f09 service nova] [instance: 8fdc4f99-9dd2-4312-9b13-f1e4170ed489] Received event network-vif-deleted-84c9dc10-8d31-4fb4-aea2-435e8940c4e4 {{(pid=61806) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1369.851267] env[61806]: INFO nova.compute.manager [req-b28a4aa8-6934-4fd2-b72d-41389f07e462 req-7734dfe4-3b54-4f10-a91c-f6f837698f09 service nova] [instance: 8fdc4f99-9dd2-4312-9b13-f1e4170ed489] Neutron deleted interface 84c9dc10-8d31-4fb4-aea2-435e8940c4e4; detaching it from the instance and deleting it from the info cache [ 1369.851432] env[61806]: DEBUG nova.network.neutron [req-b28a4aa8-6934-4fd2-b72d-41389f07e462 req-7734dfe4-3b54-4f10-a91c-f6f837698f09 service nova] [instance: 8fdc4f99-9dd2-4312-9b13-f1e4170ed489] Updating instance_info_cache with network_info: [] {{(pid=61806) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1370.328010] env[61806]: DEBUG nova.network.neutron [-] [instance: 8fdc4f99-9dd2-4312-9b13-f1e4170ed489] Updating instance_info_cache with network_info: [] {{(pid=61806) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1370.354044] env[61806]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-cc686642-cc18-4fbf-afa4-04123219262f {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1370.362801] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-667f4a99-c1f4-483a-9925-064a0c261869 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1370.386482] env[61806]: DEBUG nova.compute.manager [req-b28a4aa8-6934-4fd2-b72d-41389f07e462 req-7734dfe4-3b54-4f10-a91c-f6f837698f09 service nova] [instance: 8fdc4f99-9dd2-4312-9b13-f1e4170ed489] Detach interface failed, port_id=84c9dc10-8d31-4fb4-aea2-435e8940c4e4, reason: Instance 8fdc4f99-9dd2-4312-9b13-f1e4170ed489 could not be found. {{(pid=61806) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10963}} [ 1370.831240] env[61806]: INFO nova.compute.manager [-] [instance: 8fdc4f99-9dd2-4312-9b13-f1e4170ed489] Took 1.21 seconds to deallocate network for instance. [ 1371.018827] env[61806]: DEBUG oslo_service.periodic_task [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=61806) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1371.022500] env[61806]: DEBUG oslo_service.periodic_task [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Running periodic task ComputeManager._cleanup_expired_console_auth_tokens {{(pid=61806) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1371.337196] env[61806]: DEBUG oslo_concurrency.lockutils [None req-8757cee5-da00-4825-a9f1-6f2cb009d066 tempest-ServerGroupTestJSON-1417866451 tempest-ServerGroupTestJSON-1417866451-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1371.337500] env[61806]: DEBUG oslo_concurrency.lockutils [None req-8757cee5-da00-4825-a9f1-6f2cb009d066 tempest-ServerGroupTestJSON-1417866451 tempest-ServerGroupTestJSON-1417866451-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.001s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1371.337711] env[61806]: DEBUG nova.objects.instance [None req-8757cee5-da00-4825-a9f1-6f2cb009d066 tempest-ServerGroupTestJSON-1417866451 tempest-ServerGroupTestJSON-1417866451-project-member] Lazy-loading 'resources' on Instance uuid 8fdc4f99-9dd2-4312-9b13-f1e4170ed489 {{(pid=61806) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1371.871992] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b6f0d8f0-3e3a-4168-8ab8-bf797097dd28 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1371.879355] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ccbb5597-bc6c-4b63-8da9-b2a8acdbb3ad {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1371.907944] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-74ba97a4-7009-4096-a963-a2eb7cb0d611 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1371.914535] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-55cb5e69-288a-43e0-b1cf-ccf05eb8eef1 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1371.926998] env[61806]: DEBUG nova.compute.provider_tree [None req-8757cee5-da00-4825-a9f1-6f2cb009d066 tempest-ServerGroupTestJSON-1417866451 tempest-ServerGroupTestJSON-1417866451-project-member] Updating inventory in ProviderTree for provider a2858be1-fd22-4e08-979e-87ad25293407 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 139, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61806) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1372.445931] env[61806]: ERROR nova.scheduler.client.report [None req-8757cee5-da00-4825-a9f1-6f2cb009d066 tempest-ServerGroupTestJSON-1417866451 tempest-ServerGroupTestJSON-1417866451-project-member] [req-a09b283f-4129-413c-b60c-f5a0967bf060] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 139, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID a2858be1-fd22-4e08-979e-87ad25293407. Got 409: {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-a09b283f-4129-413c-b60c-f5a0967bf060"}]} [ 1372.461954] env[61806]: DEBUG nova.scheduler.client.report [None req-8757cee5-da00-4825-a9f1-6f2cb009d066 tempest-ServerGroupTestJSON-1417866451 tempest-ServerGroupTestJSON-1417866451-project-member] Refreshing inventories for resource provider a2858be1-fd22-4e08-979e-87ad25293407 {{(pid=61806) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:818}} [ 1372.474867] env[61806]: DEBUG nova.scheduler.client.report [None req-8757cee5-da00-4825-a9f1-6f2cb009d066 tempest-ServerGroupTestJSON-1417866451 tempest-ServerGroupTestJSON-1417866451-project-member] Updating ProviderTree inventory for provider a2858be1-fd22-4e08-979e-87ad25293407 from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 138, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61806) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:782}} [ 1372.475110] env[61806]: DEBUG nova.compute.provider_tree [None req-8757cee5-da00-4825-a9f1-6f2cb009d066 tempest-ServerGroupTestJSON-1417866451 tempest-ServerGroupTestJSON-1417866451-project-member] Updating inventory in ProviderTree for provider a2858be1-fd22-4e08-979e-87ad25293407 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 138, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61806) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1372.485657] env[61806]: DEBUG nova.scheduler.client.report [None req-8757cee5-da00-4825-a9f1-6f2cb009d066 tempest-ServerGroupTestJSON-1417866451 tempest-ServerGroupTestJSON-1417866451-project-member] Refreshing aggregate associations for resource provider a2858be1-fd22-4e08-979e-87ad25293407, aggregates: None {{(pid=61806) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:827}} [ 1372.501845] env[61806]: DEBUG nova.scheduler.client.report [None req-8757cee5-da00-4825-a9f1-6f2cb009d066 tempest-ServerGroupTestJSON-1417866451 tempest-ServerGroupTestJSON-1417866451-project-member] Refreshing trait associations for resource provider a2858be1-fd22-4e08-979e-87ad25293407, traits: COMPUTE_IMAGE_TYPE_ISO,COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_SAME_HOST_COLD_MIGRATE,COMPUTE_NET_ATTACH_INTERFACE,COMPUTE_NODE {{(pid=61806) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:839}} [ 1372.522949] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-64557551-d9ef-485e-8098-936570161cac {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1372.525563] env[61806]: DEBUG oslo_service.periodic_task [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Running periodic task ComputeManager._run_pending_deletes {{(pid=61806) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1372.525713] env[61806]: DEBUG nova.compute.manager [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Cleaning up deleted instances {{(pid=61806) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11226}} [ 1372.531889] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ed28ac16-5bad-4e72-80fb-9c11db3f9a1d {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1372.561868] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1b382222-4a7b-41be-9575-3f51e7169b63 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1372.569386] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a54e1860-6108-4920-ad6f-816a0e7f8d98 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1372.583107] env[61806]: DEBUG nova.compute.provider_tree [None req-8757cee5-da00-4825-a9f1-6f2cb009d066 tempest-ServerGroupTestJSON-1417866451 tempest-ServerGroupTestJSON-1417866451-project-member] Updating inventory in ProviderTree for provider a2858be1-fd22-4e08-979e-87ad25293407 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 139, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61806) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1373.031493] env[61806]: DEBUG nova.compute.manager [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] There are 10 instances to clean {{(pid=61806) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11235}} [ 1373.031792] env[61806]: DEBUG nova.compute.manager [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] [instance: 7afc6260-7569-44c4-9764-7aa08c662c1c] Instance has had 0 of 5 cleanup attempts {{(pid=61806) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1373.112132] env[61806]: DEBUG nova.scheduler.client.report [None req-8757cee5-da00-4825-a9f1-6f2cb009d066 tempest-ServerGroupTestJSON-1417866451 tempest-ServerGroupTestJSON-1417866451-project-member] Updated inventory for provider a2858be1-fd22-4e08-979e-87ad25293407 with generation 157 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 139, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61806) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:971}} [ 1373.112437] env[61806]: DEBUG nova.compute.provider_tree [None req-8757cee5-da00-4825-a9f1-6f2cb009d066 tempest-ServerGroupTestJSON-1417866451 tempest-ServerGroupTestJSON-1417866451-project-member] Updating resource provider a2858be1-fd22-4e08-979e-87ad25293407 generation from 157 to 158 during operation: update_inventory {{(pid=61806) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 1373.112628] env[61806]: DEBUG nova.compute.provider_tree [None req-8757cee5-da00-4825-a9f1-6f2cb009d066 tempest-ServerGroupTestJSON-1417866451 tempest-ServerGroupTestJSON-1417866451-project-member] Updating inventory in ProviderTree for provider a2858be1-fd22-4e08-979e-87ad25293407 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 139, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61806) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1373.535666] env[61806]: DEBUG nova.compute.manager [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] [instance: 4f26445f-35d0-4f53-9014-ef3e2eb1cbdf] Instance has had 0 of 5 cleanup attempts {{(pid=61806) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1373.616805] env[61806]: DEBUG oslo_concurrency.lockutils [None req-8757cee5-da00-4825-a9f1-6f2cb009d066 tempest-ServerGroupTestJSON-1417866451 tempest-ServerGroupTestJSON-1417866451-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.279s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1373.639353] env[61806]: INFO nova.scheduler.client.report [None req-8757cee5-da00-4825-a9f1-6f2cb009d066 tempest-ServerGroupTestJSON-1417866451 tempest-ServerGroupTestJSON-1417866451-project-member] Deleted allocations for instance 8fdc4f99-9dd2-4312-9b13-f1e4170ed489 [ 1374.038789] env[61806]: DEBUG nova.compute.manager [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] [instance: c66fe2e7-7af0-48fa-8509-fcc24c1a437b] Instance has had 0 of 5 cleanup attempts {{(pid=61806) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1374.146225] env[61806]: DEBUG oslo_concurrency.lockutils [None req-8757cee5-da00-4825-a9f1-6f2cb009d066 tempest-ServerGroupTestJSON-1417866451 tempest-ServerGroupTestJSON-1417866451-project-member] Lock "8fdc4f99-9dd2-4312-9b13-f1e4170ed489" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.633s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1374.542500] env[61806]: DEBUG nova.compute.manager [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] [instance: 29210bcd-0fe1-4ed9-a459-abc8016a4255] Instance has had 0 of 5 cleanup attempts {{(pid=61806) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1375.047220] env[61806]: DEBUG nova.compute.manager [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] [instance: cb9e9bac-d06c-4479-88e3-616b0f26ad7e] Instance has had 0 of 5 cleanup attempts {{(pid=61806) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1375.550818] env[61806]: DEBUG nova.compute.manager [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] [instance: ce2ad097-4c3f-4b55-b018-ac788aa5b662] Instance has had 0 of 5 cleanup attempts {{(pid=61806) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1376.054662] env[61806]: DEBUG nova.compute.manager [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] [instance: 7445d2c5-26bd-4f8d-8653-51c721ea801f] Instance has had 0 of 5 cleanup attempts {{(pid=61806) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1376.558428] env[61806]: DEBUG nova.compute.manager [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] [instance: 06658134-4c6d-4911-9a55-b6805f555c78] Instance has had 0 of 5 cleanup attempts {{(pid=61806) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1377.061817] env[61806]: DEBUG nova.compute.manager [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] [instance: d0669232-6f33-4b30-97c2-2e4239af1ad0] Instance has had 0 of 5 cleanup attempts {{(pid=61806) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1377.565226] env[61806]: DEBUG nova.compute.manager [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] [instance: fd9e520a-2e7b-43d4-b84d-8933c32cc6a6] Instance has had 0 of 5 cleanup attempts {{(pid=61806) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1379.565354] env[61806]: DEBUG oslo_service.periodic_task [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=61806) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1379.565629] env[61806]: DEBUG nova.compute.manager [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Starting heal instance info cache {{(pid=61806) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9926}} [ 1379.565827] env[61806]: DEBUG nova.compute.manager [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Rebuilding the list of instances to heal {{(pid=61806) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9930}} [ 1379.843020] env[61806]: DEBUG oslo_concurrency.lockutils [None req-1cc2ddfb-3444-4874-84d6-f013775a320a tempest-AttachVolumeTestJSON-132925025 tempest-AttachVolumeTestJSON-132925025-project-member] Acquiring lock "e3360f00-a42a-40f2-9e6b-813547ebfd2b" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1379.843020] env[61806]: DEBUG oslo_concurrency.lockutils [None req-1cc2ddfb-3444-4874-84d6-f013775a320a tempest-AttachVolumeTestJSON-132925025 tempest-AttachVolumeTestJSON-132925025-project-member] Lock "e3360f00-a42a-40f2-9e6b-813547ebfd2b" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1380.069747] env[61806]: DEBUG nova.compute.manager [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Didn't find any instances for network info cache update. {{(pid=61806) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10012}} [ 1380.069747] env[61806]: DEBUG oslo_service.periodic_task [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=61806) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1380.069747] env[61806]: DEBUG oslo_service.periodic_task [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=61806) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1380.069747] env[61806]: DEBUG oslo_service.periodic_task [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=61806) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1380.069747] env[61806]: DEBUG oslo_service.periodic_task [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=61806) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1380.069747] env[61806]: DEBUG nova.compute.manager [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=61806) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10545}} [ 1380.069747] env[61806]: DEBUG oslo_service.periodic_task [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Running periodic task ComputeManager.update_available_resource {{(pid=61806) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1380.346032] env[61806]: DEBUG nova.compute.manager [None req-1cc2ddfb-3444-4874-84d6-f013775a320a tempest-AttachVolumeTestJSON-132925025 tempest-AttachVolumeTestJSON-132925025-project-member] [instance: e3360f00-a42a-40f2-9e6b-813547ebfd2b] Starting instance... {{(pid=61806) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1380.573770] env[61806]: DEBUG oslo_concurrency.lockutils [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1380.573770] env[61806]: DEBUG oslo_concurrency.lockutils [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 0.000s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1380.573770] env[61806]: DEBUG oslo_concurrency.lockutils [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1380.574361] env[61806]: DEBUG nova.compute.resource_tracker [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=61806) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 1380.575261] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4bcefc64-7354-461f-bac0-09030af2b94c {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1380.583484] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eedcb815-eee9-42d1-a085-a5fe0850a07d {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1380.597768] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8ef2342c-9d91-4f5b-a923-d60369095e19 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1380.604529] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6735d871-a97a-42fe-840c-6a0ad57d385d {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1380.633542] env[61806]: DEBUG nova.compute.resource_tracker [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=181122MB free_disk=139GB free_vcpus=48 pci_devices=None {{(pid=61806) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 1380.633757] env[61806]: DEBUG oslo_concurrency.lockutils [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1380.633914] env[61806]: DEBUG oslo_concurrency.lockutils [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 0.000s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1380.931191] env[61806]: DEBUG oslo_concurrency.lockutils [None req-1cc2ddfb-3444-4874-84d6-f013775a320a tempest-AttachVolumeTestJSON-132925025 tempest-AttachVolumeTestJSON-132925025-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1382.159415] env[61806]: DEBUG nova.compute.resource_tracker [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Instance e3360f00-a42a-40f2-9e6b-813547ebfd2b has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61806) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 1382.159798] env[61806]: DEBUG nova.compute.resource_tracker [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Total usable vcpus: 48, total allocated vcpus: 0 {{(pid=61806) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 1382.159798] env[61806]: DEBUG nova.compute.resource_tracker [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=512MB phys_disk=200GB used_disk=0GB total_vcpus=48 used_vcpus=0 pci_stats=[] {{(pid=61806) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 1382.186599] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4fea8110-2dc0-494d-b70b-cb7ed0ff90af {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1382.193891] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e59ab1d7-ed6d-4a3f-ba60-b449738b2ce1 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1382.223520] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-615935f2-0214-4668-8c1a-9c10f48d4230 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1382.230350] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a702d428-87b5-4f05-a02f-7b2fe7f3f83a {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1382.242874] env[61806]: DEBUG nova.compute.provider_tree [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Inventory has not changed in ProviderTree for provider: a2858be1-fd22-4e08-979e-87ad25293407 {{(pid=61806) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1382.745432] env[61806]: DEBUG nova.scheduler.client.report [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Inventory has not changed for provider a2858be1-fd22-4e08-979e-87ad25293407 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 139, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61806) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1383.250597] env[61806]: DEBUG nova.compute.resource_tracker [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=61806) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 1383.250968] env[61806]: DEBUG oslo_concurrency.lockutils [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 2.617s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1383.251122] env[61806]: DEBUG oslo_concurrency.lockutils [None req-1cc2ddfb-3444-4874-84d6-f013775a320a tempest-AttachVolumeTestJSON-132925025 tempest-AttachVolumeTestJSON-132925025-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 2.320s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1383.252645] env[61806]: INFO nova.compute.claims [None req-1cc2ddfb-3444-4874-84d6-f013775a320a tempest-AttachVolumeTestJSON-132925025 tempest-AttachVolumeTestJSON-132925025-project-member] [instance: e3360f00-a42a-40f2-9e6b-813547ebfd2b] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1384.023146] env[61806]: DEBUG oslo_service.periodic_task [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=61806) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1384.023692] env[61806]: DEBUG oslo_service.periodic_task [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=61806) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1384.023889] env[61806]: DEBUG oslo_service.periodic_task [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Running periodic task ComputeManager._cleanup_incomplete_migrations {{(pid=61806) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1384.024042] env[61806]: DEBUG nova.compute.manager [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Cleaning up deleted instances with incomplete migration {{(pid=61806) _cleanup_incomplete_migrations /opt/stack/nova/nova/compute/manager.py:11264}} [ 1384.287987] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-367e63b9-cc61-4bca-8fee-6cc2f14d528e {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1384.295819] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d0d759c6-9ef2-4eda-9979-dadf55f9258b {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1384.325365] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9860f498-6c59-4237-8f8d-c0f34cd7d5d0 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1384.332177] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ee95bb52-40b6-455a-9fbf-ef11a09a1a93 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1384.344690] env[61806]: DEBUG nova.compute.provider_tree [None req-1cc2ddfb-3444-4874-84d6-f013775a320a tempest-AttachVolumeTestJSON-132925025 tempest-AttachVolumeTestJSON-132925025-project-member] Inventory has not changed in ProviderTree for provider: a2858be1-fd22-4e08-979e-87ad25293407 {{(pid=61806) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1384.848280] env[61806]: DEBUG nova.scheduler.client.report [None req-1cc2ddfb-3444-4874-84d6-f013775a320a tempest-AttachVolumeTestJSON-132925025 tempest-AttachVolumeTestJSON-132925025-project-member] Inventory has not changed for provider a2858be1-fd22-4e08-979e-87ad25293407 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 139, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61806) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1385.353093] env[61806]: DEBUG oslo_concurrency.lockutils [None req-1cc2ddfb-3444-4874-84d6-f013775a320a tempest-AttachVolumeTestJSON-132925025 tempest-AttachVolumeTestJSON-132925025-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.102s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1385.353669] env[61806]: DEBUG nova.compute.manager [None req-1cc2ddfb-3444-4874-84d6-f013775a320a tempest-AttachVolumeTestJSON-132925025 tempest-AttachVolumeTestJSON-132925025-project-member] [instance: e3360f00-a42a-40f2-9e6b-813547ebfd2b] Start building networks asynchronously for instance. {{(pid=61806) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1385.858867] env[61806]: DEBUG nova.compute.utils [None req-1cc2ddfb-3444-4874-84d6-f013775a320a tempest-AttachVolumeTestJSON-132925025 tempest-AttachVolumeTestJSON-132925025-project-member] Using /dev/sd instead of None {{(pid=61806) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1385.860386] env[61806]: DEBUG nova.compute.manager [None req-1cc2ddfb-3444-4874-84d6-f013775a320a tempest-AttachVolumeTestJSON-132925025 tempest-AttachVolumeTestJSON-132925025-project-member] [instance: e3360f00-a42a-40f2-9e6b-813547ebfd2b] Allocating IP information in the background. {{(pid=61806) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1385.861633] env[61806]: DEBUG nova.network.neutron [None req-1cc2ddfb-3444-4874-84d6-f013775a320a tempest-AttachVolumeTestJSON-132925025 tempest-AttachVolumeTestJSON-132925025-project-member] [instance: e3360f00-a42a-40f2-9e6b-813547ebfd2b] allocate_for_instance() {{(pid=61806) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 1385.907587] env[61806]: DEBUG nova.policy [None req-1cc2ddfb-3444-4874-84d6-f013775a320a tempest-AttachVolumeTestJSON-132925025 tempest-AttachVolumeTestJSON-132925025-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '4ac9f7259f12427ba22b791d5d190943', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '833defcd470d46ceacd7886d1f4fb0cf', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61806) authorize /opt/stack/nova/nova/policy.py:201}} [ 1386.155422] env[61806]: DEBUG nova.network.neutron [None req-1cc2ddfb-3444-4874-84d6-f013775a320a tempest-AttachVolumeTestJSON-132925025 tempest-AttachVolumeTestJSON-132925025-project-member] [instance: e3360f00-a42a-40f2-9e6b-813547ebfd2b] Successfully created port: 3c8061f1-58c9-49f2-bdc9-e43e9936373e {{(pid=61806) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1386.363721] env[61806]: DEBUG nova.compute.manager [None req-1cc2ddfb-3444-4874-84d6-f013775a320a tempest-AttachVolumeTestJSON-132925025 tempest-AttachVolumeTestJSON-132925025-project-member] [instance: e3360f00-a42a-40f2-9e6b-813547ebfd2b] Start building block device mappings for instance. {{(pid=61806) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1387.373015] env[61806]: DEBUG nova.compute.manager [None req-1cc2ddfb-3444-4874-84d6-f013775a320a tempest-AttachVolumeTestJSON-132925025 tempest-AttachVolumeTestJSON-132925025-project-member] [instance: e3360f00-a42a-40f2-9e6b-813547ebfd2b] Start spawning the instance on the hypervisor. {{(pid=61806) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1387.400861] env[61806]: DEBUG nova.virt.hardware [None req-1cc2ddfb-3444-4874-84d6-f013775a320a tempest-AttachVolumeTestJSON-132925025 tempest-AttachVolumeTestJSON-132925025-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-15T18:05:10Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-15T18:04:52Z,direct_url=,disk_format='vmdk',id=ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='84ba83fce2c349a6988173c5d6fc3b07',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-15T18:04:53Z,virtual_size=,visibility=), allow threads: False {{(pid=61806) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1387.401127] env[61806]: DEBUG nova.virt.hardware [None req-1cc2ddfb-3444-4874-84d6-f013775a320a tempest-AttachVolumeTestJSON-132925025 tempest-AttachVolumeTestJSON-132925025-project-member] Flavor limits 0:0:0 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1387.401300] env[61806]: DEBUG nova.virt.hardware [None req-1cc2ddfb-3444-4874-84d6-f013775a320a tempest-AttachVolumeTestJSON-132925025 tempest-AttachVolumeTestJSON-132925025-project-member] Image limits 0:0:0 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1387.401499] env[61806]: DEBUG nova.virt.hardware [None req-1cc2ddfb-3444-4874-84d6-f013775a320a tempest-AttachVolumeTestJSON-132925025 tempest-AttachVolumeTestJSON-132925025-project-member] Flavor pref 0:0:0 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1387.401648] env[61806]: DEBUG nova.virt.hardware [None req-1cc2ddfb-3444-4874-84d6-f013775a320a tempest-AttachVolumeTestJSON-132925025 tempest-AttachVolumeTestJSON-132925025-project-member] Image pref 0:0:0 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1387.401803] env[61806]: DEBUG nova.virt.hardware [None req-1cc2ddfb-3444-4874-84d6-f013775a320a tempest-AttachVolumeTestJSON-132925025 tempest-AttachVolumeTestJSON-132925025-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1387.402027] env[61806]: DEBUG nova.virt.hardware [None req-1cc2ddfb-3444-4874-84d6-f013775a320a tempest-AttachVolumeTestJSON-132925025 tempest-AttachVolumeTestJSON-132925025-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61806) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1387.402227] env[61806]: DEBUG nova.virt.hardware [None req-1cc2ddfb-3444-4874-84d6-f013775a320a tempest-AttachVolumeTestJSON-132925025 tempest-AttachVolumeTestJSON-132925025-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61806) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1387.402422] env[61806]: DEBUG nova.virt.hardware [None req-1cc2ddfb-3444-4874-84d6-f013775a320a tempest-AttachVolumeTestJSON-132925025 tempest-AttachVolumeTestJSON-132925025-project-member] Got 1 possible topologies {{(pid=61806) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1387.402595] env[61806]: DEBUG nova.virt.hardware [None req-1cc2ddfb-3444-4874-84d6-f013775a320a tempest-AttachVolumeTestJSON-132925025 tempest-AttachVolumeTestJSON-132925025-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61806) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1387.402773] env[61806]: DEBUG nova.virt.hardware [None req-1cc2ddfb-3444-4874-84d6-f013775a320a tempest-AttachVolumeTestJSON-132925025 tempest-AttachVolumeTestJSON-132925025-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61806) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1387.403671] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d4fe77b8-3daa-40a4-8c0b-8597b01e6d57 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1387.411459] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b6f56724-8f62-4cbe-b12d-5fbb9e76244e {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1387.515158] env[61806]: DEBUG nova.compute.manager [req-c6fcf947-66e3-49cb-8774-3b8b67fef738 req-ddb93aa0-e799-4ec1-944c-8f62e4253222 service nova] [instance: e3360f00-a42a-40f2-9e6b-813547ebfd2b] Received event network-vif-plugged-3c8061f1-58c9-49f2-bdc9-e43e9936373e {{(pid=61806) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1387.515557] env[61806]: DEBUG oslo_concurrency.lockutils [req-c6fcf947-66e3-49cb-8774-3b8b67fef738 req-ddb93aa0-e799-4ec1-944c-8f62e4253222 service nova] Acquiring lock "e3360f00-a42a-40f2-9e6b-813547ebfd2b-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1387.515917] env[61806]: DEBUG oslo_concurrency.lockutils [req-c6fcf947-66e3-49cb-8774-3b8b67fef738 req-ddb93aa0-e799-4ec1-944c-8f62e4253222 service nova] Lock "e3360f00-a42a-40f2-9e6b-813547ebfd2b-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1387.516252] env[61806]: DEBUG oslo_concurrency.lockutils [req-c6fcf947-66e3-49cb-8774-3b8b67fef738 req-ddb93aa0-e799-4ec1-944c-8f62e4253222 service nova] Lock "e3360f00-a42a-40f2-9e6b-813547ebfd2b-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1387.516565] env[61806]: DEBUG nova.compute.manager [req-c6fcf947-66e3-49cb-8774-3b8b67fef738 req-ddb93aa0-e799-4ec1-944c-8f62e4253222 service nova] [instance: e3360f00-a42a-40f2-9e6b-813547ebfd2b] No waiting events found dispatching network-vif-plugged-3c8061f1-58c9-49f2-bdc9-e43e9936373e {{(pid=61806) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1387.516882] env[61806]: WARNING nova.compute.manager [req-c6fcf947-66e3-49cb-8774-3b8b67fef738 req-ddb93aa0-e799-4ec1-944c-8f62e4253222 service nova] [instance: e3360f00-a42a-40f2-9e6b-813547ebfd2b] Received unexpected event network-vif-plugged-3c8061f1-58c9-49f2-bdc9-e43e9936373e for instance with vm_state building and task_state spawning. [ 1387.604298] env[61806]: DEBUG nova.network.neutron [None req-1cc2ddfb-3444-4874-84d6-f013775a320a tempest-AttachVolumeTestJSON-132925025 tempest-AttachVolumeTestJSON-132925025-project-member] [instance: e3360f00-a42a-40f2-9e6b-813547ebfd2b] Successfully updated port: 3c8061f1-58c9-49f2-bdc9-e43e9936373e {{(pid=61806) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1388.109402] env[61806]: DEBUG oslo_concurrency.lockutils [None req-1cc2ddfb-3444-4874-84d6-f013775a320a tempest-AttachVolumeTestJSON-132925025 tempest-AttachVolumeTestJSON-132925025-project-member] Acquiring lock "refresh_cache-e3360f00-a42a-40f2-9e6b-813547ebfd2b" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1388.109576] env[61806]: DEBUG oslo_concurrency.lockutils [None req-1cc2ddfb-3444-4874-84d6-f013775a320a tempest-AttachVolumeTestJSON-132925025 tempest-AttachVolumeTestJSON-132925025-project-member] Acquired lock "refresh_cache-e3360f00-a42a-40f2-9e6b-813547ebfd2b" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1388.109781] env[61806]: DEBUG nova.network.neutron [None req-1cc2ddfb-3444-4874-84d6-f013775a320a tempest-AttachVolumeTestJSON-132925025 tempest-AttachVolumeTestJSON-132925025-project-member] [instance: e3360f00-a42a-40f2-9e6b-813547ebfd2b] Building network info cache for instance {{(pid=61806) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1388.640374] env[61806]: DEBUG nova.network.neutron [None req-1cc2ddfb-3444-4874-84d6-f013775a320a tempest-AttachVolumeTestJSON-132925025 tempest-AttachVolumeTestJSON-132925025-project-member] [instance: e3360f00-a42a-40f2-9e6b-813547ebfd2b] Instance cache missing network info. {{(pid=61806) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 1388.758591] env[61806]: DEBUG nova.network.neutron [None req-1cc2ddfb-3444-4874-84d6-f013775a320a tempest-AttachVolumeTestJSON-132925025 tempest-AttachVolumeTestJSON-132925025-project-member] [instance: e3360f00-a42a-40f2-9e6b-813547ebfd2b] Updating instance_info_cache with network_info: [{"id": "3c8061f1-58c9-49f2-bdc9-e43e9936373e", "address": "fa:16:3e:d4:61:cb", "network": {"id": "0fd518d3-c254-4dba-a147-378c3c029c51", "bridge": "br-int", "label": "tempest-AttachVolumeTestJSON-538202287-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "833defcd470d46ceacd7886d1f4fb0cf", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "91b0f7e5-0d1a-46e2-bf73-09656211dea2", "external-id": "nsx-vlan-transportzone-488", "segmentation_id": 488, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3c8061f1-58", "ovs_interfaceid": "3c8061f1-58c9-49f2-bdc9-e43e9936373e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61806) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1389.261607] env[61806]: DEBUG oslo_concurrency.lockutils [None req-1cc2ddfb-3444-4874-84d6-f013775a320a tempest-AttachVolumeTestJSON-132925025 tempest-AttachVolumeTestJSON-132925025-project-member] Releasing lock "refresh_cache-e3360f00-a42a-40f2-9e6b-813547ebfd2b" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1389.261995] env[61806]: DEBUG nova.compute.manager [None req-1cc2ddfb-3444-4874-84d6-f013775a320a tempest-AttachVolumeTestJSON-132925025 tempest-AttachVolumeTestJSON-132925025-project-member] [instance: e3360f00-a42a-40f2-9e6b-813547ebfd2b] Instance network_info: |[{"id": "3c8061f1-58c9-49f2-bdc9-e43e9936373e", "address": "fa:16:3e:d4:61:cb", "network": {"id": "0fd518d3-c254-4dba-a147-378c3c029c51", "bridge": "br-int", "label": "tempest-AttachVolumeTestJSON-538202287-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "833defcd470d46ceacd7886d1f4fb0cf", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "91b0f7e5-0d1a-46e2-bf73-09656211dea2", "external-id": "nsx-vlan-transportzone-488", "segmentation_id": 488, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3c8061f1-58", "ovs_interfaceid": "3c8061f1-58c9-49f2-bdc9-e43e9936373e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61806) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1389.262504] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-1cc2ddfb-3444-4874-84d6-f013775a320a tempest-AttachVolumeTestJSON-132925025 tempest-AttachVolumeTestJSON-132925025-project-member] [instance: e3360f00-a42a-40f2-9e6b-813547ebfd2b] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:d4:61:cb', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '91b0f7e5-0d1a-46e2-bf73-09656211dea2', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '3c8061f1-58c9-49f2-bdc9-e43e9936373e', 'vif_model': 'vmxnet3'}] {{(pid=61806) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1389.270922] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [None req-1cc2ddfb-3444-4874-84d6-f013775a320a tempest-AttachVolumeTestJSON-132925025 tempest-AttachVolumeTestJSON-132925025-project-member] Creating folder: Project (833defcd470d46ceacd7886d1f4fb0cf). Parent ref: group-v277609. {{(pid=61806) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 1389.271233] env[61806]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-221106d0-f0ed-4701-8baa-4ecc04798cdc {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1389.284821] env[61806]: INFO nova.virt.vmwareapi.vm_util [None req-1cc2ddfb-3444-4874-84d6-f013775a320a tempest-AttachVolumeTestJSON-132925025 tempest-AttachVolumeTestJSON-132925025-project-member] Created folder: Project (833defcd470d46ceacd7886d1f4fb0cf) in parent group-v277609. [ 1389.285063] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [None req-1cc2ddfb-3444-4874-84d6-f013775a320a tempest-AttachVolumeTestJSON-132925025 tempest-AttachVolumeTestJSON-132925025-project-member] Creating folder: Instances. Parent ref: group-v277782. {{(pid=61806) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 1389.285315] env[61806]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-7ea823d6-15f5-43da-8ffc-5d7cbfcee6a7 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1389.296249] env[61806]: INFO nova.virt.vmwareapi.vm_util [None req-1cc2ddfb-3444-4874-84d6-f013775a320a tempest-AttachVolumeTestJSON-132925025 tempest-AttachVolumeTestJSON-132925025-project-member] Created folder: Instances in parent group-v277782. [ 1389.296515] env[61806]: DEBUG oslo.service.loopingcall [None req-1cc2ddfb-3444-4874-84d6-f013775a320a tempest-AttachVolumeTestJSON-132925025 tempest-AttachVolumeTestJSON-132925025-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61806) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1389.296744] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: e3360f00-a42a-40f2-9e6b-813547ebfd2b] Creating VM on the ESX host {{(pid=61806) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 1389.296978] env[61806]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-37b8306d-1d6e-478a-b4a9-c78610b004f1 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1389.316284] env[61806]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1389.316284] env[61806]: value = "task-1295369" [ 1389.316284] env[61806]: _type = "Task" [ 1389.316284] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1389.323770] env[61806]: DEBUG oslo_vmware.api [-] Task: {'id': task-1295369, 'name': CreateVM_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1389.545344] env[61806]: DEBUG nova.compute.manager [req-55a00fdc-1d34-48d9-bfb0-6fc437ba80ae req-1edf1248-6608-4861-bc92-e602d851e59c service nova] [instance: e3360f00-a42a-40f2-9e6b-813547ebfd2b] Received event network-changed-3c8061f1-58c9-49f2-bdc9-e43e9936373e {{(pid=61806) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1389.545597] env[61806]: DEBUG nova.compute.manager [req-55a00fdc-1d34-48d9-bfb0-6fc437ba80ae req-1edf1248-6608-4861-bc92-e602d851e59c service nova] [instance: e3360f00-a42a-40f2-9e6b-813547ebfd2b] Refreshing instance network info cache due to event network-changed-3c8061f1-58c9-49f2-bdc9-e43e9936373e. {{(pid=61806) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 1389.545873] env[61806]: DEBUG oslo_concurrency.lockutils [req-55a00fdc-1d34-48d9-bfb0-6fc437ba80ae req-1edf1248-6608-4861-bc92-e602d851e59c service nova] Acquiring lock "refresh_cache-e3360f00-a42a-40f2-9e6b-813547ebfd2b" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1389.546069] env[61806]: DEBUG oslo_concurrency.lockutils [req-55a00fdc-1d34-48d9-bfb0-6fc437ba80ae req-1edf1248-6608-4861-bc92-e602d851e59c service nova] Acquired lock "refresh_cache-e3360f00-a42a-40f2-9e6b-813547ebfd2b" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1389.546282] env[61806]: DEBUG nova.network.neutron [req-55a00fdc-1d34-48d9-bfb0-6fc437ba80ae req-1edf1248-6608-4861-bc92-e602d851e59c service nova] [instance: e3360f00-a42a-40f2-9e6b-813547ebfd2b] Refreshing network info cache for port 3c8061f1-58c9-49f2-bdc9-e43e9936373e {{(pid=61806) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1389.829561] env[61806]: DEBUG oslo_vmware.api [-] Task: {'id': task-1295369, 'name': CreateVM_Task, 'duration_secs': 0.307209} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1389.829999] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: e3360f00-a42a-40f2-9e6b-813547ebfd2b] Created VM on the ESX host {{(pid=61806) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 1389.838189] env[61806]: DEBUG oslo_concurrency.lockutils [None req-1cc2ddfb-3444-4874-84d6-f013775a320a tempest-AttachVolumeTestJSON-132925025 tempest-AttachVolumeTestJSON-132925025-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1389.838502] env[61806]: DEBUG oslo_concurrency.lockutils [None req-1cc2ddfb-3444-4874-84d6-f013775a320a tempest-AttachVolumeTestJSON-132925025 tempest-AttachVolumeTestJSON-132925025-project-member] Acquired lock "[datastore1] devstack-image-cache_base/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1389.839041] env[61806]: DEBUG oslo_concurrency.lockutils [None req-1cc2ddfb-3444-4874-84d6-f013775a320a tempest-AttachVolumeTestJSON-132925025 tempest-AttachVolumeTestJSON-132925025-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1389.839432] env[61806]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-0b90b1a1-f54d-4f30-989e-f7c9b9d2433c {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1389.845357] env[61806]: DEBUG oslo_vmware.api [None req-1cc2ddfb-3444-4874-84d6-f013775a320a tempest-AttachVolumeTestJSON-132925025 tempest-AttachVolumeTestJSON-132925025-project-member] Waiting for the task: (returnval){ [ 1389.845357] env[61806]: value = "session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]52f65b1e-441b-07af-04fa-80579d80ea17" [ 1389.845357] env[61806]: _type = "Task" [ 1389.845357] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1389.854693] env[61806]: DEBUG oslo_vmware.api [None req-1cc2ddfb-3444-4874-84d6-f013775a320a tempest-AttachVolumeTestJSON-132925025 tempest-AttachVolumeTestJSON-132925025-project-member] Task: {'id': session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]52f65b1e-441b-07af-04fa-80579d80ea17, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1390.245693] env[61806]: DEBUG nova.network.neutron [req-55a00fdc-1d34-48d9-bfb0-6fc437ba80ae req-1edf1248-6608-4861-bc92-e602d851e59c service nova] [instance: e3360f00-a42a-40f2-9e6b-813547ebfd2b] Updated VIF entry in instance network info cache for port 3c8061f1-58c9-49f2-bdc9-e43e9936373e. {{(pid=61806) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1390.246084] env[61806]: DEBUG nova.network.neutron [req-55a00fdc-1d34-48d9-bfb0-6fc437ba80ae req-1edf1248-6608-4861-bc92-e602d851e59c service nova] [instance: e3360f00-a42a-40f2-9e6b-813547ebfd2b] Updating instance_info_cache with network_info: [{"id": "3c8061f1-58c9-49f2-bdc9-e43e9936373e", "address": "fa:16:3e:d4:61:cb", "network": {"id": "0fd518d3-c254-4dba-a147-378c3c029c51", "bridge": "br-int", "label": "tempest-AttachVolumeTestJSON-538202287-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "833defcd470d46ceacd7886d1f4fb0cf", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "91b0f7e5-0d1a-46e2-bf73-09656211dea2", "external-id": "nsx-vlan-transportzone-488", "segmentation_id": 488, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3c8061f1-58", "ovs_interfaceid": "3c8061f1-58c9-49f2-bdc9-e43e9936373e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61806) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1390.357553] env[61806]: DEBUG oslo_vmware.api [None req-1cc2ddfb-3444-4874-84d6-f013775a320a tempest-AttachVolumeTestJSON-132925025 tempest-AttachVolumeTestJSON-132925025-project-member] Task: {'id': session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]52f65b1e-441b-07af-04fa-80579d80ea17, 'name': SearchDatastore_Task, 'duration_secs': 0.010335} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1390.357790] env[61806]: DEBUG oslo_concurrency.lockutils [None req-1cc2ddfb-3444-4874-84d6-f013775a320a tempest-AttachVolumeTestJSON-132925025 tempest-AttachVolumeTestJSON-132925025-project-member] Releasing lock "[datastore1] devstack-image-cache_base/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1390.358060] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-1cc2ddfb-3444-4874-84d6-f013775a320a tempest-AttachVolumeTestJSON-132925025 tempest-AttachVolumeTestJSON-132925025-project-member] [instance: e3360f00-a42a-40f2-9e6b-813547ebfd2b] Processing image ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa {{(pid=61806) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1390.358363] env[61806]: DEBUG oslo_concurrency.lockutils [None req-1cc2ddfb-3444-4874-84d6-f013775a320a tempest-AttachVolumeTestJSON-132925025 tempest-AttachVolumeTestJSON-132925025-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa.vmdk" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1390.358527] env[61806]: DEBUG oslo_concurrency.lockutils [None req-1cc2ddfb-3444-4874-84d6-f013775a320a tempest-AttachVolumeTestJSON-132925025 tempest-AttachVolumeTestJSON-132925025-project-member] Acquired lock "[datastore1] devstack-image-cache_base/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa.vmdk" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1390.358716] env[61806]: DEBUG nova.virt.vmwareapi.ds_util [None req-1cc2ddfb-3444-4874-84d6-f013775a320a tempest-AttachVolumeTestJSON-132925025 tempest-AttachVolumeTestJSON-132925025-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61806) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1390.358984] env[61806]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-e99ca845-c54b-49b7-ad13-3f6ef5fcf999 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1390.367172] env[61806]: DEBUG nova.virt.vmwareapi.ds_util [None req-1cc2ddfb-3444-4874-84d6-f013775a320a tempest-AttachVolumeTestJSON-132925025 tempest-AttachVolumeTestJSON-132925025-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61806) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1390.367326] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-1cc2ddfb-3444-4874-84d6-f013775a320a tempest-AttachVolumeTestJSON-132925025 tempest-AttachVolumeTestJSON-132925025-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61806) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 1390.368044] env[61806]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-88adfeca-2034-413b-abbd-ebba97c573da {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1390.373352] env[61806]: DEBUG oslo_vmware.api [None req-1cc2ddfb-3444-4874-84d6-f013775a320a tempest-AttachVolumeTestJSON-132925025 tempest-AttachVolumeTestJSON-132925025-project-member] Waiting for the task: (returnval){ [ 1390.373352] env[61806]: value = "session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]52d109c2-5139-e15a-7392-64926cb659fd" [ 1390.373352] env[61806]: _type = "Task" [ 1390.373352] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1390.381106] env[61806]: DEBUG oslo_vmware.api [None req-1cc2ddfb-3444-4874-84d6-f013775a320a tempest-AttachVolumeTestJSON-132925025 tempest-AttachVolumeTestJSON-132925025-project-member] Task: {'id': session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]52d109c2-5139-e15a-7392-64926cb659fd, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1390.748838] env[61806]: DEBUG oslo_concurrency.lockutils [req-55a00fdc-1d34-48d9-bfb0-6fc437ba80ae req-1edf1248-6608-4861-bc92-e602d851e59c service nova] Releasing lock "refresh_cache-e3360f00-a42a-40f2-9e6b-813547ebfd2b" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1390.884471] env[61806]: DEBUG oslo_vmware.api [None req-1cc2ddfb-3444-4874-84d6-f013775a320a tempest-AttachVolumeTestJSON-132925025 tempest-AttachVolumeTestJSON-132925025-project-member] Task: {'id': session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]52d109c2-5139-e15a-7392-64926cb659fd, 'name': SearchDatastore_Task, 'duration_secs': 0.008586} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1390.885279] env[61806]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d8033be2-1cb8-4039-86f6-7233ca55eb74 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1390.890735] env[61806]: DEBUG oslo_vmware.api [None req-1cc2ddfb-3444-4874-84d6-f013775a320a tempest-AttachVolumeTestJSON-132925025 tempest-AttachVolumeTestJSON-132925025-project-member] Waiting for the task: (returnval){ [ 1390.890735] env[61806]: value = "session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]521f2283-ba9b-23c4-f0d5-c790838754d9" [ 1390.890735] env[61806]: _type = "Task" [ 1390.890735] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1390.898529] env[61806]: DEBUG oslo_vmware.api [None req-1cc2ddfb-3444-4874-84d6-f013775a320a tempest-AttachVolumeTestJSON-132925025 tempest-AttachVolumeTestJSON-132925025-project-member] Task: {'id': session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]521f2283-ba9b-23c4-f0d5-c790838754d9, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1391.401082] env[61806]: DEBUG oslo_vmware.api [None req-1cc2ddfb-3444-4874-84d6-f013775a320a tempest-AttachVolumeTestJSON-132925025 tempest-AttachVolumeTestJSON-132925025-project-member] Task: {'id': session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]521f2283-ba9b-23c4-f0d5-c790838754d9, 'name': SearchDatastore_Task, 'duration_secs': 0.009066} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1391.401322] env[61806]: DEBUG oslo_concurrency.lockutils [None req-1cc2ddfb-3444-4874-84d6-f013775a320a tempest-AttachVolumeTestJSON-132925025 tempest-AttachVolumeTestJSON-132925025-project-member] Releasing lock "[datastore1] devstack-image-cache_base/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa.vmdk" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1391.401592] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [None req-1cc2ddfb-3444-4874-84d6-f013775a320a tempest-AttachVolumeTestJSON-132925025 tempest-AttachVolumeTestJSON-132925025-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa.vmdk to [datastore1] e3360f00-a42a-40f2-9e6b-813547ebfd2b/e3360f00-a42a-40f2-9e6b-813547ebfd2b.vmdk {{(pid=61806) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 1391.401857] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-dcf84609-fabf-4cee-b306-98a7279433af {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1391.408843] env[61806]: DEBUG oslo_vmware.api [None req-1cc2ddfb-3444-4874-84d6-f013775a320a tempest-AttachVolumeTestJSON-132925025 tempest-AttachVolumeTestJSON-132925025-project-member] Waiting for the task: (returnval){ [ 1391.408843] env[61806]: value = "task-1295370" [ 1391.408843] env[61806]: _type = "Task" [ 1391.408843] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1391.415930] env[61806]: DEBUG oslo_vmware.api [None req-1cc2ddfb-3444-4874-84d6-f013775a320a tempest-AttachVolumeTestJSON-132925025 tempest-AttachVolumeTestJSON-132925025-project-member] Task: {'id': task-1295370, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1391.919043] env[61806]: DEBUG oslo_vmware.api [None req-1cc2ddfb-3444-4874-84d6-f013775a320a tempest-AttachVolumeTestJSON-132925025 tempest-AttachVolumeTestJSON-132925025-project-member] Task: {'id': task-1295370, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.456134} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1391.919431] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [None req-1cc2ddfb-3444-4874-84d6-f013775a320a tempest-AttachVolumeTestJSON-132925025 tempest-AttachVolumeTestJSON-132925025-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa.vmdk to [datastore1] e3360f00-a42a-40f2-9e6b-813547ebfd2b/e3360f00-a42a-40f2-9e6b-813547ebfd2b.vmdk {{(pid=61806) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 1391.919517] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-1cc2ddfb-3444-4874-84d6-f013775a320a tempest-AttachVolumeTestJSON-132925025 tempest-AttachVolumeTestJSON-132925025-project-member] [instance: e3360f00-a42a-40f2-9e6b-813547ebfd2b] Extending root virtual disk to 1048576 {{(pid=61806) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1391.919746] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-8048e3ff-0a03-4ce5-be14-47ba49382f6a {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1391.925666] env[61806]: DEBUG oslo_vmware.api [None req-1cc2ddfb-3444-4874-84d6-f013775a320a tempest-AttachVolumeTestJSON-132925025 tempest-AttachVolumeTestJSON-132925025-project-member] Waiting for the task: (returnval){ [ 1391.925666] env[61806]: value = "task-1295371" [ 1391.925666] env[61806]: _type = "Task" [ 1391.925666] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1391.932299] env[61806]: DEBUG oslo_vmware.api [None req-1cc2ddfb-3444-4874-84d6-f013775a320a tempest-AttachVolumeTestJSON-132925025 tempest-AttachVolumeTestJSON-132925025-project-member] Task: {'id': task-1295371, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1392.435272] env[61806]: DEBUG oslo_vmware.api [None req-1cc2ddfb-3444-4874-84d6-f013775a320a tempest-AttachVolumeTestJSON-132925025 tempest-AttachVolumeTestJSON-132925025-project-member] Task: {'id': task-1295371, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.064364} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1392.435510] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-1cc2ddfb-3444-4874-84d6-f013775a320a tempest-AttachVolumeTestJSON-132925025 tempest-AttachVolumeTestJSON-132925025-project-member] [instance: e3360f00-a42a-40f2-9e6b-813547ebfd2b] Extended root virtual disk {{(pid=61806) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1392.436283] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-224de49e-f8fa-410b-a9be-f31111baea39 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1392.457263] env[61806]: DEBUG nova.virt.vmwareapi.volumeops [None req-1cc2ddfb-3444-4874-84d6-f013775a320a tempest-AttachVolumeTestJSON-132925025 tempest-AttachVolumeTestJSON-132925025-project-member] [instance: e3360f00-a42a-40f2-9e6b-813547ebfd2b] Reconfiguring VM instance instance-0000006f to attach disk [datastore1] e3360f00-a42a-40f2-9e6b-813547ebfd2b/e3360f00-a42a-40f2-9e6b-813547ebfd2b.vmdk or device None with type sparse {{(pid=61806) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1392.457495] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-a3106922-1b12-4554-b84c-c3077511ecef {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1392.475995] env[61806]: DEBUG oslo_vmware.api [None req-1cc2ddfb-3444-4874-84d6-f013775a320a tempest-AttachVolumeTestJSON-132925025 tempest-AttachVolumeTestJSON-132925025-project-member] Waiting for the task: (returnval){ [ 1392.475995] env[61806]: value = "task-1295372" [ 1392.475995] env[61806]: _type = "Task" [ 1392.475995] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1392.488340] env[61806]: DEBUG oslo_vmware.api [None req-1cc2ddfb-3444-4874-84d6-f013775a320a tempest-AttachVolumeTestJSON-132925025 tempest-AttachVolumeTestJSON-132925025-project-member] Task: {'id': task-1295372, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1392.986231] env[61806]: DEBUG oslo_vmware.api [None req-1cc2ddfb-3444-4874-84d6-f013775a320a tempest-AttachVolumeTestJSON-132925025 tempest-AttachVolumeTestJSON-132925025-project-member] Task: {'id': task-1295372, 'name': ReconfigVM_Task, 'duration_secs': 0.27333} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1392.986637] env[61806]: DEBUG nova.virt.vmwareapi.volumeops [None req-1cc2ddfb-3444-4874-84d6-f013775a320a tempest-AttachVolumeTestJSON-132925025 tempest-AttachVolumeTestJSON-132925025-project-member] [instance: e3360f00-a42a-40f2-9e6b-813547ebfd2b] Reconfigured VM instance instance-0000006f to attach disk [datastore1] e3360f00-a42a-40f2-9e6b-813547ebfd2b/e3360f00-a42a-40f2-9e6b-813547ebfd2b.vmdk or device None with type sparse {{(pid=61806) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1392.987241] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-f39fa82d-bf98-4e4f-bd55-1e193eb359f4 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1392.994239] env[61806]: DEBUG oslo_vmware.api [None req-1cc2ddfb-3444-4874-84d6-f013775a320a tempest-AttachVolumeTestJSON-132925025 tempest-AttachVolumeTestJSON-132925025-project-member] Waiting for the task: (returnval){ [ 1392.994239] env[61806]: value = "task-1295373" [ 1392.994239] env[61806]: _type = "Task" [ 1392.994239] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1393.002723] env[61806]: DEBUG oslo_vmware.api [None req-1cc2ddfb-3444-4874-84d6-f013775a320a tempest-AttachVolumeTestJSON-132925025 tempest-AttachVolumeTestJSON-132925025-project-member] Task: {'id': task-1295373, 'name': Rename_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1393.504026] env[61806]: DEBUG oslo_vmware.api [None req-1cc2ddfb-3444-4874-84d6-f013775a320a tempest-AttachVolumeTestJSON-132925025 tempest-AttachVolumeTestJSON-132925025-project-member] Task: {'id': task-1295373, 'name': Rename_Task, 'duration_secs': 0.130088} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1393.504338] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [None req-1cc2ddfb-3444-4874-84d6-f013775a320a tempest-AttachVolumeTestJSON-132925025 tempest-AttachVolumeTestJSON-132925025-project-member] [instance: e3360f00-a42a-40f2-9e6b-813547ebfd2b] Powering on the VM {{(pid=61806) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1393.504591] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-b175bb27-ef3e-4b0a-a74c-1a706e110a39 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1393.509912] env[61806]: DEBUG oslo_vmware.api [None req-1cc2ddfb-3444-4874-84d6-f013775a320a tempest-AttachVolumeTestJSON-132925025 tempest-AttachVolumeTestJSON-132925025-project-member] Waiting for the task: (returnval){ [ 1393.509912] env[61806]: value = "task-1295374" [ 1393.509912] env[61806]: _type = "Task" [ 1393.509912] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1393.517898] env[61806]: DEBUG oslo_vmware.api [None req-1cc2ddfb-3444-4874-84d6-f013775a320a tempest-AttachVolumeTestJSON-132925025 tempest-AttachVolumeTestJSON-132925025-project-member] Task: {'id': task-1295374, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1394.019786] env[61806]: DEBUG oslo_vmware.api [None req-1cc2ddfb-3444-4874-84d6-f013775a320a tempest-AttachVolumeTestJSON-132925025 tempest-AttachVolumeTestJSON-132925025-project-member] Task: {'id': task-1295374, 'name': PowerOnVM_Task, 'duration_secs': 0.473046} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1394.020122] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [None req-1cc2ddfb-3444-4874-84d6-f013775a320a tempest-AttachVolumeTestJSON-132925025 tempest-AttachVolumeTestJSON-132925025-project-member] [instance: e3360f00-a42a-40f2-9e6b-813547ebfd2b] Powered on the VM {{(pid=61806) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1394.020305] env[61806]: INFO nova.compute.manager [None req-1cc2ddfb-3444-4874-84d6-f013775a320a tempest-AttachVolumeTestJSON-132925025 tempest-AttachVolumeTestJSON-132925025-project-member] [instance: e3360f00-a42a-40f2-9e6b-813547ebfd2b] Took 6.65 seconds to spawn the instance on the hypervisor. [ 1394.020544] env[61806]: DEBUG nova.compute.manager [None req-1cc2ddfb-3444-4874-84d6-f013775a320a tempest-AttachVolumeTestJSON-132925025 tempest-AttachVolumeTestJSON-132925025-project-member] [instance: e3360f00-a42a-40f2-9e6b-813547ebfd2b] Checking state {{(pid=61806) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1394.021305] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5cdc0714-1ca8-4aeb-867f-482f586e2b60 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1394.539124] env[61806]: INFO nova.compute.manager [None req-1cc2ddfb-3444-4874-84d6-f013775a320a tempest-AttachVolumeTestJSON-132925025 tempest-AttachVolumeTestJSON-132925025-project-member] [instance: e3360f00-a42a-40f2-9e6b-813547ebfd2b] Took 13.69 seconds to build instance. [ 1394.908291] env[61806]: DEBUG nova.compute.manager [req-a663ea4f-f42b-4469-894f-9ea97ea449c3 req-e6af4f61-feb7-4101-b6bb-7ea6edca7522 service nova] [instance: e3360f00-a42a-40f2-9e6b-813547ebfd2b] Received event network-changed-3c8061f1-58c9-49f2-bdc9-e43e9936373e {{(pid=61806) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1394.908474] env[61806]: DEBUG nova.compute.manager [req-a663ea4f-f42b-4469-894f-9ea97ea449c3 req-e6af4f61-feb7-4101-b6bb-7ea6edca7522 service nova] [instance: e3360f00-a42a-40f2-9e6b-813547ebfd2b] Refreshing instance network info cache due to event network-changed-3c8061f1-58c9-49f2-bdc9-e43e9936373e. {{(pid=61806) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 1394.908701] env[61806]: DEBUG oslo_concurrency.lockutils [req-a663ea4f-f42b-4469-894f-9ea97ea449c3 req-e6af4f61-feb7-4101-b6bb-7ea6edca7522 service nova] Acquiring lock "refresh_cache-e3360f00-a42a-40f2-9e6b-813547ebfd2b" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1394.908826] env[61806]: DEBUG oslo_concurrency.lockutils [req-a663ea4f-f42b-4469-894f-9ea97ea449c3 req-e6af4f61-feb7-4101-b6bb-7ea6edca7522 service nova] Acquired lock "refresh_cache-e3360f00-a42a-40f2-9e6b-813547ebfd2b" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1394.908996] env[61806]: DEBUG nova.network.neutron [req-a663ea4f-f42b-4469-894f-9ea97ea449c3 req-e6af4f61-feb7-4101-b6bb-7ea6edca7522 service nova] [instance: e3360f00-a42a-40f2-9e6b-813547ebfd2b] Refreshing network info cache for port 3c8061f1-58c9-49f2-bdc9-e43e9936373e {{(pid=61806) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1395.041447] env[61806]: DEBUG oslo_concurrency.lockutils [None req-1cc2ddfb-3444-4874-84d6-f013775a320a tempest-AttachVolumeTestJSON-132925025 tempest-AttachVolumeTestJSON-132925025-project-member] Lock "e3360f00-a42a-40f2-9e6b-813547ebfd2b" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 15.199s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1395.607894] env[61806]: DEBUG nova.network.neutron [req-a663ea4f-f42b-4469-894f-9ea97ea449c3 req-e6af4f61-feb7-4101-b6bb-7ea6edca7522 service nova] [instance: e3360f00-a42a-40f2-9e6b-813547ebfd2b] Updated VIF entry in instance network info cache for port 3c8061f1-58c9-49f2-bdc9-e43e9936373e. {{(pid=61806) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1395.608298] env[61806]: DEBUG nova.network.neutron [req-a663ea4f-f42b-4469-894f-9ea97ea449c3 req-e6af4f61-feb7-4101-b6bb-7ea6edca7522 service nova] [instance: e3360f00-a42a-40f2-9e6b-813547ebfd2b] Updating instance_info_cache with network_info: [{"id": "3c8061f1-58c9-49f2-bdc9-e43e9936373e", "address": "fa:16:3e:d4:61:cb", "network": {"id": "0fd518d3-c254-4dba-a147-378c3c029c51", "bridge": "br-int", "label": "tempest-AttachVolumeTestJSON-538202287-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.229", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "833defcd470d46ceacd7886d1f4fb0cf", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "91b0f7e5-0d1a-46e2-bf73-09656211dea2", "external-id": "nsx-vlan-transportzone-488", "segmentation_id": 488, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3c8061f1-58", "ovs_interfaceid": "3c8061f1-58c9-49f2-bdc9-e43e9936373e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61806) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1396.110875] env[61806]: DEBUG oslo_concurrency.lockutils [req-a663ea4f-f42b-4469-894f-9ea97ea449c3 req-e6af4f61-feb7-4101-b6bb-7ea6edca7522 service nova] Releasing lock "refresh_cache-e3360f00-a42a-40f2-9e6b-813547ebfd2b" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1432.521274] env[61806]: DEBUG oslo_service.periodic_task [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=61806) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1434.266548] env[61806]: DEBUG oslo_concurrency.lockutils [None req-ebdb1123-e6cf-4616-83ea-f69bfb6490e8 tempest-AttachVolumeTestJSON-132925025 tempest-AttachVolumeTestJSON-132925025-project-member] Acquiring lock "e3360f00-a42a-40f2-9e6b-813547ebfd2b" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1434.266836] env[61806]: DEBUG oslo_concurrency.lockutils [None req-ebdb1123-e6cf-4616-83ea-f69bfb6490e8 tempest-AttachVolumeTestJSON-132925025 tempest-AttachVolumeTestJSON-132925025-project-member] Lock "e3360f00-a42a-40f2-9e6b-813547ebfd2b" acquired by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: waited 0.000s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1434.771184] env[61806]: DEBUG nova.compute.utils [None req-ebdb1123-e6cf-4616-83ea-f69bfb6490e8 tempest-AttachVolumeTestJSON-132925025 tempest-AttachVolumeTestJSON-132925025-project-member] Using /dev/sd instead of None {{(pid=61806) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1435.022981] env[61806]: DEBUG oslo_service.periodic_task [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=61806) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1435.274748] env[61806]: DEBUG oslo_concurrency.lockutils [None req-ebdb1123-e6cf-4616-83ea-f69bfb6490e8 tempest-AttachVolumeTestJSON-132925025 tempest-AttachVolumeTestJSON-132925025-project-member] Lock "e3360f00-a42a-40f2-9e6b-813547ebfd2b" "released" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: held 1.008s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1436.023063] env[61806]: DEBUG oslo_service.periodic_task [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=61806) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1436.023248] env[61806]: DEBUG nova.compute.manager [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Starting heal instance info cache {{(pid=61806) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9926}} [ 1436.023375] env[61806]: DEBUG nova.compute.manager [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Rebuilding the list of instances to heal {{(pid=61806) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9930}} [ 1436.331589] env[61806]: DEBUG oslo_concurrency.lockutils [None req-ebdb1123-e6cf-4616-83ea-f69bfb6490e8 tempest-AttachVolumeTestJSON-132925025 tempest-AttachVolumeTestJSON-132925025-project-member] Acquiring lock "e3360f00-a42a-40f2-9e6b-813547ebfd2b" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1436.331894] env[61806]: DEBUG oslo_concurrency.lockutils [None req-ebdb1123-e6cf-4616-83ea-f69bfb6490e8 tempest-AttachVolumeTestJSON-132925025 tempest-AttachVolumeTestJSON-132925025-project-member] Lock "e3360f00-a42a-40f2-9e6b-813547ebfd2b" acquired by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: waited 0.000s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1436.332078] env[61806]: INFO nova.compute.manager [None req-ebdb1123-e6cf-4616-83ea-f69bfb6490e8 tempest-AttachVolumeTestJSON-132925025 tempest-AttachVolumeTestJSON-132925025-project-member] [instance: e3360f00-a42a-40f2-9e6b-813547ebfd2b] Attaching volume 434af0b9-a453-4b85-9692-5eac3a926286 to /dev/sdb [ 1436.385379] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e341c535-0b93-454d-8700-f9423041c019 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1436.392615] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2305f7df-9336-485e-8754-c6ada834c63f {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1436.408885] env[61806]: DEBUG nova.virt.block_device [None req-ebdb1123-e6cf-4616-83ea-f69bfb6490e8 tempest-AttachVolumeTestJSON-132925025 tempest-AttachVolumeTestJSON-132925025-project-member] [instance: e3360f00-a42a-40f2-9e6b-813547ebfd2b] Updating existing volume attachment record: d6c531c2-a0b4-4b0b-91f9-75ccea95f4f7 {{(pid=61806) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 1436.591616] env[61806]: DEBUG oslo_concurrency.lockutils [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Acquiring lock "refresh_cache-e3360f00-a42a-40f2-9e6b-813547ebfd2b" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1436.591779] env[61806]: DEBUG oslo_concurrency.lockutils [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Acquired lock "refresh_cache-e3360f00-a42a-40f2-9e6b-813547ebfd2b" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1436.591935] env[61806]: DEBUG nova.network.neutron [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] [instance: e3360f00-a42a-40f2-9e6b-813547ebfd2b] Forcefully refreshing network info cache for instance {{(pid=61806) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2004}} [ 1436.592121] env[61806]: DEBUG nova.objects.instance [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Lazy-loading 'info_cache' on Instance uuid e3360f00-a42a-40f2-9e6b-813547ebfd2b {{(pid=61806) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1438.305824] env[61806]: DEBUG nova.network.neutron [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] [instance: e3360f00-a42a-40f2-9e6b-813547ebfd2b] Updating instance_info_cache with network_info: [{"id": "3c8061f1-58c9-49f2-bdc9-e43e9936373e", "address": "fa:16:3e:d4:61:cb", "network": {"id": "0fd518d3-c254-4dba-a147-378c3c029c51", "bridge": "br-int", "label": "tempest-AttachVolumeTestJSON-538202287-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.229", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "833defcd470d46ceacd7886d1f4fb0cf", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "91b0f7e5-0d1a-46e2-bf73-09656211dea2", "external-id": "nsx-vlan-transportzone-488", "segmentation_id": 488, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3c8061f1-58", "ovs_interfaceid": "3c8061f1-58c9-49f2-bdc9-e43e9936373e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61806) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1438.808317] env[61806]: DEBUG oslo_concurrency.lockutils [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Releasing lock "refresh_cache-e3360f00-a42a-40f2-9e6b-813547ebfd2b" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1438.808547] env[61806]: DEBUG nova.compute.manager [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] [instance: e3360f00-a42a-40f2-9e6b-813547ebfd2b] Updated the network info_cache for instance {{(pid=61806) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9997}} [ 1438.808792] env[61806]: DEBUG oslo_service.periodic_task [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Running periodic task ComputeManager.update_available_resource {{(pid=61806) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1439.312433] env[61806]: DEBUG oslo_concurrency.lockutils [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1439.312825] env[61806]: DEBUG oslo_concurrency.lockutils [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 0.000s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1439.312825] env[61806]: DEBUG oslo_concurrency.lockutils [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1439.312969] env[61806]: DEBUG nova.compute.resource_tracker [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=61806) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 1439.314269] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-59b80a8b-4b81-41de-83df-ac3e41d1d9ce {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1439.322111] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-16d03a33-f028-424e-9f7a-73c505319636 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1439.335436] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a93ea99b-8cfc-48e8-adaf-fb62defdb27e {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1439.341361] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e3503d0d-5a4d-499f-b258-3c25af0469f3 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1439.369999] env[61806]: DEBUG nova.compute.resource_tracker [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=181278MB free_disk=138GB free_vcpus=48 pci_devices=None {{(pid=61806) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 1439.370184] env[61806]: DEBUG oslo_concurrency.lockutils [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1439.370371] env[61806]: DEBUG oslo_concurrency.lockutils [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 0.000s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1440.396307] env[61806]: DEBUG nova.compute.resource_tracker [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Instance e3360f00-a42a-40f2-9e6b-813547ebfd2b actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61806) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1440.396675] env[61806]: DEBUG nova.compute.resource_tracker [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Total usable vcpus: 48, total allocated vcpus: 1 {{(pid=61806) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 1440.396675] env[61806]: DEBUG nova.compute.resource_tracker [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=704MB phys_disk=200GB used_disk=1GB total_vcpus=48 used_vcpus=1 pci_stats=[] {{(pid=61806) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 1440.432906] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-168fc443-5a48-4e50-b076-d3d7f5553257 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1440.440121] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b6c9d49a-610b-47cf-93fd-d223e3da3ddb {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1440.468112] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-80530309-8912-4efa-921d-e76b381a5431 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1440.474677] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-84266b56-8acd-4761-b5fc-e85f4da6d381 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1440.488090] env[61806]: DEBUG nova.compute.provider_tree [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Updating inventory in ProviderTree for provider a2858be1-fd22-4e08-979e-87ad25293407 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 138, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61806) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1440.950490] env[61806]: DEBUG nova.virt.vmwareapi.volumeops [None req-ebdb1123-e6cf-4616-83ea-f69bfb6490e8 tempest-AttachVolumeTestJSON-132925025 tempest-AttachVolumeTestJSON-132925025-project-member] [instance: e3360f00-a42a-40f2-9e6b-813547ebfd2b] Volume attach. Driver type: vmdk {{(pid=61806) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 1440.950765] env[61806]: DEBUG nova.virt.vmwareapi.volumeops [None req-ebdb1123-e6cf-4616-83ea-f69bfb6490e8 tempest-AttachVolumeTestJSON-132925025 tempest-AttachVolumeTestJSON-132925025-project-member] [instance: e3360f00-a42a-40f2-9e6b-813547ebfd2b] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-277786', 'volume_id': '434af0b9-a453-4b85-9692-5eac3a926286', 'name': 'volume-434af0b9-a453-4b85-9692-5eac3a926286', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'e3360f00-a42a-40f2-9e6b-813547ebfd2b', 'attached_at': '', 'detached_at': '', 'volume_id': '434af0b9-a453-4b85-9692-5eac3a926286', 'serial': '434af0b9-a453-4b85-9692-5eac3a926286'} {{(pid=61806) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 1440.951651] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c7a73e9a-c8ac-4e23-817e-7c9dc2248354 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1440.967627] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-664caf24-7878-43a9-a5a5-a3d28099c9dd {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1440.991277] env[61806]: DEBUG nova.virt.vmwareapi.volumeops [None req-ebdb1123-e6cf-4616-83ea-f69bfb6490e8 tempest-AttachVolumeTestJSON-132925025 tempest-AttachVolumeTestJSON-132925025-project-member] [instance: e3360f00-a42a-40f2-9e6b-813547ebfd2b] Reconfiguring VM instance instance-0000006f to attach disk [datastore1] volume-434af0b9-a453-4b85-9692-5eac3a926286/volume-434af0b9-a453-4b85-9692-5eac3a926286.vmdk or device None with type thin {{(pid=61806) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1440.993801] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-18d5a49f-4262-4503-b9ca-4510cf575be1 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1441.012202] env[61806]: DEBUG oslo_vmware.api [None req-ebdb1123-e6cf-4616-83ea-f69bfb6490e8 tempest-AttachVolumeTestJSON-132925025 tempest-AttachVolumeTestJSON-132925025-project-member] Waiting for the task: (returnval){ [ 1441.012202] env[61806]: value = "task-1295379" [ 1441.012202] env[61806]: _type = "Task" [ 1441.012202] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1441.019975] env[61806]: DEBUG oslo_vmware.api [None req-ebdb1123-e6cf-4616-83ea-f69bfb6490e8 tempest-AttachVolumeTestJSON-132925025 tempest-AttachVolumeTestJSON-132925025-project-member] Task: {'id': task-1295379, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1441.020745] env[61806]: ERROR nova.scheduler.client.report [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] [req-ba385e15-280a-4953-80b4-eafce606fa29] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 138, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID a2858be1-fd22-4e08-979e-87ad25293407. Got 409: {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-ba385e15-280a-4953-80b4-eafce606fa29"}]} [ 1441.037291] env[61806]: DEBUG nova.scheduler.client.report [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Refreshing inventories for resource provider a2858be1-fd22-4e08-979e-87ad25293407 {{(pid=61806) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:818}} [ 1441.051480] env[61806]: DEBUG nova.scheduler.client.report [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Updating ProviderTree inventory for provider a2858be1-fd22-4e08-979e-87ad25293407 from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 139, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61806) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:782}} [ 1441.051661] env[61806]: DEBUG nova.compute.provider_tree [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Updating inventory in ProviderTree for provider a2858be1-fd22-4e08-979e-87ad25293407 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 139, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61806) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1441.061973] env[61806]: DEBUG nova.scheduler.client.report [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Refreshing aggregate associations for resource provider a2858be1-fd22-4e08-979e-87ad25293407, aggregates: None {{(pid=61806) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:827}} [ 1441.079498] env[61806]: DEBUG nova.scheduler.client.report [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Refreshing trait associations for resource provider a2858be1-fd22-4e08-979e-87ad25293407, traits: COMPUTE_IMAGE_TYPE_ISO,COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_SAME_HOST_COLD_MIGRATE,COMPUTE_NET_ATTACH_INTERFACE,COMPUTE_NODE {{(pid=61806) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:839}} [ 1441.103934] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8f256dbf-4ac8-4463-b7f9-e2e63738cef7 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1441.111428] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-62c0fe30-2e21-4a98-ad70-8d2a6292005a {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1441.141534] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f5e96d88-f133-4836-bb53-c01abbb3887b {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1441.148646] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-981aa6e6-47b3-4801-a727-0aa35561c4f7 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1441.161560] env[61806]: DEBUG nova.compute.provider_tree [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Updating inventory in ProviderTree for provider a2858be1-fd22-4e08-979e-87ad25293407 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 138, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61806) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1441.522303] env[61806]: DEBUG oslo_vmware.api [None req-ebdb1123-e6cf-4616-83ea-f69bfb6490e8 tempest-AttachVolumeTestJSON-132925025 tempest-AttachVolumeTestJSON-132925025-project-member] Task: {'id': task-1295379, 'name': ReconfigVM_Task, 'duration_secs': 0.336506} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1441.522684] env[61806]: DEBUG nova.virt.vmwareapi.volumeops [None req-ebdb1123-e6cf-4616-83ea-f69bfb6490e8 tempest-AttachVolumeTestJSON-132925025 tempest-AttachVolumeTestJSON-132925025-project-member] [instance: e3360f00-a42a-40f2-9e6b-813547ebfd2b] Reconfigured VM instance instance-0000006f to attach disk [datastore1] volume-434af0b9-a453-4b85-9692-5eac3a926286/volume-434af0b9-a453-4b85-9692-5eac3a926286.vmdk or device None with type thin {{(pid=61806) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1441.527230] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-c120b785-5e3c-4372-8c3a-c0c51d31cbc5 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1441.541530] env[61806]: DEBUG oslo_vmware.api [None req-ebdb1123-e6cf-4616-83ea-f69bfb6490e8 tempest-AttachVolumeTestJSON-132925025 tempest-AttachVolumeTestJSON-132925025-project-member] Waiting for the task: (returnval){ [ 1441.541530] env[61806]: value = "task-1295380" [ 1441.541530] env[61806]: _type = "Task" [ 1441.541530] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1441.548570] env[61806]: DEBUG oslo_vmware.api [None req-ebdb1123-e6cf-4616-83ea-f69bfb6490e8 tempest-AttachVolumeTestJSON-132925025 tempest-AttachVolumeTestJSON-132925025-project-member] Task: {'id': task-1295380, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1441.693471] env[61806]: DEBUG nova.scheduler.client.report [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Updated inventory for provider a2858be1-fd22-4e08-979e-87ad25293407 with generation 159 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 138, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61806) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:971}} [ 1441.693710] env[61806]: DEBUG nova.compute.provider_tree [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Updating resource provider a2858be1-fd22-4e08-979e-87ad25293407 generation from 159 to 160 during operation: update_inventory {{(pid=61806) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 1441.693863] env[61806]: DEBUG nova.compute.provider_tree [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Updating inventory in ProviderTree for provider a2858be1-fd22-4e08-979e-87ad25293407 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 138, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61806) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1442.051342] env[61806]: DEBUG oslo_vmware.api [None req-ebdb1123-e6cf-4616-83ea-f69bfb6490e8 tempest-AttachVolumeTestJSON-132925025 tempest-AttachVolumeTestJSON-132925025-project-member] Task: {'id': task-1295380, 'name': ReconfigVM_Task, 'duration_secs': 0.142446} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1442.051675] env[61806]: DEBUG nova.virt.vmwareapi.volumeops [None req-ebdb1123-e6cf-4616-83ea-f69bfb6490e8 tempest-AttachVolumeTestJSON-132925025 tempest-AttachVolumeTestJSON-132925025-project-member] [instance: e3360f00-a42a-40f2-9e6b-813547ebfd2b] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-277786', 'volume_id': '434af0b9-a453-4b85-9692-5eac3a926286', 'name': 'volume-434af0b9-a453-4b85-9692-5eac3a926286', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'e3360f00-a42a-40f2-9e6b-813547ebfd2b', 'attached_at': '', 'detached_at': '', 'volume_id': '434af0b9-a453-4b85-9692-5eac3a926286', 'serial': '434af0b9-a453-4b85-9692-5eac3a926286'} {{(pid=61806) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 1442.198243] env[61806]: DEBUG nova.compute.resource_tracker [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=61806) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 1442.198432] env[61806]: DEBUG oslo_concurrency.lockutils [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 2.828s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1443.087985] env[61806]: DEBUG nova.objects.instance [None req-ebdb1123-e6cf-4616-83ea-f69bfb6490e8 tempest-AttachVolumeTestJSON-132925025 tempest-AttachVolumeTestJSON-132925025-project-member] Lazy-loading 'flavor' on Instance uuid e3360f00-a42a-40f2-9e6b-813547ebfd2b {{(pid=61806) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1443.412595] env[61806]: DEBUG oslo_service.periodic_task [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Running periodic task ComputeManager._sync_scheduler_instance_info {{(pid=61806) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1443.594817] env[61806]: DEBUG oslo_concurrency.lockutils [None req-ebdb1123-e6cf-4616-83ea-f69bfb6490e8 tempest-AttachVolumeTestJSON-132925025 tempest-AttachVolumeTestJSON-132925025-project-member] Lock "e3360f00-a42a-40f2-9e6b-813547ebfd2b" "released" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: held 7.263s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1443.652650] env[61806]: DEBUG oslo_concurrency.lockutils [None req-0faf01e0-9a2e-4e7d-bbbd-71f87bf60864 tempest-AttachVolumeTestJSON-132925025 tempest-AttachVolumeTestJSON-132925025-project-member] Acquiring lock "e3360f00-a42a-40f2-9e6b-813547ebfd2b" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1443.653306] env[61806]: DEBUG oslo_concurrency.lockutils [None req-0faf01e0-9a2e-4e7d-bbbd-71f87bf60864 tempest-AttachVolumeTestJSON-132925025 tempest-AttachVolumeTestJSON-132925025-project-member] Lock "e3360f00-a42a-40f2-9e6b-813547ebfd2b" acquired by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: waited 0.001s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1443.653509] env[61806]: DEBUG nova.compute.manager [None req-0faf01e0-9a2e-4e7d-bbbd-71f87bf60864 tempest-AttachVolumeTestJSON-132925025 tempest-AttachVolumeTestJSON-132925025-project-member] [instance: e3360f00-a42a-40f2-9e6b-813547ebfd2b] Checking state {{(pid=61806) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1443.654396] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-60986d32-54c4-4886-9f5a-f9847199f6fd {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1443.660993] env[61806]: DEBUG nova.compute.manager [None req-0faf01e0-9a2e-4e7d-bbbd-71f87bf60864 tempest-AttachVolumeTestJSON-132925025 tempest-AttachVolumeTestJSON-132925025-project-member] [instance: e3360f00-a42a-40f2-9e6b-813547ebfd2b] Stopping instance; current vm_state: active, current task_state: powering-off, current DB power_state: 1, current VM power_state: 1 {{(pid=61806) do_stop_instance /opt/stack/nova/nova/compute/manager.py:3368}} [ 1443.661504] env[61806]: DEBUG nova.objects.instance [None req-0faf01e0-9a2e-4e7d-bbbd-71f87bf60864 tempest-AttachVolumeTestJSON-132925025 tempest-AttachVolumeTestJSON-132925025-project-member] Lazy-loading 'flavor' on Instance uuid e3360f00-a42a-40f2-9e6b-813547ebfd2b {{(pid=61806) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1443.918228] env[61806]: DEBUG oslo_service.periodic_task [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=61806) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1443.918422] env[61806]: DEBUG oslo_service.periodic_task [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=61806) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1443.918562] env[61806]: DEBUG oslo_service.periodic_task [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=61806) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1443.918723] env[61806]: DEBUG oslo_service.periodic_task [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=61806) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1443.918866] env[61806]: DEBUG nova.compute.manager [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=61806) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10545}} [ 1444.166334] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [None req-0faf01e0-9a2e-4e7d-bbbd-71f87bf60864 tempest-AttachVolumeTestJSON-132925025 tempest-AttachVolumeTestJSON-132925025-project-member] [instance: e3360f00-a42a-40f2-9e6b-813547ebfd2b] Powering off the VM {{(pid=61806) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1444.166712] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-ac704654-73b4-45d5-8b37-630f33f315b0 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1444.173549] env[61806]: DEBUG oslo_vmware.api [None req-0faf01e0-9a2e-4e7d-bbbd-71f87bf60864 tempest-AttachVolumeTestJSON-132925025 tempest-AttachVolumeTestJSON-132925025-project-member] Waiting for the task: (returnval){ [ 1444.173549] env[61806]: value = "task-1295381" [ 1444.173549] env[61806]: _type = "Task" [ 1444.173549] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1444.181780] env[61806]: DEBUG oslo_vmware.api [None req-0faf01e0-9a2e-4e7d-bbbd-71f87bf60864 tempest-AttachVolumeTestJSON-132925025 tempest-AttachVolumeTestJSON-132925025-project-member] Task: {'id': task-1295381, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1444.683326] env[61806]: DEBUG oslo_vmware.api [None req-0faf01e0-9a2e-4e7d-bbbd-71f87bf60864 tempest-AttachVolumeTestJSON-132925025 tempest-AttachVolumeTestJSON-132925025-project-member] Task: {'id': task-1295381, 'name': PowerOffVM_Task, 'duration_secs': 0.175588} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1444.683608] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [None req-0faf01e0-9a2e-4e7d-bbbd-71f87bf60864 tempest-AttachVolumeTestJSON-132925025 tempest-AttachVolumeTestJSON-132925025-project-member] [instance: e3360f00-a42a-40f2-9e6b-813547ebfd2b] Powered off the VM {{(pid=61806) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1444.683794] env[61806]: DEBUG nova.compute.manager [None req-0faf01e0-9a2e-4e7d-bbbd-71f87bf60864 tempest-AttachVolumeTestJSON-132925025 tempest-AttachVolumeTestJSON-132925025-project-member] [instance: e3360f00-a42a-40f2-9e6b-813547ebfd2b] Checking state {{(pid=61806) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1444.684557] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9d7d3f01-e543-4443-bcb8-3924f9622d20 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1445.196048] env[61806]: DEBUG oslo_concurrency.lockutils [None req-0faf01e0-9a2e-4e7d-bbbd-71f87bf60864 tempest-AttachVolumeTestJSON-132925025 tempest-AttachVolumeTestJSON-132925025-project-member] Lock "e3360f00-a42a-40f2-9e6b-813547ebfd2b" "released" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: held 1.542s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1446.023167] env[61806]: DEBUG oslo_service.periodic_task [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=61806) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1446.088434] env[61806]: DEBUG nova.objects.instance [None req-956e03b1-edef-4f70-8cce-b1aea76cfd8b tempest-AttachVolumeTestJSON-132925025 tempest-AttachVolumeTestJSON-132925025-project-member] Lazy-loading 'flavor' on Instance uuid e3360f00-a42a-40f2-9e6b-813547ebfd2b {{(pid=61806) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1446.594699] env[61806]: DEBUG oslo_concurrency.lockutils [None req-956e03b1-edef-4f70-8cce-b1aea76cfd8b tempest-AttachVolumeTestJSON-132925025 tempest-AttachVolumeTestJSON-132925025-project-member] Acquiring lock "refresh_cache-e3360f00-a42a-40f2-9e6b-813547ebfd2b" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1446.594699] env[61806]: DEBUG oslo_concurrency.lockutils [None req-956e03b1-edef-4f70-8cce-b1aea76cfd8b tempest-AttachVolumeTestJSON-132925025 tempest-AttachVolumeTestJSON-132925025-project-member] Acquired lock "refresh_cache-e3360f00-a42a-40f2-9e6b-813547ebfd2b" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1446.595104] env[61806]: DEBUG nova.network.neutron [None req-956e03b1-edef-4f70-8cce-b1aea76cfd8b tempest-AttachVolumeTestJSON-132925025 tempest-AttachVolumeTestJSON-132925025-project-member] [instance: e3360f00-a42a-40f2-9e6b-813547ebfd2b] Building network info cache for instance {{(pid=61806) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1446.595104] env[61806]: DEBUG nova.objects.instance [None req-956e03b1-edef-4f70-8cce-b1aea76cfd8b tempest-AttachVolumeTestJSON-132925025 tempest-AttachVolumeTestJSON-132925025-project-member] Lazy-loading 'info_cache' on Instance uuid e3360f00-a42a-40f2-9e6b-813547ebfd2b {{(pid=61806) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1447.099113] env[61806]: DEBUG nova.objects.base [None req-956e03b1-edef-4f70-8cce-b1aea76cfd8b tempest-AttachVolumeTestJSON-132925025 tempest-AttachVolumeTestJSON-132925025-project-member] Object Instance lazy-loaded attributes: flavor,info_cache {{(pid=61806) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 1447.802284] env[61806]: DEBUG nova.network.neutron [None req-956e03b1-edef-4f70-8cce-b1aea76cfd8b tempest-AttachVolumeTestJSON-132925025 tempest-AttachVolumeTestJSON-132925025-project-member] [instance: e3360f00-a42a-40f2-9e6b-813547ebfd2b] Updating instance_info_cache with network_info: [{"id": "3c8061f1-58c9-49f2-bdc9-e43e9936373e", "address": "fa:16:3e:d4:61:cb", "network": {"id": "0fd518d3-c254-4dba-a147-378c3c029c51", "bridge": "br-int", "label": "tempest-AttachVolumeTestJSON-538202287-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.229", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "833defcd470d46ceacd7886d1f4fb0cf", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "91b0f7e5-0d1a-46e2-bf73-09656211dea2", "external-id": "nsx-vlan-transportzone-488", "segmentation_id": 488, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3c8061f1-58", "ovs_interfaceid": "3c8061f1-58c9-49f2-bdc9-e43e9936373e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61806) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1448.305081] env[61806]: DEBUG oslo_concurrency.lockutils [None req-956e03b1-edef-4f70-8cce-b1aea76cfd8b tempest-AttachVolumeTestJSON-132925025 tempest-AttachVolumeTestJSON-132925025-project-member] Releasing lock "refresh_cache-e3360f00-a42a-40f2-9e6b-813547ebfd2b" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1448.808478] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [None req-956e03b1-edef-4f70-8cce-b1aea76cfd8b tempest-AttachVolumeTestJSON-132925025 tempest-AttachVolumeTestJSON-132925025-project-member] [instance: e3360f00-a42a-40f2-9e6b-813547ebfd2b] Powering on the VM {{(pid=61806) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1448.808935] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-3d276b3c-360b-4163-b23c-f90cc937b717 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1448.816748] env[61806]: DEBUG oslo_vmware.api [None req-956e03b1-edef-4f70-8cce-b1aea76cfd8b tempest-AttachVolumeTestJSON-132925025 tempest-AttachVolumeTestJSON-132925025-project-member] Waiting for the task: (returnval){ [ 1448.816748] env[61806]: value = "task-1295382" [ 1448.816748] env[61806]: _type = "Task" [ 1448.816748] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1448.823897] env[61806]: DEBUG oslo_vmware.api [None req-956e03b1-edef-4f70-8cce-b1aea76cfd8b tempest-AttachVolumeTestJSON-132925025 tempest-AttachVolumeTestJSON-132925025-project-member] Task: {'id': task-1295382, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1449.327654] env[61806]: DEBUG oslo_vmware.api [None req-956e03b1-edef-4f70-8cce-b1aea76cfd8b tempest-AttachVolumeTestJSON-132925025 tempest-AttachVolumeTestJSON-132925025-project-member] Task: {'id': task-1295382, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1449.827185] env[61806]: DEBUG oslo_vmware.api [None req-956e03b1-edef-4f70-8cce-b1aea76cfd8b tempest-AttachVolumeTestJSON-132925025 tempest-AttachVolumeTestJSON-132925025-project-member] Task: {'id': task-1295382, 'name': PowerOnVM_Task, 'duration_secs': 0.967032} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1449.827596] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [None req-956e03b1-edef-4f70-8cce-b1aea76cfd8b tempest-AttachVolumeTestJSON-132925025 tempest-AttachVolumeTestJSON-132925025-project-member] [instance: e3360f00-a42a-40f2-9e6b-813547ebfd2b] Powered on the VM {{(pid=61806) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1449.827648] env[61806]: DEBUG nova.compute.manager [None req-956e03b1-edef-4f70-8cce-b1aea76cfd8b tempest-AttachVolumeTestJSON-132925025 tempest-AttachVolumeTestJSON-132925025-project-member] [instance: e3360f00-a42a-40f2-9e6b-813547ebfd2b] Checking state {{(pid=61806) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1449.828420] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f1e366c3-276b-4b35-8aa5-dcd71047278b {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1486.335535] env[61806]: DEBUG oslo_concurrency.lockutils [None req-f6a63672-6e24-4b1e-ba8c-d4e8874da949 tempest-AttachVolumeTestJSON-132925025 tempest-AttachVolumeTestJSON-132925025-project-member] Acquiring lock "e3360f00-a42a-40f2-9e6b-813547ebfd2b" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1486.335944] env[61806]: DEBUG oslo_concurrency.lockutils [None req-f6a63672-6e24-4b1e-ba8c-d4e8874da949 tempest-AttachVolumeTestJSON-132925025 tempest-AttachVolumeTestJSON-132925025-project-member] Lock "e3360f00-a42a-40f2-9e6b-813547ebfd2b" acquired by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: waited 0.001s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1486.839080] env[61806]: INFO nova.compute.manager [None req-f6a63672-6e24-4b1e-ba8c-d4e8874da949 tempest-AttachVolumeTestJSON-132925025 tempest-AttachVolumeTestJSON-132925025-project-member] [instance: e3360f00-a42a-40f2-9e6b-813547ebfd2b] Detaching volume 434af0b9-a453-4b85-9692-5eac3a926286 [ 1486.870122] env[61806]: INFO nova.virt.block_device [None req-f6a63672-6e24-4b1e-ba8c-d4e8874da949 tempest-AttachVolumeTestJSON-132925025 tempest-AttachVolumeTestJSON-132925025-project-member] [instance: e3360f00-a42a-40f2-9e6b-813547ebfd2b] Attempting to driver detach volume 434af0b9-a453-4b85-9692-5eac3a926286 from mountpoint /dev/sdb [ 1486.870473] env[61806]: DEBUG nova.virt.vmwareapi.volumeops [None req-f6a63672-6e24-4b1e-ba8c-d4e8874da949 tempest-AttachVolumeTestJSON-132925025 tempest-AttachVolumeTestJSON-132925025-project-member] [instance: e3360f00-a42a-40f2-9e6b-813547ebfd2b] Volume detach. Driver type: vmdk {{(pid=61806) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 1486.870715] env[61806]: DEBUG nova.virt.vmwareapi.volumeops [None req-f6a63672-6e24-4b1e-ba8c-d4e8874da949 tempest-AttachVolumeTestJSON-132925025 tempest-AttachVolumeTestJSON-132925025-project-member] [instance: e3360f00-a42a-40f2-9e6b-813547ebfd2b] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-277786', 'volume_id': '434af0b9-a453-4b85-9692-5eac3a926286', 'name': 'volume-434af0b9-a453-4b85-9692-5eac3a926286', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'e3360f00-a42a-40f2-9e6b-813547ebfd2b', 'attached_at': '', 'detached_at': '', 'volume_id': '434af0b9-a453-4b85-9692-5eac3a926286', 'serial': '434af0b9-a453-4b85-9692-5eac3a926286'} {{(pid=61806) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 1486.871612] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-19cb121d-2270-47c7-8dd9-911c7219aecb {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1486.893188] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0b91dab2-3e3b-452b-86d2-224b4afc96e6 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1486.899912] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a2f51e86-86b2-4e3f-9898-76318bd32462 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1486.920498] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-79cfcbae-f6f4-4c5f-9ed5-4c67b0d47d0a {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1486.934641] env[61806]: DEBUG nova.virt.vmwareapi.volumeops [None req-f6a63672-6e24-4b1e-ba8c-d4e8874da949 tempest-AttachVolumeTestJSON-132925025 tempest-AttachVolumeTestJSON-132925025-project-member] The volume has not been displaced from its original location: [datastore1] volume-434af0b9-a453-4b85-9692-5eac3a926286/volume-434af0b9-a453-4b85-9692-5eac3a926286.vmdk. No consolidation needed. {{(pid=61806) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 1486.939856] env[61806]: DEBUG nova.virt.vmwareapi.volumeops [None req-f6a63672-6e24-4b1e-ba8c-d4e8874da949 tempest-AttachVolumeTestJSON-132925025 tempest-AttachVolumeTestJSON-132925025-project-member] [instance: e3360f00-a42a-40f2-9e6b-813547ebfd2b] Reconfiguring VM instance instance-0000006f to detach disk 2001 {{(pid=61806) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1486.940155] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-20f5a23d-0efa-4266-a26a-5d2897152783 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1486.958050] env[61806]: DEBUG oslo_vmware.api [None req-f6a63672-6e24-4b1e-ba8c-d4e8874da949 tempest-AttachVolumeTestJSON-132925025 tempest-AttachVolumeTestJSON-132925025-project-member] Waiting for the task: (returnval){ [ 1486.958050] env[61806]: value = "task-1295383" [ 1486.958050] env[61806]: _type = "Task" [ 1486.958050] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1486.965539] env[61806]: DEBUG oslo_vmware.api [None req-f6a63672-6e24-4b1e-ba8c-d4e8874da949 tempest-AttachVolumeTestJSON-132925025 tempest-AttachVolumeTestJSON-132925025-project-member] Task: {'id': task-1295383, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1487.468174] env[61806]: DEBUG oslo_vmware.api [None req-f6a63672-6e24-4b1e-ba8c-d4e8874da949 tempest-AttachVolumeTestJSON-132925025 tempest-AttachVolumeTestJSON-132925025-project-member] Task: {'id': task-1295383, 'name': ReconfigVM_Task, 'duration_secs': 0.20117} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1487.468570] env[61806]: DEBUG nova.virt.vmwareapi.volumeops [None req-f6a63672-6e24-4b1e-ba8c-d4e8874da949 tempest-AttachVolumeTestJSON-132925025 tempest-AttachVolumeTestJSON-132925025-project-member] [instance: e3360f00-a42a-40f2-9e6b-813547ebfd2b] Reconfigured VM instance instance-0000006f to detach disk 2001 {{(pid=61806) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1487.473112] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-cd096af0-edc1-4953-b699-c36bc07c3d10 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1487.487367] env[61806]: DEBUG oslo_vmware.api [None req-f6a63672-6e24-4b1e-ba8c-d4e8874da949 tempest-AttachVolumeTestJSON-132925025 tempest-AttachVolumeTestJSON-132925025-project-member] Waiting for the task: (returnval){ [ 1487.487367] env[61806]: value = "task-1295384" [ 1487.487367] env[61806]: _type = "Task" [ 1487.487367] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1487.494785] env[61806]: DEBUG oslo_vmware.api [None req-f6a63672-6e24-4b1e-ba8c-d4e8874da949 tempest-AttachVolumeTestJSON-132925025 tempest-AttachVolumeTestJSON-132925025-project-member] Task: {'id': task-1295384, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1487.997317] env[61806]: DEBUG oslo_vmware.api [None req-f6a63672-6e24-4b1e-ba8c-d4e8874da949 tempest-AttachVolumeTestJSON-132925025 tempest-AttachVolumeTestJSON-132925025-project-member] Task: {'id': task-1295384, 'name': ReconfigVM_Task, 'duration_secs': 0.127011} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1487.997617] env[61806]: DEBUG nova.virt.vmwareapi.volumeops [None req-f6a63672-6e24-4b1e-ba8c-d4e8874da949 tempest-AttachVolumeTestJSON-132925025 tempest-AttachVolumeTestJSON-132925025-project-member] [instance: e3360f00-a42a-40f2-9e6b-813547ebfd2b] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-277786', 'volume_id': '434af0b9-a453-4b85-9692-5eac3a926286', 'name': 'volume-434af0b9-a453-4b85-9692-5eac3a926286', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'e3360f00-a42a-40f2-9e6b-813547ebfd2b', 'attached_at': '', 'detached_at': '', 'volume_id': '434af0b9-a453-4b85-9692-5eac3a926286', 'serial': '434af0b9-a453-4b85-9692-5eac3a926286'} {{(pid=61806) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 1488.539564] env[61806]: DEBUG nova.objects.instance [None req-f6a63672-6e24-4b1e-ba8c-d4e8874da949 tempest-AttachVolumeTestJSON-132925025 tempest-AttachVolumeTestJSON-132925025-project-member] Lazy-loading 'flavor' on Instance uuid e3360f00-a42a-40f2-9e6b-813547ebfd2b {{(pid=61806) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1489.496198] env[61806]: DEBUG oslo_concurrency.lockutils [None req-01235003-f58b-43f8-a92a-f61c3583436d tempest-AttachVolumeTestJSON-132925025 tempest-AttachVolumeTestJSON-132925025-project-member] Acquiring lock "e3360f00-a42a-40f2-9e6b-813547ebfd2b" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1489.548836] env[61806]: DEBUG oslo_concurrency.lockutils [None req-f6a63672-6e24-4b1e-ba8c-d4e8874da949 tempest-AttachVolumeTestJSON-132925025 tempest-AttachVolumeTestJSON-132925025-project-member] Lock "e3360f00-a42a-40f2-9e6b-813547ebfd2b" "released" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: held 3.213s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1489.549929] env[61806]: DEBUG oslo_concurrency.lockutils [None req-01235003-f58b-43f8-a92a-f61c3583436d tempest-AttachVolumeTestJSON-132925025 tempest-AttachVolumeTestJSON-132925025-project-member] Lock "e3360f00-a42a-40f2-9e6b-813547ebfd2b" acquired by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: waited 0.054s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1489.550142] env[61806]: DEBUG nova.compute.manager [None req-01235003-f58b-43f8-a92a-f61c3583436d tempest-AttachVolumeTestJSON-132925025 tempest-AttachVolumeTestJSON-132925025-project-member] [instance: e3360f00-a42a-40f2-9e6b-813547ebfd2b] Checking state {{(pid=61806) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1489.551238] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e9752e2d-70e8-446a-8f60-b8a1c2e10dbd {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1489.558330] env[61806]: DEBUG nova.compute.manager [None req-01235003-f58b-43f8-a92a-f61c3583436d tempest-AttachVolumeTestJSON-132925025 tempest-AttachVolumeTestJSON-132925025-project-member] [instance: e3360f00-a42a-40f2-9e6b-813547ebfd2b] Stopping instance; current vm_state: active, current task_state: powering-off, current DB power_state: 1, current VM power_state: 1 {{(pid=61806) do_stop_instance /opt/stack/nova/nova/compute/manager.py:3368}} [ 1489.558919] env[61806]: DEBUG nova.objects.instance [None req-01235003-f58b-43f8-a92a-f61c3583436d tempest-AttachVolumeTestJSON-132925025 tempest-AttachVolumeTestJSON-132925025-project-member] Lazy-loading 'flavor' on Instance uuid e3360f00-a42a-40f2-9e6b-813547ebfd2b {{(pid=61806) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1490.063856] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [None req-01235003-f58b-43f8-a92a-f61c3583436d tempest-AttachVolumeTestJSON-132925025 tempest-AttachVolumeTestJSON-132925025-project-member] [instance: e3360f00-a42a-40f2-9e6b-813547ebfd2b] Powering off the VM {{(pid=61806) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1490.064163] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-fda99351-0e3a-43bd-aeac-2bbe3896d41f {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1490.071917] env[61806]: DEBUG oslo_vmware.api [None req-01235003-f58b-43f8-a92a-f61c3583436d tempest-AttachVolumeTestJSON-132925025 tempest-AttachVolumeTestJSON-132925025-project-member] Waiting for the task: (returnval){ [ 1490.071917] env[61806]: value = "task-1295385" [ 1490.071917] env[61806]: _type = "Task" [ 1490.071917] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1490.082222] env[61806]: DEBUG oslo_vmware.api [None req-01235003-f58b-43f8-a92a-f61c3583436d tempest-AttachVolumeTestJSON-132925025 tempest-AttachVolumeTestJSON-132925025-project-member] Task: {'id': task-1295385, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1490.581384] env[61806]: DEBUG oslo_vmware.api [None req-01235003-f58b-43f8-a92a-f61c3583436d tempest-AttachVolumeTestJSON-132925025 tempest-AttachVolumeTestJSON-132925025-project-member] Task: {'id': task-1295385, 'name': PowerOffVM_Task, 'duration_secs': 0.178586} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1490.581784] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [None req-01235003-f58b-43f8-a92a-f61c3583436d tempest-AttachVolumeTestJSON-132925025 tempest-AttachVolumeTestJSON-132925025-project-member] [instance: e3360f00-a42a-40f2-9e6b-813547ebfd2b] Powered off the VM {{(pid=61806) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1490.581832] env[61806]: DEBUG nova.compute.manager [None req-01235003-f58b-43f8-a92a-f61c3583436d tempest-AttachVolumeTestJSON-132925025 tempest-AttachVolumeTestJSON-132925025-project-member] [instance: e3360f00-a42a-40f2-9e6b-813547ebfd2b] Checking state {{(pid=61806) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1490.582626] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2965463a-c039-4165-af44-be718f0afbe3 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1491.093663] env[61806]: DEBUG oslo_concurrency.lockutils [None req-01235003-f58b-43f8-a92a-f61c3583436d tempest-AttachVolumeTestJSON-132925025 tempest-AttachVolumeTestJSON-132925025-project-member] Lock "e3360f00-a42a-40f2-9e6b-813547ebfd2b" "released" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: held 1.544s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1491.892830] env[61806]: DEBUG nova.objects.instance [None req-be3aca29-0ae9-4294-89d0-2dd86f528805 tempest-AttachVolumeTestJSON-132925025 tempest-AttachVolumeTestJSON-132925025-project-member] Lazy-loading 'flavor' on Instance uuid e3360f00-a42a-40f2-9e6b-813547ebfd2b {{(pid=61806) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1492.398580] env[61806]: DEBUG oslo_concurrency.lockutils [None req-be3aca29-0ae9-4294-89d0-2dd86f528805 tempest-AttachVolumeTestJSON-132925025 tempest-AttachVolumeTestJSON-132925025-project-member] Acquiring lock "refresh_cache-e3360f00-a42a-40f2-9e6b-813547ebfd2b" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1492.398770] env[61806]: DEBUG oslo_concurrency.lockutils [None req-be3aca29-0ae9-4294-89d0-2dd86f528805 tempest-AttachVolumeTestJSON-132925025 tempest-AttachVolumeTestJSON-132925025-project-member] Acquired lock "refresh_cache-e3360f00-a42a-40f2-9e6b-813547ebfd2b" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1492.398933] env[61806]: DEBUG nova.network.neutron [None req-be3aca29-0ae9-4294-89d0-2dd86f528805 tempest-AttachVolumeTestJSON-132925025 tempest-AttachVolumeTestJSON-132925025-project-member] [instance: e3360f00-a42a-40f2-9e6b-813547ebfd2b] Building network info cache for instance {{(pid=61806) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1492.399139] env[61806]: DEBUG nova.objects.instance [None req-be3aca29-0ae9-4294-89d0-2dd86f528805 tempest-AttachVolumeTestJSON-132925025 tempest-AttachVolumeTestJSON-132925025-project-member] Lazy-loading 'info_cache' on Instance uuid e3360f00-a42a-40f2-9e6b-813547ebfd2b {{(pid=61806) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1492.903067] env[61806]: DEBUG nova.objects.base [None req-be3aca29-0ae9-4294-89d0-2dd86f528805 tempest-AttachVolumeTestJSON-132925025 tempest-AttachVolumeTestJSON-132925025-project-member] Object Instance lazy-loaded attributes: flavor,info_cache {{(pid=61806) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 1493.599107] env[61806]: DEBUG nova.network.neutron [None req-be3aca29-0ae9-4294-89d0-2dd86f528805 tempest-AttachVolumeTestJSON-132925025 tempest-AttachVolumeTestJSON-132925025-project-member] [instance: e3360f00-a42a-40f2-9e6b-813547ebfd2b] Updating instance_info_cache with network_info: [{"id": "3c8061f1-58c9-49f2-bdc9-e43e9936373e", "address": "fa:16:3e:d4:61:cb", "network": {"id": "0fd518d3-c254-4dba-a147-378c3c029c51", "bridge": "br-int", "label": "tempest-AttachVolumeTestJSON-538202287-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.229", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "833defcd470d46ceacd7886d1f4fb0cf", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "91b0f7e5-0d1a-46e2-bf73-09656211dea2", "external-id": "nsx-vlan-transportzone-488", "segmentation_id": 488, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3c8061f1-58", "ovs_interfaceid": "3c8061f1-58c9-49f2-bdc9-e43e9936373e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61806) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1494.018073] env[61806]: DEBUG oslo_service.periodic_task [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=61806) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1494.103053] env[61806]: DEBUG oslo_concurrency.lockutils [None req-be3aca29-0ae9-4294-89d0-2dd86f528805 tempest-AttachVolumeTestJSON-132925025 tempest-AttachVolumeTestJSON-132925025-project-member] Releasing lock "refresh_cache-e3360f00-a42a-40f2-9e6b-813547ebfd2b" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1494.605843] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [None req-be3aca29-0ae9-4294-89d0-2dd86f528805 tempest-AttachVolumeTestJSON-132925025 tempest-AttachVolumeTestJSON-132925025-project-member] [instance: e3360f00-a42a-40f2-9e6b-813547ebfd2b] Powering on the VM {{(pid=61806) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1494.606202] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-d481ba81-feea-40f8-96a8-3482f8c423a3 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1494.613539] env[61806]: DEBUG oslo_vmware.api [None req-be3aca29-0ae9-4294-89d0-2dd86f528805 tempest-AttachVolumeTestJSON-132925025 tempest-AttachVolumeTestJSON-132925025-project-member] Waiting for the task: (returnval){ [ 1494.613539] env[61806]: value = "task-1295386" [ 1494.613539] env[61806]: _type = "Task" [ 1494.613539] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1494.620977] env[61806]: DEBUG oslo_vmware.api [None req-be3aca29-0ae9-4294-89d0-2dd86f528805 tempest-AttachVolumeTestJSON-132925025 tempest-AttachVolumeTestJSON-132925025-project-member] Task: {'id': task-1295386, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1495.122889] env[61806]: DEBUG oslo_vmware.api [None req-be3aca29-0ae9-4294-89d0-2dd86f528805 tempest-AttachVolumeTestJSON-132925025 tempest-AttachVolumeTestJSON-132925025-project-member] Task: {'id': task-1295386, 'name': PowerOnVM_Task, 'duration_secs': 0.378647} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1495.123238] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [None req-be3aca29-0ae9-4294-89d0-2dd86f528805 tempest-AttachVolumeTestJSON-132925025 tempest-AttachVolumeTestJSON-132925025-project-member] [instance: e3360f00-a42a-40f2-9e6b-813547ebfd2b] Powered on the VM {{(pid=61806) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1495.123369] env[61806]: DEBUG nova.compute.manager [None req-be3aca29-0ae9-4294-89d0-2dd86f528805 tempest-AttachVolumeTestJSON-132925025 tempest-AttachVolumeTestJSON-132925025-project-member] [instance: e3360f00-a42a-40f2-9e6b-813547ebfd2b] Checking state {{(pid=61806) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1495.124116] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-65648892-3590-432d-8065-e47e99f92945 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1497.023298] env[61806]: DEBUG oslo_service.periodic_task [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=61806) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1498.023726] env[61806]: DEBUG oslo_service.periodic_task [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=61806) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1498.023953] env[61806]: DEBUG nova.compute.manager [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Starting heal instance info cache {{(pid=61806) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9926}} [ 1498.024097] env[61806]: DEBUG nova.compute.manager [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Rebuilding the list of instances to heal {{(pid=61806) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9930}} [ 1498.562511] env[61806]: DEBUG oslo_concurrency.lockutils [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Acquiring lock "refresh_cache-e3360f00-a42a-40f2-9e6b-813547ebfd2b" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1498.562700] env[61806]: DEBUG oslo_concurrency.lockutils [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Acquired lock "refresh_cache-e3360f00-a42a-40f2-9e6b-813547ebfd2b" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1498.562855] env[61806]: DEBUG nova.network.neutron [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] [instance: e3360f00-a42a-40f2-9e6b-813547ebfd2b] Forcefully refreshing network info cache for instance {{(pid=61806) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2004}} [ 1498.563024] env[61806]: DEBUG nova.objects.instance [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Lazy-loading 'info_cache' on Instance uuid e3360f00-a42a-40f2-9e6b-813547ebfd2b {{(pid=61806) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1500.271967] env[61806]: DEBUG nova.network.neutron [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] [instance: e3360f00-a42a-40f2-9e6b-813547ebfd2b] Updating instance_info_cache with network_info: [{"id": "3c8061f1-58c9-49f2-bdc9-e43e9936373e", "address": "fa:16:3e:d4:61:cb", "network": {"id": "0fd518d3-c254-4dba-a147-378c3c029c51", "bridge": "br-int", "label": "tempest-AttachVolumeTestJSON-538202287-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.229", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "833defcd470d46ceacd7886d1f4fb0cf", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "91b0f7e5-0d1a-46e2-bf73-09656211dea2", "external-id": "nsx-vlan-transportzone-488", "segmentation_id": 488, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3c8061f1-58", "ovs_interfaceid": "3c8061f1-58c9-49f2-bdc9-e43e9936373e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61806) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1500.775169] env[61806]: DEBUG oslo_concurrency.lockutils [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Releasing lock "refresh_cache-e3360f00-a42a-40f2-9e6b-813547ebfd2b" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1500.775378] env[61806]: DEBUG nova.compute.manager [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] [instance: e3360f00-a42a-40f2-9e6b-813547ebfd2b] Updated the network info_cache for instance {{(pid=61806) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9997}} [ 1500.775573] env[61806]: DEBUG oslo_service.periodic_task [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=61806) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1500.775756] env[61806]: DEBUG oslo_service.periodic_task [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Running periodic task ComputeManager.update_available_resource {{(pid=61806) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1501.279228] env[61806]: DEBUG oslo_concurrency.lockutils [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1501.279615] env[61806]: DEBUG oslo_concurrency.lockutils [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 0.000s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1501.279660] env[61806]: DEBUG oslo_concurrency.lockutils [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1501.279786] env[61806]: DEBUG nova.compute.resource_tracker [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=61806) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 1501.281189] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a8a6309f-7e90-4acb-9a2f-ddbb4fd4159e {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1501.290326] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-02c91446-15f0-4dae-bc7f-91559c2ae1e9 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1501.303521] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cfcb4fd1-48e7-4746-9092-858a26d2a1c3 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1501.309665] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8c50a3c4-2d2a-41fe-a2b7-528acb820106 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1501.338282] env[61806]: DEBUG nova.compute.resource_tracker [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=181104MB free_disk=138GB free_vcpus=48 pci_devices=None {{(pid=61806) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 1501.338430] env[61806]: DEBUG oslo_concurrency.lockutils [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1501.338616] env[61806]: DEBUG oslo_concurrency.lockutils [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 0.000s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1502.364619] env[61806]: DEBUG nova.compute.resource_tracker [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Instance e3360f00-a42a-40f2-9e6b-813547ebfd2b actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61806) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1502.364842] env[61806]: DEBUG nova.compute.resource_tracker [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Total usable vcpus: 48, total allocated vcpus: 1 {{(pid=61806) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 1502.365017] env[61806]: DEBUG nova.compute.resource_tracker [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=704MB phys_disk=200GB used_disk=1GB total_vcpus=48 used_vcpus=1 pci_stats=[] {{(pid=61806) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 1502.390502] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8eb31fb9-10f8-4333-b119-ad31acf82d8c {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1502.397878] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ad9bca1d-8c8a-4441-b1fa-9865db640994 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1502.426106] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c1fbefc9-7688-4672-b639-1ea19a24adb4 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1502.432485] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-893a27f8-c3d9-49f2-94e0-c529098cfb71 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1502.444654] env[61806]: DEBUG nova.compute.provider_tree [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Inventory has not changed in ProviderTree for provider: a2858be1-fd22-4e08-979e-87ad25293407 {{(pid=61806) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1502.947243] env[61806]: DEBUG nova.scheduler.client.report [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Inventory has not changed for provider a2858be1-fd22-4e08-979e-87ad25293407 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 138, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61806) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1502.948604] env[61806]: DEBUG nova.compute.resource_tracker [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=61806) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 1502.948850] env[61806]: DEBUG oslo_concurrency.lockutils [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 1.610s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1503.196450] env[61806]: DEBUG oslo_service.periodic_task [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=61806) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1503.196680] env[61806]: DEBUG oslo_service.periodic_task [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=61806) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1503.196866] env[61806]: DEBUG oslo_service.periodic_task [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=61806) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1503.197021] env[61806]: DEBUG nova.compute.manager [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=61806) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10545}} [ 1507.023902] env[61806]: DEBUG oslo_service.periodic_task [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=61806) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1531.292564] env[61806]: DEBUG oslo_concurrency.lockutils [None req-ae9b29d4-744a-4bb4-8a0e-9ca5ba248a92 tempest-AttachVolumeTestJSON-132925025 tempest-AttachVolumeTestJSON-132925025-project-member] Acquiring lock "e3360f00-a42a-40f2-9e6b-813547ebfd2b" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1531.292959] env[61806]: DEBUG oslo_concurrency.lockutils [None req-ae9b29d4-744a-4bb4-8a0e-9ca5ba248a92 tempest-AttachVolumeTestJSON-132925025 tempest-AttachVolumeTestJSON-132925025-project-member] Lock "e3360f00-a42a-40f2-9e6b-813547ebfd2b" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1531.293813] env[61806]: DEBUG oslo_concurrency.lockutils [None req-ae9b29d4-744a-4bb4-8a0e-9ca5ba248a92 tempest-AttachVolumeTestJSON-132925025 tempest-AttachVolumeTestJSON-132925025-project-member] Acquiring lock "e3360f00-a42a-40f2-9e6b-813547ebfd2b-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1531.293813] env[61806]: DEBUG oslo_concurrency.lockutils [None req-ae9b29d4-744a-4bb4-8a0e-9ca5ba248a92 tempest-AttachVolumeTestJSON-132925025 tempest-AttachVolumeTestJSON-132925025-project-member] Lock "e3360f00-a42a-40f2-9e6b-813547ebfd2b-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1531.293813] env[61806]: DEBUG oslo_concurrency.lockutils [None req-ae9b29d4-744a-4bb4-8a0e-9ca5ba248a92 tempest-AttachVolumeTestJSON-132925025 tempest-AttachVolumeTestJSON-132925025-project-member] Lock "e3360f00-a42a-40f2-9e6b-813547ebfd2b-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1531.296946] env[61806]: INFO nova.compute.manager [None req-ae9b29d4-744a-4bb4-8a0e-9ca5ba248a92 tempest-AttachVolumeTestJSON-132925025 tempest-AttachVolumeTestJSON-132925025-project-member] [instance: e3360f00-a42a-40f2-9e6b-813547ebfd2b] Terminating instance [ 1531.298728] env[61806]: DEBUG nova.compute.manager [None req-ae9b29d4-744a-4bb4-8a0e-9ca5ba248a92 tempest-AttachVolumeTestJSON-132925025 tempest-AttachVolumeTestJSON-132925025-project-member] [instance: e3360f00-a42a-40f2-9e6b-813547ebfd2b] Start destroying the instance on the hypervisor. {{(pid=61806) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1531.298965] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-ae9b29d4-744a-4bb4-8a0e-9ca5ba248a92 tempest-AttachVolumeTestJSON-132925025 tempest-AttachVolumeTestJSON-132925025-project-member] [instance: e3360f00-a42a-40f2-9e6b-813547ebfd2b] Destroying instance {{(pid=61806) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1531.299850] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d976e45c-59eb-4d65-a4d6-d414ba8cc108 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1531.307938] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [None req-ae9b29d4-744a-4bb4-8a0e-9ca5ba248a92 tempest-AttachVolumeTestJSON-132925025 tempest-AttachVolumeTestJSON-132925025-project-member] [instance: e3360f00-a42a-40f2-9e6b-813547ebfd2b] Powering off the VM {{(pid=61806) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1531.308178] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-08877bcc-007e-4f04-bd3e-2dabbcd9d62f {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1531.314052] env[61806]: DEBUG oslo_vmware.api [None req-ae9b29d4-744a-4bb4-8a0e-9ca5ba248a92 tempest-AttachVolumeTestJSON-132925025 tempest-AttachVolumeTestJSON-132925025-project-member] Waiting for the task: (returnval){ [ 1531.314052] env[61806]: value = "task-1295387" [ 1531.314052] env[61806]: _type = "Task" [ 1531.314052] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1531.321404] env[61806]: DEBUG oslo_vmware.api [None req-ae9b29d4-744a-4bb4-8a0e-9ca5ba248a92 tempest-AttachVolumeTestJSON-132925025 tempest-AttachVolumeTestJSON-132925025-project-member] Task: {'id': task-1295387, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1531.823987] env[61806]: DEBUG oslo_vmware.api [None req-ae9b29d4-744a-4bb4-8a0e-9ca5ba248a92 tempest-AttachVolumeTestJSON-132925025 tempest-AttachVolumeTestJSON-132925025-project-member] Task: {'id': task-1295387, 'name': PowerOffVM_Task, 'duration_secs': 0.165321} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1531.824297] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [None req-ae9b29d4-744a-4bb4-8a0e-9ca5ba248a92 tempest-AttachVolumeTestJSON-132925025 tempest-AttachVolumeTestJSON-132925025-project-member] [instance: e3360f00-a42a-40f2-9e6b-813547ebfd2b] Powered off the VM {{(pid=61806) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1531.824475] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-ae9b29d4-744a-4bb4-8a0e-9ca5ba248a92 tempest-AttachVolumeTestJSON-132925025 tempest-AttachVolumeTestJSON-132925025-project-member] [instance: e3360f00-a42a-40f2-9e6b-813547ebfd2b] Unregistering the VM {{(pid=61806) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1531.824734] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-d5d55df3-cf25-4148-a75e-9c6deb930089 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1531.884055] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-ae9b29d4-744a-4bb4-8a0e-9ca5ba248a92 tempest-AttachVolumeTestJSON-132925025 tempest-AttachVolumeTestJSON-132925025-project-member] [instance: e3360f00-a42a-40f2-9e6b-813547ebfd2b] Unregistered the VM {{(pid=61806) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1531.884284] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-ae9b29d4-744a-4bb4-8a0e-9ca5ba248a92 tempest-AttachVolumeTestJSON-132925025 tempest-AttachVolumeTestJSON-132925025-project-member] [instance: e3360f00-a42a-40f2-9e6b-813547ebfd2b] Deleting contents of the VM from datastore datastore1 {{(pid=61806) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1531.884477] env[61806]: DEBUG nova.virt.vmwareapi.ds_util [None req-ae9b29d4-744a-4bb4-8a0e-9ca5ba248a92 tempest-AttachVolumeTestJSON-132925025 tempest-AttachVolumeTestJSON-132925025-project-member] Deleting the datastore file [datastore1] e3360f00-a42a-40f2-9e6b-813547ebfd2b {{(pid=61806) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1531.884750] env[61806]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-b725963f-fa90-430c-9bc9-f6f92a25bd78 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1531.891793] env[61806]: DEBUG oslo_vmware.api [None req-ae9b29d4-744a-4bb4-8a0e-9ca5ba248a92 tempest-AttachVolumeTestJSON-132925025 tempest-AttachVolumeTestJSON-132925025-project-member] Waiting for the task: (returnval){ [ 1531.891793] env[61806]: value = "task-1295389" [ 1531.891793] env[61806]: _type = "Task" [ 1531.891793] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1531.899421] env[61806]: DEBUG oslo_vmware.api [None req-ae9b29d4-744a-4bb4-8a0e-9ca5ba248a92 tempest-AttachVolumeTestJSON-132925025 tempest-AttachVolumeTestJSON-132925025-project-member] Task: {'id': task-1295389, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1532.402090] env[61806]: DEBUG oslo_vmware.api [None req-ae9b29d4-744a-4bb4-8a0e-9ca5ba248a92 tempest-AttachVolumeTestJSON-132925025 tempest-AttachVolumeTestJSON-132925025-project-member] Task: {'id': task-1295389, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.145851} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1532.402485] env[61806]: DEBUG nova.virt.vmwareapi.ds_util [None req-ae9b29d4-744a-4bb4-8a0e-9ca5ba248a92 tempest-AttachVolumeTestJSON-132925025 tempest-AttachVolumeTestJSON-132925025-project-member] Deleted the datastore file {{(pid=61806) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1532.402570] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-ae9b29d4-744a-4bb4-8a0e-9ca5ba248a92 tempest-AttachVolumeTestJSON-132925025 tempest-AttachVolumeTestJSON-132925025-project-member] [instance: e3360f00-a42a-40f2-9e6b-813547ebfd2b] Deleted contents of the VM from datastore datastore1 {{(pid=61806) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1532.402726] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-ae9b29d4-744a-4bb4-8a0e-9ca5ba248a92 tempest-AttachVolumeTestJSON-132925025 tempest-AttachVolumeTestJSON-132925025-project-member] [instance: e3360f00-a42a-40f2-9e6b-813547ebfd2b] Instance destroyed {{(pid=61806) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1532.402906] env[61806]: INFO nova.compute.manager [None req-ae9b29d4-744a-4bb4-8a0e-9ca5ba248a92 tempest-AttachVolumeTestJSON-132925025 tempest-AttachVolumeTestJSON-132925025-project-member] [instance: e3360f00-a42a-40f2-9e6b-813547ebfd2b] Took 1.10 seconds to destroy the instance on the hypervisor. [ 1532.403209] env[61806]: DEBUG oslo.service.loopingcall [None req-ae9b29d4-744a-4bb4-8a0e-9ca5ba248a92 tempest-AttachVolumeTestJSON-132925025 tempest-AttachVolumeTestJSON-132925025-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61806) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1532.403414] env[61806]: DEBUG nova.compute.manager [-] [instance: e3360f00-a42a-40f2-9e6b-813547ebfd2b] Deallocating network for instance {{(pid=61806) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1532.403519] env[61806]: DEBUG nova.network.neutron [-] [instance: e3360f00-a42a-40f2-9e6b-813547ebfd2b] deallocate_for_instance() {{(pid=61806) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1532.854202] env[61806]: DEBUG nova.compute.manager [req-fd613287-9ee3-4395-a2b9-f21e8110a3cb req-e2da1dd4-6c44-4909-bcfd-5e587293035e service nova] [instance: e3360f00-a42a-40f2-9e6b-813547ebfd2b] Received event network-vif-deleted-3c8061f1-58c9-49f2-bdc9-e43e9936373e {{(pid=61806) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1532.854434] env[61806]: INFO nova.compute.manager [req-fd613287-9ee3-4395-a2b9-f21e8110a3cb req-e2da1dd4-6c44-4909-bcfd-5e587293035e service nova] [instance: e3360f00-a42a-40f2-9e6b-813547ebfd2b] Neutron deleted interface 3c8061f1-58c9-49f2-bdc9-e43e9936373e; detaching it from the instance and deleting it from the info cache [ 1532.855169] env[61806]: DEBUG nova.network.neutron [req-fd613287-9ee3-4395-a2b9-f21e8110a3cb req-e2da1dd4-6c44-4909-bcfd-5e587293035e service nova] [instance: e3360f00-a42a-40f2-9e6b-813547ebfd2b] Updating instance_info_cache with network_info: [] {{(pid=61806) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1533.299255] env[61806]: DEBUG nova.network.neutron [-] [instance: e3360f00-a42a-40f2-9e6b-813547ebfd2b] Updating instance_info_cache with network_info: [] {{(pid=61806) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1533.357279] env[61806]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-d3e5b542-0b8e-4aeb-a35f-b47d54807c3f {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1533.366730] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5a00538c-271b-4560-a066-b9943e33094c {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1533.389299] env[61806]: DEBUG nova.compute.manager [req-fd613287-9ee3-4395-a2b9-f21e8110a3cb req-e2da1dd4-6c44-4909-bcfd-5e587293035e service nova] [instance: e3360f00-a42a-40f2-9e6b-813547ebfd2b] Detach interface failed, port_id=3c8061f1-58c9-49f2-bdc9-e43e9936373e, reason: Instance e3360f00-a42a-40f2-9e6b-813547ebfd2b could not be found. {{(pid=61806) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10963}} [ 1533.803489] env[61806]: INFO nova.compute.manager [-] [instance: e3360f00-a42a-40f2-9e6b-813547ebfd2b] Took 1.40 seconds to deallocate network for instance. [ 1534.310229] env[61806]: DEBUG oslo_concurrency.lockutils [None req-ae9b29d4-744a-4bb4-8a0e-9ca5ba248a92 tempest-AttachVolumeTestJSON-132925025 tempest-AttachVolumeTestJSON-132925025-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1534.310530] env[61806]: DEBUG oslo_concurrency.lockutils [None req-ae9b29d4-744a-4bb4-8a0e-9ca5ba248a92 tempest-AttachVolumeTestJSON-132925025 tempest-AttachVolumeTestJSON-132925025-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.001s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1534.310761] env[61806]: DEBUG nova.objects.instance [None req-ae9b29d4-744a-4bb4-8a0e-9ca5ba248a92 tempest-AttachVolumeTestJSON-132925025 tempest-AttachVolumeTestJSON-132925025-project-member] Lazy-loading 'resources' on Instance uuid e3360f00-a42a-40f2-9e6b-813547ebfd2b {{(pid=61806) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1534.849414] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-43c0229e-606f-4d99-8446-de98034fc165 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1534.856943] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-44fdb423-9e87-4183-8a4c-8791a36b8588 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1534.885724] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a165fd7f-b292-4253-826f-7b8614bf167a {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1534.892900] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-236e1725-925f-4bc6-a2f7-487985e5e2e1 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1534.905694] env[61806]: DEBUG nova.compute.provider_tree [None req-ae9b29d4-744a-4bb4-8a0e-9ca5ba248a92 tempest-AttachVolumeTestJSON-132925025 tempest-AttachVolumeTestJSON-132925025-project-member] Updating inventory in ProviderTree for provider a2858be1-fd22-4e08-979e-87ad25293407 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 139, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61806) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1535.436107] env[61806]: DEBUG nova.scheduler.client.report [None req-ae9b29d4-744a-4bb4-8a0e-9ca5ba248a92 tempest-AttachVolumeTestJSON-132925025 tempest-AttachVolumeTestJSON-132925025-project-member] Updated inventory for provider a2858be1-fd22-4e08-979e-87ad25293407 with generation 160 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 139, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61806) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:971}} [ 1535.436397] env[61806]: DEBUG nova.compute.provider_tree [None req-ae9b29d4-744a-4bb4-8a0e-9ca5ba248a92 tempest-AttachVolumeTestJSON-132925025 tempest-AttachVolumeTestJSON-132925025-project-member] Updating resource provider a2858be1-fd22-4e08-979e-87ad25293407 generation from 160 to 161 during operation: update_inventory {{(pid=61806) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 1535.436585] env[61806]: DEBUG nova.compute.provider_tree [None req-ae9b29d4-744a-4bb4-8a0e-9ca5ba248a92 tempest-AttachVolumeTestJSON-132925025 tempest-AttachVolumeTestJSON-132925025-project-member] Updating inventory in ProviderTree for provider a2858be1-fd22-4e08-979e-87ad25293407 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 139, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61806) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1535.941684] env[61806]: DEBUG oslo_concurrency.lockutils [None req-ae9b29d4-744a-4bb4-8a0e-9ca5ba248a92 tempest-AttachVolumeTestJSON-132925025 tempest-AttachVolumeTestJSON-132925025-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.631s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1535.964851] env[61806]: INFO nova.scheduler.client.report [None req-ae9b29d4-744a-4bb4-8a0e-9ca5ba248a92 tempest-AttachVolumeTestJSON-132925025 tempest-AttachVolumeTestJSON-132925025-project-member] Deleted allocations for instance e3360f00-a42a-40f2-9e6b-813547ebfd2b [ 1536.474184] env[61806]: DEBUG oslo_concurrency.lockutils [None req-ae9b29d4-744a-4bb4-8a0e-9ca5ba248a92 tempest-AttachVolumeTestJSON-132925025 tempest-AttachVolumeTestJSON-132925025-project-member] Lock "e3360f00-a42a-40f2-9e6b-813547ebfd2b" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.181s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1538.122770] env[61806]: DEBUG oslo_concurrency.lockutils [None req-4c79c0a2-ac11-403f-8d39-52af9c1b1bfb tempest-AttachVolumeTestJSON-132925025 tempest-AttachVolumeTestJSON-132925025-project-member] Acquiring lock "ef96151b-c908-4108-92ca-8a7868312436" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1538.123075] env[61806]: DEBUG oslo_concurrency.lockutils [None req-4c79c0a2-ac11-403f-8d39-52af9c1b1bfb tempest-AttachVolumeTestJSON-132925025 tempest-AttachVolumeTestJSON-132925025-project-member] Lock "ef96151b-c908-4108-92ca-8a7868312436" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1538.625864] env[61806]: DEBUG nova.compute.manager [None req-4c79c0a2-ac11-403f-8d39-52af9c1b1bfb tempest-AttachVolumeTestJSON-132925025 tempest-AttachVolumeTestJSON-132925025-project-member] [instance: ef96151b-c908-4108-92ca-8a7868312436] Starting instance... {{(pid=61806) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1539.148898] env[61806]: DEBUG oslo_concurrency.lockutils [None req-4c79c0a2-ac11-403f-8d39-52af9c1b1bfb tempest-AttachVolumeTestJSON-132925025 tempest-AttachVolumeTestJSON-132925025-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1539.149226] env[61806]: DEBUG oslo_concurrency.lockutils [None req-4c79c0a2-ac11-403f-8d39-52af9c1b1bfb tempest-AttachVolumeTestJSON-132925025 tempest-AttachVolumeTestJSON-132925025-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.000s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1539.150797] env[61806]: INFO nova.compute.claims [None req-4c79c0a2-ac11-403f-8d39-52af9c1b1bfb tempest-AttachVolumeTestJSON-132925025 tempest-AttachVolumeTestJSON-132925025-project-member] [instance: ef96151b-c908-4108-92ca-8a7868312436] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1540.187602] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4fcfb856-ed14-4e6c-bf31-6111977da6c2 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1540.195043] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c9c0c8b9-8f9e-433c-8383-d51069a0902b {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1540.225440] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e823ecae-e2fa-4714-9ba4-68d35a0b962f {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1540.231999] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2bdb59c6-9b7f-4d7b-9b4e-67a9ca0b3b84 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1540.244767] env[61806]: DEBUG nova.compute.provider_tree [None req-4c79c0a2-ac11-403f-8d39-52af9c1b1bfb tempest-AttachVolumeTestJSON-132925025 tempest-AttachVolumeTestJSON-132925025-project-member] Inventory has not changed in ProviderTree for provider: a2858be1-fd22-4e08-979e-87ad25293407 {{(pid=61806) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1540.747462] env[61806]: DEBUG nova.scheduler.client.report [None req-4c79c0a2-ac11-403f-8d39-52af9c1b1bfb tempest-AttachVolumeTestJSON-132925025 tempest-AttachVolumeTestJSON-132925025-project-member] Inventory has not changed for provider a2858be1-fd22-4e08-979e-87ad25293407 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 139, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61806) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1541.252344] env[61806]: DEBUG oslo_concurrency.lockutils [None req-4c79c0a2-ac11-403f-8d39-52af9c1b1bfb tempest-AttachVolumeTestJSON-132925025 tempest-AttachVolumeTestJSON-132925025-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.103s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1541.252881] env[61806]: DEBUG nova.compute.manager [None req-4c79c0a2-ac11-403f-8d39-52af9c1b1bfb tempest-AttachVolumeTestJSON-132925025 tempest-AttachVolumeTestJSON-132925025-project-member] [instance: ef96151b-c908-4108-92ca-8a7868312436] Start building networks asynchronously for instance. {{(pid=61806) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1541.758313] env[61806]: DEBUG nova.compute.utils [None req-4c79c0a2-ac11-403f-8d39-52af9c1b1bfb tempest-AttachVolumeTestJSON-132925025 tempest-AttachVolumeTestJSON-132925025-project-member] Using /dev/sd instead of None {{(pid=61806) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1541.759773] env[61806]: DEBUG nova.compute.manager [None req-4c79c0a2-ac11-403f-8d39-52af9c1b1bfb tempest-AttachVolumeTestJSON-132925025 tempest-AttachVolumeTestJSON-132925025-project-member] [instance: ef96151b-c908-4108-92ca-8a7868312436] Allocating IP information in the background. {{(pid=61806) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1541.759960] env[61806]: DEBUG nova.network.neutron [None req-4c79c0a2-ac11-403f-8d39-52af9c1b1bfb tempest-AttachVolumeTestJSON-132925025 tempest-AttachVolumeTestJSON-132925025-project-member] [instance: ef96151b-c908-4108-92ca-8a7868312436] allocate_for_instance() {{(pid=61806) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 1541.807682] env[61806]: DEBUG nova.policy [None req-4c79c0a2-ac11-403f-8d39-52af9c1b1bfb tempest-AttachVolumeTestJSON-132925025 tempest-AttachVolumeTestJSON-132925025-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '4ac9f7259f12427ba22b791d5d190943', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '833defcd470d46ceacd7886d1f4fb0cf', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61806) authorize /opt/stack/nova/nova/policy.py:201}} [ 1542.057937] env[61806]: DEBUG nova.network.neutron [None req-4c79c0a2-ac11-403f-8d39-52af9c1b1bfb tempest-AttachVolumeTestJSON-132925025 tempest-AttachVolumeTestJSON-132925025-project-member] [instance: ef96151b-c908-4108-92ca-8a7868312436] Successfully created port: 8092e872-d167-4a57-8986-f823cb3d99b3 {{(pid=61806) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1542.263491] env[61806]: DEBUG nova.compute.manager [None req-4c79c0a2-ac11-403f-8d39-52af9c1b1bfb tempest-AttachVolumeTestJSON-132925025 tempest-AttachVolumeTestJSON-132925025-project-member] [instance: ef96151b-c908-4108-92ca-8a7868312436] Start building block device mappings for instance. {{(pid=61806) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1543.273695] env[61806]: DEBUG nova.compute.manager [None req-4c79c0a2-ac11-403f-8d39-52af9c1b1bfb tempest-AttachVolumeTestJSON-132925025 tempest-AttachVolumeTestJSON-132925025-project-member] [instance: ef96151b-c908-4108-92ca-8a7868312436] Start spawning the instance on the hypervisor. {{(pid=61806) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1543.298924] env[61806]: DEBUG nova.virt.hardware [None req-4c79c0a2-ac11-403f-8d39-52af9c1b1bfb tempest-AttachVolumeTestJSON-132925025 tempest-AttachVolumeTestJSON-132925025-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-15T18:05:10Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-15T18:04:52Z,direct_url=,disk_format='vmdk',id=ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='84ba83fce2c349a6988173c5d6fc3b07',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-15T18:04:53Z,virtual_size=,visibility=), allow threads: False {{(pid=61806) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1543.299238] env[61806]: DEBUG nova.virt.hardware [None req-4c79c0a2-ac11-403f-8d39-52af9c1b1bfb tempest-AttachVolumeTestJSON-132925025 tempest-AttachVolumeTestJSON-132925025-project-member] Flavor limits 0:0:0 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1543.299427] env[61806]: DEBUG nova.virt.hardware [None req-4c79c0a2-ac11-403f-8d39-52af9c1b1bfb tempest-AttachVolumeTestJSON-132925025 tempest-AttachVolumeTestJSON-132925025-project-member] Image limits 0:0:0 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1543.299955] env[61806]: DEBUG nova.virt.hardware [None req-4c79c0a2-ac11-403f-8d39-52af9c1b1bfb tempest-AttachVolumeTestJSON-132925025 tempest-AttachVolumeTestJSON-132925025-project-member] Flavor pref 0:0:0 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1543.299955] env[61806]: DEBUG nova.virt.hardware [None req-4c79c0a2-ac11-403f-8d39-52af9c1b1bfb tempest-AttachVolumeTestJSON-132925025 tempest-AttachVolumeTestJSON-132925025-project-member] Image pref 0:0:0 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1543.299955] env[61806]: DEBUG nova.virt.hardware [None req-4c79c0a2-ac11-403f-8d39-52af9c1b1bfb tempest-AttachVolumeTestJSON-132925025 tempest-AttachVolumeTestJSON-132925025-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61806) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1543.300197] env[61806]: DEBUG nova.virt.hardware [None req-4c79c0a2-ac11-403f-8d39-52af9c1b1bfb tempest-AttachVolumeTestJSON-132925025 tempest-AttachVolumeTestJSON-132925025-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61806) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1543.300303] env[61806]: DEBUG nova.virt.hardware [None req-4c79c0a2-ac11-403f-8d39-52af9c1b1bfb tempest-AttachVolumeTestJSON-132925025 tempest-AttachVolumeTestJSON-132925025-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61806) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1543.300483] env[61806]: DEBUG nova.virt.hardware [None req-4c79c0a2-ac11-403f-8d39-52af9c1b1bfb tempest-AttachVolumeTestJSON-132925025 tempest-AttachVolumeTestJSON-132925025-project-member] Got 1 possible topologies {{(pid=61806) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1543.300654] env[61806]: DEBUG nova.virt.hardware [None req-4c79c0a2-ac11-403f-8d39-52af9c1b1bfb tempest-AttachVolumeTestJSON-132925025 tempest-AttachVolumeTestJSON-132925025-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61806) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1543.300837] env[61806]: DEBUG nova.virt.hardware [None req-4c79c0a2-ac11-403f-8d39-52af9c1b1bfb tempest-AttachVolumeTestJSON-132925025 tempest-AttachVolumeTestJSON-132925025-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61806) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1543.301721] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fe01756b-64ae-4384-9cbe-1e7e135eb70a {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1543.309279] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b17334f8-45fb-4963-be20-4d985afc0d2e {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1543.425286] env[61806]: DEBUG nova.compute.manager [req-c289d902-2d36-45b1-a961-7046b65e91cd req-e242653e-27f2-41c2-b239-c76405875636 service nova] [instance: ef96151b-c908-4108-92ca-8a7868312436] Received event network-vif-plugged-8092e872-d167-4a57-8986-f823cb3d99b3 {{(pid=61806) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1543.425662] env[61806]: DEBUG oslo_concurrency.lockutils [req-c289d902-2d36-45b1-a961-7046b65e91cd req-e242653e-27f2-41c2-b239-c76405875636 service nova] Acquiring lock "ef96151b-c908-4108-92ca-8a7868312436-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1543.425822] env[61806]: DEBUG oslo_concurrency.lockutils [req-c289d902-2d36-45b1-a961-7046b65e91cd req-e242653e-27f2-41c2-b239-c76405875636 service nova] Lock "ef96151b-c908-4108-92ca-8a7868312436-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1543.426195] env[61806]: DEBUG oslo_concurrency.lockutils [req-c289d902-2d36-45b1-a961-7046b65e91cd req-e242653e-27f2-41c2-b239-c76405875636 service nova] Lock "ef96151b-c908-4108-92ca-8a7868312436-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1543.426442] env[61806]: DEBUG nova.compute.manager [req-c289d902-2d36-45b1-a961-7046b65e91cd req-e242653e-27f2-41c2-b239-c76405875636 service nova] [instance: ef96151b-c908-4108-92ca-8a7868312436] No waiting events found dispatching network-vif-plugged-8092e872-d167-4a57-8986-f823cb3d99b3 {{(pid=61806) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1543.426678] env[61806]: WARNING nova.compute.manager [req-c289d902-2d36-45b1-a961-7046b65e91cd req-e242653e-27f2-41c2-b239-c76405875636 service nova] [instance: ef96151b-c908-4108-92ca-8a7868312436] Received unexpected event network-vif-plugged-8092e872-d167-4a57-8986-f823cb3d99b3 for instance with vm_state building and task_state spawning. [ 1543.508080] env[61806]: DEBUG nova.network.neutron [None req-4c79c0a2-ac11-403f-8d39-52af9c1b1bfb tempest-AttachVolumeTestJSON-132925025 tempest-AttachVolumeTestJSON-132925025-project-member] [instance: ef96151b-c908-4108-92ca-8a7868312436] Successfully updated port: 8092e872-d167-4a57-8986-f823cb3d99b3 {{(pid=61806) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1544.010921] env[61806]: DEBUG oslo_concurrency.lockutils [None req-4c79c0a2-ac11-403f-8d39-52af9c1b1bfb tempest-AttachVolumeTestJSON-132925025 tempest-AttachVolumeTestJSON-132925025-project-member] Acquiring lock "refresh_cache-ef96151b-c908-4108-92ca-8a7868312436" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1544.011184] env[61806]: DEBUG oslo_concurrency.lockutils [None req-4c79c0a2-ac11-403f-8d39-52af9c1b1bfb tempest-AttachVolumeTestJSON-132925025 tempest-AttachVolumeTestJSON-132925025-project-member] Acquired lock "refresh_cache-ef96151b-c908-4108-92ca-8a7868312436" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1544.011354] env[61806]: DEBUG nova.network.neutron [None req-4c79c0a2-ac11-403f-8d39-52af9c1b1bfb tempest-AttachVolumeTestJSON-132925025 tempest-AttachVolumeTestJSON-132925025-project-member] [instance: ef96151b-c908-4108-92ca-8a7868312436] Building network info cache for instance {{(pid=61806) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1544.542622] env[61806]: DEBUG nova.network.neutron [None req-4c79c0a2-ac11-403f-8d39-52af9c1b1bfb tempest-AttachVolumeTestJSON-132925025 tempest-AttachVolumeTestJSON-132925025-project-member] [instance: ef96151b-c908-4108-92ca-8a7868312436] Instance cache missing network info. {{(pid=61806) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 1544.657848] env[61806]: DEBUG nova.network.neutron [None req-4c79c0a2-ac11-403f-8d39-52af9c1b1bfb tempest-AttachVolumeTestJSON-132925025 tempest-AttachVolumeTestJSON-132925025-project-member] [instance: ef96151b-c908-4108-92ca-8a7868312436] Updating instance_info_cache with network_info: [{"id": "8092e872-d167-4a57-8986-f823cb3d99b3", "address": "fa:16:3e:03:83:c0", "network": {"id": "0fd518d3-c254-4dba-a147-378c3c029c51", "bridge": "br-int", "label": "tempest-AttachVolumeTestJSON-538202287-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "833defcd470d46ceacd7886d1f4fb0cf", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "91b0f7e5-0d1a-46e2-bf73-09656211dea2", "external-id": "nsx-vlan-transportzone-488", "segmentation_id": 488, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap8092e872-d1", "ovs_interfaceid": "8092e872-d167-4a57-8986-f823cb3d99b3", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61806) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1545.160794] env[61806]: DEBUG oslo_concurrency.lockutils [None req-4c79c0a2-ac11-403f-8d39-52af9c1b1bfb tempest-AttachVolumeTestJSON-132925025 tempest-AttachVolumeTestJSON-132925025-project-member] Releasing lock "refresh_cache-ef96151b-c908-4108-92ca-8a7868312436" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1545.161193] env[61806]: DEBUG nova.compute.manager [None req-4c79c0a2-ac11-403f-8d39-52af9c1b1bfb tempest-AttachVolumeTestJSON-132925025 tempest-AttachVolumeTestJSON-132925025-project-member] [instance: ef96151b-c908-4108-92ca-8a7868312436] Instance network_info: |[{"id": "8092e872-d167-4a57-8986-f823cb3d99b3", "address": "fa:16:3e:03:83:c0", "network": {"id": "0fd518d3-c254-4dba-a147-378c3c029c51", "bridge": "br-int", "label": "tempest-AttachVolumeTestJSON-538202287-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "833defcd470d46ceacd7886d1f4fb0cf", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "91b0f7e5-0d1a-46e2-bf73-09656211dea2", "external-id": "nsx-vlan-transportzone-488", "segmentation_id": 488, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap8092e872-d1", "ovs_interfaceid": "8092e872-d167-4a57-8986-f823cb3d99b3", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61806) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1545.161686] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-4c79c0a2-ac11-403f-8d39-52af9c1b1bfb tempest-AttachVolumeTestJSON-132925025 tempest-AttachVolumeTestJSON-132925025-project-member] [instance: ef96151b-c908-4108-92ca-8a7868312436] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:03:83:c0', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '91b0f7e5-0d1a-46e2-bf73-09656211dea2', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '8092e872-d167-4a57-8986-f823cb3d99b3', 'vif_model': 'vmxnet3'}] {{(pid=61806) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1545.170057] env[61806]: DEBUG oslo.service.loopingcall [None req-4c79c0a2-ac11-403f-8d39-52af9c1b1bfb tempest-AttachVolumeTestJSON-132925025 tempest-AttachVolumeTestJSON-132925025-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61806) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1545.170279] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: ef96151b-c908-4108-92ca-8a7868312436] Creating VM on the ESX host {{(pid=61806) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 1545.170508] env[61806]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-a7132c96-5ad0-48d4-ac3e-9d61bb102cac {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1545.192373] env[61806]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1545.192373] env[61806]: value = "task-1295390" [ 1545.192373] env[61806]: _type = "Task" [ 1545.192373] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1545.201715] env[61806]: DEBUG oslo_vmware.api [-] Task: {'id': task-1295390, 'name': CreateVM_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1545.454281] env[61806]: DEBUG nova.compute.manager [req-f74fd3fd-d107-4599-9418-3a12cc25e53c req-20e5bfa3-af2a-45aa-8c3d-7dd86fd4c560 service nova] [instance: ef96151b-c908-4108-92ca-8a7868312436] Received event network-changed-8092e872-d167-4a57-8986-f823cb3d99b3 {{(pid=61806) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1545.454281] env[61806]: DEBUG nova.compute.manager [req-f74fd3fd-d107-4599-9418-3a12cc25e53c req-20e5bfa3-af2a-45aa-8c3d-7dd86fd4c560 service nova] [instance: ef96151b-c908-4108-92ca-8a7868312436] Refreshing instance network info cache due to event network-changed-8092e872-d167-4a57-8986-f823cb3d99b3. {{(pid=61806) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 1545.454475] env[61806]: DEBUG oslo_concurrency.lockutils [req-f74fd3fd-d107-4599-9418-3a12cc25e53c req-20e5bfa3-af2a-45aa-8c3d-7dd86fd4c560 service nova] Acquiring lock "refresh_cache-ef96151b-c908-4108-92ca-8a7868312436" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1545.454588] env[61806]: DEBUG oslo_concurrency.lockutils [req-f74fd3fd-d107-4599-9418-3a12cc25e53c req-20e5bfa3-af2a-45aa-8c3d-7dd86fd4c560 service nova] Acquired lock "refresh_cache-ef96151b-c908-4108-92ca-8a7868312436" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1545.454773] env[61806]: DEBUG nova.network.neutron [req-f74fd3fd-d107-4599-9418-3a12cc25e53c req-20e5bfa3-af2a-45aa-8c3d-7dd86fd4c560 service nova] [instance: ef96151b-c908-4108-92ca-8a7868312436] Refreshing network info cache for port 8092e872-d167-4a57-8986-f823cb3d99b3 {{(pid=61806) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1545.702710] env[61806]: DEBUG oslo_vmware.api [-] Task: {'id': task-1295390, 'name': CreateVM_Task} progress is 25%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1546.132094] env[61806]: DEBUG nova.network.neutron [req-f74fd3fd-d107-4599-9418-3a12cc25e53c req-20e5bfa3-af2a-45aa-8c3d-7dd86fd4c560 service nova] [instance: ef96151b-c908-4108-92ca-8a7868312436] Updated VIF entry in instance network info cache for port 8092e872-d167-4a57-8986-f823cb3d99b3. {{(pid=61806) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1546.132517] env[61806]: DEBUG nova.network.neutron [req-f74fd3fd-d107-4599-9418-3a12cc25e53c req-20e5bfa3-af2a-45aa-8c3d-7dd86fd4c560 service nova] [instance: ef96151b-c908-4108-92ca-8a7868312436] Updating instance_info_cache with network_info: [{"id": "8092e872-d167-4a57-8986-f823cb3d99b3", "address": "fa:16:3e:03:83:c0", "network": {"id": "0fd518d3-c254-4dba-a147-378c3c029c51", "bridge": "br-int", "label": "tempest-AttachVolumeTestJSON-538202287-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "833defcd470d46ceacd7886d1f4fb0cf", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "91b0f7e5-0d1a-46e2-bf73-09656211dea2", "external-id": "nsx-vlan-transportzone-488", "segmentation_id": 488, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap8092e872-d1", "ovs_interfaceid": "8092e872-d167-4a57-8986-f823cb3d99b3", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61806) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1546.203169] env[61806]: DEBUG oslo_vmware.api [-] Task: {'id': task-1295390, 'name': CreateVM_Task, 'duration_secs': 0.768736} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1546.203365] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: ef96151b-c908-4108-92ca-8a7868312436] Created VM on the ESX host {{(pid=61806) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 1546.204057] env[61806]: DEBUG oslo_concurrency.lockutils [None req-4c79c0a2-ac11-403f-8d39-52af9c1b1bfb tempest-AttachVolumeTestJSON-132925025 tempest-AttachVolumeTestJSON-132925025-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1546.204247] env[61806]: DEBUG oslo_concurrency.lockutils [None req-4c79c0a2-ac11-403f-8d39-52af9c1b1bfb tempest-AttachVolumeTestJSON-132925025 tempest-AttachVolumeTestJSON-132925025-project-member] Acquired lock "[datastore1] devstack-image-cache_base/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1546.204588] env[61806]: DEBUG oslo_concurrency.lockutils [None req-4c79c0a2-ac11-403f-8d39-52af9c1b1bfb tempest-AttachVolumeTestJSON-132925025 tempest-AttachVolumeTestJSON-132925025-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1546.204847] env[61806]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-cb7dcd34-1bb4-488e-8a48-e777867f8160 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1546.209256] env[61806]: DEBUG oslo_vmware.api [None req-4c79c0a2-ac11-403f-8d39-52af9c1b1bfb tempest-AttachVolumeTestJSON-132925025 tempest-AttachVolumeTestJSON-132925025-project-member] Waiting for the task: (returnval){ [ 1546.209256] env[61806]: value = "session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]5293ac4d-9e2c-6ced-6b9a-99fa0ac96822" [ 1546.209256] env[61806]: _type = "Task" [ 1546.209256] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1546.216108] env[61806]: DEBUG oslo_vmware.api [None req-4c79c0a2-ac11-403f-8d39-52af9c1b1bfb tempest-AttachVolumeTestJSON-132925025 tempest-AttachVolumeTestJSON-132925025-project-member] Task: {'id': session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]5293ac4d-9e2c-6ced-6b9a-99fa0ac96822, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1546.634844] env[61806]: DEBUG oslo_concurrency.lockutils [req-f74fd3fd-d107-4599-9418-3a12cc25e53c req-20e5bfa3-af2a-45aa-8c3d-7dd86fd4c560 service nova] Releasing lock "refresh_cache-ef96151b-c908-4108-92ca-8a7868312436" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1546.719228] env[61806]: DEBUG oslo_vmware.api [None req-4c79c0a2-ac11-403f-8d39-52af9c1b1bfb tempest-AttachVolumeTestJSON-132925025 tempest-AttachVolumeTestJSON-132925025-project-member] Task: {'id': session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]5293ac4d-9e2c-6ced-6b9a-99fa0ac96822, 'name': SearchDatastore_Task, 'duration_secs': 0.013571} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1546.719569] env[61806]: DEBUG oslo_concurrency.lockutils [None req-4c79c0a2-ac11-403f-8d39-52af9c1b1bfb tempest-AttachVolumeTestJSON-132925025 tempest-AttachVolumeTestJSON-132925025-project-member] Releasing lock "[datastore1] devstack-image-cache_base/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1546.719788] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-4c79c0a2-ac11-403f-8d39-52af9c1b1bfb tempest-AttachVolumeTestJSON-132925025 tempest-AttachVolumeTestJSON-132925025-project-member] [instance: ef96151b-c908-4108-92ca-8a7868312436] Processing image ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa {{(pid=61806) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1546.720034] env[61806]: DEBUG oslo_concurrency.lockutils [None req-4c79c0a2-ac11-403f-8d39-52af9c1b1bfb tempest-AttachVolumeTestJSON-132925025 tempest-AttachVolumeTestJSON-132925025-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa.vmdk" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1546.720198] env[61806]: DEBUG oslo_concurrency.lockutils [None req-4c79c0a2-ac11-403f-8d39-52af9c1b1bfb tempest-AttachVolumeTestJSON-132925025 tempest-AttachVolumeTestJSON-132925025-project-member] Acquired lock "[datastore1] devstack-image-cache_base/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa.vmdk" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1546.720388] env[61806]: DEBUG nova.virt.vmwareapi.ds_util [None req-4c79c0a2-ac11-403f-8d39-52af9c1b1bfb tempest-AttachVolumeTestJSON-132925025 tempest-AttachVolumeTestJSON-132925025-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61806) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1546.720648] env[61806]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-5946d873-907a-490d-adc0-1bbb9ea06869 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1546.728221] env[61806]: DEBUG nova.virt.vmwareapi.ds_util [None req-4c79c0a2-ac11-403f-8d39-52af9c1b1bfb tempest-AttachVolumeTestJSON-132925025 tempest-AttachVolumeTestJSON-132925025-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61806) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1546.728398] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-4c79c0a2-ac11-403f-8d39-52af9c1b1bfb tempest-AttachVolumeTestJSON-132925025 tempest-AttachVolumeTestJSON-132925025-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61806) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 1546.729077] env[61806]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-542f1441-8383-49bf-ac9f-0007bb2f3650 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1546.733418] env[61806]: DEBUG oslo_vmware.api [None req-4c79c0a2-ac11-403f-8d39-52af9c1b1bfb tempest-AttachVolumeTestJSON-132925025 tempest-AttachVolumeTestJSON-132925025-project-member] Waiting for the task: (returnval){ [ 1546.733418] env[61806]: value = "session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]520fa321-1bf2-67a2-307a-0e351e600d19" [ 1546.733418] env[61806]: _type = "Task" [ 1546.733418] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1546.740123] env[61806]: DEBUG oslo_vmware.api [None req-4c79c0a2-ac11-403f-8d39-52af9c1b1bfb tempest-AttachVolumeTestJSON-132925025 tempest-AttachVolumeTestJSON-132925025-project-member] Task: {'id': session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]520fa321-1bf2-67a2-307a-0e351e600d19, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1547.244958] env[61806]: DEBUG oslo_vmware.api [None req-4c79c0a2-ac11-403f-8d39-52af9c1b1bfb tempest-AttachVolumeTestJSON-132925025 tempest-AttachVolumeTestJSON-132925025-project-member] Task: {'id': session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]520fa321-1bf2-67a2-307a-0e351e600d19, 'name': SearchDatastore_Task, 'duration_secs': 0.007876} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1547.245726] env[61806]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-39bd848d-7ce5-4aeb-b9c3-b2fe6dfe971d {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1547.250768] env[61806]: DEBUG oslo_vmware.api [None req-4c79c0a2-ac11-403f-8d39-52af9c1b1bfb tempest-AttachVolumeTestJSON-132925025 tempest-AttachVolumeTestJSON-132925025-project-member] Waiting for the task: (returnval){ [ 1547.250768] env[61806]: value = "session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]521cc52c-f069-675e-df7e-228bd45d0a64" [ 1547.250768] env[61806]: _type = "Task" [ 1547.250768] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1547.257721] env[61806]: DEBUG oslo_vmware.api [None req-4c79c0a2-ac11-403f-8d39-52af9c1b1bfb tempest-AttachVolumeTestJSON-132925025 tempest-AttachVolumeTestJSON-132925025-project-member] Task: {'id': session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]521cc52c-f069-675e-df7e-228bd45d0a64, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1547.760572] env[61806]: DEBUG oslo_vmware.api [None req-4c79c0a2-ac11-403f-8d39-52af9c1b1bfb tempest-AttachVolumeTestJSON-132925025 tempest-AttachVolumeTestJSON-132925025-project-member] Task: {'id': session[520ec6a8-ed8f-4556-5484-7e5ecf52b2b3]521cc52c-f069-675e-df7e-228bd45d0a64, 'name': SearchDatastore_Task, 'duration_secs': 0.00904} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1547.760957] env[61806]: DEBUG oslo_concurrency.lockutils [None req-4c79c0a2-ac11-403f-8d39-52af9c1b1bfb tempest-AttachVolumeTestJSON-132925025 tempest-AttachVolumeTestJSON-132925025-project-member] Releasing lock "[datastore1] devstack-image-cache_base/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa.vmdk" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1547.761116] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [None req-4c79c0a2-ac11-403f-8d39-52af9c1b1bfb tempest-AttachVolumeTestJSON-132925025 tempest-AttachVolumeTestJSON-132925025-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa.vmdk to [datastore1] ef96151b-c908-4108-92ca-8a7868312436/ef96151b-c908-4108-92ca-8a7868312436.vmdk {{(pid=61806) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 1547.762579] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-6da58863-38d3-41c3-bad5-9fe3de789d8c {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1547.767820] env[61806]: DEBUG oslo_vmware.api [None req-4c79c0a2-ac11-403f-8d39-52af9c1b1bfb tempest-AttachVolumeTestJSON-132925025 tempest-AttachVolumeTestJSON-132925025-project-member] Waiting for the task: (returnval){ [ 1547.767820] env[61806]: value = "task-1295391" [ 1547.767820] env[61806]: _type = "Task" [ 1547.767820] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1547.775028] env[61806]: DEBUG oslo_vmware.api [None req-4c79c0a2-ac11-403f-8d39-52af9c1b1bfb tempest-AttachVolumeTestJSON-132925025 tempest-AttachVolumeTestJSON-132925025-project-member] Task: {'id': task-1295391, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1548.276951] env[61806]: DEBUG oslo_vmware.api [None req-4c79c0a2-ac11-403f-8d39-52af9c1b1bfb tempest-AttachVolumeTestJSON-132925025 tempest-AttachVolumeTestJSON-132925025-project-member] Task: {'id': task-1295391, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.473345} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1548.277255] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [None req-4c79c0a2-ac11-403f-8d39-52af9c1b1bfb tempest-AttachVolumeTestJSON-132925025 tempest-AttachVolumeTestJSON-132925025-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa/ed5bd0ca-ebdf-4db4-949c-b10c7341f9aa.vmdk to [datastore1] ef96151b-c908-4108-92ca-8a7868312436/ef96151b-c908-4108-92ca-8a7868312436.vmdk {{(pid=61806) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 1548.277456] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-4c79c0a2-ac11-403f-8d39-52af9c1b1bfb tempest-AttachVolumeTestJSON-132925025 tempest-AttachVolumeTestJSON-132925025-project-member] [instance: ef96151b-c908-4108-92ca-8a7868312436] Extending root virtual disk to 1048576 {{(pid=61806) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1548.277718] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-7bc3d320-ff51-4def-bb27-1442fb9e48fa {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1548.284158] env[61806]: DEBUG oslo_vmware.api [None req-4c79c0a2-ac11-403f-8d39-52af9c1b1bfb tempest-AttachVolumeTestJSON-132925025 tempest-AttachVolumeTestJSON-132925025-project-member] Waiting for the task: (returnval){ [ 1548.284158] env[61806]: value = "task-1295392" [ 1548.284158] env[61806]: _type = "Task" [ 1548.284158] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1548.291533] env[61806]: DEBUG oslo_vmware.api [None req-4c79c0a2-ac11-403f-8d39-52af9c1b1bfb tempest-AttachVolumeTestJSON-132925025 tempest-AttachVolumeTestJSON-132925025-project-member] Task: {'id': task-1295392, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1548.794255] env[61806]: DEBUG oslo_vmware.api [None req-4c79c0a2-ac11-403f-8d39-52af9c1b1bfb tempest-AttachVolumeTestJSON-132925025 tempest-AttachVolumeTestJSON-132925025-project-member] Task: {'id': task-1295392, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.07004} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1548.794603] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-4c79c0a2-ac11-403f-8d39-52af9c1b1bfb tempest-AttachVolumeTestJSON-132925025 tempest-AttachVolumeTestJSON-132925025-project-member] [instance: ef96151b-c908-4108-92ca-8a7868312436] Extended root virtual disk {{(pid=61806) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1548.795334] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6261114b-5b74-408c-9eee-ca383273637b {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1548.816286] env[61806]: DEBUG nova.virt.vmwareapi.volumeops [None req-4c79c0a2-ac11-403f-8d39-52af9c1b1bfb tempest-AttachVolumeTestJSON-132925025 tempest-AttachVolumeTestJSON-132925025-project-member] [instance: ef96151b-c908-4108-92ca-8a7868312436] Reconfiguring VM instance instance-00000070 to attach disk [datastore1] ef96151b-c908-4108-92ca-8a7868312436/ef96151b-c908-4108-92ca-8a7868312436.vmdk or device None with type sparse {{(pid=61806) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1548.816504] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-b0b45218-321d-4608-a624-6812f2562cbf {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1548.835535] env[61806]: DEBUG oslo_vmware.api [None req-4c79c0a2-ac11-403f-8d39-52af9c1b1bfb tempest-AttachVolumeTestJSON-132925025 tempest-AttachVolumeTestJSON-132925025-project-member] Waiting for the task: (returnval){ [ 1548.835535] env[61806]: value = "task-1295393" [ 1548.835535] env[61806]: _type = "Task" [ 1548.835535] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1548.842716] env[61806]: DEBUG oslo_vmware.api [None req-4c79c0a2-ac11-403f-8d39-52af9c1b1bfb tempest-AttachVolumeTestJSON-132925025 tempest-AttachVolumeTestJSON-132925025-project-member] Task: {'id': task-1295393, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1549.345140] env[61806]: DEBUG oslo_vmware.api [None req-4c79c0a2-ac11-403f-8d39-52af9c1b1bfb tempest-AttachVolumeTestJSON-132925025 tempest-AttachVolumeTestJSON-132925025-project-member] Task: {'id': task-1295393, 'name': ReconfigVM_Task, 'duration_secs': 0.278803} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1549.345431] env[61806]: DEBUG nova.virt.vmwareapi.volumeops [None req-4c79c0a2-ac11-403f-8d39-52af9c1b1bfb tempest-AttachVolumeTestJSON-132925025 tempest-AttachVolumeTestJSON-132925025-project-member] [instance: ef96151b-c908-4108-92ca-8a7868312436] Reconfigured VM instance instance-00000070 to attach disk [datastore1] ef96151b-c908-4108-92ca-8a7868312436/ef96151b-c908-4108-92ca-8a7868312436.vmdk or device None with type sparse {{(pid=61806) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1549.346062] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-12fff6ff-03b2-4091-8591-ea18b34061c5 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1549.353146] env[61806]: DEBUG oslo_vmware.api [None req-4c79c0a2-ac11-403f-8d39-52af9c1b1bfb tempest-AttachVolumeTestJSON-132925025 tempest-AttachVolumeTestJSON-132925025-project-member] Waiting for the task: (returnval){ [ 1549.353146] env[61806]: value = "task-1295394" [ 1549.353146] env[61806]: _type = "Task" [ 1549.353146] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1549.363750] env[61806]: DEBUG oslo_vmware.api [None req-4c79c0a2-ac11-403f-8d39-52af9c1b1bfb tempest-AttachVolumeTestJSON-132925025 tempest-AttachVolumeTestJSON-132925025-project-member] Task: {'id': task-1295394, 'name': Rename_Task} progress is 6%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1549.862952] env[61806]: DEBUG oslo_vmware.api [None req-4c79c0a2-ac11-403f-8d39-52af9c1b1bfb tempest-AttachVolumeTestJSON-132925025 tempest-AttachVolumeTestJSON-132925025-project-member] Task: {'id': task-1295394, 'name': Rename_Task, 'duration_secs': 0.151446} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1549.863342] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [None req-4c79c0a2-ac11-403f-8d39-52af9c1b1bfb tempest-AttachVolumeTestJSON-132925025 tempest-AttachVolumeTestJSON-132925025-project-member] [instance: ef96151b-c908-4108-92ca-8a7868312436] Powering on the VM {{(pid=61806) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1549.863525] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-57561c5e-2ad3-40e1-ab30-12541eaae130 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1549.869237] env[61806]: DEBUG oslo_vmware.api [None req-4c79c0a2-ac11-403f-8d39-52af9c1b1bfb tempest-AttachVolumeTestJSON-132925025 tempest-AttachVolumeTestJSON-132925025-project-member] Waiting for the task: (returnval){ [ 1549.869237] env[61806]: value = "task-1295395" [ 1549.869237] env[61806]: _type = "Task" [ 1549.869237] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1549.875945] env[61806]: DEBUG oslo_vmware.api [None req-4c79c0a2-ac11-403f-8d39-52af9c1b1bfb tempest-AttachVolumeTestJSON-132925025 tempest-AttachVolumeTestJSON-132925025-project-member] Task: {'id': task-1295395, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1550.379174] env[61806]: DEBUG oslo_vmware.api [None req-4c79c0a2-ac11-403f-8d39-52af9c1b1bfb tempest-AttachVolumeTestJSON-132925025 tempest-AttachVolumeTestJSON-132925025-project-member] Task: {'id': task-1295395, 'name': PowerOnVM_Task, 'duration_secs': 0.458785} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1550.379472] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [None req-4c79c0a2-ac11-403f-8d39-52af9c1b1bfb tempest-AttachVolumeTestJSON-132925025 tempest-AttachVolumeTestJSON-132925025-project-member] [instance: ef96151b-c908-4108-92ca-8a7868312436] Powered on the VM {{(pid=61806) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1550.379732] env[61806]: INFO nova.compute.manager [None req-4c79c0a2-ac11-403f-8d39-52af9c1b1bfb tempest-AttachVolumeTestJSON-132925025 tempest-AttachVolumeTestJSON-132925025-project-member] [instance: ef96151b-c908-4108-92ca-8a7868312436] Took 7.11 seconds to spawn the instance on the hypervisor. [ 1550.379932] env[61806]: DEBUG nova.compute.manager [None req-4c79c0a2-ac11-403f-8d39-52af9c1b1bfb tempest-AttachVolumeTestJSON-132925025 tempest-AttachVolumeTestJSON-132925025-project-member] [instance: ef96151b-c908-4108-92ca-8a7868312436] Checking state {{(pid=61806) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1550.380709] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-afd06a3a-fdef-44c7-8fc6-50f81f64cb4a {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1550.903971] env[61806]: INFO nova.compute.manager [None req-4c79c0a2-ac11-403f-8d39-52af9c1b1bfb tempest-AttachVolumeTestJSON-132925025 tempest-AttachVolumeTestJSON-132925025-project-member] [instance: ef96151b-c908-4108-92ca-8a7868312436] Took 11.77 seconds to build instance. [ 1551.069104] env[61806]: DEBUG nova.compute.manager [req-e2abb713-9ebb-43f4-90c8-42cc952a2985 req-67554753-4494-46e9-ad74-e412b5b925ae service nova] [instance: ef96151b-c908-4108-92ca-8a7868312436] Received event network-changed-8092e872-d167-4a57-8986-f823cb3d99b3 {{(pid=61806) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1551.069260] env[61806]: DEBUG nova.compute.manager [req-e2abb713-9ebb-43f4-90c8-42cc952a2985 req-67554753-4494-46e9-ad74-e412b5b925ae service nova] [instance: ef96151b-c908-4108-92ca-8a7868312436] Refreshing instance network info cache due to event network-changed-8092e872-d167-4a57-8986-f823cb3d99b3. {{(pid=61806) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 1551.069475] env[61806]: DEBUG oslo_concurrency.lockutils [req-e2abb713-9ebb-43f4-90c8-42cc952a2985 req-67554753-4494-46e9-ad74-e412b5b925ae service nova] Acquiring lock "refresh_cache-ef96151b-c908-4108-92ca-8a7868312436" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1551.069631] env[61806]: DEBUG oslo_concurrency.lockutils [req-e2abb713-9ebb-43f4-90c8-42cc952a2985 req-67554753-4494-46e9-ad74-e412b5b925ae service nova] Acquired lock "refresh_cache-ef96151b-c908-4108-92ca-8a7868312436" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1551.069797] env[61806]: DEBUG nova.network.neutron [req-e2abb713-9ebb-43f4-90c8-42cc952a2985 req-67554753-4494-46e9-ad74-e412b5b925ae service nova] [instance: ef96151b-c908-4108-92ca-8a7868312436] Refreshing network info cache for port 8092e872-d167-4a57-8986-f823cb3d99b3 {{(pid=61806) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1551.405598] env[61806]: DEBUG oslo_concurrency.lockutils [None req-4c79c0a2-ac11-403f-8d39-52af9c1b1bfb tempest-AttachVolumeTestJSON-132925025 tempest-AttachVolumeTestJSON-132925025-project-member] Lock "ef96151b-c908-4108-92ca-8a7868312436" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 13.282s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1551.792080] env[61806]: DEBUG nova.network.neutron [req-e2abb713-9ebb-43f4-90c8-42cc952a2985 req-67554753-4494-46e9-ad74-e412b5b925ae service nova] [instance: ef96151b-c908-4108-92ca-8a7868312436] Updated VIF entry in instance network info cache for port 8092e872-d167-4a57-8986-f823cb3d99b3. {{(pid=61806) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1551.792526] env[61806]: DEBUG nova.network.neutron [req-e2abb713-9ebb-43f4-90c8-42cc952a2985 req-67554753-4494-46e9-ad74-e412b5b925ae service nova] [instance: ef96151b-c908-4108-92ca-8a7868312436] Updating instance_info_cache with network_info: [{"id": "8092e872-d167-4a57-8986-f823cb3d99b3", "address": "fa:16:3e:03:83:c0", "network": {"id": "0fd518d3-c254-4dba-a147-378c3c029c51", "bridge": "br-int", "label": "tempest-AttachVolumeTestJSON-538202287-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.200", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "833defcd470d46ceacd7886d1f4fb0cf", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "91b0f7e5-0d1a-46e2-bf73-09656211dea2", "external-id": "nsx-vlan-transportzone-488", "segmentation_id": 488, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap8092e872-d1", "ovs_interfaceid": "8092e872-d167-4a57-8986-f823cb3d99b3", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61806) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1552.295355] env[61806]: DEBUG oslo_concurrency.lockutils [req-e2abb713-9ebb-43f4-90c8-42cc952a2985 req-67554753-4494-46e9-ad74-e412b5b925ae service nova] Releasing lock "refresh_cache-ef96151b-c908-4108-92ca-8a7868312436" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1554.018956] env[61806]: DEBUG oslo_service.periodic_task [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=61806) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1558.023618] env[61806]: DEBUG oslo_service.periodic_task [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=61806) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1558.023977] env[61806]: DEBUG nova.compute.manager [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Starting heal instance info cache {{(pid=61806) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9926}} [ 1558.023977] env[61806]: DEBUG nova.compute.manager [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Rebuilding the list of instances to heal {{(pid=61806) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9930}} [ 1558.555762] env[61806]: DEBUG oslo_concurrency.lockutils [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Acquiring lock "refresh_cache-ef96151b-c908-4108-92ca-8a7868312436" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1558.555928] env[61806]: DEBUG oslo_concurrency.lockutils [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Acquired lock "refresh_cache-ef96151b-c908-4108-92ca-8a7868312436" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1558.556420] env[61806]: DEBUG nova.network.neutron [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] [instance: ef96151b-c908-4108-92ca-8a7868312436] Forcefully refreshing network info cache for instance {{(pid=61806) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2004}} [ 1558.556420] env[61806]: DEBUG nova.objects.instance [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Lazy-loading 'info_cache' on Instance uuid ef96151b-c908-4108-92ca-8a7868312436 {{(pid=61806) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1560.282422] env[61806]: DEBUG nova.network.neutron [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] [instance: ef96151b-c908-4108-92ca-8a7868312436] Updating instance_info_cache with network_info: [{"id": "8092e872-d167-4a57-8986-f823cb3d99b3", "address": "fa:16:3e:03:83:c0", "network": {"id": "0fd518d3-c254-4dba-a147-378c3c029c51", "bridge": "br-int", "label": "tempest-AttachVolumeTestJSON-538202287-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.200", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "833defcd470d46ceacd7886d1f4fb0cf", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "91b0f7e5-0d1a-46e2-bf73-09656211dea2", "external-id": "nsx-vlan-transportzone-488", "segmentation_id": 488, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap8092e872-d1", "ovs_interfaceid": "8092e872-d167-4a57-8986-f823cb3d99b3", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61806) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1560.785508] env[61806]: DEBUG oslo_concurrency.lockutils [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Releasing lock "refresh_cache-ef96151b-c908-4108-92ca-8a7868312436" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1560.785720] env[61806]: DEBUG nova.compute.manager [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] [instance: ef96151b-c908-4108-92ca-8a7868312436] Updated the network info_cache for instance {{(pid=61806) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9997}} [ 1560.785930] env[61806]: DEBUG oslo_service.periodic_task [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=61806) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1560.786120] env[61806]: DEBUG oslo_service.periodic_task [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=61806) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1560.786288] env[61806]: DEBUG oslo_service.periodic_task [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Running periodic task ComputeManager.update_available_resource {{(pid=61806) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1561.289576] env[61806]: DEBUG oslo_concurrency.lockutils [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1561.289871] env[61806]: DEBUG oslo_concurrency.lockutils [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 0.000s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1561.290067] env[61806]: DEBUG oslo_concurrency.lockutils [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1561.290166] env[61806]: DEBUG nova.compute.resource_tracker [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=61806) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 1561.291071] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-22601f34-6c07-457f-9116-d2a83ce97717 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1561.298862] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-264ccdd6-d85a-45e5-88e2-78e68de79661 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1561.312896] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9936ea97-812f-4617-aad2-5577470dd463 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1561.318854] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-799958c6-3777-4698-966a-66a2b365b76a {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1561.346405] env[61806]: DEBUG nova.compute.resource_tracker [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=181135MB free_disk=138GB free_vcpus=48 pci_devices=None {{(pid=61806) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 1561.346584] env[61806]: DEBUG oslo_concurrency.lockutils [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1561.346725] env[61806]: DEBUG oslo_concurrency.lockutils [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 0.000s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1562.370792] env[61806]: DEBUG nova.compute.resource_tracker [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Instance ef96151b-c908-4108-92ca-8a7868312436 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61806) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1562.371069] env[61806]: DEBUG nova.compute.resource_tracker [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Total usable vcpus: 48, total allocated vcpus: 1 {{(pid=61806) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 1562.371148] env[61806]: DEBUG nova.compute.resource_tracker [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=704MB phys_disk=200GB used_disk=1GB total_vcpus=48 used_vcpus=1 pci_stats=[] {{(pid=61806) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 1562.395759] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1e9f86e7-9b2e-46e8-8aad-1dec95b6ad56 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1562.403353] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1a51de50-40ae-4223-8b9e-0edd6f5503f2 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1562.433258] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e868d21d-7fa8-4073-9e52-d249fde258fd {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1562.439882] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f10427b5-b070-40c4-a7ca-fbdce1682478 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1562.452844] env[61806]: DEBUG nova.compute.provider_tree [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Updating inventory in ProviderTree for provider a2858be1-fd22-4e08-979e-87ad25293407 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 138, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61806) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1562.972505] env[61806]: ERROR nova.scheduler.client.report [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] [req-e94c70b3-515c-4b07-bf7c-8b96b397840e] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 138, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID a2858be1-fd22-4e08-979e-87ad25293407. Got 409: {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-e94c70b3-515c-4b07-bf7c-8b96b397840e"}]} [ 1562.988408] env[61806]: DEBUG nova.scheduler.client.report [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Refreshing inventories for resource provider a2858be1-fd22-4e08-979e-87ad25293407 {{(pid=61806) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:818}} [ 1563.000445] env[61806]: DEBUG nova.scheduler.client.report [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Updating ProviderTree inventory for provider a2858be1-fd22-4e08-979e-87ad25293407 from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 139, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61806) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:782}} [ 1563.000620] env[61806]: DEBUG nova.compute.provider_tree [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Updating inventory in ProviderTree for provider a2858be1-fd22-4e08-979e-87ad25293407 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 139, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61806) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1563.010325] env[61806]: DEBUG nova.scheduler.client.report [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Refreshing aggregate associations for resource provider a2858be1-fd22-4e08-979e-87ad25293407, aggregates: None {{(pid=61806) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:827}} [ 1563.026645] env[61806]: DEBUG nova.scheduler.client.report [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Refreshing trait associations for resource provider a2858be1-fd22-4e08-979e-87ad25293407, traits: COMPUTE_IMAGE_TYPE_ISO,COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_SAME_HOST_COLD_MIGRATE,COMPUTE_NET_ATTACH_INTERFACE,COMPUTE_NODE {{(pid=61806) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:839}} [ 1563.048647] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c20cc251-8624-4ac5-b133-978b56e31e04 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1563.055769] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-126141c0-549e-4c22-ba7b-7391616482af {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1563.083983] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-85c05c3b-2f3f-4ae3-b3b8-ba601c4f8abe {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1563.090525] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-54b68c59-aa31-4af9-8372-2ff075c43418 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1563.103210] env[61806]: DEBUG nova.compute.provider_tree [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Updating inventory in ProviderTree for provider a2858be1-fd22-4e08-979e-87ad25293407 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 138, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61806) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1563.632550] env[61806]: DEBUG nova.scheduler.client.report [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Updated inventory for provider a2858be1-fd22-4e08-979e-87ad25293407 with generation 162 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 138, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61806) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:971}} [ 1563.632852] env[61806]: DEBUG nova.compute.provider_tree [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Updating resource provider a2858be1-fd22-4e08-979e-87ad25293407 generation from 162 to 163 during operation: update_inventory {{(pid=61806) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 1563.632921] env[61806]: DEBUG nova.compute.provider_tree [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Updating inventory in ProviderTree for provider a2858be1-fd22-4e08-979e-87ad25293407 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 138, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61806) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1564.137343] env[61806]: DEBUG nova.compute.resource_tracker [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=61806) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 1564.137571] env[61806]: DEBUG oslo_concurrency.lockutils [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 2.791s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1566.374552] env[61806]: DEBUG oslo_service.periodic_task [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Running periodic task ComputeManager._sync_scheduler_instance_info {{(pid=61806) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1566.879645] env[61806]: DEBUG oslo_service.periodic_task [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=61806) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1566.879883] env[61806]: DEBUG oslo_service.periodic_task [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=61806) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1566.880035] env[61806]: DEBUG oslo_service.periodic_task [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=61806) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1566.880199] env[61806]: DEBUG nova.compute.manager [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=61806) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10545}} [ 1569.023302] env[61806]: DEBUG oslo_service.periodic_task [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=61806) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1589.316632] env[61806]: DEBUG oslo_concurrency.lockutils [None req-6101df2d-6f13-4c52-8e7f-afb42e1a0ecc tempest-AttachVolumeTestJSON-132925025 tempest-AttachVolumeTestJSON-132925025-project-member] Acquiring lock "ef96151b-c908-4108-92ca-8a7868312436" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1589.316632] env[61806]: DEBUG oslo_concurrency.lockutils [None req-6101df2d-6f13-4c52-8e7f-afb42e1a0ecc tempest-AttachVolumeTestJSON-132925025 tempest-AttachVolumeTestJSON-132925025-project-member] Lock "ef96151b-c908-4108-92ca-8a7868312436" acquired by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: waited 0.000s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1589.820488] env[61806]: DEBUG nova.compute.utils [None req-6101df2d-6f13-4c52-8e7f-afb42e1a0ecc tempest-AttachVolumeTestJSON-132925025 tempest-AttachVolumeTestJSON-132925025-project-member] Using /dev/sd instead of None {{(pid=61806) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1590.323444] env[61806]: DEBUG oslo_concurrency.lockutils [None req-6101df2d-6f13-4c52-8e7f-afb42e1a0ecc tempest-AttachVolumeTestJSON-132925025 tempest-AttachVolumeTestJSON-132925025-project-member] Lock "ef96151b-c908-4108-92ca-8a7868312436" "released" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: held 1.007s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1591.383253] env[61806]: DEBUG oslo_concurrency.lockutils [None req-6101df2d-6f13-4c52-8e7f-afb42e1a0ecc tempest-AttachVolumeTestJSON-132925025 tempest-AttachVolumeTestJSON-132925025-project-member] Acquiring lock "ef96151b-c908-4108-92ca-8a7868312436" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1591.383617] env[61806]: DEBUG oslo_concurrency.lockutils [None req-6101df2d-6f13-4c52-8e7f-afb42e1a0ecc tempest-AttachVolumeTestJSON-132925025 tempest-AttachVolumeTestJSON-132925025-project-member] Lock "ef96151b-c908-4108-92ca-8a7868312436" acquired by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: waited 0.001s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1591.383781] env[61806]: INFO nova.compute.manager [None req-6101df2d-6f13-4c52-8e7f-afb42e1a0ecc tempest-AttachVolumeTestJSON-132925025 tempest-AttachVolumeTestJSON-132925025-project-member] [instance: ef96151b-c908-4108-92ca-8a7868312436] Attaching volume 5be1c0ea-d62b-49b8-8933-8afef0e140cb to /dev/sdb [ 1591.413719] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c7dc15e4-bde5-421e-9da7-d2458542229e {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1591.420937] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9e5f35e6-35eb-4664-abab-50c3b7b0ae51 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1591.433686] env[61806]: DEBUG nova.virt.block_device [None req-6101df2d-6f13-4c52-8e7f-afb42e1a0ecc tempest-AttachVolumeTestJSON-132925025 tempest-AttachVolumeTestJSON-132925025-project-member] [instance: ef96151b-c908-4108-92ca-8a7868312436] Updating existing volume attachment record: e79bfc2b-ebc2-41e7-9510-e91a43a47c6a {{(pid=61806) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 1595.976166] env[61806]: DEBUG nova.virt.vmwareapi.volumeops [None req-6101df2d-6f13-4c52-8e7f-afb42e1a0ecc tempest-AttachVolumeTestJSON-132925025 tempest-AttachVolumeTestJSON-132925025-project-member] [instance: ef96151b-c908-4108-92ca-8a7868312436] Volume attach. Driver type: vmdk {{(pid=61806) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 1595.976445] env[61806]: DEBUG nova.virt.vmwareapi.volumeops [None req-6101df2d-6f13-4c52-8e7f-afb42e1a0ecc tempest-AttachVolumeTestJSON-132925025 tempest-AttachVolumeTestJSON-132925025-project-member] [instance: ef96151b-c908-4108-92ca-8a7868312436] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-277788', 'volume_id': '5be1c0ea-d62b-49b8-8933-8afef0e140cb', 'name': 'volume-5be1c0ea-d62b-49b8-8933-8afef0e140cb', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'ef96151b-c908-4108-92ca-8a7868312436', 'attached_at': '', 'detached_at': '', 'volume_id': '5be1c0ea-d62b-49b8-8933-8afef0e140cb', 'serial': '5be1c0ea-d62b-49b8-8933-8afef0e140cb'} {{(pid=61806) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 1595.977332] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dd599569-3b4e-49be-9141-49858052cab4 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1595.994915] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-64a80199-bf70-4fb4-88f7-2ca7335b7804 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1596.018803] env[61806]: DEBUG nova.virt.vmwareapi.volumeops [None req-6101df2d-6f13-4c52-8e7f-afb42e1a0ecc tempest-AttachVolumeTestJSON-132925025 tempest-AttachVolumeTestJSON-132925025-project-member] [instance: ef96151b-c908-4108-92ca-8a7868312436] Reconfiguring VM instance instance-00000070 to attach disk [datastore1] volume-5be1c0ea-d62b-49b8-8933-8afef0e140cb/volume-5be1c0ea-d62b-49b8-8933-8afef0e140cb.vmdk or device None with type thin {{(pid=61806) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1596.019062] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-3bb4a19d-5e7e-400b-b664-35dc9653b698 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1596.036794] env[61806]: DEBUG oslo_vmware.api [None req-6101df2d-6f13-4c52-8e7f-afb42e1a0ecc tempest-AttachVolumeTestJSON-132925025 tempest-AttachVolumeTestJSON-132925025-project-member] Waiting for the task: (returnval){ [ 1596.036794] env[61806]: value = "task-1295398" [ 1596.036794] env[61806]: _type = "Task" [ 1596.036794] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1596.044594] env[61806]: DEBUG oslo_vmware.api [None req-6101df2d-6f13-4c52-8e7f-afb42e1a0ecc tempest-AttachVolumeTestJSON-132925025 tempest-AttachVolumeTestJSON-132925025-project-member] Task: {'id': task-1295398, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1596.546783] env[61806]: DEBUG oslo_vmware.api [None req-6101df2d-6f13-4c52-8e7f-afb42e1a0ecc tempest-AttachVolumeTestJSON-132925025 tempest-AttachVolumeTestJSON-132925025-project-member] Task: {'id': task-1295398, 'name': ReconfigVM_Task, 'duration_secs': 0.3434} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1596.546783] env[61806]: DEBUG nova.virt.vmwareapi.volumeops [None req-6101df2d-6f13-4c52-8e7f-afb42e1a0ecc tempest-AttachVolumeTestJSON-132925025 tempest-AttachVolumeTestJSON-132925025-project-member] [instance: ef96151b-c908-4108-92ca-8a7868312436] Reconfigured VM instance instance-00000070 to attach disk [datastore1] volume-5be1c0ea-d62b-49b8-8933-8afef0e140cb/volume-5be1c0ea-d62b-49b8-8933-8afef0e140cb.vmdk or device None with type thin {{(pid=61806) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1596.551253] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-9be92e90-a8c2-4b96-a1f6-dbe0b94c2fa6 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1596.565414] env[61806]: DEBUG oslo_vmware.api [None req-6101df2d-6f13-4c52-8e7f-afb42e1a0ecc tempest-AttachVolumeTestJSON-132925025 tempest-AttachVolumeTestJSON-132925025-project-member] Waiting for the task: (returnval){ [ 1596.565414] env[61806]: value = "task-1295399" [ 1596.565414] env[61806]: _type = "Task" [ 1596.565414] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1596.572797] env[61806]: DEBUG oslo_vmware.api [None req-6101df2d-6f13-4c52-8e7f-afb42e1a0ecc tempest-AttachVolumeTestJSON-132925025 tempest-AttachVolumeTestJSON-132925025-project-member] Task: {'id': task-1295399, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1597.075240] env[61806]: DEBUG oslo_vmware.api [None req-6101df2d-6f13-4c52-8e7f-afb42e1a0ecc tempest-AttachVolumeTestJSON-132925025 tempest-AttachVolumeTestJSON-132925025-project-member] Task: {'id': task-1295399, 'name': ReconfigVM_Task, 'duration_secs': 0.140383} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1597.075588] env[61806]: DEBUG nova.virt.vmwareapi.volumeops [None req-6101df2d-6f13-4c52-8e7f-afb42e1a0ecc tempest-AttachVolumeTestJSON-132925025 tempest-AttachVolumeTestJSON-132925025-project-member] [instance: ef96151b-c908-4108-92ca-8a7868312436] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-277788', 'volume_id': '5be1c0ea-d62b-49b8-8933-8afef0e140cb', 'name': 'volume-5be1c0ea-d62b-49b8-8933-8afef0e140cb', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'ef96151b-c908-4108-92ca-8a7868312436', 'attached_at': '', 'detached_at': '', 'volume_id': '5be1c0ea-d62b-49b8-8933-8afef0e140cb', 'serial': '5be1c0ea-d62b-49b8-8933-8afef0e140cb'} {{(pid=61806) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 1598.112898] env[61806]: DEBUG nova.objects.instance [None req-6101df2d-6f13-4c52-8e7f-afb42e1a0ecc tempest-AttachVolumeTestJSON-132925025 tempest-AttachVolumeTestJSON-132925025-project-member] Lazy-loading 'flavor' on Instance uuid ef96151b-c908-4108-92ca-8a7868312436 {{(pid=61806) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1598.620142] env[61806]: DEBUG oslo_concurrency.lockutils [None req-6101df2d-6f13-4c52-8e7f-afb42e1a0ecc tempest-AttachVolumeTestJSON-132925025 tempest-AttachVolumeTestJSON-132925025-project-member] Lock "ef96151b-c908-4108-92ca-8a7868312436" "released" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: held 7.236s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1599.427712] env[61806]: DEBUG oslo_concurrency.lockutils [None req-6edbfd5a-950f-487f-b437-2027a7da7d5b tempest-AttachVolumeTestJSON-132925025 tempest-AttachVolumeTestJSON-132925025-project-member] Acquiring lock "ef96151b-c908-4108-92ca-8a7868312436" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1599.428049] env[61806]: DEBUG oslo_concurrency.lockutils [None req-6edbfd5a-950f-487f-b437-2027a7da7d5b tempest-AttachVolumeTestJSON-132925025 tempest-AttachVolumeTestJSON-132925025-project-member] Lock "ef96151b-c908-4108-92ca-8a7868312436" acquired by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: waited 0.000s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1599.931128] env[61806]: DEBUG nova.compute.utils [None req-6edbfd5a-950f-487f-b437-2027a7da7d5b tempest-AttachVolumeTestJSON-132925025 tempest-AttachVolumeTestJSON-132925025-project-member] Using /dev/sd instead of None {{(pid=61806) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1600.434664] env[61806]: DEBUG oslo_concurrency.lockutils [None req-6edbfd5a-950f-487f-b437-2027a7da7d5b tempest-AttachVolumeTestJSON-132925025 tempest-AttachVolumeTestJSON-132925025-project-member] Lock "ef96151b-c908-4108-92ca-8a7868312436" "released" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: held 1.006s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1601.488227] env[61806]: DEBUG oslo_concurrency.lockutils [None req-6edbfd5a-950f-487f-b437-2027a7da7d5b tempest-AttachVolumeTestJSON-132925025 tempest-AttachVolumeTestJSON-132925025-project-member] Acquiring lock "ef96151b-c908-4108-92ca-8a7868312436" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1601.488607] env[61806]: DEBUG oslo_concurrency.lockutils [None req-6edbfd5a-950f-487f-b437-2027a7da7d5b tempest-AttachVolumeTestJSON-132925025 tempest-AttachVolumeTestJSON-132925025-project-member] Lock "ef96151b-c908-4108-92ca-8a7868312436" acquired by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: waited 0.001s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1601.488750] env[61806]: INFO nova.compute.manager [None req-6edbfd5a-950f-487f-b437-2027a7da7d5b tempest-AttachVolumeTestJSON-132925025 tempest-AttachVolumeTestJSON-132925025-project-member] [instance: ef96151b-c908-4108-92ca-8a7868312436] Attaching volume 067647b8-9a04-41d3-910d-e279dd278064 to /dev/sdc [ 1601.519176] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-afef57f9-8417-4b1a-b55e-ca3b42cad84f {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1601.526410] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fd520ed9-3a7c-4f39-a107-294dfa1fddc4 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1601.539326] env[61806]: DEBUG nova.virt.block_device [None req-6edbfd5a-950f-487f-b437-2027a7da7d5b tempest-AttachVolumeTestJSON-132925025 tempest-AttachVolumeTestJSON-132925025-project-member] [instance: ef96151b-c908-4108-92ca-8a7868312436] Updating existing volume attachment record: 7cc57224-4738-4b68-87ed-691d0d95e2c3 {{(pid=61806) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 1606.081321] env[61806]: DEBUG nova.virt.vmwareapi.volumeops [None req-6edbfd5a-950f-487f-b437-2027a7da7d5b tempest-AttachVolumeTestJSON-132925025 tempest-AttachVolumeTestJSON-132925025-project-member] [instance: ef96151b-c908-4108-92ca-8a7868312436] Volume attach. Driver type: vmdk {{(pid=61806) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 1606.081583] env[61806]: DEBUG nova.virt.vmwareapi.volumeops [None req-6edbfd5a-950f-487f-b437-2027a7da7d5b tempest-AttachVolumeTestJSON-132925025 tempest-AttachVolumeTestJSON-132925025-project-member] [instance: ef96151b-c908-4108-92ca-8a7868312436] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-277789', 'volume_id': '067647b8-9a04-41d3-910d-e279dd278064', 'name': 'volume-067647b8-9a04-41d3-910d-e279dd278064', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'ef96151b-c908-4108-92ca-8a7868312436', 'attached_at': '', 'detached_at': '', 'volume_id': '067647b8-9a04-41d3-910d-e279dd278064', 'serial': '067647b8-9a04-41d3-910d-e279dd278064'} {{(pid=61806) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 1606.082515] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d5f45249-a6cd-4046-ac86-603716ea3eaf {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1606.098804] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2a415312-6d0d-4e6f-8baf-dd0e391904c3 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1606.124633] env[61806]: DEBUG nova.virt.vmwareapi.volumeops [None req-6edbfd5a-950f-487f-b437-2027a7da7d5b tempest-AttachVolumeTestJSON-132925025 tempest-AttachVolumeTestJSON-132925025-project-member] [instance: ef96151b-c908-4108-92ca-8a7868312436] Reconfiguring VM instance instance-00000070 to attach disk [datastore1] volume-067647b8-9a04-41d3-910d-e279dd278064/volume-067647b8-9a04-41d3-910d-e279dd278064.vmdk or device None with type thin {{(pid=61806) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1606.124862] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-5c7f7712-ca68-4eb5-b51d-74985eed99d1 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1606.142314] env[61806]: DEBUG oslo_vmware.api [None req-6edbfd5a-950f-487f-b437-2027a7da7d5b tempest-AttachVolumeTestJSON-132925025 tempest-AttachVolumeTestJSON-132925025-project-member] Waiting for the task: (returnval){ [ 1606.142314] env[61806]: value = "task-1295402" [ 1606.142314] env[61806]: _type = "Task" [ 1606.142314] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1606.149595] env[61806]: DEBUG oslo_vmware.api [None req-6edbfd5a-950f-487f-b437-2027a7da7d5b tempest-AttachVolumeTestJSON-132925025 tempest-AttachVolumeTestJSON-132925025-project-member] Task: {'id': task-1295402, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1606.652231] env[61806]: DEBUG oslo_vmware.api [None req-6edbfd5a-950f-487f-b437-2027a7da7d5b tempest-AttachVolumeTestJSON-132925025 tempest-AttachVolumeTestJSON-132925025-project-member] Task: {'id': task-1295402, 'name': ReconfigVM_Task, 'duration_secs': 0.321321} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1606.652522] env[61806]: DEBUG nova.virt.vmwareapi.volumeops [None req-6edbfd5a-950f-487f-b437-2027a7da7d5b tempest-AttachVolumeTestJSON-132925025 tempest-AttachVolumeTestJSON-132925025-project-member] [instance: ef96151b-c908-4108-92ca-8a7868312436] Reconfigured VM instance instance-00000070 to attach disk [datastore1] volume-067647b8-9a04-41d3-910d-e279dd278064/volume-067647b8-9a04-41d3-910d-e279dd278064.vmdk or device None with type thin {{(pid=61806) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1606.657119] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-5ea772d3-49ff-49df-9106-966efc20bf21 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1606.672582] env[61806]: DEBUG oslo_vmware.api [None req-6edbfd5a-950f-487f-b437-2027a7da7d5b tempest-AttachVolumeTestJSON-132925025 tempest-AttachVolumeTestJSON-132925025-project-member] Waiting for the task: (returnval){ [ 1606.672582] env[61806]: value = "task-1295403" [ 1606.672582] env[61806]: _type = "Task" [ 1606.672582] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1606.682156] env[61806]: DEBUG oslo_vmware.api [None req-6edbfd5a-950f-487f-b437-2027a7da7d5b tempest-AttachVolumeTestJSON-132925025 tempest-AttachVolumeTestJSON-132925025-project-member] Task: {'id': task-1295403, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1607.182671] env[61806]: DEBUG oslo_vmware.api [None req-6edbfd5a-950f-487f-b437-2027a7da7d5b tempest-AttachVolumeTestJSON-132925025 tempest-AttachVolumeTestJSON-132925025-project-member] Task: {'id': task-1295403, 'name': ReconfigVM_Task, 'duration_secs': 0.130377} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1607.182998] env[61806]: DEBUG nova.virt.vmwareapi.volumeops [None req-6edbfd5a-950f-487f-b437-2027a7da7d5b tempest-AttachVolumeTestJSON-132925025 tempest-AttachVolumeTestJSON-132925025-project-member] [instance: ef96151b-c908-4108-92ca-8a7868312436] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-277789', 'volume_id': '067647b8-9a04-41d3-910d-e279dd278064', 'name': 'volume-067647b8-9a04-41d3-910d-e279dd278064', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'ef96151b-c908-4108-92ca-8a7868312436', 'attached_at': '', 'detached_at': '', 'volume_id': '067647b8-9a04-41d3-910d-e279dd278064', 'serial': '067647b8-9a04-41d3-910d-e279dd278064'} {{(pid=61806) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 1608.218661] env[61806]: DEBUG nova.objects.instance [None req-6edbfd5a-950f-487f-b437-2027a7da7d5b tempest-AttachVolumeTestJSON-132925025 tempest-AttachVolumeTestJSON-132925025-project-member] Lazy-loading 'flavor' on Instance uuid ef96151b-c908-4108-92ca-8a7868312436 {{(pid=61806) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1608.723723] env[61806]: DEBUG oslo_concurrency.lockutils [None req-6edbfd5a-950f-487f-b437-2027a7da7d5b tempest-AttachVolumeTestJSON-132925025 tempest-AttachVolumeTestJSON-132925025-project-member] Lock "ef96151b-c908-4108-92ca-8a7868312436" "released" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: held 7.235s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1609.001942] env[61806]: DEBUG oslo_concurrency.lockutils [None req-f7f197b2-67ed-4881-9889-58406e9bff06 tempest-AttachVolumeTestJSON-132925025 tempest-AttachVolumeTestJSON-132925025-project-member] Acquiring lock "ef96151b-c908-4108-92ca-8a7868312436" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1609.002250] env[61806]: DEBUG oslo_concurrency.lockutils [None req-f7f197b2-67ed-4881-9889-58406e9bff06 tempest-AttachVolumeTestJSON-132925025 tempest-AttachVolumeTestJSON-132925025-project-member] Lock "ef96151b-c908-4108-92ca-8a7868312436" acquired by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: waited 0.001s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1609.505123] env[61806]: INFO nova.compute.manager [None req-f7f197b2-67ed-4881-9889-58406e9bff06 tempest-AttachVolumeTestJSON-132925025 tempest-AttachVolumeTestJSON-132925025-project-member] [instance: ef96151b-c908-4108-92ca-8a7868312436] Detaching volume 5be1c0ea-d62b-49b8-8933-8afef0e140cb [ 1609.535226] env[61806]: INFO nova.virt.block_device [None req-f7f197b2-67ed-4881-9889-58406e9bff06 tempest-AttachVolumeTestJSON-132925025 tempest-AttachVolumeTestJSON-132925025-project-member] [instance: ef96151b-c908-4108-92ca-8a7868312436] Attempting to driver detach volume 5be1c0ea-d62b-49b8-8933-8afef0e140cb from mountpoint /dev/sdb [ 1609.535478] env[61806]: DEBUG nova.virt.vmwareapi.volumeops [None req-f7f197b2-67ed-4881-9889-58406e9bff06 tempest-AttachVolumeTestJSON-132925025 tempest-AttachVolumeTestJSON-132925025-project-member] [instance: ef96151b-c908-4108-92ca-8a7868312436] Volume detach. Driver type: vmdk {{(pid=61806) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 1609.535781] env[61806]: DEBUG nova.virt.vmwareapi.volumeops [None req-f7f197b2-67ed-4881-9889-58406e9bff06 tempest-AttachVolumeTestJSON-132925025 tempest-AttachVolumeTestJSON-132925025-project-member] [instance: ef96151b-c908-4108-92ca-8a7868312436] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-277788', 'volume_id': '5be1c0ea-d62b-49b8-8933-8afef0e140cb', 'name': 'volume-5be1c0ea-d62b-49b8-8933-8afef0e140cb', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'ef96151b-c908-4108-92ca-8a7868312436', 'attached_at': '', 'detached_at': '', 'volume_id': '5be1c0ea-d62b-49b8-8933-8afef0e140cb', 'serial': '5be1c0ea-d62b-49b8-8933-8afef0e140cb'} {{(pid=61806) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 1609.536738] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8a136d50-c3da-46f8-97ff-0eb2f0a47dcb {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1609.560428] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-313224ad-a52a-4567-824c-f5c0155e4018 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1609.566853] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3521eda3-7e01-46ec-b1c7-b50a59abde98 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1609.591130] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3cb489ba-676d-465c-b82c-e8be00f2a311 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1609.604612] env[61806]: DEBUG nova.virt.vmwareapi.volumeops [None req-f7f197b2-67ed-4881-9889-58406e9bff06 tempest-AttachVolumeTestJSON-132925025 tempest-AttachVolumeTestJSON-132925025-project-member] The volume has not been displaced from its original location: [datastore1] volume-5be1c0ea-d62b-49b8-8933-8afef0e140cb/volume-5be1c0ea-d62b-49b8-8933-8afef0e140cb.vmdk. No consolidation needed. {{(pid=61806) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 1609.609802] env[61806]: DEBUG nova.virt.vmwareapi.volumeops [None req-f7f197b2-67ed-4881-9889-58406e9bff06 tempest-AttachVolumeTestJSON-132925025 tempest-AttachVolumeTestJSON-132925025-project-member] [instance: ef96151b-c908-4108-92ca-8a7868312436] Reconfiguring VM instance instance-00000070 to detach disk 2001 {{(pid=61806) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1609.610092] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-f857df12-7eba-40c6-acdf-ca99c3d3249a {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1609.627499] env[61806]: DEBUG oslo_vmware.api [None req-f7f197b2-67ed-4881-9889-58406e9bff06 tempest-AttachVolumeTestJSON-132925025 tempest-AttachVolumeTestJSON-132925025-project-member] Waiting for the task: (returnval){ [ 1609.627499] env[61806]: value = "task-1295404" [ 1609.627499] env[61806]: _type = "Task" [ 1609.627499] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1609.634668] env[61806]: DEBUG oslo_vmware.api [None req-f7f197b2-67ed-4881-9889-58406e9bff06 tempest-AttachVolumeTestJSON-132925025 tempest-AttachVolumeTestJSON-132925025-project-member] Task: {'id': task-1295404, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1610.136461] env[61806]: DEBUG oslo_vmware.api [None req-f7f197b2-67ed-4881-9889-58406e9bff06 tempest-AttachVolumeTestJSON-132925025 tempest-AttachVolumeTestJSON-132925025-project-member] Task: {'id': task-1295404, 'name': ReconfigVM_Task, 'duration_secs': 0.212822} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1610.136750] env[61806]: DEBUG nova.virt.vmwareapi.volumeops [None req-f7f197b2-67ed-4881-9889-58406e9bff06 tempest-AttachVolumeTestJSON-132925025 tempest-AttachVolumeTestJSON-132925025-project-member] [instance: ef96151b-c908-4108-92ca-8a7868312436] Reconfigured VM instance instance-00000070 to detach disk 2001 {{(pid=61806) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1610.141317] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-ddacbc8c-7858-416f-b44d-800de17f2f47 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1610.155361] env[61806]: DEBUG oslo_vmware.api [None req-f7f197b2-67ed-4881-9889-58406e9bff06 tempest-AttachVolumeTestJSON-132925025 tempest-AttachVolumeTestJSON-132925025-project-member] Waiting for the task: (returnval){ [ 1610.155361] env[61806]: value = "task-1295405" [ 1610.155361] env[61806]: _type = "Task" [ 1610.155361] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1610.162273] env[61806]: DEBUG oslo_vmware.api [None req-f7f197b2-67ed-4881-9889-58406e9bff06 tempest-AttachVolumeTestJSON-132925025 tempest-AttachVolumeTestJSON-132925025-project-member] Task: {'id': task-1295405, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1610.665524] env[61806]: DEBUG oslo_vmware.api [None req-f7f197b2-67ed-4881-9889-58406e9bff06 tempest-AttachVolumeTestJSON-132925025 tempest-AttachVolumeTestJSON-132925025-project-member] Task: {'id': task-1295405, 'name': ReconfigVM_Task, 'duration_secs': 0.127801} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1610.665861] env[61806]: DEBUG nova.virt.vmwareapi.volumeops [None req-f7f197b2-67ed-4881-9889-58406e9bff06 tempest-AttachVolumeTestJSON-132925025 tempest-AttachVolumeTestJSON-132925025-project-member] [instance: ef96151b-c908-4108-92ca-8a7868312436] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-277788', 'volume_id': '5be1c0ea-d62b-49b8-8933-8afef0e140cb', 'name': 'volume-5be1c0ea-d62b-49b8-8933-8afef0e140cb', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'ef96151b-c908-4108-92ca-8a7868312436', 'attached_at': '', 'detached_at': '', 'volume_id': '5be1c0ea-d62b-49b8-8933-8afef0e140cb', 'serial': '5be1c0ea-d62b-49b8-8933-8afef0e140cb'} {{(pid=61806) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 1611.206139] env[61806]: DEBUG nova.objects.instance [None req-f7f197b2-67ed-4881-9889-58406e9bff06 tempest-AttachVolumeTestJSON-132925025 tempest-AttachVolumeTestJSON-132925025-project-member] Lazy-loading 'flavor' on Instance uuid ef96151b-c908-4108-92ca-8a7868312436 {{(pid=61806) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1612.212924] env[61806]: DEBUG oslo_concurrency.lockutils [None req-f7f197b2-67ed-4881-9889-58406e9bff06 tempest-AttachVolumeTestJSON-132925025 tempest-AttachVolumeTestJSON-132925025-project-member] Lock "ef96151b-c908-4108-92ca-8a7868312436" "released" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: held 3.210s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1612.231235] env[61806]: DEBUG oslo_concurrency.lockutils [None req-a1cb657a-0863-47d9-b972-9dccdcce0800 tempest-AttachVolumeTestJSON-132925025 tempest-AttachVolumeTestJSON-132925025-project-member] Acquiring lock "ef96151b-c908-4108-92ca-8a7868312436" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1612.231485] env[61806]: DEBUG oslo_concurrency.lockutils [None req-a1cb657a-0863-47d9-b972-9dccdcce0800 tempest-AttachVolumeTestJSON-132925025 tempest-AttachVolumeTestJSON-132925025-project-member] Lock "ef96151b-c908-4108-92ca-8a7868312436" acquired by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: waited 0.000s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1612.734565] env[61806]: INFO nova.compute.manager [None req-a1cb657a-0863-47d9-b972-9dccdcce0800 tempest-AttachVolumeTestJSON-132925025 tempest-AttachVolumeTestJSON-132925025-project-member] [instance: ef96151b-c908-4108-92ca-8a7868312436] Detaching volume 067647b8-9a04-41d3-910d-e279dd278064 [ 1612.764389] env[61806]: INFO nova.virt.block_device [None req-a1cb657a-0863-47d9-b972-9dccdcce0800 tempest-AttachVolumeTestJSON-132925025 tempest-AttachVolumeTestJSON-132925025-project-member] [instance: ef96151b-c908-4108-92ca-8a7868312436] Attempting to driver detach volume 067647b8-9a04-41d3-910d-e279dd278064 from mountpoint /dev/sdc [ 1612.764634] env[61806]: DEBUG nova.virt.vmwareapi.volumeops [None req-a1cb657a-0863-47d9-b972-9dccdcce0800 tempest-AttachVolumeTestJSON-132925025 tempest-AttachVolumeTestJSON-132925025-project-member] [instance: ef96151b-c908-4108-92ca-8a7868312436] Volume detach. Driver type: vmdk {{(pid=61806) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 1612.764827] env[61806]: DEBUG nova.virt.vmwareapi.volumeops [None req-a1cb657a-0863-47d9-b972-9dccdcce0800 tempest-AttachVolumeTestJSON-132925025 tempest-AttachVolumeTestJSON-132925025-project-member] [instance: ef96151b-c908-4108-92ca-8a7868312436] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-277789', 'volume_id': '067647b8-9a04-41d3-910d-e279dd278064', 'name': 'volume-067647b8-9a04-41d3-910d-e279dd278064', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'ef96151b-c908-4108-92ca-8a7868312436', 'attached_at': '', 'detached_at': '', 'volume_id': '067647b8-9a04-41d3-910d-e279dd278064', 'serial': '067647b8-9a04-41d3-910d-e279dd278064'} {{(pid=61806) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 1612.765740] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8f1f7a91-3e26-4f4a-85df-4caf39da5bbc {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1612.787783] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-477262fb-936b-41b7-88ed-668b473f5f11 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1612.794103] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-347485f6-29dd-4f03-a3f1-0788f6303ab6 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1612.813153] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c2191e7e-8815-4ad9-895f-98455cde96d0 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1612.826735] env[61806]: DEBUG nova.virt.vmwareapi.volumeops [None req-a1cb657a-0863-47d9-b972-9dccdcce0800 tempest-AttachVolumeTestJSON-132925025 tempest-AttachVolumeTestJSON-132925025-project-member] The volume has not been displaced from its original location: [datastore1] volume-067647b8-9a04-41d3-910d-e279dd278064/volume-067647b8-9a04-41d3-910d-e279dd278064.vmdk. No consolidation needed. {{(pid=61806) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 1612.831814] env[61806]: DEBUG nova.virt.vmwareapi.volumeops [None req-a1cb657a-0863-47d9-b972-9dccdcce0800 tempest-AttachVolumeTestJSON-132925025 tempest-AttachVolumeTestJSON-132925025-project-member] [instance: ef96151b-c908-4108-92ca-8a7868312436] Reconfiguring VM instance instance-00000070 to detach disk 2002 {{(pid=61806) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1612.832101] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-97706a8c-2ea1-4df9-9bd1-05f9bf7177f4 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1612.848846] env[61806]: DEBUG oslo_vmware.api [None req-a1cb657a-0863-47d9-b972-9dccdcce0800 tempest-AttachVolumeTestJSON-132925025 tempest-AttachVolumeTestJSON-132925025-project-member] Waiting for the task: (returnval){ [ 1612.848846] env[61806]: value = "task-1295406" [ 1612.848846] env[61806]: _type = "Task" [ 1612.848846] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1612.855798] env[61806]: DEBUG oslo_vmware.api [None req-a1cb657a-0863-47d9-b972-9dccdcce0800 tempest-AttachVolumeTestJSON-132925025 tempest-AttachVolumeTestJSON-132925025-project-member] Task: {'id': task-1295406, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1613.358119] env[61806]: DEBUG oslo_vmware.api [None req-a1cb657a-0863-47d9-b972-9dccdcce0800 tempest-AttachVolumeTestJSON-132925025 tempest-AttachVolumeTestJSON-132925025-project-member] Task: {'id': task-1295406, 'name': ReconfigVM_Task, 'duration_secs': 0.219925} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1613.358508] env[61806]: DEBUG nova.virt.vmwareapi.volumeops [None req-a1cb657a-0863-47d9-b972-9dccdcce0800 tempest-AttachVolumeTestJSON-132925025 tempest-AttachVolumeTestJSON-132925025-project-member] [instance: ef96151b-c908-4108-92ca-8a7868312436] Reconfigured VM instance instance-00000070 to detach disk 2002 {{(pid=61806) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1613.363616] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-3f83c8ee-584b-4f09-88b8-627b6a23311f {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1613.377850] env[61806]: DEBUG oslo_vmware.api [None req-a1cb657a-0863-47d9-b972-9dccdcce0800 tempest-AttachVolumeTestJSON-132925025 tempest-AttachVolumeTestJSON-132925025-project-member] Waiting for the task: (returnval){ [ 1613.377850] env[61806]: value = "task-1295407" [ 1613.377850] env[61806]: _type = "Task" [ 1613.377850] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1613.385338] env[61806]: DEBUG oslo_vmware.api [None req-a1cb657a-0863-47d9-b972-9dccdcce0800 tempest-AttachVolumeTestJSON-132925025 tempest-AttachVolumeTestJSON-132925025-project-member] Task: {'id': task-1295407, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1613.887528] env[61806]: DEBUG oslo_vmware.api [None req-a1cb657a-0863-47d9-b972-9dccdcce0800 tempest-AttachVolumeTestJSON-132925025 tempest-AttachVolumeTestJSON-132925025-project-member] Task: {'id': task-1295407, 'name': ReconfigVM_Task, 'duration_secs': 0.136748} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1613.887828] env[61806]: DEBUG nova.virt.vmwareapi.volumeops [None req-a1cb657a-0863-47d9-b972-9dccdcce0800 tempest-AttachVolumeTestJSON-132925025 tempest-AttachVolumeTestJSON-132925025-project-member] [instance: ef96151b-c908-4108-92ca-8a7868312436] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-277789', 'volume_id': '067647b8-9a04-41d3-910d-e279dd278064', 'name': 'volume-067647b8-9a04-41d3-910d-e279dd278064', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'ef96151b-c908-4108-92ca-8a7868312436', 'attached_at': '', 'detached_at': '', 'volume_id': '067647b8-9a04-41d3-910d-e279dd278064', 'serial': '067647b8-9a04-41d3-910d-e279dd278064'} {{(pid=61806) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 1614.430884] env[61806]: DEBUG nova.objects.instance [None req-a1cb657a-0863-47d9-b972-9dccdcce0800 tempest-AttachVolumeTestJSON-132925025 tempest-AttachVolumeTestJSON-132925025-project-member] Lazy-loading 'flavor' on Instance uuid ef96151b-c908-4108-92ca-8a7868312436 {{(pid=61806) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1615.018545] env[61806]: DEBUG oslo_service.periodic_task [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=61806) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1615.438515] env[61806]: DEBUG oslo_concurrency.lockutils [None req-a1cb657a-0863-47d9-b972-9dccdcce0800 tempest-AttachVolumeTestJSON-132925025 tempest-AttachVolumeTestJSON-132925025-project-member] Lock "ef96151b-c908-4108-92ca-8a7868312436" "released" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: held 3.207s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1616.592717] env[61806]: DEBUG oslo_concurrency.lockutils [None req-bd496b58-a40f-49e2-a64b-e5f5e99348a0 tempest-AttachVolumeTestJSON-132925025 tempest-AttachVolumeTestJSON-132925025-project-member] Acquiring lock "ef96151b-c908-4108-92ca-8a7868312436" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1616.593145] env[61806]: DEBUG oslo_concurrency.lockutils [None req-bd496b58-a40f-49e2-a64b-e5f5e99348a0 tempest-AttachVolumeTestJSON-132925025 tempest-AttachVolumeTestJSON-132925025-project-member] Lock "ef96151b-c908-4108-92ca-8a7868312436" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1616.593431] env[61806]: DEBUG oslo_concurrency.lockutils [None req-bd496b58-a40f-49e2-a64b-e5f5e99348a0 tempest-AttachVolumeTestJSON-132925025 tempest-AttachVolumeTestJSON-132925025-project-member] Acquiring lock "ef96151b-c908-4108-92ca-8a7868312436-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1616.593748] env[61806]: DEBUG oslo_concurrency.lockutils [None req-bd496b58-a40f-49e2-a64b-e5f5e99348a0 tempest-AttachVolumeTestJSON-132925025 tempest-AttachVolumeTestJSON-132925025-project-member] Lock "ef96151b-c908-4108-92ca-8a7868312436-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1616.593978] env[61806]: DEBUG oslo_concurrency.lockutils [None req-bd496b58-a40f-49e2-a64b-e5f5e99348a0 tempest-AttachVolumeTestJSON-132925025 tempest-AttachVolumeTestJSON-132925025-project-member] Lock "ef96151b-c908-4108-92ca-8a7868312436-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1616.596285] env[61806]: INFO nova.compute.manager [None req-bd496b58-a40f-49e2-a64b-e5f5e99348a0 tempest-AttachVolumeTestJSON-132925025 tempest-AttachVolumeTestJSON-132925025-project-member] [instance: ef96151b-c908-4108-92ca-8a7868312436] Terminating instance [ 1616.597996] env[61806]: DEBUG nova.compute.manager [None req-bd496b58-a40f-49e2-a64b-e5f5e99348a0 tempest-AttachVolumeTestJSON-132925025 tempest-AttachVolumeTestJSON-132925025-project-member] [instance: ef96151b-c908-4108-92ca-8a7868312436] Start destroying the instance on the hypervisor. {{(pid=61806) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1616.598223] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-bd496b58-a40f-49e2-a64b-e5f5e99348a0 tempest-AttachVolumeTestJSON-132925025 tempest-AttachVolumeTestJSON-132925025-project-member] [instance: ef96151b-c908-4108-92ca-8a7868312436] Destroying instance {{(pid=61806) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1616.599056] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-54fc5a31-2edc-40f2-8ce1-ab4a6322b331 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1616.607880] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [None req-bd496b58-a40f-49e2-a64b-e5f5e99348a0 tempest-AttachVolumeTestJSON-132925025 tempest-AttachVolumeTestJSON-132925025-project-member] [instance: ef96151b-c908-4108-92ca-8a7868312436] Powering off the VM {{(pid=61806) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1616.608122] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-0352b050-602d-42bf-9970-d4ba2df1c013 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1616.614489] env[61806]: DEBUG oslo_vmware.api [None req-bd496b58-a40f-49e2-a64b-e5f5e99348a0 tempest-AttachVolumeTestJSON-132925025 tempest-AttachVolumeTestJSON-132925025-project-member] Waiting for the task: (returnval){ [ 1616.614489] env[61806]: value = "task-1295408" [ 1616.614489] env[61806]: _type = "Task" [ 1616.614489] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1616.621360] env[61806]: DEBUG oslo_vmware.api [None req-bd496b58-a40f-49e2-a64b-e5f5e99348a0 tempest-AttachVolumeTestJSON-132925025 tempest-AttachVolumeTestJSON-132925025-project-member] Task: {'id': task-1295408, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1617.124636] env[61806]: DEBUG oslo_vmware.api [None req-bd496b58-a40f-49e2-a64b-e5f5e99348a0 tempest-AttachVolumeTestJSON-132925025 tempest-AttachVolumeTestJSON-132925025-project-member] Task: {'id': task-1295408, 'name': PowerOffVM_Task, 'duration_secs': 0.189635} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1617.124909] env[61806]: DEBUG nova.virt.vmwareapi.vm_util [None req-bd496b58-a40f-49e2-a64b-e5f5e99348a0 tempest-AttachVolumeTestJSON-132925025 tempest-AttachVolumeTestJSON-132925025-project-member] [instance: ef96151b-c908-4108-92ca-8a7868312436] Powered off the VM {{(pid=61806) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1617.125101] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-bd496b58-a40f-49e2-a64b-e5f5e99348a0 tempest-AttachVolumeTestJSON-132925025 tempest-AttachVolumeTestJSON-132925025-project-member] [instance: ef96151b-c908-4108-92ca-8a7868312436] Unregistering the VM {{(pid=61806) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1617.125350] env[61806]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-c71f4688-a2cf-4ce2-afbc-9c6dfa5c4c05 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1617.634702] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-bd496b58-a40f-49e2-a64b-e5f5e99348a0 tempest-AttachVolumeTestJSON-132925025 tempest-AttachVolumeTestJSON-132925025-project-member] [instance: ef96151b-c908-4108-92ca-8a7868312436] Unregistered the VM {{(pid=61806) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1617.635095] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-bd496b58-a40f-49e2-a64b-e5f5e99348a0 tempest-AttachVolumeTestJSON-132925025 tempest-AttachVolumeTestJSON-132925025-project-member] [instance: ef96151b-c908-4108-92ca-8a7868312436] Deleting contents of the VM from datastore datastore1 {{(pid=61806) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1617.635212] env[61806]: DEBUG nova.virt.vmwareapi.ds_util [None req-bd496b58-a40f-49e2-a64b-e5f5e99348a0 tempest-AttachVolumeTestJSON-132925025 tempest-AttachVolumeTestJSON-132925025-project-member] Deleting the datastore file [datastore1] ef96151b-c908-4108-92ca-8a7868312436 {{(pid=61806) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1617.635546] env[61806]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-1448ae54-2412-4dc7-bcb1-89cc7df3b65a {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1617.643132] env[61806]: DEBUG oslo_vmware.api [None req-bd496b58-a40f-49e2-a64b-e5f5e99348a0 tempest-AttachVolumeTestJSON-132925025 tempest-AttachVolumeTestJSON-132925025-project-member] Waiting for the task: (returnval){ [ 1617.643132] env[61806]: value = "task-1295410" [ 1617.643132] env[61806]: _type = "Task" [ 1617.643132] env[61806]: } to complete. {{(pid=61806) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1617.650563] env[61806]: DEBUG oslo_vmware.api [None req-bd496b58-a40f-49e2-a64b-e5f5e99348a0 tempest-AttachVolumeTestJSON-132925025 tempest-AttachVolumeTestJSON-132925025-project-member] Task: {'id': task-1295410, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1618.154341] env[61806]: DEBUG oslo_vmware.api [None req-bd496b58-a40f-49e2-a64b-e5f5e99348a0 tempest-AttachVolumeTestJSON-132925025 tempest-AttachVolumeTestJSON-132925025-project-member] Task: {'id': task-1295410, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.14028} completed successfully. {{(pid=61806) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1618.154612] env[61806]: DEBUG nova.virt.vmwareapi.ds_util [None req-bd496b58-a40f-49e2-a64b-e5f5e99348a0 tempest-AttachVolumeTestJSON-132925025 tempest-AttachVolumeTestJSON-132925025-project-member] Deleted the datastore file {{(pid=61806) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1618.154808] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-bd496b58-a40f-49e2-a64b-e5f5e99348a0 tempest-AttachVolumeTestJSON-132925025 tempest-AttachVolumeTestJSON-132925025-project-member] [instance: ef96151b-c908-4108-92ca-8a7868312436] Deleted contents of the VM from datastore datastore1 {{(pid=61806) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1618.154993] env[61806]: DEBUG nova.virt.vmwareapi.vmops [None req-bd496b58-a40f-49e2-a64b-e5f5e99348a0 tempest-AttachVolumeTestJSON-132925025 tempest-AttachVolumeTestJSON-132925025-project-member] [instance: ef96151b-c908-4108-92ca-8a7868312436] Instance destroyed {{(pid=61806) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1618.155195] env[61806]: INFO nova.compute.manager [None req-bd496b58-a40f-49e2-a64b-e5f5e99348a0 tempest-AttachVolumeTestJSON-132925025 tempest-AttachVolumeTestJSON-132925025-project-member] [instance: ef96151b-c908-4108-92ca-8a7868312436] Took 1.56 seconds to destroy the instance on the hypervisor. [ 1618.155444] env[61806]: DEBUG oslo.service.loopingcall [None req-bd496b58-a40f-49e2-a64b-e5f5e99348a0 tempest-AttachVolumeTestJSON-132925025 tempest-AttachVolumeTestJSON-132925025-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61806) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1618.155642] env[61806]: DEBUG nova.compute.manager [-] [instance: ef96151b-c908-4108-92ca-8a7868312436] Deallocating network for instance {{(pid=61806) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1618.155736] env[61806]: DEBUG nova.network.neutron [-] [instance: ef96151b-c908-4108-92ca-8a7868312436] deallocate_for_instance() {{(pid=61806) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1618.565976] env[61806]: DEBUG nova.compute.manager [req-9be23fb5-b5e5-4be6-9298-4687f179838d req-f15c6b43-c795-42b4-8404-117741db4da1 service nova] [instance: ef96151b-c908-4108-92ca-8a7868312436] Received event network-vif-deleted-8092e872-d167-4a57-8986-f823cb3d99b3 {{(pid=61806) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1618.566188] env[61806]: INFO nova.compute.manager [req-9be23fb5-b5e5-4be6-9298-4687f179838d req-f15c6b43-c795-42b4-8404-117741db4da1 service nova] [instance: ef96151b-c908-4108-92ca-8a7868312436] Neutron deleted interface 8092e872-d167-4a57-8986-f823cb3d99b3; detaching it from the instance and deleting it from the info cache [ 1618.566396] env[61806]: DEBUG nova.network.neutron [req-9be23fb5-b5e5-4be6-9298-4687f179838d req-f15c6b43-c795-42b4-8404-117741db4da1 service nova] [instance: ef96151b-c908-4108-92ca-8a7868312436] Updating instance_info_cache with network_info: [] {{(pid=61806) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1619.044976] env[61806]: DEBUG nova.network.neutron [-] [instance: ef96151b-c908-4108-92ca-8a7868312436] Updating instance_info_cache with network_info: [] {{(pid=61806) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1619.068973] env[61806]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-c6a3a7f4-06da-43ca-ae10-2857f0c501ea {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1619.078448] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-beec0a51-66ab-4e22-a342-4e359eb7c22a {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1619.100255] env[61806]: DEBUG nova.compute.manager [req-9be23fb5-b5e5-4be6-9298-4687f179838d req-f15c6b43-c795-42b4-8404-117741db4da1 service nova] [instance: ef96151b-c908-4108-92ca-8a7868312436] Detach interface failed, port_id=8092e872-d167-4a57-8986-f823cb3d99b3, reason: Instance ef96151b-c908-4108-92ca-8a7868312436 could not be found. {{(pid=61806) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10963}} [ 1619.547515] env[61806]: INFO nova.compute.manager [-] [instance: ef96151b-c908-4108-92ca-8a7868312436] Took 1.39 seconds to deallocate network for instance. [ 1620.022592] env[61806]: DEBUG oslo_service.periodic_task [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=61806) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1620.022796] env[61806]: DEBUG nova.compute.manager [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Starting heal instance info cache {{(pid=61806) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9926}} [ 1620.022904] env[61806]: DEBUG nova.compute.manager [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Rebuilding the list of instances to heal {{(pid=61806) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9930}} [ 1620.054187] env[61806]: DEBUG oslo_concurrency.lockutils [None req-bd496b58-a40f-49e2-a64b-e5f5e99348a0 tempest-AttachVolumeTestJSON-132925025 tempest-AttachVolumeTestJSON-132925025-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1620.054524] env[61806]: DEBUG oslo_concurrency.lockutils [None req-bd496b58-a40f-49e2-a64b-e5f5e99348a0 tempest-AttachVolumeTestJSON-132925025 tempest-AttachVolumeTestJSON-132925025-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.000s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1620.054706] env[61806]: DEBUG nova.objects.instance [None req-bd496b58-a40f-49e2-a64b-e5f5e99348a0 tempest-AttachVolumeTestJSON-132925025 tempest-AttachVolumeTestJSON-132925025-project-member] Lazy-loading 'resources' on Instance uuid ef96151b-c908-4108-92ca-8a7868312436 {{(pid=61806) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1620.542955] env[61806]: DEBUG oslo_concurrency.lockutils [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Acquiring lock "refresh_cache-ef96151b-c908-4108-92ca-8a7868312436" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1620.543105] env[61806]: DEBUG oslo_concurrency.lockutils [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Acquired lock "refresh_cache-ef96151b-c908-4108-92ca-8a7868312436" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1620.543227] env[61806]: DEBUG nova.network.neutron [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] [instance: ef96151b-c908-4108-92ca-8a7868312436] Forcefully refreshing network info cache for instance {{(pid=61806) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2004}} [ 1620.543384] env[61806]: DEBUG nova.objects.instance [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Lazy-loading 'info_cache' on Instance uuid ef96151b-c908-4108-92ca-8a7868312436 {{(pid=61806) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1620.587615] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bb12647c-af17-4391-a34d-46021b5abfab {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1620.595374] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f7830af6-3032-4f3d-921c-b249a5b35924 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1620.624236] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f0eaeb8d-6dac-4040-98f3-f146e5fe2e6f {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1620.630994] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a19a37c4-7bc1-4d54-8a57-6f5b19dce94a {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1620.643373] env[61806]: DEBUG nova.compute.provider_tree [None req-bd496b58-a40f-49e2-a64b-e5f5e99348a0 tempest-AttachVolumeTestJSON-132925025 tempest-AttachVolumeTestJSON-132925025-project-member] Inventory has not changed in ProviderTree for provider: a2858be1-fd22-4e08-979e-87ad25293407 {{(pid=61806) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1621.145997] env[61806]: DEBUG nova.scheduler.client.report [None req-bd496b58-a40f-49e2-a64b-e5f5e99348a0 tempest-AttachVolumeTestJSON-132925025 tempest-AttachVolumeTestJSON-132925025-project-member] Inventory has not changed for provider a2858be1-fd22-4e08-979e-87ad25293407 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 138, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61806) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1621.565000] env[61806]: DEBUG nova.network.neutron [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] [instance: ef96151b-c908-4108-92ca-8a7868312436] Instance cache missing network info. {{(pid=61806) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 1621.650211] env[61806]: DEBUG oslo_concurrency.lockutils [None req-bd496b58-a40f-49e2-a64b-e5f5e99348a0 tempest-AttachVolumeTestJSON-132925025 tempest-AttachVolumeTestJSON-132925025-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.596s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1621.671238] env[61806]: INFO nova.scheduler.client.report [None req-bd496b58-a40f-49e2-a64b-e5f5e99348a0 tempest-AttachVolumeTestJSON-132925025 tempest-AttachVolumeTestJSON-132925025-project-member] Deleted allocations for instance ef96151b-c908-4108-92ca-8a7868312436 [ 1622.139084] env[61806]: DEBUG nova.network.neutron [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] [instance: ef96151b-c908-4108-92ca-8a7868312436] Updating instance_info_cache with network_info: [] {{(pid=61806) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1622.178606] env[61806]: DEBUG oslo_concurrency.lockutils [None req-bd496b58-a40f-49e2-a64b-e5f5e99348a0 tempest-AttachVolumeTestJSON-132925025 tempest-AttachVolumeTestJSON-132925025-project-member] Lock "ef96151b-c908-4108-92ca-8a7868312436" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.585s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1622.641380] env[61806]: DEBUG oslo_concurrency.lockutils [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Releasing lock "refresh_cache-ef96151b-c908-4108-92ca-8a7868312436" {{(pid=61806) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1622.641636] env[61806]: DEBUG nova.compute.manager [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] [instance: ef96151b-c908-4108-92ca-8a7868312436] Updated the network info_cache for instance {{(pid=61806) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9997}} [ 1622.641919] env[61806]: DEBUG oslo_service.periodic_task [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=61806) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1622.642206] env[61806]: DEBUG oslo_service.periodic_task [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=61806) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1622.642446] env[61806]: DEBUG oslo_service.periodic_task [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Running periodic task ComputeManager.update_available_resource {{(pid=61806) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1623.145355] env[61806]: DEBUG oslo_concurrency.lockutils [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1623.145591] env[61806]: DEBUG oslo_concurrency.lockutils [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 0.000s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1623.145745] env[61806]: DEBUG oslo_concurrency.lockutils [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1623.145902] env[61806]: DEBUG nova.compute.resource_tracker [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=61806) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 1623.146837] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f99af6c8-fdc6-4563-bbc8-3174ecb16aa0 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1623.154787] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a22863f8-829f-467a-b6ba-a64a5006bc86 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1623.168096] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9510f041-cc10-49a5-8c88-4f9383f59584 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1623.173846] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fe9843a3-1c50-475d-8a66-80e63c041817 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1623.201583] env[61806]: DEBUG nova.compute.resource_tracker [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=181073MB free_disk=138GB free_vcpus=48 pci_devices=None {{(pid=61806) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 1623.201939] env[61806]: DEBUG oslo_concurrency.lockutils [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1623.201939] env[61806]: DEBUG oslo_concurrency.lockutils [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 0.000s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1624.224223] env[61806]: DEBUG nova.compute.resource_tracker [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Total usable vcpus: 48, total allocated vcpus: 0 {{(pid=61806) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 1624.224537] env[61806]: DEBUG nova.compute.resource_tracker [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=512MB phys_disk=200GB used_disk=0GB total_vcpus=48 used_vcpus=0 pci_stats=[] {{(pid=61806) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 1624.238024] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9386f823-96eb-400b-a699-d9ae24bd618a {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1624.245418] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c36def80-2361-4977-b672-775f7c8acc42 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1624.275522] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1f1380a1-d6f0-4526-9990-1a5b094fd05f {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1624.282281] env[61806]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1c90a95f-3646-4fa1-a384-76de9b24ff21 {{(pid=61806) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1624.294860] env[61806]: DEBUG nova.compute.provider_tree [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Inventory has not changed in ProviderTree for provider: a2858be1-fd22-4e08-979e-87ad25293407 {{(pid=61806) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1624.797645] env[61806]: DEBUG nova.scheduler.client.report [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Inventory has not changed for provider a2858be1-fd22-4e08-979e-87ad25293407 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 138, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61806) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1625.302906] env[61806]: DEBUG nova.compute.resource_tracker [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=61806) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 1625.303334] env[61806]: DEBUG oslo_concurrency.lockutils [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 2.101s {{(pid=61806) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1626.684020] env[61806]: DEBUG oslo_service.periodic_task [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=61806) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1626.684448] env[61806]: DEBUG oslo_service.periodic_task [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=61806) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1626.684448] env[61806]: DEBUG oslo_service.periodic_task [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=61806) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1626.684614] env[61806]: DEBUG nova.compute.manager [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=61806) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10545}} [ 1631.023679] env[61806]: DEBUG oslo_service.periodic_task [None req-b449a4f6-e826-48b7-abe7-8bd4159ad0c3 None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=61806) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}}